This repository consists of methods to run Transformers in PyTorch and ONNX with operators dispatch to NPU.
Open Anaconda prompt on Windows PC.
git clone https://github.com/amd/RyzenAI-SW.git
cd RyzenAI-SW\example\transformers
conda env create --file=env.yaml
conda activate ryzenai-transformers
AWQ Model zoo has precomputed scales, clips and zeros for various LLMs including OPT, Llama. Get the precomputed results:
git lfs install
cd RyzenAI-SW\example\transformers\ext
git clone https://huggingface.co/datasets/mit-han-lab/awq-model-zoo awq_cache
cd RyzenAI-SW\example\transformers\
setup.bat
pip install ops\cpp --force-reinstall
Download Ryzen-AI Software package and extract
NOTE: This step is not required for Pytorch based flows
pip install onnxruntime
cd ryzen-ai-sw-1.1\ryzen-ai-sw-1.1\voe-4.0-win_amd64
pip install voe-0.1.0-cp39-cp39-win_amd64.whl
pip install onnxruntime_vitisai-1.15.1-cp39-cp39-win_amd64.whl
python installer.py