Skip to content

Commit 21846e0

Browse files
authored
Update oga_model_prepare.rst
1 parent b3852b0 commit 21846e0

File tree

1 file changed

+6
-2
lines changed

1 file changed

+6
-2
lines changed

docs/oga_model_prepare.rst

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -93,6 +93,9 @@ Copy the quantized model to the Windows PC with Ryzen AI installed, activate the
9393
9494
conda activate ryzen-ai-<version>
9595
pip install torch
96+
pip uninstall onnx -y
97+
conda clean --all
98+
pip install onnx==1.17.0
9699
97100
Generate the final model for Hybrid execution mode:
98101

@@ -129,8 +132,9 @@ In Release 1.5.1 there is a new option added to generate prefill fused version o
129132
130133
After the model is generated, locate the ``genai_config.json`` file inside the model folder. Edit it as follows:
131134

132-
1. Set ``"custom_ops_library"`` to ``"C:/Program Files/RyzenAI/<release version>/deployment/onnx_custom_ops.dll"``
133-
2. Delete these two entries from ``"amd_options"``: ``"dd_cache"`` and ``"compile_fusion_rt"``.
135+
1. Set ``"custom_ops_library"`` to ``"C:\\Program Files\\RyzenAI\\<release version>\\deployment\\onnx_custom_ops.dll"``
136+
2. Delete ``"compile_fusion_rt"`` entry from ``"amd_options"``
137+
3. Set ``dd_cache`` to ``.cache`` folder path inside the generated model folder, for example ``"dd_cache": "C:\\Users\\user\\<model folder>\\.cache"
134138
3. For ``Phi-3.5-mini-instruct``, ``Llama-2-7b-chat-hf model``
135139

136140
- Set ``"hybrid_opt_disable_npu_ops": "1"`` inside ``"amd_options"``.

0 commit comments

Comments
 (0)