Update README.md
Browse files
README.md
CHANGED
@@ -164,12 +164,14 @@ print(make_table(results))
|
|
164 |
|
165 |
# Exporting to ExecuTorch
|
166 |
|
167 |
-
|
|
|
168 |
|
169 |
|
170 |
## Convert quantized checkpoint to ExecuTorch's format
|
171 |
|
172 |
-
|
|
|
173 |
```
|
174 |
python -m executorch.examples.models.phi_4_mini.convert_weights phi4-mini-8dq4w.bin phi4-mini-8dq4w-converted.bin
|
175 |
```
|
|
|
164 |
|
165 |
# Exporting to ExecuTorch
|
166 |
|
167 |
+
We can run the quantized model on a mobile phone using [ExecuTorch](https://github.com/pytorch/executorch).
|
168 |
+
Once ExecuTorch is [set-up](https://pytorch.org/executorch/main/getting-started.html), exporting and running the model on device is a breeze.
|
169 |
|
170 |
|
171 |
## Convert quantized checkpoint to ExecuTorch's format
|
172 |
|
173 |
+
We first convert the quantized checkpoint to one ExecuTorch's LLM export script expects by renaming some of the checkpoint keys.
|
174 |
+
The following script does this for you.
|
175 |
```
|
176 |
python -m executorch.examples.models.phi_4_mini.convert_weights phi4-mini-8dq4w.bin phi4-mini-8dq4w-converted.bin
|
177 |
```
|