You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/models/llama2/README.md
+2-1
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,5 @@
1
1
# Summary
2
-
This example demonstrates how to run a [Llama 2](https://llama.meta.com/llama2/) 7B or [Llama 3](https://ai.meta.com/llama/) 8B model on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on a phone.
2
+
This example demonstrates how to run a [Llama 2](https://llama.meta.com/llama2/) 7B or [Llama 3](https://ai.meta.com/llama/) 8B model on mobile via ExecuTorch. We use XNNPACK to accelerate the performance and 4-bit groupwise PTQ quantization to fit the model on a phone.
3
3
4
4
For more details, see [Llama 2 repo](https://github.com/facebookresearch/llama) or [Llama 3 repo](https://github.com/facebookresearch/llama3).
5
5
@@ -107,6 +107,7 @@ If you want to deploy and run a smaller model for educational purposes. From `ex
107
107
### Option C: Download and export Llama 3 8B instruct model
108
108
109
109
You can export and run the original Llama 3 8B instruct model.
110
+
> :warning: **use the main branch**: Llama 3 is only supported on the ExecuTorch main branch (not release 2.0)
110
111
111
112
1. Llama 3 pretrained parameters can be downloaded from [Meta's official Llama 3 repository](https://github.com/meta-llama/llama3/).
0 commit comments