You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+78-20Lines changed: 78 additions & 20 deletions
Original file line number
Diff line number
Diff line change
@@ -20,7 +20,14 @@ Optimum ExecuTorch enables efficient deployment of transformer models using Meta
20
20
21
21
## ⚡ Quick Installation
22
22
23
-
Install from source:
23
+
### 1. Create a virtual environment:
24
+
Install [conda](https://docs.conda.io/projects/conda/en/latest/user-guide/install/index.html) on your machine. Then, create a virtual environment to manage our dependencies.
prompt="Simply put, the theory of relativity states that",
62
119
max_seq_len=128
63
120
)
121
+
print(generated_text)
64
122
```
65
123
66
-
### Option 2: Python API
67
-
```python
68
-
from optimum.executorch import ExecuTorchModelForCausalLM
69
-
from transformers import AutoTokenizer
124
+
## Supported Models
70
125
71
-
# Load and export the model on-the-fly
72
-
model_id ="meta-llama/Llama-3.2-1B"
73
-
model = ExecuTorchModelForCausalLM.from_pretrained(model_id, recipe="xnnpack")
126
+
Optimum ExecuTorch currently supports the following transformer models:
127
+
128
+
-**meta-llama/Llama-3.2-1B (and its variants):** A robust large language model designed for a wide range of natural language tasks.
129
+
-**HuggingFaceTB/SmolLM2-135M (and its variants):** A lightweight model optimized for rapid inference with a smaller computational footprint.
130
+
-**Qwen/Qwen2.5-0.5B (and its variants):** An efficient model delivering balanced performance, especially suited for resource-constrained environments.
131
+
-**deepseek-ai/DeepSeek-R1-Distill-Llama-8B:** A distilled version of the Llama model, offering faster inference while retaining strong performance.
132
+
-**google/gemma-2-2b (and its variants):** A cutting-edge model from Google, optimized for diverse deployment scenarios.
133
+
-**allenai/OLMo-1B-hf:** A specialized model from Allen AI, tailored for advanced language understanding tasks.
134
+
135
+
*Note: This list is continuously expanding. As we continue to expand support, more models and variants will be added.*
136
+
137
+
138
+
## Supported Recipes
139
+
140
+
Optimum ExecuTorch currently only supports [`XNNPACK` Backend](https://pytorch.org/executorch/main/backends-xnnpack.html).
0 commit comments