ExecuTorch documentation

Quickstart

Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

Quickstart

Export

You can export your πŸ€— Transformers models to ExecuTorch easily:

optimum-cli export executorch --model meta-llama/Llama-3.2-1B --recipe xnnpack --output_dir meta_llama3_2_1b_executorch

Inference

To load a model and run inference, you can just replace your AutoModelForCausalLM class with the corresponding ExecuTorchModelForCausalLM class. You can also load a PyTorch checkpoint and convert it to ExecuTorch on-the-fly when loading your model.

- from transformers import AutoModelForCausalLM
+ from optimum.executorch import ExecuTorchModelForCausalLM
  from transformers import AutoTokenizer

  model_id = "meta-llama/Llama-3.2-1B"
  tokenizer = AutoTokenizer.from_pretrained(model_id)
- model = AutoModelForCausalLM.from_pretrained(model_id)
+ model = ExecuTorchModelForCausalLM.from_pretrained(model_id)