Running LLMs
React Native ExecuTorch supports Llama 3.2 models, including quantized versions. Before getting started, you’ll need to obtain the .pte binary—a serialized model—and the tokenizer. There are various ways to accomplish this:
Exporting Llama
In order to make the process of export as simple as possible for you, we created a script that runs a Docker container and exports the model.