Skip to main content
Version: 0.2.x

Loading models

There are three different methods available for loading model files, depending on their size and location.

1. Load from React-Native assets folder (For Files < 512MB)

modelSource: require('../assets/llama3_2.pte');

2. Load from Remote URL:

For files larger than 512MB or when you want to keep size of the app smaller, you can load the model from a remote URL (e.g. HuggingFace).

modelSource: 'https://.../llama3_2.pte';

3. Load from local file system:

If you prefer to delegate the process of obtaining and loading model and tokenizer files to the user, you can use the following method:

modelSource: 'file::///var/mobile/.../llama3_2.pte',
info

The downloaded files are stored in documents directory of your application.

Example

The following code snippet demonstrates how to load model and tokenizer files using useLLM hook:

import { useLLM } from 'react-native-executorch';

const llama = useLLM({
modelSource: 'https://.../llama3_2.pte',
tokenizer: require('../assets/tokenizer.bin'),
});

We are Software Mansion.