useTextToSpeech
Text to speech is a task that allows to transform written text into spoken language. It is commonly used to implement features such as voice assistants, accessibility tools, or audiobooks.
It is recommended to use models provided by us, which are available at our Hugging Face repository. You can also use constants shipped with our library.
Reference
You can play the generated waveform in any way most suitable to you; however, in the snippet below we utilize the react-native-audio-api library to play synthesized speech.
import {
useTextToSpeech,
KOKORO_MEDIUM,
KOKORO_VOICE_AF_HEART,
} from 'react-native-executorch';
import { AudioContext } from 'react-native-audio-api';
const model = useTextToSpeech({
model: KOKORO_MEDIUM,
voice: KOKORO_VOICE_AF_HEART,
});
const audioContext = new AudioContext({ sampleRate: 24000 });
const handleSpeech = async (text: string) => {
const speed = 1.0;
const waveform = await model.forward(text, speed);
const audioBuffer = audioContext.createBuffer(1, waveform.length, 24000);
audioBuffer.getChannelData(0).set(waveform);
const source = audioContext.createBufferSource();
source.buffer = audioBuffer;
source.connect(audioContext.destination);
source.start();
};
Arguments
model (KokoroConfig) - Object specifying the source files for the Kokoro TTS model (duration predictor, synthesizer).
voice (VoiceConfig) - Object specifying the voice data and phonemizer assets (tagger and lexicon).
preventLoad? - Boolean that can prevent automatic model loading after running the hook.
For more information on loading resources, take a look at loading models page.
Returns
| Field | Type | Description |
|---|---|---|
forward | (text: string, speed?: number) => Promise<Float32Array> | Synthesizes a full text into speech. Returns a promise resolving to the full audio waveform as a Float32Array. |
stream | (input: TextToSpeechStreamingInput) => Promise<void> | Starts a streaming synthesis session. Takes a text input and callbacks to handle audio chunks as they are generated. Ideal for reducing the "time to first audio" for long sentences |
streamStop | (): void | Stops the streaming process if there is any ongoing. |
error | RnExecutorchError | null | Contains the error message if the model failed to load or synthesis failed. |
isGenerating | boolean | Indicates whether the model is currently processing a synthesis. |
isReady | boolean | Indicates whether the model has successfully loaded and is ready for synthesis. |
downloadProgress | number | Tracks the progress of the model and voice assets download process. |


Type definitions
interface TextToSpeechStreamingInput {
text: string;
speed?: number;
onBegin?: () => void | Promise<void>;
onNext?: (chunk: Float32Array) => Promise<void> | void;
onEnd?: () => Promise<void> | void;
}
interface KokoroConfig {
durationSource: ResourceSource;
synthesizerSource: ResourceSource;
}
interface VoiceConfig {
voiceSource: ResourceSource;
extra: {
taggerSource: ResourceSource;
lexiconSource: ResourceSource;
};
}
Running the model
The module provides two ways to generate speech:
forward(text, speed): Generates the complete audio waveform at once. Returns a promise resolving to aFloat32Array.
Since it processes the entire text at once, it might take a significant amount of time to produce an audio for long text inputs.
stream({ text, speed }): An async generator that yields chunks of audio as they are computed. This is ideal for reducing the "time to first audio" for long sentences.
Example
Speech Synthesis
import React from 'react';
import { Button, View } from 'react-native';
import {
useTextToSpeech,
KOKORO_MEDIUM,
KOKORO_VOICE_AF_HEART,
} from 'react-native-executorch';
import { AudioContext } from 'react-native-audio-api';
export default function App() {
const tts = useTextToSpeech({
model: KOKORO_MEDIUM,
voice: KOKORO_VOICE_AF_HEART,
});
const generateAudio = async () => {
const audioData = await tts.forward({
text: 'Hello world! This is a sample text.',
});
// Playback example
const ctx = new AudioContext({ sampleRate: 24000 });
const buffer = ctx.createBuffer(1, audioData.length, 24000);
buffer.getChannelData(0).set(audioData);
const source = ctx.createBufferSource();
source.buffer = buffer;
source.connect(ctx.destination);
source.start();
};
return (
<View style={{ flex: 1, justifyContent: 'center', alignItems: 'center' }}>
<Button title="Speak" onPress={generateAudio} disabled={!tts.isReady} />
</View>
);
}
Streaming Synthesis
import React, { useRef } from 'react';
import { Button, View } from 'react-native';
import {
useTextToSpeech,
KOKORO_MEDIUM,
KOKORO_VOICE_AF_HEART,
} from 'react-native-executorch';
import { AudioContext } from 'react-native-audio-api';
export default function App() {
const tts = useTextToSpeech({
model: KOKORO_MEDIUM,
voice: KOKORO_VOICE_AF_HEART,
});
const contextRef = useRef(new AudioContext({ sampleRate: 24000 }));
const generateStream = async () => {
const ctx = contextRef.current;
await tts.stream({
text: "This is a longer text, which is being streamed chunk by chunk. Let's see how it works!",
onNext: async (chunk) => {
return new Promise((resolve) => {
const buffer = ctx.createBuffer(1, chunk.length, 24000);
buffer.getChannelData(0).set(chunk);
const source = ctx.createBufferSource();
source.buffer = buffer;
source.connect(ctx.destination);
source.onEnded = () => resolve();
source.start();
});
},
});
};
return (
<View style={{ flex: 1, justifyContent: 'center', alignItems: 'center' }}>
<Button title="Stream" onPress={generateStream} disabled={!tts.isReady} />
</View>
);
}
Supported models
| Model | Language |
|---|---|
| Kokoro | English |