Skip to main content

ImageSegmentationModule

TypeScript API implementation of the useImageSegmentation hook.

Reference

import {
ImageSegmentationModule,
DEEPLAB_V3_RESNET50,
} from 'react-native-executorch';

const imageUri = 'path/to/image.png';

// Loading the model
await ImageSegmentationModule.load(DEEPLAB_V3_RESNET50);

// Running the model
const outputDict = await ImageSegmentationModule.forward(imageUri);

Methods

MethodTypeDescription
load(modelSource: ResourceSource): Promise<void>Loads the model, where modelSource is a string that specifies the location of the model binary.
forward(input: string, classesOfInterest?: DeeplabLabel[], resize?: boolean) => Promise<{[key in DeeplabLabel]?: number[]}>Executes the model's forward pass, where :
* input can be a fetchable resource or a Base64-encoded string.
* classesOfInterest is an optional list of DeeplabLabel used to indicate additional arrays of probabilities to output (see section "Running the model"). The default is an empty list.
* resize is an optional boolean to indicate whether the output should be resized to the original image dimensions, or left in the size of the model (see section "Running the model"). The default is false.

The return is a dictionary containing:
* for the key DeeplabLabel.ARGMAX an array of integers corresponding to the most probable class for each pixel
* an array of floats for each class from classesOfInterest corresponding to the probabilities for this class.
onDownloadProgress(callback: (downloadProgress: number) => void): anySubscribe to the download progress event.

Type definitions

type ResourceSource = string | number | object;

Loading the model

To load the model, use the load method. It accepts the modelSource which is a string that specifies the location of the model binary. For more information, take a look at loading models page. This method returns a promise, which can resolve to an error or void.

Running the model

To run the model, you can use the forward method. It accepts three arguments: a required image, an optional list of classes, and an optional flag whether to resize the output to the original dimensions.

  • The image can be a remote URL, a local file URI, or a base64-encoded image.
  • The classesOfInterest list contains classes for which to output the full results. By default the list is empty, and only the most probable classes are returned (essentially an arg max for each pixel). Look at DeeplabLabel enum for possible classes.
  • The resize flag says whether the output will be rescaled back to the size of the image you put in. The default is false. The model runs inference on a scaled (probably smaller) version of your image (224x224 for the DEEPLAB_V3_RESNET50). If you choose to resize, the output will be number[] of size width * height of your original image.
caution

Setting resize to true will make forward slower.

forward returns a promise which can resolve either to an error or a dictionary containing number arrays with size depending on resize:

  • For the key DeeplabLabel.ARGMAX the array contains for each pixel an integer corresponding to the class with the highest probability.
  • For every other key from DeeplabLabel, if the label was included in classesOfInterest the dictionary will contain an array of floats corresponding to the probability of this class for every pixel.