Introduction

This repository hosts the Qwen 3 models for the React Native ExecuTorch library. It includes unquantized and quantized versions of the Qwen model in .pte format, ready for use in the ExecuTorch runtime.

If you'd like to run these models in your own ExecuTorch runtime, refer to the official documentation for setup instructions.

Compatibility

If you intend to use this model outside of React Native ExecuTorch, make sure your runtime is compatible with the ExecuTorch version used to export the .pte files. For more details, see the compatibility note in the ExecuTorch GitHub repository. If you work with React Native ExecuTorch, the constants from the library will guarantee compatibility with runtime used behind the scenes.

These models were exported using v0.6.0 version and no forward compatibility is guaranteed. Older versions of the runtime may not work with these files.

Repository Structure

The repository is organized into three main directories:

  • qwen-3-0.6B
  • qwen-3-1.7B
  • qwen-3-4B

Each directory contains different versions of the model, including quantized, and the original models.

  • The .pte file should be passed to the modelSource parameter.
  • The tokenizer for the models is available within the repo root, under tokenizer.json and tokenizer_config.json
Downloads last month
1,275
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support