Qualcomm NPU
Collection
Latest SOTA models supported on Qualcomm NPU.
•
13 items
•
Updated
•
1
Run Phi-3.5-Mini optimized for Qualcomm NPUs with nexaSDK.
Install nexaSDK and create a free account at sdk.nexa.ai
Activate your device with your access token:
nexa config set license '<access_token>'
Run the model on Qualcomm NPU in one line:
nexa infer NexaAI/phi3.5-mini-npu
Phi-3.5-Mini is a ~3.8B-parameter instruction-tuned language model from Microsoft’s Phi family. It’s designed to deliver strong reasoning and instruction-following quality within a compact footprint, making it ideal for on-device and latency-sensitive applications. This Turbo build uses Nexa’s Qualcomm NPU path for faster inference and higher throughput while preserving model quality.
Input:
Output: