Intel NPU
Collection
Latest SOTA models supported on Intel NPU
•
6 items
•
Updated
Run Llama-3.2-3B optimized for Intel NPUs with nexaSDK.
Install nexaSDK and create a free account at sdk.nexa.ai
Activate your device with your access token:
nexa config set license '<access_token>'
Run the model on Qualcomm NPU in one line:
nexa infer NexaAI/llama3.2-3B-intel-npu
Llama-3.2-3B is a compact member of the Llama 3.2 family, designed to provide strong general-purpose language modeling in a lightweight 3B parameter footprint.
It balances efficiency with capability, making it well-suited for edge devices, prototyping, and applications where latency and resource constraints are critical.
Input: Text prompts—questions, commands, or code snippets.
Output: Natural language responses including answers, explanations, or structured outputs.