YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

DeepSeek-r1-distill-qwen-1.5B

Run DeepSeek-r1-distill-qwen-1.5B optimized for Intel NPUs with nexaSDK.

Quickstart

  1. Install nexaSDK and create a free account at sdk.nexa.ai

  2. Activate your device with your access token:

    nexa config set license '<access_token>'
    
  3. Run the model on NPU in one line:

    nexa infer NexaAI/deepSeek-r1-distill-qwen-1.5B-intel-npu
    

Model Description

deepSeek-r1-distill-qwen-1.5B is a distilled variant of DeepSeek-R1, built on the Qwen-1.5B architecture.
It compresses the reasoning and instruction-following capabilities of larger DeepSeek models into an ultra-lightweight 1.5B parameter model—ideal for fast, efficient deployment on constrained devices while retaining strong performance for its size.

Features

  • Distilled from DeepSeek-R1: Maintains core reasoning and comprehension strengths in a smaller model.
  • Instruction-tuned: Optimized for Q&A, task completion, and logical reasoning.
  • Compact footprint: 1.5B parameters enable deployment in edge and mobile contexts.
  • Multilingual support: Handles a wide range of global languages with efficiency.

Use Cases

  • Lightweight conversational agents and personal assistants.
  • Coding help and small-scale algorithmic reasoning.
  • Multilingual Q&A or translation in resource-limited environments.
  • Edge, mobile, and offline applications where compute or memory is limited.

Inputs and Outputs

Input: Text prompts including natural language queries, tasks, or code snippets.
Output: Direct responses—answers, explanations, or code—without extra reasoning annotations.

License

  • Licensed under Apache-2.0

References

Downloads last month
29
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including NexaAI/deepSeek-r1-distill-qwen-1.5B-intel-npu