DeepSeek-R1-Distill-Qwen-1.5B

Generate text responses to prompts, enabling natural language understanding, multilingual support, content creation and advanced reasoning

Model Properties

Utilizes a transformer-based architecture with instruction tuning to enhance logical reasoning, natural language understanding, and content generation

License name: MIT License
Number of parameters: 1.5B
Model Size: 2.37 GB
Select device..

Technical Details

Operations: 29.4 GOPs per input token
Context Length: 2048
Numerical Scheme: A8W4, symmetric, group-wise
Inference API: C++, Python, Hailo-Ollama
Compiled Model:

Performance Metrics

Load Time In Sec 4.79
Time To First Token In Sec 0.74
TPS 6.98

Explore Related Models

Qwen2.5-Coder-1.5B-Instruct
GenAI Models
Qwen2.5-Coder-1.5B
Generate text responses to prompts, enabling natural language understanding, multilingual support, and code generation
LLM
GenAI Models
Qwen2.5-1.5B-Instruct
Generate text responses to prompts, enabling natural language understanding, multilingual support, and content creation