DeepSeek-R1-Distill-Qwen-1.5B

Generate text responses to prompts, enabling natural language understanding, multilingual support, content creation and advanced reasoning

Model Properties

Utilizes a transformer-based architecture with instruction tuning to enhance logical reasoning, natural language understanding, and content
generation.

License name: MIT License
Number of parameters: 1.5B
Model Size: 2.37 GB
Select device..

Technical Details

Operations: 29.4 GOPs per input token
Context Length: 2048
Numerical Scheme: A8W4, symmetric, channel-wise, symmetric, channel-wise
Inference API: CPP, Hailo-Ollama
Compiled Model:

Performance Metrics

Load Time In Sec 8.82
Time To First Token In Sec 0.66
TPS 7.05

Explore Related Models

Qwen2.5-Coder-1.5B-Instruct
GenAI Models
Qwen2.5-Coder-1.5B
Generate text responses to prompts, enabling natural language understanding, multilingual support, and code generation
LLM
GenAI Models
Qwen2.5 1.5B-Instruct
Generate text responses to prompts, enabling natural language understanding, multilingual support, and content creation