DeepSeek R1 Distill Llama 8B
Overview
Overview
DeepSeek-R1 is the first-generation reasoning model built atop DeepSeek-V3 (671B total parameters, 37B activated per token). It incorporates large-scale reinforcement learning (RL) to enhance its chain-of-thought and reasoning capabilities, delivering strong performance in math, code, and multi-step reasoning tasks.
DeepSeek R1 Distill Llama 8B was released on January 20, 2025.
Performance
Timeline
ReleasedUnknown
Knowledge CutoffUnknown
Specifications
Parameters
8.0B
License
MIT
Training Data
Unknown
Benchmarks
Benchmarks
DeepSeek R1 Distill Llama 8B Performance Across Datasets
Scores sourced from the model's scorecard, paper, or official blog posts
Notice missing or incorrect data?Start an Issue discussion→
Pricing
Pricing
Pricing, performance, and capabilities for DeepSeek R1 Distill Llama 8B across different providers:
No pricing information available for this model.
API Access
API Access Coming Soon
API access for DeepSeek R1 Distill Llama 8B will be available soon through our gateway.
Recent Posts
Recent Reviews
FAQ
Common questions about DeepSeek R1 Distill Llama 8B
DeepSeek R1 Distill Llama 8B was released on January 20, 2025 by DeepSeek.
DeepSeek R1 Distill Llama 8B was created by DeepSeek.
DeepSeek R1 Distill Llama 8B has 8.0 billion parameters.
DeepSeek R1 Distill Llama 8B is released under the MIT license. This is an open-source/open-weight license.