--- license: mit library_name: transformers base_model: - deepseek-ai/DeepSeek-V3-0324 - deepseek-ai/DeepSeek-R1 pipeline_tag: text-generation --- # DeepSeek-R1T-Chimera
TNG Logo

License

License
**Model merge of DeepSeek-R1 and DeepSeek-V3 (0324)** An open weights model combining the intelligence of R1 with the token efficiency of V3. [Announcement on X](https://x.com/tngtech/status/1916284566127444468) | [LinkedIn post](https://www.linkedin.com/posts/tng-technology-consulting_on-the-weekend-we-released-deepseek-r1t-chimera-activity-7323008947236290560-Cf2m) ## Model Details - **Architecture**: DeepSeek-MoE Transformer-based language model - **Combination Method**: Merged model weights from DeepSeek-R1 and DeepSeek-V3 (0324) - **Release Date**: 2025-04-27 ## Contact - Email: research@tngtech.com - X.com: @tngtech