DeepSeek Papers

DeepSeek Papers

1. DeepSeek LLM: Scaling Open-Source Language Models with Longtermism

Description: Scaling open-source language models with a focus on longtermism.

2. DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models

Description: Exploring expert specialization in Mixture-of-Experts language models.

3. DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence

Description: Investigating the intersection of large language models and programming.

17. Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention

Description: Hardware-Aligned and Natively Trainable Sparse Attention.

Related Links

There's a lot of excellent work being done in the field of AI and machine learning. For more information, check out these resources:

BibTeX


@article{deepseek2024papers,
  author    = {DeepSeek Research Team},
  title     = {DeepSeek Papers: Advancements in Language Models and Multimodal Understanding},
  journal   = {DeepSeek Publications},
  year      = {2024-2025},
}