Spaces:
Running
Running
File size: 3,112 Bytes
85bcb47 ecc40f6 85bcb47 ecc40f6 85bcb47 ecc40f6 85bcb47 ecc40f6 63c5240 ecc40f6 63c5240 ecc40f6 5852be1 ecc40f6 63c5240 ecc40f6 63c5240 ecc40f6 bf1e8f3 ecc40f6 89f340d ecc40f6 89f340d ecc40f6 85bcb47 94b76d5 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 |
<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>DeepSeek Papers</title>
<link rel="stylesheet" href="styles.css">
</head>
<body>
<header>
<h1>DeepSeek Papers</h1>
</header>
<main>
<section>
<h2>DeepSeek Papers</h2>
<article>
<h3>1. <strong>DeepSeek LLM: Scaling Open-Source Language Models with Longtermism</strong></h3>
<p>Description: Scaling open-source language models with a focus on longtermism.</p>
<ul>
<li><a href="https://huggingface.co/papers/2401.02954">Link to Paper</a> {Jan 6, 2024}</li>
</ul>
</article>
<article>
<h3>2. <strong>DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models</strong></h3>
<p>Description: Exploring expert specialization in Mixture-of-Experts language models.</p>
<ul>
<li><a href="https://arxiv.org/abs/2401.06066">Link to Paper</a> {Jan 11, 2024}</li>
</ul>
</article>
<article>
<h3>3. <strong>DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence</strong></h3>
<p>Description: Investigating the intersection of large language models and programming.</p>
<ul>
<li><a href="https://arxiv.org/abs/2401.14196">Link to Paper</a> {Jan 25, 2024}</li>
</ul>
</article>
<!-- Repeat for the remaining papers -->
<article>
<h3>17. <strong>Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention</strong></h3>
<p>Description: Hardware-Aligned and Natively Trainable Sparse Attention.</p>
<ul>
<li><a href="https://arxiv.org/abs/2502.11089">Link to Paper</a> {Feb 16, 2025}</li>
</ul>
</article>
</section>
<section>
<h2>Related Links</h2>
<p>There's a lot of excellent work being done in the field of AI and machine learning. For more information, check out these resources:</p>
<ul>
<li><a href="https://huggingface.co/papers">Hugging Face Papers</a></li>
<li><a href="https://arxiv.org">arXiv.org</a></li>
</ul>
</section>
<section>
<h2>BibTeX</h2>
<pre><code class="bibtex">
@article{deepseek2024papers,
author = {DeepSeek Research Team},
title = {DeepSeek Papers: Advancements in Language Models and Multimodal Understanding},
journal = {DeepSeek Publications},
year = {2024-2025},
}
</code></pre>
</section>
</main>
<footer>
<p>This website is licensed under a <a href="https://creativecommons.org/licenses/by-sa/4.0/">Creative Commons Attribution-ShareAlike 4.0 International License</a>.</p>
<p>This means you are free to borrow the <a href="source-code-link">source code</a> of this website; we just ask that you link back to this page in the footer. Please remember to remove the analytics code included in the header of the website which you do not want on your website.</p>
</footer>
</body>
</html> |