mirror of
https://github.com/deepseek-ai/DeepSeek-R1.git
synced 2025-02-22 21:59:00 -05:00
Add table of contents to README for better navigation
This commit is contained in:
parent
ed4409d2a9
commit
e6668d558a
20
README.md
20
README.md
@ -37,11 +37,19 @@
|
||||
</a>
|
||||
</div>
|
||||
|
||||
|
||||
<p align="center">
|
||||
<a href="https://github.com/deepseek-ai/DeepSeek-R1/blob/main/DeepSeek_R1.pdf"><b>Paper Link</b>👁️</a>
|
||||
</p>
|
||||
|
||||
1. [Introduction](#1-introduction)
|
||||
2. [Model Summary](#2-model-summary)
|
||||
3. [Model Downloads](#3-model-downloads)
|
||||
4. [Evaluation Results](#4-evaluation-results)
|
||||
5. [Chat Website & API Platform](#5-chat-website--api-platform)
|
||||
6. [How to Run Locally](#6-how-to-run-locally)
|
||||
7. [License](#7-license)
|
||||
8. [Citation](#8-citation)
|
||||
9. [Contact](#9-contact)
|
||||
|
||||
## 1. Introduction
|
||||
|
||||
@ -114,10 +122,10 @@ We slightly change their configs and tokenizers. Please use our setting to run t
|
||||
## 4. Evaluation Results
|
||||
|
||||
### DeepSeek-R1-Evaluation
|
||||
|
||||
For all our models, the maximum generation length is set to 32,768 tokens. For benchmarks requiring sampling, we use a temperature of $0.6$, a top-p value of $0.95$, and generate 64 responses per query to estimate pass@1.
|
||||
<div align="center">
|
||||
|
||||
|
||||
| Category | Benchmark (Metric) | Claude-3.5-Sonnet-1022 | GPT-4o 0513 | DeepSeek V3 | OpenAI o1-mini | OpenAI o1-1217 | DeepSeek R1 |
|
||||
|----------|-------------------|----------------------|------------|--------------|----------------|------------|--------------|
|
||||
| | Architecture | - | - | MoE | - | - | MoE |
|
||||
@ -147,10 +155,8 @@ We slightly change their configs and tokenizers. Please use our setting to run t
|
||||
|
||||
</div>
|
||||
|
||||
|
||||
### Distilled Model Evaluation
|
||||
|
||||
|
||||
<div align="center">
|
||||
|
||||
| Model | AIME 2024 pass@1 | AIME 2024 cons@64 | MATH-500 pass@1 | GPQA Diamond pass@1 | LiveCodeBench pass@1 | CodeForces rating |
|
||||
@ -168,8 +174,8 @@ We slightly change their configs and tokenizers. Please use our setting to run t
|
||||
|
||||
</div>
|
||||
|
||||
|
||||
## 5. Chat Website & API Platform
|
||||
|
||||
You can chat with DeepSeek-R1 on DeepSeek's official website: [chat.deepseek.com](https://chat.deepseek.com), and switch on the button "DeepThink"
|
||||
|
||||
We also provide OpenAI-Compatible API at DeepSeek Platform: [platform.deepseek.com](https://platform.deepseek.com/)
|
||||
@ -208,13 +214,16 @@ python3 -m sglang.launch_server --model deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
|
||||
4. When evaluating model performance, it is recommended to conduct multiple tests and average the results.
|
||||
|
||||
## 7. License
|
||||
|
||||
This code repository and the model weights are licensed under the [MIT License](https://github.com/deepseek-ai/DeepSeek-R1/blob/main/LICENSE).
|
||||
DeepSeek-R1 series support commercial use, allow for any modifications and derivative works, including, but not limited to, distillation for training other LLMs. Please note that:
|
||||
|
||||
- DeepSeek-R1-Distill-Qwen-1.5B, DeepSeek-R1-Distill-Qwen-7B, DeepSeek-R1-Distill-Qwen-14B and DeepSeek-R1-Distill-Qwen-32B are derived from [Qwen-2.5 series](https://github.com/QwenLM/Qwen2.5), which are originally licensed under [Apache 2.0 License](https://huggingface.co/Qwen/Qwen2.5-1.5B/blob/main/LICENSE), and now finetuned with 800k samples curated with DeepSeek-R1.
|
||||
- DeepSeek-R1-Distill-Llama-8B is derived from Llama3.1-8B-Base and is originally licensed under [llama3.1 license](https://huggingface.co/meta-llama/Llama-3.1-8B/blob/main/LICENSE).
|
||||
- DeepSeek-R1-Distill-Llama-70B is derived from Llama3.3-70B-Instruct and is originally licensed under [llama3.3 license](https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct/blob/main/LICENSE).
|
||||
|
||||
## 8. Citation
|
||||
|
||||
```
|
||||
@misc{deepseekai2025deepseekr1incentivizingreasoningcapability,
|
||||
title={DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning},
|
||||
@ -229,4 +238,5 @@ DeepSeek-R1 series support commercial use, allow for any modifications and deriv
|
||||
```
|
||||
|
||||
## 9. Contact
|
||||
|
||||
If you have any questions, please raise an issue or contact us at [service@deepseek.com](service@deepseek.com).
|
||||
|
Loading…
Reference in New Issue
Block a user