From e2d4c31b6fffb2dc125b69ad8f369a9bb5691a80 Mon Sep 17 00:00:00 2001 From: musvaage <112724366+musvaage@users.noreply.github.com> Date: Thu, 30 Jan 2025 16:01:12 -0600 Subject: [PATCH] alert --- README.md | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index c6ea85a..8fcef26 100644 --- a/README.md +++ b/README.md @@ -53,7 +53,8 @@ we introduce DeepSeek-R1, which incorporates cold-start data before RL. DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. To support the research community, we have open-sourced DeepSeek-R1-Zero, DeepSeek-R1, and six dense models distilled from DeepSeek-R1 based on Llama and Qwen. DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. -**NOTE: Before running DeepSeek-R1 series models locally, we kindly recommend reviewing the [Usage Recommendation](#usage-recommendations) section.** +> [!NOTE] +> Before running DeepSeek-R1 series models locally, we kindly recommend reviewing the [Usage Recommendation](#usage-recommendations) section.

@@ -180,7 +181,8 @@ We also provide OpenAI-Compatible API at DeepSeek Platform: [platform.deepseek.c Please visit [DeepSeek-V3](https://github.com/deepseek-ai/DeepSeek-V3) repo for more information about running DeepSeek-R1 locally. -**NOTE: Hugging Face's Transformers has not been directly supported yet.** +> [!NOTE] +> Hugging Face's Transformers has not been directly supported yet. ### DeepSeek-R1-Distill Models