About DeepSeek-R1

About DeepSeek-R1

A quick overview

Pasted image 20250205031941.png


Pasted image 20250205032023.png

Based on the abstract and introduction, this paper essentially proves that obtaining a post-training model solely through a pure reinforcement learning process is possible. Although R1-Zero exhibits poor readability and language mixing, it still demonstrates remarkable reasoning capabilities. The remaining question is how to address the abnormal outputs. The text underlined in red in the picture outlines their entire process.

Keyword:

  1. Pure RL.
  2. CoT, it is like gpt-o1.
  3. Group Relative Policy Optimization (GRPO), they adopt GRPO as the RL framework.

Worth noticing:

  1. Reward modeling
    1. Accuracy.
    2. Format. - format matters. (like the tag </think>)
  2. our distilled 14 B model outperforms state-of-the-art open-source QwQ-32 B-Preview.

  3. Their unsuccessful Attempts
    1. Process Reward Model
    2. Monte Carlo Tree Search

Run it on my device...

Running 8b model on M4 mac mini is practical, the number of token/s is about 20.

While it is around 7 on my Windows laptop without a graphics card.

As for the 14B model...

Screenshot 2025-02-05 035837.png