Is deepseek’s r1 model distillated from ChatGPT?

Yes, there is evidence suggesting that DeepSeek’s R1 model was developed using a technique called “distillation� from OpenAI’s models. Distillation involves training a new model by leveraging the outputs of a pre-existing model, effectively transferring knowledge from the original to the new model. OpenAI has indicated that it found evidence linking DeepSeek to the use of distillation—a common technique developers use to train models by leveraging the outputs of existing ones. 

 

Additionally, discussions within the AI community have raised concerns about DeepSeek’s methods. For instance, a thread on the OpenAI Developer Forum titled “Looks like Deep Seek R1/V3 was distilled from GPT-4/3.5 - Can anyone confirm?� delves into this topic. 

 

Therefore, it appears that DeepSeek’s R1 model was indeed developed through distillation from OpenAI’s models.

 

所有跟帖: 

Answer from ChatGPT 4o -QualityWithoutName- 給 QualityWithoutName 發送悄悄話 QualityWithoutName 的博客首頁 (0 bytes) () 01/30/2025 postreply 10:35:14

他們為什麽不自己蒸餾?還是技不如人 -sxyz- 給 sxyz 發送悄悄話 (0 bytes) () 01/30/2025 postreply 10:49:09

蒸餾技 -Bob007- 給 Bob007 發送悄悄話 (0 bytes) () 01/30/2025 postreply 11:19:43

我剛剛問了ChatGPT 同樣的問題,回答正相反。奇怪! -方外居士- 給 方外居士 發送悄悄話 (5444 bytes) () 01/30/2025 postreply 10:50:31

Can anyone confirm? -監考老師- 給 監考老師 發送悄悄話 監考老師 的博客首頁 (0 bytes) () 01/30/2025 postreply 11:00:55

上麵附上了截屏。 -QualityWithoutName- 給 QualityWithoutName 發送悄悄話 QualityWithoutName 的博客首頁 (0 bytes) () 01/30/2025 postreply 11:21:00

請您先登陸,再發跟帖!