Discussion about this post

User's avatar
Nimish Sanghi's avatar

Like the balanced analysis of the issue.

Jordan Schneider's avatar

Or deepseek was just sneakier about it and only 150k they could attribute with confidence

“In the scale of training a language model, 150K samples is only scratching the surface as a substantive experiment. It looks like they were experimenting with some rubrics, which could’ve been for an online RL run, but that’s extremely unlikely with how distributed the access was, and then some minor stuff on completions for sensitive queries. This usage of Anthropic’s API will have a negligible impact on DeepSeek’s long-rumored V4 model (or whichever model the data here contributed to). This was also very likely a small team at DeepSeek and unknown to much of the broader training organization.”

12 more comments...

No posts

Ready for more?