
Server He
Add a review FollowOverview
-
Founded Date July 19, 1993
-
Sectors Accounting / Finance
-
Posted Jobs 0
-
Viewed 9
Company Description
Open-R1: a Totally Open Reproduction Of DeepSeek-R1
Hey there! This post is an introduction to the job, not a claim that we’ve replicated R1 yet. We’re constructing in the open, so as quickly as we have assessment numbers, we’ll share them. You can follow our development on Hugging Face and GitHub.
True, however it looks like there’s absolutely nothing to be examined as of today. I assume the ultimate goal is to train a brand-new thinking design and then utilize the very same examination metrics as o1 and the DeepSeek-R1.
Well, there need to be at least some peace of mind check and validation to guarantee the model was trained properly.
Oh yes, if you are speaking about the assessment variety of deepseek’s design it’s coming really soon!
As mentioned in the post there is no design called Open-R1 to evaluate at all … not yet anyway. This is a blog describing that Hugging face will take the R1 Deepseek design, exercise how it was developed as laid out in the paper and from what they launched, and then reproduce that .
in truth this is basically how science works … A develops a strategy, discovery or innovation and it is checked by B, C and D to see if it is reproduceable. Thats been the cornerstone of research now for a few centuries.
This blog site is not stating they have actually already done so … Its a blog site outlining an intent to begin training a model like R1 and calling it Open-R1.
Also DeepSeek-R1 was just released recently, and even in their paper they laid out the compute hours required. While those are low compute hours for a SOTA model this does not indicate you can train stated design in a week. I ‘d personally love to be able to train a transformer design in a week, but we might need to wait a while for that level of compute innovation.
So there are no standards for a model that has not been developed yet right? As described in the blog site, and once again in reply to your question.
However fear not, there is a GitHub Repo currently and factors (hell I may join myself), some prelim work done, and a strategy of attack. A great beginning position.
n
@edbeeching
has actually assessed the released designs currently
( src: https://x.com/edwardbeeching/status/1884273209136275742)
R1 simply trained on o1 outputs, so jointly …/ s. This is what the new AI czars are saying
Hi! This post is an introduction to the job, not a claim that we’ve reproduced R1 yet. We will completely share the missing piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo
That’s nice and crucial to understand this significant buzz that lacks technical understanding and explanation. Science has to do with recreation, and if they declare to be open, let them fullfill the open part.
Please do release the training expense.
We will!
Excalidraw Hi n
@bojan2501
thanks, we will indeed be working hard to make sure this training dish can work for small language designs on customer hardware given that not everybody has a cluster of H100s in your home:-RRB- The tool we utilized for the images was Excalidraw! https://excalidraw.com
eagerly anticipating it! WTF are your talking about?
need to be a joke
It’s really cool to see how the whole open source community comes together!
Ops …
5.5 M is number press reporter in the deepseekv3 tech report (simply the training, not the experiment afaik), for R1 hard to estimate tbh however much less than 5.5 M imo
Historically, they have never released code or datasets of their LLM training, so I wouldn’t expect this time to be various. If they would release it that would be fantastic of course!
Yes naturally!
So generally you’re asking to change existing censorship with another flavour of censorship?
The code for the models are inside the design repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py
Hello Team, I’m Ray Bernard, the author and developer of EQUATOR. My research study group will be working on a paper concentrated on reproducing particular parts of DeepSeek R1. Our goal is to recreate the cold start and offer your group with a dataset that consists of COT and other strategies to support these efforts. We like to contribute our work to assist. Please let me understand if you find this useful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/
Where is the assessment numbers? without it you can’t call it recreation.
8 replies
True, however it looks like there’s absolutely nothing to be assessed since today. I presume the ultimate goal is to train a new reasoning model and after that use the very same evaluation metrics as o1 and the DeepSeek-R1.
That’s quite fascinating, I was asking myself why the concerns the author exposed here are not being asked by others? I think the work they have actually done is unforgettable however at the very same time I question why they wouldn’t put these missing out on pieces on if they are expected to be totally open.
Why even without recreation and understanding of the innovation they could impact so much the market in this method?
4 replies
Hi! This blog site post is an intro to the task, not a claim that we have actually reproduced R1 yet. We will absolutely share the missing piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo
Interesting read, and it is excellent that we see more effort into this direction: more optimization and less brute force.
Also wonder what tool did the author usage for creating step diagram.
2 replies
Excalidraw I’m so glad that initiative like this already exist, I’m gon na attempt to contribute:-RRB- 1 reply
eagerly anticipating it! So racist articel
2 replies
WTF are your talking about?
Awesome to have this open reproduction began!
For Step # 1 check out https://github.com/open-thoughts/open-thoughts!
https://x.com/ryanmart3n/status/1884284101265612856
Let’s do this thing!
1 reply
It’s truly cool to see how the entire open source community comes together!
Does anyone understand the real training cost of r1? I can’t find it in the paper or the announcement post. Is the 6M cost reported by media just the number drawn from v3’s training cost?
2 replies
Ops …
Has anyone asked the DeepSeek group to publish their training information and code, or a minimum of share them privately with an independent replication task like this? Have they declined such a demand?
A devoted duplication depends upon utilizing the exact same dataset and hyperparameters. Otherwise, any significant inconsistencies with the published benchmarks would be tough to pin down-whether due to training data differences or the replication approach itself.
1 reply
Historically, they have never ever released code or datasets of their LLM training, so I would not anticipate this time to be various. If they would launch it that would be remarkable obviously!
In the meantime we have to make best guess estimates and see if we can get there ourselves.
You provide excellent replication process of Deepseek reasoning training. I will attempt something comparable to it.
This is really excellent info, can we tweak with particular use case when code is released?
1 reply
Yes naturally!
Please think about getting rid of prejudiced, polluted or unaligned training information and make an effort to eliminate copyrighted works from the crawl from intake. This will make the design more functional. If you reused anthropic curation checks, this might likewise help, get rid of obviouslybiased data will likely include a great deal of worth. We do not desire another polluted, unaligned open source design, right? And no corporate would ever use deepseek or a model that recycles it, right?
We value your work for the advantage of humankind, we hope.
Miike C from NJ
1 reply
So essentially you’re asking to replace existing censorship with another flavour of censorship?
Can’t wait! Hopefully the design will be uncensored but whatever you can do is alright! Love seeing open source structure itself up. I’m not smart enough to in fact assist however I can contribute support lol
Hello guys, I am even simply looking for code for DeepSeek-V2, in order to totally comprehend multi-head hidden attention. You do not appear to have code in Hugging Face even for that. Or am I missing out on something? Don’t see anything in src/transformers/models. MLA is not effectively described in their paper, so it would be very important to have code for this.