
Internetagentur Aus Hamburg
Add a review FollowOverview
-
Founded Date 18 September 1980
-
Sectors Accounting
-
Posted Jobs 0
-
Viewed 41
Company Description
Open-R1: a Fully Open Reproduction Of DeepSeek-R1
Hey there! This blog site post is an intro to the job, not a claim that we have actually recreated R1 yet. We’re integrating in the open, so as soon as we have examination numbers, we’ll share them. You can follow our development on Hugging Face and GitHub.
True, however it looks like there’s absolutely nothing to be evaluated since right now. I assume the supreme goal is to train a new thinking model and then utilize the exact same examination metrics as o1 and the DeepSeek-R1.
Well, there ought to be at least some sanity check and validation to guarantee the design was trained correctly.
Oh yes, if you are talking about the evaluation number of deepseek’s model it’s coming extremely quickly!
As discussed in the post there is no model called Open-R1 to test at all … not yet anyhow. This is a blog site detailing that Hugging face will take the R1 Deepseek design, exercise how it was developed as detailed in the paper and from what they released, and after that reproduce that process.
in reality this is quite much how science works … A creates a strategy, discovery or innovation and it is checked by B, C and D to see if it is reproduceable. Thats been the cornerstone of research study now for a couple of centuries.
This blog is not stating they have already done so … Its a blog site detailing an intent to start training a model like R1 and calling it Open-R1.
Also DeepSeek-R1 was only launched recently, and even in their paper they described the compute hours needed. While those are low compute hours for a SOTA model this does not indicate you can train stated model in a week. I ‘d personally like to be able to train a transformer design in a week, however we might need to wait a while for that level of calculate technology.
So there are no for a design that has not been built yet right? As laid out in the blog site, and once again in reply to your concern.
However fear not, there is a GitHub Repo already and factors (hell I may join myself), some prelim work done, and a strategy of attack. A good beginning position.
n
@edbeeching
has actually assessed the launched models already
( src: https://x.com/edwardbeeching/status/1884273209136275742)
R1 simply trained on o1 outputs, so jointly …/ s. This is what the new AI czars are stating
Hi! This blog post is an introduction to the job, not a claim that we have actually replicated R1 yet. We will totally share the missing out on piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo
That’s great and crucial to understand this tremendous buzz that does not have technical comprehension and description. Science is about recreation, and if they claim to be open, let them fullfill the open part.
Please do publish the training cost.
We will!
Excalidraw Hi n
@bojan2501
thanks, we will indeed be striving to make sure this training recipe can work for small language models on customer hardware because not everybody has a cluster of H100s in the house:-RRB- The tool we utilized for the images was Excalidraw! https://excalidraw.com
anticipating it! WTF are your talking about?
need to be a joke
It’s actually cool to see how the whole open source community comes together!
Ops …
5.5 M is number press reporter in the deepseekv3 tech report (just the training, not the experiment afaik), for R1 difficult to estimate tbh however much less than 5.5 M imo
Historically, they have never launched code or datasets of their LLM training, so I would not anticipate this time to be various. If they would launch it that would be amazing obviously!
Yes obviously!
So basically you’re asking to change existing censorship with another flavour of censorship?
The code for the designs are inside the design repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py
Hello Team, I’m Ray Bernard, the author and developer of EQUATOR. My research study group will be dealing with a paper focused on reproducing certain elements of DeepSeek R1. Our objective is to reproduce the cold start and provide your group with a dataset that includes COT and other methods to support these efforts. We like to contribute our work to assist. Please let me know if you find this useful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/
Where is the examination numbers? without it you can’t call it reproduction.
8 replies
True, however it appears like there’s nothing to be examined since today. I assume the supreme goal is to train a new thinking design and after that use the exact same evaluation metrics as o1 and the DeepSeek-R1.
That’s quite fascinating, I was asking myself why the questions the author exposed here are not being asked by others? I believe the work they have actually done is remarkable but at the exact same time I question why they would not put these missing out on pieces on if they are supposed to be fully open.
Why even without reproduction and understanding of the development they could affect a lot the market in this method?
4 replies
Hi! This blog post is an intro to the task, not a claim that we’ve reproduced R1 yet. We will totally share the missing piece when we have them, you can anticipate the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo
Interesting read, and it is great that we see more effort into this direction: more optimization and less strength.
Also question what tool did the author use for creating step diagram.
2 replies
Excalidraw I’m so pleased that effort like this currently exist, I’m gon na try to contribute:-RRB- 1 reply
looking forward to it! So racist articel
2 replies
WTF are your speaking about?
Awesome to have this open reproduction began!
For Step # 1 check out https://github.com/open-thoughts/open-thoughts!
https://x.com/ryanmart3n/status/1884284101265612856
Let’s do this thing!
1 reply
It’s actually cool to see how the entire open source neighborhood comes together!
Does anybody know the real training cost of r1? I can’t discover it in the paper or the announcement post. Is the 6M cost reported by media simply the number taken from v3’s training cost?
2 replies
Ops …
Has anyone asked the DeepSeek team to publish their training data and code, or at least share them privately with an independent duplication project like this? Have they rejected such a request?
A faithful replication depends on using the same dataset and hyperparameters. Otherwise, any major discrepancies with the released standards would be hard to pin down-whether due to training information distinctions or the duplication approach itself.
1 reply
Historically, they have actually never launched code or datasets of their LLM training, so I wouldn’t expect this time to be different. If they would launch it that would be remarkable naturally!
In the meantime we need to make best guess estimates and see if we can arrive ourselves.
You offer excellent duplication procedure of Deepseek reasoning training. I will attempt something similar to it.
This is actually great info, can we fine tune with particular usage case when code is released?
1 reply
Yes of course!
Please consider eliminating biased, polluted or unaligned training data and make an effort to remove copyrighted works from the crawl from consumption. This will make the design more functional. If you reused anthropic curation checks, this might also help, eliminate obviouslybiased data will likely add a lot of worth. We don’t want another tainted, unaligned open source design, right? And no corporate would ever use deepseek or a model that recycles it, right?
We appreciate your work for the benefit of humankind, we hope.
Miike C from NJ
1 reply
So generally you’re asking to replace existing censorship with another flavour of censorship?
Can’t wait! Hopefully the design will be uncensored however whatever you can do is alright! Love seeing open source structure itself up. I’m not wise sufficient to in fact assist but I can contribute support lol
Hello guys, I am even simply searching for code for DeepSeek-V2, in order to completely understand multi-head hidden attention. You do not appear to have code in Hugging Face even for that. Or am I missing out on something? Don’t see anything in src/transformers/models. MLA is not properly described in their paper, so it would be essential to have code for this.