Amycherryphoto

Overview

  • Founded Date December 25, 1929
  • Posted Jobs 0
  • Viewed 28

Company Description

Open-R1: a Fully Open Reproduction Of DeepSeek-R1

Hey there! This article is an introduction to the project, not a claim that we’ve replicated R1 yet. We’re integrating in the open, so as quickly as we have examination numbers, we’ll share them. You can follow our development on Hugging Face and GitHub.

True, but it appears like there’s absolutely nothing to be evaluated as of today. I assume the supreme goal is to train a new reasoning design and after that use the very same assessment metrics as o1 and the DeepSeek-R1.

Well, there should be at least some sanity check and recognition to make sure the design was trained correctly.

Oh yes, if you are talking about the examination number of deepseek’s design it’s coming very quickly!

As discussed in the post there is no model called Open-R1 to test at all … not yet anyhow. This is a blog detailing that Hugging face will take the R1 Deepseek design, exercise how it was constructed as detailed in the paper and from what they launched, and after that reproduce that process.

in truth this is pretty much how science works … A develops a plan, discovery or innovation and it is tested by B, C and D to see if it is reproduceable. Thats been the foundation of research now for a couple of centuries.

This blog is not stating they have actually already done so … Its a blog site describing an intent to begin training a model like R1 and calling it Open-R1.

Also DeepSeek-R1 was only launched last week, and even in their paper they described the calculate hours needed. While those are low calculate hours for a SOTA model this does not suggest you can train stated model in a week. I ‘d personally enjoy to be able to train a transformer design in a week, however we might need to wait a while for that level of compute innovation.

So there are no benchmarks for a design that has not been built yet right? As detailed in the blog, and once again in reply to your question.

However fear not, there is a GitHub Repo currently and contributors (hell I might join myself), some prelim work done, and a master plan. An excellent starting position.

n
@edbeeching
has actually evaluated the launched models already

( src: https://x.com/edwardbeeching/status/1884273209136275742)

R1 simply trained on o1 outputs, so jointly …/ s. This is what the new AI czars are stating

Hi! This blog site post is an intro to the job, not a claim that we have actually reproduced R1 yet. We will totally share the missing piece when we have them, you can expect the designs and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

That’s great and essential to understand this significant hype that does not have technical understanding and description. Science is about recreation, and if they claim to be open, let them fullfill the open part.

Please do release the training expense.

We will!

Excalidraw Hi n
@bojan2501
thanks, we will certainly be working hard to make certain this training recipe can work for little language designs on consumer hardware because not everybody has a cluster of H100s in the house:-RRB- The tool we used for the images was Excalidraw! https://excalidraw.com

eagerly anticipating it! WTF are your discussing?

need to be a joke

It’s truly cool to see how the entire open source community comes together!

Ops …

5.5 M is number reporter in the deepseekv3 tech report (just the training, not the experiment afaik), for R1 difficult to approximate tbh however much less than 5.5 M imo

Historically, they have actually never ever released code or datasets of their LLM training, so I would not anticipate this time to be different. If they would release it that would be remarkable naturally!

Yes of course!

So basically you’re asking to replace existing censorship with another flavour of censorship?

The code for the models are inside the model repositories, e.g. for V3: https://huggingface.co/deepseek-ai/DeepSeek-V3/blob/main/modeling_deepseek.py

Hello Team, I’m Ray Bernard, the author and developer of EQUATOR. My research team will be working on a paper concentrated on reproducing particular elements of DeepSeek R1. Our objective is to reproduce the cold start and supply your group with a dataset that consists of COT and other strategies to support these efforts. We like to contribute our work to help. Please let me know if you find this useful. Best, Ray Bernard https://www.facebook.com/groups/1186310571520299/

Where is the examination numbers? without it you can’t call it reproduction.

8 replies

True, but it appears like there’s nothing to be evaluated since today. I assume the supreme objective is to train a brand-new reasoning design and then use the exact same examination metrics as o1 and the DeepSeek-R1.

That’s rather intriguing, I was asking myself why the questions the author exposed here are not being asked by others? I believe the work they have done is unforgettable but at the exact same time I wonder why they wouldn’t put these missing pieces on if they are expected to be fully open.
Why even without recreation and understanding of the development they could impact so much the marketplace in this method?

4 replies

Hi! This post is an intro to the project, not a claim that we have actually recreated R1 yet. We will completely share the missing out on piece when we have them, you can expect the models and datasets to be upload in this Hugging Face org and the code to be in this GitHub repo

Interesting read, and it is excellent that we see more effort into this instructions: more optimization and less strength.
Also wonder what tool did the author use for producing action diagram.

2 replies

Excalidraw I’m so grateful that initiative like this currently exist, I’m gon na try to contribute:-RRB- 1 reply

eagerly anticipating it! So racist articel

2 replies

WTF are your speaking about?

Awesome to have this open reproduction began!

For Step # 1 check out https://github.com/open-thoughts/!

https://x.com/ryanmart3n/status/1884284101265612856

Let’s do this thing!

1 reply

It’s really cool to see how the whole open source neighborhood comes together!

Does anyone know the actual training cost of r1? I can’t discover it in the paper or the announcement post. Is the 6M expense reported by media simply the number taken from v3’s training cost?

2 replies

Ops …

Has anybody asked the DeepSeek team to release their training information and code, or at least share them privately with an independent replication job like this? Have they rejected such a demand?

A devoted duplication depends upon utilizing the very same dataset and hyperparameters. Otherwise, any significant inconsistencies with the published standards would be tough to pin down-whether due to training information differences or the replication technique itself.

1 reply

Historically, they have never ever released code or datasets of their LLM training, so I wouldn’t expect this time to be different. If they would release it that would be remarkable of course!

In the meantime we need to make finest guess price quotes and see if we can get there ourselves.

You supply great duplication procedure of Deepseek reasoning training. I will try something comparable to it.

This is truly good information, can we tweak with specific use case when code is launched?

1 reply

Yes obviously!

Please think about removing prejudiced, polluted or unaligned training data and make an effort to remove copyrighted works from the crawl from consumption. This will make the design more usable. If you reused anthropic curation checks, this might also assist, eliminate obviouslybiased data will likely include a great deal of worth. We do not want another tainted, unaligned open source design, right? And no corporate would ever use deepseek or a design that recycles it, right?
We appreciate your work for the benefit of humanity, we hope.
Miike C from NJ

1 reply

So essentially you’re asking to replace existing censorship with another flavour of censorship?

Can’t wait! Hopefully the model will be uncensored but whatever you can do is alright! Love seeing open source building itself up. I’m not smart adequate to actually assist however I can contribute ethical support lol

Hello guys, I am even just searching for code for DeepSeek-V2, in order to totally comprehend multi-head latent attention. You do not seem to have code in Hugging Face even for that. Or am I missing out on something? Don’t see anything in src/transformers/models. MLA is not effectively explained in their paper, so it would be essential to have code for this.