jackminong Profile Banner
Jackmin Profile
Jackmin

@jackminong

Followers
710
Following
6K
Statuses
341

πŸ‡²πŸ‡Ύ. Making GPUs go brr @PrimeIntellect πŸ‡ΊπŸ‡Έ. Previously @JinaAI_ πŸ‡©πŸ‡ͺ.

San Francisco, CA
Joined October 2021
Don't wanna be here? Send us removal request.
@jackminong
Jackmin
4 months
In order to support dynamically on and off-boarding compute to the run, we introduced a new distributed abstraction `ElasticDeviceMesh` which manages the resizing of process groups without requiring a cold restart. More info on this, along with some other interesting innovations we did to pull this off in our blog post:
@PrimeIntellect
Prime Intellect
4 months
Announcing INTELLECT-1: the first-ever decentralized training of a 10B model Scaling decentralized training 10x beyond prior efforts. Anyone can join us to build open-source AGI πŸ¦‹
4
9
57
@jackminong
Jackmin
21 hours
This seems like a good idea more people should be aware of. The choice of hyperparameters for the relative heights of the correct and wrong piece are interesting though. Is there an accuracy difference between Reward B and C? If not, does that mean you should always pick larger correct reward > wrong?
Tweet media one
@xiangyue96
Xiang Yue
4 days
Takeaway 4: Reward shaping can be used to stabilize and control CoT length while improving accuracy. We designed a reward function (Cosine Reward) to use CoT length as an additional input to stabilize emergent length scaling.
Tweet media one
0
0
4
@jackminong
Jackmin
22 hours
@nooriefyi you dont differentiate through the reward?
0
0
0
@jackminong
Jackmin
2 days
@tenderizzation sacril3ge
0
0
2
@jackminong
Jackmin
2 days
@Ar_Douillard has schmidhuber tried to claim your work yet?
1
0
5
@jackminong
Jackmin
2 days
@cloneofsimo @main_horse might be interesting for @wanchao_ to comment abit on this
1
0
1
@jackminong
Jackmin
3 days
@Ar_Douillard @kalomaze do they scale outside of google?
1
0
1
@jackminong
Jackmin
3 days
@tenderizzation this ham is pretty cunning
0
0
2
@jackminong
Jackmin
3 days
@sachoslks Yup! I think open source should allow for derivatives that are closed.
0
0
0
@jackminong
Jackmin
3 days
@afurgs bigger and better than before!
0
0
2
@jackminong
Jackmin
3 days
@bronzeagepapi We will be sure to inform you guys if it discovers any quantum mechanic glitches during the run 🫑
1
0
2
@jackminong
Jackmin
3 days
RT @MatternJustus: The path to our first reasoning model consists of three steps: 1. Generating cold-start reasoning data 2. SFT on cold-s…
0
11
0
@jackminong
Jackmin
4 days
@shxf0072 has there been any papers using GRPO for non LLM tasks? why should it only be great at optimizing LLM?
2
0
4
@jackminong
Jackmin
4 days
@teortaxesTex We need more memes! Memes allow people to say what they actually want to say without the full commitment of having said it explicitly
0
0
1
@jackminong
Jackmin
4 days
@manveerxyz lfg 🫑
0
0
2
@jackminong
Jackmin
5 days
RT @srush_nlp: Got talked into giving a DeepSeek talk this afternoon Not sure I have anything new to say here! But…
0
53
0
@jackminong
Jackmin
6 days
@CamutoDante i dont quite get what u mean here. intermediate activations are deterministic given the same input sequence and model. there isnt much a malicious or a benevolent actor can do to alter them
0
0
0
@jackminong
Jackmin
6 days
RT @huseinzol05: @mesolitica released Malaysian TTS models including dataset! Special thanks to @jackminong and @PrimeIntellect for the com…
0
3
0
@jackminong
Jackmin
7 days
0
0
2
@jackminong
Jackmin
7 days
@huseinzol05 tldr; hash activations good
1
0
1