kexun_zhang Profile Banner
Kexun Zhang Profile
Kexun Zhang

@kexun_zhang

Followers
1K
Following
2K
Statuses
528

PhD student at @LTIatCMU. Previously at @ucsbNLP, @ZJU_china. language lover.

Joined December 2021
Don't wanna be here? Send us removal request.
@kexun_zhang
Kexun Zhang
3 months
Everyone talks about scaling inference compute after o1. But how exactly should we do that? We studied compute allocation for sampling -- a basic operation in most LLM meta-generators, and found that optimized allocation can save as much as 128x compute!
Tweet media one
Tweet media two
Tweet media three
6
22
105
@kexun_zhang
Kexun Zhang
2 days
The most challenging part of o1/r1 replications is that people will soon find there are only 26 letters in the alphabet. I'm pretty sure we'll start to see code-switching names like 阿1 or ℵ1.
1
0
11
@kexun_zhang
Kexun Zhang
3 days
RT @isdownapp: 🚨 Users are reporting problems with Docker Hub Registry. Is Docker Hub Registry down for you? RT if you are having issues. h…
0
14
0
@kexun_zhang
Kexun Zhang
3 days
@teortaxesTex do you think it's still gonna be a single model generation just controlled with a different decoding algorithm, or it's a meta generation strategy like best-of-n?
1
0
2
@kexun_zhang
Kexun Zhang
3 days
RT @xiangyue96: Demystifying Long CoT Reasoning in LLMs Reasoning models like R1 / O1 / O3 have gained massive atte…
0
189
0
@kexun_zhang
Kexun Zhang
3 days
@ZeyuanAllenZhu @DOGE @dogecoin You should submit this to sigbovik
1
0
2
@kexun_zhang
Kexun Zhang
3 days
太抽🐘了
0
0
2
@kexun_zhang
Kexun Zhang
3 days
Who’s getting fired? 🤪
@markchen90
Mark Chen
3 days
Leadership is forged through fire.
0
0
2
@kexun_zhang
Kexun Zhang
4 days
@dhadfieldmenell how about formal verification as verifiable rewards? Is that considered a perfect reward?
1
0
2
@kexun_zhang
Kexun Zhang
4 days
@DongfuJiang off the top of my head: the fine-tuning part of code llama: I think there are many more.
1
1
5
@kexun_zhang
Kexun Zhang
5 days
@Muennighoff @hahahahohohe look how their idea is the same as yours.
1
0
4
@kexun_zhang
Kexun Zhang
7 days
RT @Yoshua_Bengio: A few reflections I had while watching this interview featuring @geoffreyhinton: It does not (or should not) really mat…
0
8
0
@kexun_zhang
Kexun Zhang
10 days
@hahahahohohe Interesting! I would really love to see an ACC vs cot length plot
0
0
1
@kexun_zhang
Kexun Zhang
10 days
Thanks David, now is there a way I can run openai o1 without sharing any data with an American company? 😅
@DavidSacks
David Sacks
12 days
This is one of several ways that you can try DeepSeek R1 without downloading the app or sharing any data with a Chinese company.
0
3
15
@kexun_zhang
Kexun Zhang
10 days
@anton_iades R1’s ability to write in all sorts of styles is way too overlooked. The story I heard is they hired lots of students from the top humanities programs in china to help with their data.
1
0
4
@kexun_zhang
Kexun Zhang
12 days
RT @lateinteraction: There are four types of research problems involving "natural language processing" that I find really fascinating. The…
0
8
0
@kexun_zhang
Kexun Zhang
12 days
RT @yuxiangw_cs: It's interesting to see people entering panic mode over #DeepSeekR1 and going bearish in computing. It's like going bearis…
0
5
0
@kexun_zhang
Kexun Zhang
13 days
RT @pthangeda_: @DanHendrycks @Miles_Brundage @dwarkesh_sp @polynoamial has been publicly saying that in his talks. I don’t know what can b…
0
1
0
@kexun_zhang
Kexun Zhang
14 days
this↓
@lateinteraction
Omar Khattab
14 days
@DimitrisPapail @rm_rafailov But if it's about the model, not the algorithm/process, that conversation/ablation would be very illuminating. Someone should just re-do basic STaR on a few different recent base/instruct models and show what happens.
0
0
3