Yusen Zhang Profile
Yusen Zhang

@YusenZhangNLP

Followers
330
Following
217
Statuses
79

PhD Candidate @PennStateEECS | NLP Lab @NLP_PennState #NLProc | Prev Research Intern @MSFTResearch, @AmazonScience @GoogleAI

State College, PA
Joined November 2022
Don't wanna be here? Send us removal request.
@YusenZhangNLP
Yusen Zhang
5 months
Chain-of-Agents has been accepted by NeurIPS 2024! 📣We propose Chain-of-Agents, a training-free, task-agnostic, highly interpretable framework for Long Context. CoA improves significantly over RAG/Long LLM/Multi-Agent. Link:
@YusenZhangNLP
Yusen Zhang
8 months
Chain-of-Agents (CoA) consists of multiple worker agents who sequentially communicate to handle different segmented portions of the text, followed by a manager agent who synthesizes these contributions into a coherent final output.
Tweet media one
3
24
124
@YusenZhangNLP
Yusen Zhang
7 days
Very interesting founding! More interestingly, in our recent paper: we found a similar behavior termed "Verbosity Compensation" that the recall of verbose - concise > 20%! To explain, we found this behavior is related to the model uncertainty.
@AlexGDimakis
Alex Dimakis
8 days
Discovered a very interesting thing about DeepSeek-R1 and all reasoning models: The wrong answers are much longer while the correct answers are much shorter. Even on the same question, when we re-run the model, it sometimes produces a short (usually correct) answer or a wrong verbose one. Based on this, I'd like to propose a simple idea called Laconic decoding: Run the model 5 times (in parallel) and pick the answer with the smallest number of tokens. Our preliminary results show that this decoding gives +6-7% on AIME24 with only a few parallel runs. I think this is better (and faster) than consensus decoding.
Tweet media one
0
1
6
@YusenZhangNLP
Yusen Zhang
7 days
@AlexGDimakis Very interesting founding! Interestingly, we found a similar behavior in our recent paper: We found that the recall of verbose responses can be more than 20 percent lower than that of concise responses! And this is connected to the uncertain of the model.
0
0
1
@YusenZhangNLP
Yusen Zhang
8 days
RT @vipul_1011: Wohhoo, our work got accepted at NAACL 2025! 🥳 TLDR: we built a methodology to improve reliability of any dataset by selec…
0
10
0
@YusenZhangNLP
Yusen Zhang
8 days
RT @GoogleAI: Large language models are often limited by restrictions on the length of their inputs. To address this, we propose Chain-of-A…
0
195
0
@YusenZhangNLP
Yusen Zhang
11 days
RT @NanZhangNLP: Thanks @SFResearch for sharing our SiReRAG paper ( accepted by #ICLR2025! We proposed a RAG indexi…
0
6
0
@YusenZhangNLP
Yusen Zhang
2 months
RT @deedydas: OpenAI o3 is 2727 on Codeforces which is equivalent to the #175 best human competitive coder on the planet. This is an absol…
0
727
0
@YusenZhangNLP
Yusen Zhang
2 months
RT @RyoKamoi: 📢 New preprint! Do LVLMs have strong visual perception capabilities? Not quite yet... We introduce VisOnlyQA, a new dataset…
0
19
0
@YusenZhangNLP
Yusen Zhang
2 months
RT @RyoKamoi: Curious about LLM self-correction? Check out our reading list! 📚 We feature papers & blogs in * Key…
0
28
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @xiye_nlp: 🔔 I'm recruiting multiple fully funded MSc/PhD students @UAlberta for Fall 2025! Join my lab working on NLP, especially reaso…
0
162
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @vipul_1011: The fan-boy in me was really happy (and a bit nervous) while organizing the talk by @srush_nlp and having discussions with…
0
4
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @momergul_: This still feels very surreal! I would like to thank @emnlpmeeting for this great honor, @yoavartzi and my labmates for all…
0
9
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @vipul_1011: The code and SMART-Filtered datasets are now open-sourced! 🚀✨ 🔗Code: 🤗SMART-Filtered datasets on…
0
11
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @RyoKamoi: We will present our survey on self-correction of LLMs (TACL) at #EMNLP2024 in person! Oral: Nov 12 (Tue) 11:00- (Language Mo…
0
11
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @PennStateEECS: NSF CAREER award: Centering people while advancing artificial intelligence Computer science researcher Rui Zhang to buil…
0
11
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @Reza0843: 🚨New Paper Alert! As AI/ML researchers, we often juggle "brain-intensive" research tasks in our daily work 🤯, like crafting…
0
40
0
@YusenZhangNLP
Yusen Zhang
3 months
RT @vipul_1011: 🚨 New paper alert 🚨 Ever struggled with quick saturation or unreliability in benchmark datasets? Introducing SMART Filteri…
0
13
0
@YusenZhangNLP
Yusen Zhang
4 months
I will be there as well! Welcome to join and talk.
@RyoKamoi
Ryo Kamoi
4 months
Our poster presentation @COLM_conf is today 11am-1pm (Poster #50)! Just outside the theater. Come to our poster if you are interested in LLMs safety, self-correction, evaluation, or any related topics! Evaluating LLMs at Detecting Errors in LLM Responses
0
1
4
@YusenZhangNLP
Yusen Zhang
4 months
RT @RyoKamoi: I'll be at @COLM_conf and present our work on Tuesday 11am-1pm (Poster #50)! Looking forward to chatting about LLM safety, se…
0
8
0
@YusenZhangNLP
Yusen Zhang
4 months
RT @MingZhong_: Excited to share our recent work! We define and benchmark cross capabilities in LLMs, revealing the "Law of the Weakest Lin…
0
19
0
@YusenZhangNLP
Yusen Zhang
4 months
@adridder Thanks, Alexander!
0
0
0