Tu Vu Profile
Tu Vu

@tuvllms

Followers
4K
Following
3K
Statuses
1K

Research Scientist @GoogleDeepMind & Assistant Professor @VT_CS. PhD from @UMass_NLP. Google FLAMe/FreshLLMs/Flan-T5 Collection/SPoT #NLProc

California, USA
Joined April 2017
Don't wanna be here? Send us removal request.
@tuvllms
Tu Vu
7 months
šŸšØ New @GoogleDeepMind paper šŸšØ We trained Foundational Large Autorater Models (FLAMe) on extensive human evaluations, achieving the best RewardBench perf. among generative models trained solely on permissive data, surpassing both GPT-4 & 4o. šŸ“°: šŸ§µ:šŸ‘‡
Tweet media one
Tweet media two
26
96
562
@tuvllms
Tu Vu
3 days
RT @lmarena_ai: News: @GoogleDeepMind Gemini-2.0 family (Pro, Flash, and Flash-lite) is now live in Arena! - Gemini-2.0-Pro takes #1 spotā€¦
0
171
0
@tuvllms
Tu Vu
3 days
RT @sundarpichai: 1/ New Gemini 2.0 updates, here we go! Gemini 2.0 Flash is now GA, so devs can now build production applications. Findā€¦
0
535
0
@tuvllms
Tu Vu
3 days
RT @jacobaustin132: Making LLMs run efficiently can feel scary, but scaling isnā€™t magic, itā€™s math! We wanted to demystify the ā€œsystems vieā€¦
0
357
0
@tuvllms
Tu Vu
3 days
RT @Muennighoff: DeepSeek r1 is exciting but misses OpenAIā€™s test-time scaling plot and needs lots of data. We introduce s1 reproducing o1ā€¦
0
179
0
@tuvllms
Tu Vu
3 days
RT @srush_nlp: What to know about DeepSeek In which we aim to understand MoE, o1, scaling, tech reporting, moderā€¦
0
105
0
@tuvllms
Tu Vu
9 days
RT @WenhuChen: Everyone is talking about RL these days. But are we done with SFT? The answer is NO. If we revive SFT in another form, it caā€¦
0
100
0
@tuvllms
Tu Vu
9 days
RT @karpathy: We have to take the LLMs to school. When you open any textbook, you'll see three major types of information: 1. Backgroundā€¦
0
2K
0
@tuvllms
Tu Vu
9 days
RT @allen_ai: Here is TĆ¼lu 3 405B šŸ« our open-source post-training model that surpasses the performance of DeepSeek-V3! The last member of tā€¦
0
390
0
@tuvllms
Tu Vu
10 days
RT @Azaliamirh: We are releasing CodeMonkeys, a system for solving SWE-bench problems with a focus on careful parallel and serial scaling oā€¦
0
30
0
@tuvllms
Tu Vu
10 days
RT @wzihanw: šŸš€ Introducing RAGENā€”the worldā€™s first reproduction of DeepSeek-R1(-Zero) methods for training agentic AI models! Weā€™re bettinā€¦
0
249
0
@tuvllms
Tu Vu
11 days
RT @arankomatsuzaki: SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training Shows that: - RL generalizes inā€¦
0
151
0
@tuvllms
Tu Vu
11 days
RT @JayAlammar: The Illustrated DeepSeek-R1 Spent the weekend reading the paper and sorting through the intuitions. Here's a visual guideā€¦
0
228
0
@tuvllms
Tu Vu
11 days
RT @mar_kar_: Microsoft Translator group is looking for a Ph.D. student #intern this summer to work with us in Redmond on machine translatiā€¦
0
14
0
@tuvllms
Tu Vu
11 days
RT @jennajrussell: People often claim they know when ChatGPT wrote something, but are they as accurate as they think? Turns out that whileā€¦
0
148
0
@tuvllms
Tu Vu
14 days
RT @junxian_he: We replicated the DeepSeek-R1-Zero and DeepSeek-R1 training on 7B model with only 8K examples, the results are surprisinglyā€¦
0
668
0
@tuvllms
Tu Vu
15 days
RT @ClementDelangue: Our science team has started working on fully reproducing and open-sourcing R1 including training data, training scripā€¦
0
549
0
@tuvllms
Tu Vu
16 days
RT @lmarena_ai: Breaking News: DeepSeek-R1 surges to the top-3 in ArenašŸ³! Now ranked #3 Overall, matching the top reasoning model, o1, whiā€¦
0
386
0
@tuvllms
Tu Vu
16 days
RT @DanHendrycks: Weā€™re releasing Humanityā€™s Last Exam, a dataset with 3,000 questions developed with hundreds of subject matter experts toā€¦
0
802
0
@tuvllms
Tu Vu
17 days
@profnaren Congratulations, Naren!! šŸŽ‰
0
0
2
@tuvllms
Tu Vu
17 days
RT @lmarena_ai: Breaking news from Text-to-Image Arena! šŸ–¼ļøāœØ @GoogleDeepMindā€™s Imagen 3 debuts at #1, surpassing Recraft-v3 with a remarkabā€¦
0
134
0