Zorik Gekhman Profile
Zorik Gekhman

@zorikgekhman

Followers
242
Following
325
Statuses
132

PhD student @ Technion | Research intern @Google

Joined August 2019
Don't wanna be here? Send us removal request.
@zorikgekhman
Zorik Gekhman
9 months
Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? New preprint!๐Ÿ“ฃ - LLMs struggle to integrate new factual knowledge through fine-tuning - As the model eventually learns new knowledge, it becomes more prone to hallucinations๐Ÿ˜ตโ€๐Ÿ’ซ ๐Ÿ“œ ๐Ÿงต1/12๐Ÿ‘‡
@arankomatsuzaki
Aran Komatsuzaki
9 months
Google presents Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? Highlights the risk in introducing new factual knowledge through fine-tuning, which leads to hallucinations
Tweet media one
5
57
187
@zorikgekhman
Zorik Gekhman
2 days
Our work on abductive reasoning across multiple images has been accepted to #ICLR2025! Congrats, @mor_ventura95 !
@mor_ventura95
Mor Ventura
6 days
Weโ€™re excited to share that our paper, "NL-EYE: Abductive NLI for Images" has been accepted to ICLR 2025!๐Ÿฅณ Our benchmark reveals that VLMs struggle with abductive reasoning across multiple images (worse than random!) Think you can solve it? Check out these examples! Linkโฌ‡๏ธ
0
4
10
@zorikgekhman
Zorik Gekhman
2 days
Our work on the intrinsic representation of hallucinations in LLMs has been accepted to #ICLR2025! Congrats, @OrgadHadas
@OrgadHadas
Hadas Orgad
2 days
Our work "LLMs Know More Than They Show: On the Intrinsic Representation of LLM Hallucinations" was accepted to #ICLR2025! Looking forward to discussing our findings in person. Project page >> Code >>
0
2
21
@zorikgekhman
Zorik Gekhman
3 days
RT @NitCal: Our paper: "On Behalf of the Stakeholders: Trends in NLP Model Interpretability in the Era of LLMs" has been accepted to NAACLโ€ฆ
0
25
0
@zorikgekhman
Zorik Gekhman
14 days
Cool work by @NitCal. If you ever wonder if LLM-based labels work for your use case, now youโ€™ve got a (simple) way to find out.
@NitCal
Nitay Calderon
14 days
Do you use LLM-as-a-judge or LLM annotations in your research? Thereโ€™s a growing trend of replacing human annotators with LLMs in researchโ€”they're fast, cheap, and require less effort. But can we trust them?๐Ÿค” Well, we need a rigorous procedure to answer this. ๐ŸšจNew preprint๐Ÿ‘‡
Tweet media one
0
1
10
@zorikgekhman
Zorik Gekhman
2 months
RT @goldshtn: Today we published FACTS Grounding, a benchmark and leaderboard for evaluating the factuality of LLMs when grounding to the iโ€ฆ
0
8
0
@zorikgekhman
Zorik Gekhman
2 months
@_jasonwei Great work! I couldnโ€™t find the QA prompt used for evaluating the models in Table 3. Was it the same for all models? If possible, could you share it? Thanks!
0
0
1
@zorikgekhman
Zorik Gekhman
2 months
RT @YonatanBitton: ๐Ÿšจ Happening NOW at #NeurIPS2024 with @nitzanguetta ! ๐ŸŽญ #VisualRiddles: A Commonsense and World Knowledge Challenge for Vโ€ฆ
0
7
0
@zorikgekhman
Zorik Gekhman
3 months
Thank you @MilaNLProc for featuring our work!
@MilaNLProc
MilaNLP
3 months
For this week's @MilaNLProc reading group, Yujie presented "Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?" by @zorikgekhman et al. Paper: #NLProc
Tweet media one
0
0
11
@zorikgekhman
Zorik Gekhman
3 months
RT @roireichart: My students, collaborators and I have just published blog posts describing our work on NLP for the human sciences (pursuedโ€ฆ
0
4
0
@zorikgekhman
Zorik Gekhman
3 months
RT @NitCal: Do you think LLMs could win a Nobel Prize one day? ๐Ÿค” Can NLP predict heroin addiction outcomes, uncover suicide risks, or simuโ€ฆ
0
8
0
@zorikgekhman
Zorik Gekhman
3 months
At #EMNLP2024? Join me in the Language Modeling 1 session tomorrow, 11:00-11:15, for a talk on how fine-tuning with new knowledge impacts hallucinations.
@zorikgekhman
Zorik Gekhman
3 months
I'll be at #EMNLP2024 next week to give an oral presentation on our work about how fine-tuning with new knowledge affects hallucinations ๐Ÿ˜ตโ€๐Ÿ’ซ ๐Ÿ“… Nov 12 (Tue) 11:00-12:30, Language Modeling 1 Hope to see you there. If you're interested in factuality, letโ€™s talk!
0
5
14
@zorikgekhman
Zorik Gekhman
3 months
And now, you can also use our implementation to probe LLMs for truthfulness in your research
@OrgadHadas
Hadas Orgad
3 months
Our code for "LLMs Know More Than They Show: On the Intrinsic Representation of LLM Hallucinations" is now available! Utilize our implementation to probe the internal representations of LLMs and explore the insights found in our work. Check it out here:
0
0
1
@zorikgekhman
Zorik Gekhman
3 months
RT @mor_ventura95: ๐ŸŽ‰ Our paper "๐๐š๐ฏ๐ข๐ ๐š๐ญ๐ข๐ง๐  ๐‚๐ฎ๐ฅ๐ญ๐ฎ๐ซ๐š๐ฅ ๐‚๐ก๐š๐ฌ๐ฆ๐ฌ: ๐„๐ฑ๐ฉ๐ฅ๐จ๐ซ๐ข๐ง๐  ๐š๐ง๐ ๐”๐ง๐ฅ๐จ๐œ๐ค๐ข๐ง๐  ๐ญ๐ก๐ž ๐‚๐ฎ๐ฅ๐ญ๐ฎ๐ซ๐š๐ฅ ๐๐Ž๐• ๐จ๐Ÿ ๐“๐ž๐ฑ๐ญ-๐ญ๐จ-๐ˆ๐ฆ๐š๐ ๐ž ๐Œ๐จ๐๐ž๐ฅ๐ฌ" is accepted tโ€ฆ
0
12
0
@zorikgekhman
Zorik Gekhman
3 months
I'll be at #EMNLP2024 next week to give an oral presentation on our work about how fine-tuning with new knowledge affects hallucinations ๐Ÿ˜ตโ€๐Ÿ’ซ ๐Ÿ“… Nov 12 (Tue) 11:00-12:30, Language Modeling 1 Hope to see you there. If you're interested in factuality, letโ€™s talk!
@zorikgekhman
Zorik Gekhman
9 months
Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations? New preprint!๐Ÿ“ฃ - LLMs struggle to integrate new factual knowledge through fine-tuning - As the model eventually learns new knowledge, it becomes more prone to hallucinations๐Ÿ˜ตโ€๐Ÿ’ซ ๐Ÿ“œ ๐Ÿงต1/12๐Ÿ‘‡
0
6
16
@zorikgekhman
Zorik Gekhman
3 months
RT @alon_jacovi: "Is It Really Long Context if All You Need Is Retrieval? Towards Genuinely Difficult Long Context NLP" was accepted to EMNโ€ฆ
0
20
0
@zorikgekhman
Zorik Gekhman
3 months
RT @AdiSimhi: LLMs often "hallucinate". But not all hallucinations are the same! This paper reveals two distinct types: (1) due to lack ofโ€ฆ
0
187
0
@zorikgekhman
Zorik Gekhman
3 months
RT @itay__nakash: Breaking ReAct Agents: Foot-in-the-Door Attack ๐Ÿšช๐Ÿง‘โ€๐Ÿ’ปโš ๏ธ ๐Ÿšจ New preprint! ๐Ÿšจ FITD attack subtly misleads LLM agents ๐Ÿง . Our reโ€ฆ
0
12
0
@zorikgekhman
Zorik Gekhman
3 months
RT @omer6nahum: ๐Ÿš€ Excited to share our research: "Are LLMs Better than Reported? Detecting Label Errors and Mitigating Their Effect on Modeโ€ฆ
0
11
0
@zorikgekhman
Zorik Gekhman
4 months
Repeated model sampling boosts the chance of producing the right answer, which motivates scaling the inference compute. Yet this work shows that repeated guessing achieves the same, raising questions about whether the model is "right for the right reasons" in these improvements.
@OHonovich
Or Honovich
4 months
Scaling inference compute by repeated sampling boosts coverage (% problems solved), but could this be due to lucky guesses, rather than correct reasoning? We show that sometimes, guessing beats repeated sampling ๐ŸŽฒ @_galyo @omerlevy_ @roeeaharoni
Tweet media one
0
0
2
@zorikgekhman
Zorik Gekhman
4 months
RT @omarsar0: LLMs Know More Than They Show We know very little about how and why LLMs "hallucinate" but it's an important topic nonetheleโ€ฆ
0
134
0