ShiyangLi6 Profile Banner
Shiyang Li Profile
Shiyang Li

@ShiyangLi6

Followers
48
Following
81
Statuses
9

Ph.D in Computer Science & Working on Large Language Models.

California, USA
Joined December 2018
Don't wanna be here? Send us removal request.
@ShiyangLi6
Shiyang Li
2 years
Thanks for sharing our work!
@_akhaliq
AK
2 years
Instruction-following Evaluation through Verbalizer Manipulation paper page: While instruction-tuned models have shown remarkable success in various natural language processing tasks, accurately evaluating their ability to follow instructions remains challenging. Existing benchmarks primarily focus on common instructions that align well with what the model learned during training. However, proficiency in responding to these instructions does not necessarily imply strong ability in instruction following. In this paper, we propose a novel instruction-following evaluation protocol called verbalizer manipulation. It instructs the model to verbalize the task label with words aligning with model priors to different extents, adopting verbalizers from highly aligned (e.g., outputting ``postive'' for positive sentiment), to minimally aligned (e.g., outputting ``negative'' for positive sentiment). Verbalizer manipulation can be seamlessly integrated with any classification benchmark to examine the model's reliance on priors and its ability to override them to accurately follow the instructions. We conduct a comprehensive evaluation of four major model families across nine datasets, employing twelve sets of verbalizers for each of them. We observe that the instruction-following abilities of models, across different families and scales, are significantly distinguished by their performance on less natural verbalizers. Even the strongest GPT-4 model struggles to perform better than random guessing on the most challenging verbalizer, emphasizing the need for continued advancements to improve their instruction-following abilities.
Tweet media one
0
0
1
@ShiyangLi6
Shiyang Li
2 years
Thanks for sharing our work!
@arankomatsuzaki
Aran Komatsuzaki
2 years
AlpaGasus: Training A Better Alpaca with Fewer Data Significantly outperforms the original Alpaca and reaches 90% of davinci-003 w/ 5.7x faster training. proj: abs:
Tweet media one
0
0
9
@ShiyangLi6
Shiyang Li
2 years
RT @arankomatsuzaki: AlpaGasus: Training A Better Alpaca with Fewer Data Significantly outperforms the original Alpaca and reaches 90% of…
0
30
0
@ShiyangLi6
Shiyang Li
2 years
RT @_akhaliq: AlpaGasus: Training A Better Alpaca with Fewer Data paper page: Large language models~(LLMs) obtain…
0
26
0
@ShiyangLi6
Shiyang Li
5 years
@amitness @Thom_Wolf @colinraffel @PatrickPlaten @GoogleAI Hi Amit, do you find anything to finish this? I didn't find it. I want to do something like a circle translation, en-->fr-->en
1
0
1
@ShiyangLi6
Shiyang Li
6 years
RT @ValaAfshar: Family first Keep in touch with friends You are not your job Do not respond to negativity Fight against entitlement Be hone…
0
154
0
@ShiyangLi6
Shiyang Li
6 years
RT @jenheemstra: As academics, we’re surrounded by people, yet our jobs can be extremely isolating. Put people in your life with whom you c…
0
238
0
@ShiyangLi6
Shiyang Li
6 years
RT @OriolVinyalsML: "Failing" is too often portrayed as something negative. Failures, and not successes, have been a higher drive for me bo…
0
87
0
@ShiyangLi6
Shiyang Li
6 years
Here’s my year in emoji: #HappyNewYear #MyEmojiYear ☺️
Tweet media one
0
0
1