Daniella Ye Profile Banner
Daniella Ye Profile
Daniella Ye

@Daniella_yz

Followers
357
Following
158
Media
4
Statuses
12

PhD NLP student @UniofOxford Prev: intern @Cohere

Oxford, England
Joined February 2021
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
@Daniella_yz
Daniella Ye
3 months
Beyond their use in assisting human evaluation (e.g. CriticGPT), can critiques directly enhance preference learning? During my @Cohere internship, we explored using synthetic critiques from large language models to improve reward models. 📑Preprint:
Tweet media one
5
57
322
@Daniella_yz
Daniella Ye
3 months
Scaling experiments show that high-quality critiques significantly enhance data efficiency, especially with limited data. 1 critique roughly equals 40 vanilla preference pairs. Our method, using open-source models, is accessible and budget-friendly.
Tweet media one
1
0
15
@Daniella_yz
Daniella Ye
3 months
@cohere Instead of solely relying on better/worse annotations, we enrich reward models with synthetic critiques from LLMs, dissecting completion features. We then train the model to predict scalar rewards, conditioned on these critiques.
Tweet media one
1
0
8
@Daniella_yz
Daniella Ye
3 months
Results show that adding critiques improves reward model accuracy. We found that critique quality matters: high-quality critiques boost performance, while low-quality ones can hinder it.
Tweet media one
1
0
8
@Daniella_yz
Daniella Ye
4 years
Thank you @stefan_fee , @JinlanFu , and Professor @gneubig for all your guidance and support! I'm interested in exploring the use of performance prediction in many more scenarios.
@stefan_fee
Pengfei Liu
4 years
How can we reliably estimate a system's performance without performing experiments? Check out our work (EACL) 1. Formulate performance prediction as Tensor Completion problem 2. Establish a set of reliability analysis mechanisms (confidence, calibration)
Tweet media one
3
6
41
0
0
5
@Daniella_yz
Daniella Ye
3 months
@FangruLin99 Thanks a lot Fangru!
0
0
2
@Daniella_yz
Daniella Ye
3 months
@Zhuang_Li_NLP Thank you for your comment! Yes, the RM is trained on instruction-response-critique triplets. We haven't done experiments to train LLM using the critiques yet, but in practice yes this means passing the critique context.
0
0
1