Chulin Xie Profile Banner
Chulin Xie Profile
Chulin Xie

@ChulinXie

Followers
667
Following
687
Media
6
Statuses
53

CS PhD student at UIUC and student researcher @GoogleAI ; Ex research intern @MSFTResearch @NvidiaAI

Joined March 2019
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@ChulinXie
Chulin Xie
5 months
Some text data is private & cannot be shared... Can we generate synthetic replicas with privacy guarantees?🤔 Instead of DP-SGD finetuning, use Aug-PE with inference APIs! Compatible with strong LLMs (GPT-3.5, Mistral), where DP-SGD is infeasible. 🔗 [1/n]
Tweet media one
4
25
88
@ChulinXie
Chulin Xie
6 months
Excited to see the release of the book 🥳 and grateful for the opportunity to contribute a chapter. Big thanks to the three editors! @pinyuchenTW @LamMNguyen3 @nghiaht87
@pinyuchenTW
Pin-Yu Chen
6 months
Happy to share the release of the book "Federated Learning: Theory and Practice" that I co-edited with @LamMNguyen3 @nghiaht87 , covering fundamentals, emerging topics, and applications. Kudos to the amazing contributors to make this book happen! @ElsevierNews @sciencedirect
Tweet media one
Tweet media two
2
10
61
1
2
30
@ChulinXie
Chulin Xie
7 months
Our benchmark is on @huggingface leaderboard!!
@uiuc_aisecure
Bo Li
7 months
Super excited to set up the LLM safety & trustworthiness leaderboard on Huggingface, and we will keep adding new safety perspectives. Here we evaluate (open & close) LLMs and compressed LLMs. Looking forward to more exciting evaluations to assess and enhance LLM safety!!! 🥳
1
8
51
0
4
27
@ChulinXie
Chulin Xie
1 year
Please consider submitting! 🚀🚀
@nmervegurel
Nezihe Merve Gürel
1 year
📣 Call for Papers 📣 We are now accepting submissions for our #KLR workshop at #ICML2023 ! Submit your completed research or work-in-progress at the intersection of knowledge reasoning🤔 and ML🤖 by May 26, 2023 (AOE)! Stay tuned for the schedule🔥🚀
3
7
39
0
0
9
@ChulinXie
Chulin Xie
5 months
Our adaptive sequence length mechanism can enable GPT-3.5 to match the length of private data across iterations. [6/n]
Tweet media one
2
0
4
@ChulinXie
Chulin Xie
5 months
@liliang_ren Congrats!!!
0
0
3
@ChulinXie
Chulin Xie
11 months
@XinyuTang7 Congrats!
0
0
3
@ChulinXie
Chulin Xie
5 months
4 steps: 1) first draw random samples from an LLM, and then iteratively improve them by 2) computing the nearest neighbor, 3) selecting (with DP) the most similar ones to the private dataset, and 4) querying the LLM to generate more of such related samples. [2/n]
Tweet media one
1
0
3
@ChulinXie
Chulin Xie
5 months
Aug-PE is compatible with open-source LLMs, where DP finetuning is hard to implement due to the need to calculate the per-sample gradient in DP-SGD. [4/n]
Tweet media one
1
0
2
@ChulinXie
Chulin Xie
5 months
Aug-PE (w/ inference only) can match DP finetuning with the same model under privacy budget epsilon = 4, 2, 1 in some cases, and outperform it with stronger closed-source LLMs like GPT-3.5, where DP finetuning is infeasible! [3/n]
Tweet media one
1
0
3
@ChulinXie
Chulin Xie
5 months
Aug-PE (w/ inference APIs) is more efficient than DP-SGD finetuning on the Yelp dataset. The running time of Aug-PE mainly depends on the number of API calls (associated with the hyperparameter L). [5/n]
Tweet media one
2
0
2
@ChulinXie
Chulin Xie
9 months
@gorkaabad_ @acm_ccs Thank you Gorka!
0
0
0
@ChulinXie
Chulin Xie
5 months
@Albertyusun Thank you!!
0
0
1
@ChulinXie
Chulin Xie
7 months
@DongfuJiang Thank you Dongfu!
0
0
1
@ChulinXie
Chulin Xie
5 months
@PandaAshwinee Thanks! We exactly used downstream objectives to evaluate syn data. For Yelp/OpenReview syn data, we evaluate accuracy on downstream classification tasks. For PubMed syn data, we evaluate next-word prediction accuracy of a downstream model, inspired by
2
0
1