![Nan Zhang Profile](https://pbs.twimg.com/profile_images/1712939762435162112/2cxzDp3E_x96.jpg)
Nan Zhang
@NanZhangNLP
Followers
113
Following
212
Statuses
61
PhD Student @ISTatPENNSTATE, NLP #NLProc, ML, AI. Ex-intern @SFResearch, @NECLabsAmerica
State College, PA
Joined December 2015
📢Training efficient LLMs toward a knowledge-intensive domain? Consider domain-specific and task-agnostic compression! Code: Paper: Excited to present our #NAACL findings paper for LLMs pruning during Virtual Poster Session 2!🧵
2
5
9
RT @heyshrutimishra: NVIDIA just dropped a bombshell. 671-billion-parameter AI model DeepSeek-R1 is here! 🤯🚀
0
211
0
Thanks @SFResearch for sharing our SiReRAG paper ( accepted by #ICLR2025! We proposed a RAG indexing method that models both similarity and relatedness signals. Stay tuned for more updates🚨...
📣 From efficient key caches and multimodal embeddings to self-improving reasoning and faithful context adherence... we're thrilled to present a broad range of powerful new research at #ICLR2025! 🎉 Bookmark our accepted papers below, and we'll see you in Singapore, @iclr_conf ! 🔖 REGENESIS: LLMs can grow into reasoning generalists via self improvement 👉 🧠Becky Xiangyu Peng Congying Xia Xinyi Yang Caiming Xiong Jason Wu Chen Xing 🔖SiReRAG: Indexing Similar and Related Information for Multihop Reasoning 👉 🧠 Nan Zhang, Prafulla Choubey, Alexander. Fabbri, Gabriel Bernadett-Shapiro, Jason Wu 🔖FaithEval: Can Your Language Model Stay Faithful to Context, Even If “The Moon is Made of Marshmallows'' 👉 🧠 Yifei Ming, Senthil Purushwalkam, Shrey Pandit, Zixuan Ke, Xuan Phi Nguyen, Caiming Xiong, Shafiq Joty 🔖Preference Optimization for Reasoning with Pseudo Feedback 👉 🧠Fangkai Jiao, Geyang Guo, Xingxing Zhang, Nancy F. Chen, Shafiq Joty, Furu Wei 🔖ThinK: Thinner Key Cache by Query-Driven Pruning 👉 🧠Yuhui Xu; Zhanming Jie; Hanze Dong; Lei Wang; Xudong Lu; Aojun Zhou; Amrita Saha; Caiming Xiong; Doyen Sahoo 🔖Automatic Curriculum Expert Iteration for Reliable LLM Reasoning 👉 🧠Zirui Zhao, Hanze Dong, Amrita Saha, Caiming Xiong, Doyen Sahoo 🔖VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks 👉 🧠Ziyan Jiang, Rui Meng, Xinyi Yang, Semih Yavuz, Yingbo Zhou, Wenhu Chen 🔖Integrating Expertise of Software Engineering Agents 👉 🧠Kexun Zhang, Weiran Yao, Zuxin Liu, Yihao Feng, Zhiwei Liu, Rithesh Murthy, Tian Lan, Lei Li, Renze Lou, Jiacheng Xu, Bo Pang, Yingbo Zhou, Shelby Heinecke, Silvio Savarese, Huan Wang, Caiming Xiong Congrats to our researchers for the incredible body of work! #MachineLearning #AIResearch
0
6
13
@vipul_1011 I heard CVPR scores were out today, not to mention ICML abstract deadline (of interest to many rejected submissions).
1
0
0
RT @ManlingLi_: [Long Tweet Ahead] Faculty Interview Tips & Common Questions: 🧘♀️0. Firstly, do not be nervous - Almost everything can…
0
78
0
RT @_philschmid: The only fine-tuning guide you need for 2025 ‼️ Excited to share “How to fine-tune open LLMs in 2025 with @huggingface” co…
0
229
0
RT @RyoKamoi: 📢 New preprint! Do LVLMs have strong visual perception capabilities? Not quite yet... We introduce VisOnlyQA, a new dataset…
0
19
0
RT @RyoKamoi: Curious about LLM self-correction? Check out our reading list! 📚 We feature papers & blogs in * Key…
0
28
0
RT @steeve__huang: 📈 Chart-tastic News Flash 🚀 Our survey paper has just been accepted by #TKDE 🎉 It's my first ever journal paper! This…
0
11
0
RT @Wenpeng_Yin: @emnlpmeeting : To all EMNLPer’s 2024, we warmly invite you to join our tutorial on November 15 at 9:00 AM, titled 'Enhanc…
0
11
0
RT @PennStateEECS: NSF CAREER award: Centering people while advancing artificial intelligence Computer science researcher Rui Zhang to buil…
0
11
0
RT @HaoyiQiu: 🌐 Are LLM agents prepared to navigate the rich diversity of cultural and social norms? 🏠 CASA tests them on real-world tasks…
0
32
0
RT @PranavVenkit: 🥳New Paper Alert🥳 Excited to share my work from @salesforce —where we audited answer engines (aka generative search) like…
0
12
0
RT @phuguo: We used mechanistic interpretability to supercharge model editing and unlearning! 🚀 Our new method removes unwanted knowledge m…
0
39
0
RT @RyoKamoi: I'll be at @COLM_conf and present our work on Tuesday 11am-1pm (Poster #50)! Looking forward to chatting about LLM safety, se…
0
8
0
RT @weijie444: 10 years ago, ML papers were math-heavy. Advice I got: less math, more empirics. Today, many ML/AI papers lack even a single…
0
174
0
RT @YusenZhangNLP: Chain-of-Agents has been accepted by NeurIPS 2024! 📣We propose Chain-of-Agents, a training-free, task-agnostic, high…
0
24
0
RT @RyoKamoi: We will present our survey on self-correction (TACL, to appear) at #EMNLP2024! Let's discuss the future of inference-time sca…
0
14
0