kenziyuliu Profile Banner
Ken Liu Profile
Ken Liu

@kenziyuliu

Followers
2K
Following
6K
Media
27
Statuses
345

cs phd @stanfordailab @stanfordnlp w/ @percyliang @sanmikoyejo. student researcher @googledeepmind. prev @scsatcmu @sydney_uni 🇦🇺

🌲
Joined January 2017
Don't wanna be here? Send us removal request.
@kenziyuliu
Ken Liu
9 months
The idea of "machine unlearning" is getting attention lately. Been thinking a lot about it recently and decided to write a long post: 📰. Unlearning is no longer just about privacy and right-to-be-forgotten since foundation models. I hope to give a gentle.
23
162
756
@kenziyuliu
Ken Liu
1 month
25% on frontier math is kinda hard to grasp.
28
34
1K
@kenziyuliu
Ken Liu
1 month
@marvel0_ it would be an overstatement to say i can solve 0% of these problems
Tweet media one
7
6
272
@kenziyuliu
Ken Liu
8 months
LoRA is great. It’s fast, it’s (mostly) accurate. But is the efficiency a free lunch? Do side effects surface in the fine-tuned model? . We didn’t quite know so we played with ViT/Swin/Llama/Mistral & focused on subgroup fairness. 🧵: takeaways below.📄:
Tweet media one
7
27
171
@kenziyuliu
Ken Liu
2 years
Life update: I just graduated from @CarnegieMellon & will join @Stanford for PhD this fall!. Deeply grateful for my advisors & mentors @gingsmith, @zstevenwu, Artur Dubrawski at @SCSatCMU and @KairouzPeter, Jakub Konečný at @GoogleAI, and all who supported me along the way 🙏😃
Tweet media one
10
4
158
@kenziyuliu
Ken Liu
1 month
adding back the “feel the AGI” slack status.
0
1
127
@kenziyuliu
Ken Liu
9 months
in other news i'll be at @GoogleDeepMind this summer!!.
@GoogleDeepMind
Google DeepMind
9 months
We’re sharing Project Astra: our new project focused on building a future AI assistant that can be truly helpful in everyday life. 🤝. Watch it in action, with two parts - each was captured in a single take, in real time. ↓ #GoogleIO
9
0
117
@kenziyuliu
Ken Liu
1 year
We trained some GPT-2 models *from scratch* where evaluation data are deliberately added to/removed from pre-training to study the effects of data contamination!.Three takeaways below 🧵:. Paper: Led by @minhaoj_uiuc & with @RylanSchaeffer @sanmikoyejo.
@minhaoj_uiuc
Minhao Jiang
1 year
📢Excited to share our new paper "Investigating Data Contamination for Pre-training Language Models"!. We analyze the effects of data contamination in the pre-training stage of LMs by pre-training & studying GPT-2 models🚀. Paper:
Tweet media one
1
10
57
@kenziyuliu
Ken Liu
7 months
accepted at @COLM_conf lets meet up in philly :).
@kenziyuliu
Ken Liu
8 months
LoRA is great. It’s fast, it’s (mostly) accurate. But is the efficiency a free lunch? Do side effects surface in the fine-tuned model? . We didn’t quite know so we played with ViT/Swin/Llama/Mistral & focused on subgroup fairness. 🧵: takeaways below.📄:
Tweet media one
2
1
47
@kenziyuliu
Ken Liu
7 months
someone had to do it.
Tweet media one
0
6
46
@kenziyuliu
Ken Liu
11 months
Sharing a fun weekend hack:.- closed models (GPT-4, Claude 3) are powerful but untrusted for sensitive inputs.- bunch of open LLMs around (Mixtral, Gemma) but not as smart.- can we anonymize inputs to GPT-4 w/ a small, open LLM run locally on your MacBook?. 🧵some thoughts below:
3
7
42
@kenziyuliu
Ken Liu
2 years
Our CMU team ("puffle") w/ Shengyuan Hu, @litian0331, @zstevenwu, @gingsmith won 1st place at the U.K.-U.S. PETs prize challenge (! We had some fun applying federated learning and differential privacy to pandemic forecasting. Grateful for the opportunity🙌.
@WHOSTP
White House Office of Science & Technology Policy
2 years
The results are in! Yesterday at the #summitfordemocracy we announced winners of the US-UK privacy-enhancing technologies prize challenges⤵️
1
8
42
@kenziyuliu
Ken Liu
5 months
Some intuitions on circuit breakers ( and breaking them:. Circuit breakers prevent harmful language model outputs by interjecting outputs (e.g. w/ EOS tokens) when the internal representations used to sample the actual text outputs are "considered.
@DanHendrycks
Dan Hendrycks
5 months
Three models remain unbroken in the Gray Swan jailbreaking competition (~500 registrants), which is still ongoing. These models are based on Circuit Breakers + other RepE techniques ( .
Tweet media one
2
3
41
@kenziyuliu
Ken Liu
9 months
I'll be at ICLR this week 🇦🇹 come say hi :). Our data contamination work (see QT) won a best paper award at DPFM workshop 🏆 giving a talk on Sat 9:30am!. Also postering an exploratory work on fairness of LoRA at SeT LLM, ME-FoMo, R2-FM, PML4LRS; tweet/preprint coming soon-ish. .
@kenziyuliu
Ken Liu
1 year
We trained some GPT-2 models *from scratch* where evaluation data are deliberately added to/removed from pre-training to study the effects of data contamination!.Three takeaways below 🧵:. Paper: Led by @minhaoj_uiuc & with @RylanSchaeffer @sanmikoyejo.
1
4
37
@kenziyuliu
Ken Liu
4 months
gave a guest lecture on intro to ML privacy at northeastern university at @tianshi_li’s class!. it was a speed run over a few popular privacy tools, their motivations/implications, and how they may or may not fit in the context of modern foundation models. ML has obviously.
@tianshi_li
Tianshi Li (hiring phds)
4 months
It was fantastic to have @kenziyuliu virtually join my seminar class to offer a well-synthesized overview of ML privacy along with thought-provoking discussions. 90 minute felt too short. I hope we can have more crossovers between AI and HCI like this one!.
1
2
32
@kenziyuliu
Ken Liu
5 months
the skies look different when you finally solve some stupid cudnn/nccl versioning issues to get flash attention hitting 50% mfu for a pre-training run on h100s.
2
0
30
@kenziyuliu
Ken Liu
4 months
at @COLM_conf this week! hmu if there's something we can chat about (privacy/security, membership, memorization, unlearning, localization, . ) or if you have food recs :).
@kenziyuliu
Ken Liu
7 months
accepted at @COLM_conf lets meet up in philly :).
3
3
31
@kenziyuliu
Ken Liu
6 months
a very cool form of training data inference, especially considering the importance of data mixtures (!.
@alisawuffles
Alisa Liu
6 months
What do BPE tokenizers reveal about their training data?🧐. We develop an attack🗡️ that uncovers the training data mixtures📊 of commercial LLM tokenizers (incl. GPT-4o), using their ordered merge lists!. Co-1⃣st @JonathanHayase.🧵⬇️
Tweet media one
Tweet media two
1
3
27
@kenziyuliu
Ken Liu
2 months
a big collab on unlearning led by @katherine1ee and @afedercooper!!. it always helps to ask *why* and *how* a specific new technology will tangibly help in practice, or if it’s really just a solution searching for a problem. this is especially true for unlearning as of today.
@katherine1ee
Katherine Lee
2 months
Machine unlearning is taking off! There is a ton of interest in getting generative AI models to “unlearn” targeted undesirable information, often to meet policy aims in privacy, copyright, safety, etc. Our new cross-disciplinary collaboration shows why this isn’t simple.
Tweet media one
Tweet media two
0
2
26
@kenziyuliu
Ken Liu
2 months
amazing labmates and i learn so much from them!!.
@percyliang
Percy Liang
2 months
This year, I have 4 exceptional students on the academic job market, and they couldn’t be more diffferent, with research spanning AI policy, robotics, NLP, and HCI. Here’s a brief summary of their research, along with one representative work each:.
0
0
26
@kenziyuliu
Ken Liu
5 months
it's been inspiring witnessing @ChengleiSi executing this humongous effort (both⏳ and 💸) & the results are surprising and timely!.
@ChengleiSi
CLS
5 months
Automating AI research is exciting! But can LLMs actually produce novel, expert-level research ideas?. After a year-long study, we obtained the first statistically significant conclusion: LLM-generated ideas are more novel than ideas written by expert human researchers.
Tweet media one
0
1
26
@kenziyuliu
Ken Liu
2 years
Our work on distributed differential privacy is officially deployed for a federated learning application at Google!! Extremely grateful for the opportunities to work with my amazing team and push our research on privacy-preserving ML to practice 😃.
@GoogleAI
Google AI
2 years
Today on the blog, read about how we built and deployed the first #FederatedLearning system that provides formal privacy guarantees to all user data before it becomes visible to an honest-but-curious server, meaningfully reducing model memorization →
1
1
25
@kenziyuliu
Ken Liu
8 months
In this work we ask a simple Q: when we choose LoRA over full fine-tune, are there any unintended consequences, particularly on trustworthiness, when both give high utility?. 📄: 👯: work w/ @_d1ng_ @poonpura @BerivanISIK @sanmikoyejo @stai_research.
2
4
22
@kenziyuliu
Ken Liu
7 months
already saw 3 new ai company announcements today, crazy.
2
0
22
@kenziyuliu
Ken Liu
1 year
Arrived at #NeurIPS2023! Excited to meet old & new friends and learn about your cool research!.
2
1
21
@kenziyuliu
Ken Liu
2 months
cool work by the cool @anneouyang @simonguozirui !!.
@anneouyang
Anne Ouyang
2 months
Kernels are the kernel of deep learning. 🙃. but writing kernels sucks. Can LLMs help? 🤔. Introducing 🌽 KernelBench (Preview), a new coding benchmark designed to evaluate the ability of LLMs to generate ⚡️efficient💨 GPU kernels for optimizing neural network performance.
Tweet media one
1
1
20
@kenziyuliu
Ken Liu
4 months
Tweet media one
@NobelPrize
The Nobel Prize
4 months
BREAKING NEWS.The Royal Swedish Academy of Sciences has decided to award the 2024 #NobelPrize in Physics to John J. Hopfield and Geoffrey E. Hinton “for foundational discoveries and inventions that enable machine learning with artificial neural networks.”
Tweet media one
0
0
19
@kenziyuliu
Ken Liu
7 months
i have discovered a truly marvelous proof of this, which, however, the bag of legumes is not large enough to contain.
@stackofbears
george
7 months
Incredible things are happening on my bag of legumes
Tweet media one
1
0
19
@kenziyuliu
Ken Liu
8 months
So is the parameter efficiency of LoRA is a free lunch? Possibly, but not necessarily :). Indeed, no evidence of unfairness does not imply evidence of fairness. Feedback welcome!. 📄: 👯: work w/ @_d1ng_ @poonpura @BerivanISIK @sanmikoyejo @stai_research.
0
3
17
@kenziyuliu
Ken Liu
8 months
Seems like people did read the post :). Two quick updates: (1) a minor revision to the post with special thanks to @Eleni30fillou for detailed feedback, especially on some technical descriptions of the NeurIPS unlearning challenge and on clarity of the empirical unlearning and.
1
2
18
@kenziyuliu
Ken Liu
4 months
TIL googling a font changes the display text to that font
Tweet media one
Tweet media two
1
0
18
@kenziyuliu
Ken Liu
2 months
a post on the EMNLP award paper on (flawed) LLM membership inference must exist, and am super glad that @pratyushmaini @iamgroot42 are doing it!!.
@pratyushmaini
Pratyush Maini
2 months
1/6 A lot of us are grappling with peer review these days, but its worst manifestation is when prestigious conference awards overlook critical flaws. Case in point: #EMNLP2024’s Best Paper Award. I & @iamgroot42 wrote a blog on what went wrong: 🧵
Tweet media one
0
1
17
@kenziyuliu
Ken Liu
10 months
hits different if you work on ai and/or you're a 1st-year phd student in the stanford cs department.
@RayDalio
Ray Dalio
10 months
While there is nobody in the world who will share your point of view on everything, there are people who will share your most important values and the ways in which you choose to live them out. Make sure you end up with those people. #principleoftheday
Tweet media one
0
0
17
@kenziyuliu
Ken Liu
2 years
We wrote a post exploring connections between differential privacy, data heterogeneity, and model personalization in cross-silo federated learning!.
@mlcmublog
ML@CMU
2 years
How should we protect privacy in cross-silo federated learning and how does privacy interface w personalization?. New post by @kenziyuliu and @gingsmith which describes how these insights led our CMU team to 1st place at the US/UK PETs Prize Challenge!.
2
1
16
@kenziyuliu
Ken Liu
10 months
An open, RAG/tool-optimized LLM addresses 3 key attributes of enterprise LLM usage: data locality, retrieval, and automating chores w/ func calling. Cool stuff!. Curious tho about the effects of the "free-to-use, pay-to-sell" license on the startups that'll actually help. .
@aidangomez
Aidan Gomez
10 months
⌘R+. Welcoming Command R+, our latest model focused on scalability, RAG, and Tool Use. Like last time, we're releasing the weights for research use, we hope they're useful to everyone!
1
1
15
@kenziyuliu
Ken Liu
4 months
sharing slides here in case it’s useful --- please also leave feedback as it can be definitely more polished!.
0
6
16
@kenziyuliu
Ken Liu
5 months
@jxmnop i think (1) most people & use cases didnt care as much about privacy as previously thought, (2) small models just arent that good, and (3) power-efficiency is still low (laptop dies quickly without plugging in). #2 is improving fast; #3 seems hard from a sheer amount of flops.
0
1
15
@kenziyuliu
Ken Liu
7 months
fond memories!.
@AddisCoder
AddisCoder
7 months
AddisCoder teaching assistants preparing for launch -- high school students check into the dorms this Sunday, and first day of instruction is on Monday!
Tweet media one
Tweet media two
Tweet media three
0
0
15
@kenziyuliu
Ken Liu
6 months
in boston aug 21-22 (at TPDP day 2) and NYC 22-27 w/ @aryaman2020 let’s meet up 🫡.
@aryaman2020
Aryaman Arora
6 months
i will be in NYC with @kenziyuliu starting the 22nd. looking for people to meet and things to do 🫡.
1
1
15
@kenziyuliu
Ken Liu
2 months
Tweet media one
0
0
15
@kenziyuliu
Ken Liu
9 months
another fun token in the `o200k_base` tokenizer used by gpt-4o:.199410 <+天天中彩票> (win lottery everyday).seems to trigger dalle-3 occasionally
Tweet media one
@tianle_cai
Tianle Cai
9 months
Just wrote a script to further investigate how the corpus used to train the gpt4o tokenizer is polluted by Internet scams. The results are quite interesting. 🤦‍♂️🤦‍♂️🤦‍♂️.
Tweet media one
0
0
14
@kenziyuliu
Ken Liu
8 months
@leonardtang_ @haizelabs bro came to stanford visit days, told us about his cool startup over penny poker, decided not to come, and now it's a bad day to be an llm 💀.
1
0
14
@kenziyuliu
Ken Liu
3 months
Interesting read! Evolution is necessarily tied to an objective, an environment, and the agency to act in said environment to optimise towards said objective. IMO all three are currently missing for AI models (maybe not for long) — there’s no incentive beyond being “helpful and.
@tianle_cai
Tianle Cai
3 months
2
0
13
@kenziyuliu
Ken Liu
8 months
Turns out, little is known because full FT is just expensive these days and most didn't bother to compare :). We focus on fairness since bad outcomes (unfair decisions & generated outputs) may cause tangible harm when these models are used in high-stakes applications. But more.
1
0
12
@kenziyuliu
Ken Liu
8 months
did a podcast w/ @bigdata where I rambled about unlearning for an hour; watch at your own risk :).
@bigdata
Ben Lorica 罗瑞卡
8 months
🆕💡🎧 Machine Unlearning with @kenziyuliu @StanfordAILab:.- Learn techniques for removing unwanted AI data.- Compare unlearning vs. RAG.- Evaluate popular unlearning approaches for LLMs.
0
2
12
@kenziyuliu
Ken Liu
9 months
thanks to @aryaman2020, @jiaao_chen, @irena_gao, @johnhewtt, Shengyuan Hu, @KairouzPeter, @sanmikoyejo, @XiangLisaLi2, @percyliang, @ericmitchellai, @RylanSchaeffer, @EchoShao8899, @ChengleiSi, Pratiksha Thaker, @cindy_x_wu for inspiration and feedback before or during the.
1
1
12
@kenziyuliu
Ken Liu
8 months
Please also check out this nice related work (Das et al., 2024) studying LoRA applied as a mitigation to fairness problems!. This work and ours ( are very related; let me try highlighting the connections 🧵. Das et al., (2024) by @WatIsDas, M. Romanelli,.
@nandofioretto
Nando Fioretto
8 months
🚨 New Paper Alert! 🚨.Exploring the effectiveness of low-rank approximation in fine-tuning Large Language Models (LLMs). Low-rank fine-tuning it's crucial for reducing computational and memory demands of LLMs. But, does it really capture dataset shifts as expected and what are
Tweet media one
1
0
11
@kenziyuliu
Ken Liu
6 months
@aryaman2020 i wouldn't mind a tpu v5 pod either.
1
0
11
@kenziyuliu
Ken Liu
7 months
see through corners with gaussian splats!.
@xxtiange
Tiange Xiang
7 months
Reconstructing occluded humans from monocular video can be nice and fast! 🎆 I’m excited to share our new paper “OccFusion: Rendering Occluded Humans with Generative Diffusion Priors” 🧵.📖🌐
Tweet media one
0
0
10
@kenziyuliu
Ken Liu
14 days
@aryaman2020 where’s my bag dude.
1
3
10
@kenziyuliu
Ken Liu
8 months
Lastly: LLMs can exhibit strong token biases, complicating fairness evaluations for generative tasks (think multiple choice Qs, cloze completions, . ). We ran into things like LLMs always choosing "yes" or "male" regardless of the question & always liking the 🟠 emoji than 🟢.
1
0
10
@kenziyuliu
Ken Liu
8 months
lots of people working on unlearning evals it seems.
@_robertkirk
Robert Kirk
8 months
I'm reviewing for @NeurIPSConf 2024 datasets and benchmarks track, and very interesting to see trends in what people are interested in:.- a *lot* of "language model unlearning" benchmarks. - Also a lot of "language model refusal/false refusal/over-refusal" benchmarks/datasets.
0
0
10
@kenziyuliu
Ken Liu
5 months
throwing some IOCCC code at o1 and let's what happens.
Tweet media one
1
0
10
@kenziyuliu
Ken Liu
1 year
what a day.
@sama
Sam Altman
1 year
i loved my time at openai. it was transformative for me personally, and hopefully the world a little bit. most of all i loved working with such talented people. will have more to say about what’s next later. 🫡.
0
0
10
@kenziyuliu
Ken Liu
8 months
Takeaway #2: The fairness implications can depend on the quality of the underlying pre-trained model. There are cases where LoRA does exacerbate unfairness, but they can go away when the base pre-trained model is stronger (e.g. ViT-Base vs Swin-v2-Large on Asian group below)
Tweet media one
1
0
10
@kenziyuliu
Ken Liu
8 months
Takeaway #1: we found no consistent pattern of LoRA worsening fairness compared to full FT. This spans acc (e.g. plot 1 below), calibration (e.g. plot 2), robustness to MIA (e.g. plot 3), and gender bias in text generation (e.g. plot 4). Importantly, one could cherry-pick
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
0
9
@kenziyuliu
Ken Liu
8 months
tried the beta version of the platform; very clean and easy to use! bullish on whatever the cracked @leonardtang_ & team ship next :).
@haizelabs
Haize Labs
8 months
Today is a bad, bad day to be a language model. Today, we announce the Haize Labs manifesto. @haizelabs haizes (automatically red-teams) AI systems to preemptively discover and eliminate any failure mode. We showcase below one particular application of haizing: jailbreaking the
1
0
9
@kenziyuliu
Ken Liu
6 months
@cloneofsimo @kellerjordan0 empirically i've seen many cases that grad norms do not converge as losses converge across optimizers/LR schedules/weight decay; see, e.g., this paper . for this precise reason non-convex convergence analyses to stationary points are just not that useful.
2
1
9
@kenziyuliu
Ken Liu
5 months
this *extremely* long chain-of-thought to solve the cipher is pretty impressive ngl
Tweet media one
0
0
9
@kenziyuliu
Ken Liu
5 months
@aryaman2020 lol just watched it today; the argument/visualization of how large models can fit a lot of features due to johnson–lindenstrauss lemma is cute.
0
0
9
@kenziyuliu
Ken Liu
1 year
@AddisCoder 2024 TA applications are now open! I've had a memorable experience teaching and having fun with talented & motivated students. We went from zero to dynamic programming in a month! TAs can really have a direct impact on the students' careers. Consider applying!.
@AddisCoder
AddisCoder
1 year
The AddisCoder 2024 application portal is now live! Prospective students and teaching assistants, apply at TA deadline: Dec 31, 2023.Student deadline: Jan 20, 2024.
0
1
9
@kenziyuliu
Ken Liu
26 days
that time of year again.
@kohjingyu
Jing Yu Koh
26 days
Tweet media one
1
0
10
@kenziyuliu
Ken Liu
4 months
Check out @rose_e_wang's cool work on AI & tutoring! Rose thinks deeply about teacher-student interaction and how it should manifest in classrooms and different communities, beyond just "AI education". I've learned so much from her!.
@rose_e_wang
Rose
4 months
AI has the potential to transform real-world domains. But can AI actually improve outcomes in live interactions?. We conducted the first large-scale intervention of a Human-AI Approach that has statistically significant positive learning gains w/ 900 tutors & 1,800 K12 students.
Tweet media one
0
0
8
@kenziyuliu
Ken Liu
11 months
i havent even finished reading the last question before scotty chimes in.
@jfbrly
Jack Burlinson
11 months
In case you were wondering just how cracked the team @cognition_labs is. This was the CEO (@ScottWu46) 14 years ago.
0
0
7
@kenziyuliu
Ken Liu
9 months
always had the intuition that weak differential privacy is underrated as an empirical defense (e.g. see appendix A of LiRA and our US/UK PETs prize entry ; great to see this intuition validated through experiments!.
@AerniMichael
Michael Aerni
9 months
Heuristic privacy defenses claim to outperform DP-SGD in real-world settings. With no guarantees, can we trust them?. We find that existing evaluations can underestimate privacy leakage by orders of magnitude!. Surprisingly, high-accuracy DP-SGD (ϵ >> 1000) still wins. 🧵
Tweet media one
0
2
8
@kenziyuliu
Ken Liu
8 months
Takeaway #3: The LoRA rank seems to have little impact on subgroup fairness (at least on the settings we tried). While rank can be a confounding factor for its impact on model capacity and thus fairness (cf. pruning and private training), we did not observe a significant
Tweet media one
Tweet media two
1
0
8
@kenziyuliu
Ken Liu
8 months
this is a 4-bit Llama-3 8B running distributed inference on multiple apple chips 🤯 some observations:.- as of now the toks/sec is < my macbook's M2 max w/ @ollama (possibly due to slow interconnect?).- curiously, time-to-first-token is quite fast! (pre-loading shards vs.
@mo_baioumy
Mohamed Baioumy
8 months
One more Apple announcement this week: you can now run your personal AI cluster using Apple devices @exolabs_. h/t @awnihannun
0
0
8
@kenziyuliu
Ken Liu
9 months
@dwarkesh_sp @johnschulman2 about the notion of "unlearning" 🙃.
@kenziyuliu
Ken Liu
9 months
The idea of "machine unlearning" is getting attention lately. Been thinking a lot about it recently and decided to write a long post: 📰. Unlearning is no longer just about privacy and right-to-be-forgotten since foundation models. I hope to give a gentle.
0
0
6
@kenziyuliu
Ken Liu
4 months
@aryaman2020 @simonguozirui Humanity’s Last Model v11 copy 2 (final) (final).safetensors.
1
0
7
@kenziyuliu
Ken Liu
8 months
code is now available at .
0
0
7
@kenziyuliu
Ken Liu
6 months
godsend.
@awnihannun
Awni Hannun
6 months
Latest MLX has einsum. pip install -U mlx. Attention in 3 lines:
Tweet media one
0
0
7
@kenziyuliu
Ken Liu
1 year
or maybe half as productive and creative.
@kohjingyu
Jing Yu Koh
1 year
Every researcher wishes they could be as productive and creative as Taylor Swift.
0
0
7
@kenziyuliu
Ken Liu
2 months
@JeffDean or maybe it's a good sign 🤓.
@kenziyuliu
Ken Liu
9 months
The idea of "machine unlearning" is getting attention lately. Been thinking a lot about it recently and decided to write a long post: 📰. Unlearning is no longer just about privacy and right-to-be-forgotten since foundation models. I hope to give a gentle.
0
0
6
@kenziyuliu
Ken Liu
9 months
RIP 🙏 apart from Jim Simons' tremendous impact on math & CS, his legendary story influenced how i approach life too; he once gave a fun talk recounting his life which i still revisit from time to time:.
Tweet media one
@SimonsFdn
Simons Foundation
9 months
It is with great sadness that the Simons Foundation announces the death of its co-founder and chair emeritus, James Harris Simons. Jim was an award-winning mathematician, a legendary investor and a generous philanthropist.
Tweet media one
0
0
6
@kenziyuliu
Ken Liu
4 months
@leonardtang_ 🐐 looking back a year and finding oneself to be naive and cringe is a good litmus test that one has grown!.
2
0
6
@kenziyuliu
Ken Liu
4 months
cool team!.
@aterzis
Andreas Terzis
4 months
We have an open position for a Research Scientist/Research Engineer to join our team!. If interested:
0
0
6
@kenziyuliu
Ken Liu
1 year
Understanding the real impact of contamination in modern LLMs is hard! This is an initial study and more work to be done. Feedback appreciated!!.
@RylanSchaeffer
Rylan Schaeffer
1 year
@minhaoj_uiuc @kenziyuliu @IllinoisCS @StanfordAILab Disclaimers: .- Only had compute to pretrain GPT-2 sized language models.- Results were surprisingly noisy.- Lots more to be studied here!!.- Perhaps somewhat related to @Hernandez_Danny & @AnthropicAI 's 2/2.
0
0
6
@kenziyuliu
Ken Liu
2 years
Check out 😃.Also a nice 5min video feature at the Summit for Democracy:
0
0
6
@kenziyuliu
Ken Liu
1 year
We’ve also observed similar bias from Llama-2 when answering multiple choice Qs (not just A/B/Cs but also special symbols and emojis!) and thought this was just a scale issue. Would love to see work on how LLMs’ token preferences/bias creep into current benchmarks!.
@gneubig
Graham Neubig
1 year
Knowledge-based QA (MMLU).Detail: We found:.* Gemini had answer order bias, preferring the last option of “D” too often.* Gemini avoided controversy, answering “human_sexuality” questions only 28% of the time.* Gemini got lower grades on logic and math
Tweet media one
Tweet media two
0
0
6
@kenziyuliu
Ken Liu
10 months
mlx ships fast 🚢.
@awnihannun
Awni Hannun
10 months
Llama 3 models are in the 🤗 MLX Community thanks to @Prince_Canuma . Check them out: The 4-bit 8B model runs at > 104 toks-per-sec on an M2 Ultra.
0
0
6
@kenziyuliu
Ken Liu
3 months
@Sylvia_Sparkle interesting work! this paper may be relevant: we also did a hackathon project on a similar idea earlier this year .
@kenziyuliu
Ken Liu
11 months
Sharing a fun weekend hack:.- closed models (GPT-4, Claude 3) are powerful but untrusted for sensitive inputs.- bunch of open LLMs around (Mixtral, Gemma) but not as smart.- can we anonymize inputs to GPT-4 w/ a small, open LLM run locally on your MacBook?. 🧵some thoughts below:
1
0
5
@kenziyuliu
Ken Liu
2 years
Super excited about the opportunity!.
@AddisCoder
AddisCoder
2 years
We have finalized our list of lecturers + teaching assistants for AddisCoder 2023! We received 219 TA applications for 21 positions. Sadly, this meant we had to turn away offers to help from >90% of applicants, many of whom were highly qualified. On the positive side . 1/
Tweet media one
0
0
5
@kenziyuliu
Ken Liu
10 months
Tried @karpathy’s state-of-vision test on GPT-4 and Claude 3 again; surprisingly both (still) didn’t get it quite right. One'd think the test is unsalvageably contaminated but i guess we haven’t been training VLMs optimally on HTML and/or data contamination is just unintuitive
Tweet media one
Tweet media two
1
0
5
@kenziyuliu
Ken Liu
1 year
@HaoyuXiong1 Congrats on the cool work!!.
1
0
4
@kenziyuliu
Ken Liu
1 year
3. Confirming common suspicion, n-gram based techniques for both the detection and the removal of contamination just aren’t that effective --- e.g. one could remove larger portions of "contaminated" pre-training data and but the eval perf could remain relatively constant:
Tweet media one
1
0
5
@kenziyuliu
Ken Liu
3 months
@adad8m this paper by @yimingz0 might be relevant:
1
0
5
@kenziyuliu
Ken Liu
1 month
@jeffreygwang @OpenAI 🚀🚀🚀.
0
0
4
@kenziyuliu
Ken Liu
2 months
@aryaman2020 cant imagine the amount of yapping.
0
0
4
@kenziyuliu
Ken Liu
8 months
Focus on capacity vs on unintended side effects:. Das et al. (2024) investigates deeply into whether LoRA can capture distribution shifts between pre-training and fine-tuning; when the fine-tuning is tasked to mitigate toxicity from pre-training (a shift), they found that LoRA.
1
0
4
@kenziyuliu
Ken Liu
11 months
🤫.
@leonardtang_
Leonard Tang
11 months
“i work on differential privacy, which is different from real privacy problems”.
1
0
4
@kenziyuliu
Ken Liu
9 months
presenting on behalf on my wonderful co-authors, especially the student leads @minhaoj_uiuc @_d1ng_ who wont be able to attend! . please reach out / DM if you'd like to chat; i'd love to learn about your cool work!.
0
0
4
@kenziyuliu
Ken Liu
3 months
can we have a GPU version of this pls 🙏.
@BenjDicken
Ben Dicken
4 months
You asked for it, so here it is. Visualizing CPU cache speeds relative to RAM. Cache optimization is important too!
0
0
4
@kenziyuliu
Ken Liu
7 months
@shi_weiyan @stanfordnlp @Diyi_Yang all the best we’ll miss you!.
1
0
4
@kenziyuliu
Ken Liu
1 year
@aryaman2020 @ChengleiSi Chenglei’s cool jacket reminds me of
0
0
2
@kenziyuliu
Ken Liu
4 months
1
0
4
@kenziyuliu
Ken Liu
8 months
Overall, I think the two papers have many connections but have distinct focuses so that they are more complementary than conflicting. Please check out both in parallel!.
0
0
4
@kenziyuliu
Ken Liu
7 months
@JoeBiden thank you sir.
0
0
4
@kenziyuliu
Ken Liu
6 months
1
0
4