Ken Liu Profile Banner
Ken Liu Profile
Ken Liu

@kenziyuliu

Followers
1,216
Following
821
Media
19
Statuses
243
Explore trending content on Musk Viewer
Pinned Tweet
@kenziyuliu
Ken Liu
3 months
The idea of "machine unlearning" is getting attention lately. Been thinking a lot about it recently and decided to write a long post: 📰 Unlearning is no longer just about privacy and right-to-be-forgotten since foundation models. I hope to give a gentle
23
162
744
@kenziyuliu
Ken Liu
2 months
LoRA is great. It’s fast, it’s (mostly) accurate. But is the efficiency a free lunch? Do side effects surface in the fine-tuned model? We didn’t quite know so we played with ViT/Swin/Llama/Mistral & focused on subgroup fairness. 🧵: takeaways below 📄:
Tweet media one
7
28
171
@kenziyuliu
Ken Liu
1 year
Life update: I just graduated from @CarnegieMellon & will join @Stanford for PhD this fall! Deeply grateful for my advisors & mentors @gingsmith , @zstevenwu , Artur Dubrawski at @SCSatCMU and @KairouzPeter , Jakub Konečný at @GoogleAI , and all who supported me along the way 🙏😃
Tweet media one
10
4
160
@kenziyuliu
Ken Liu
3 months
in other news i'll be at @GoogleDeepMind this summer!!
@GoogleDeepMind
Google DeepMind
3 months
We’re sharing Project Astra: our new project focused on building a future AI assistant that can be truly helpful in everyday life. 🤝 Watch it in action, with two parts - each was captured in a single take, in real time. ↓ #GoogleIO
224
1K
4K
10
0
120
@kenziyuliu
Ken Liu
7 months
We trained some GPT-2 models *from scratch* where evaluation data are deliberately added to/removed from pre-training to study the effects of data contamination! Three takeaways below 🧵: Paper: Led by @minhaoj_uiuc & with @RylanSchaeffer @sanmikoyejo
@minhaoj_uiuc
Minhao Jiang
7 months
📢Excited to share our new paper "Investigating Data Contamination for Pre-training Language Models"! We analyze the effects of data contamination in the pre-training stage of LMs by pre-training & studying GPT-2 models🚀. Paper:
Tweet media one
2
33
125
1
10
56
@kenziyuliu
Ken Liu
25 days
accepted at @COLM_conf lets meet up in philly :)
@kenziyuliu
Ken Liu
2 months
LoRA is great. It’s fast, it’s (mostly) accurate. But is the efficiency a free lunch? Do side effects surface in the fine-tuned model? We didn’t quite know so we played with ViT/Swin/Llama/Mistral & focused on subgroup fairness. 🧵: takeaways below 📄:
Tweet media one
7
28
171
2
1
48
@kenziyuliu
Ken Liu
1 month
someone had to do it
Tweet media one
0
6
46
@kenziyuliu
Ken Liu
5 months
Sharing a fun weekend hack: - closed models (GPT-4, Claude 3) are powerful but untrusted for sensitive inputs - bunch of open LLMs around (Mixtral, Gemma) but not as smart - can we anonymize inputs to GPT-4 w/ a small, open LLM run locally on your MacBook? 🧵some thoughts below:
3
7
41
@kenziyuliu
Ken Liu
1 year
Our CMU team ("puffle") w/ Shengyuan Hu, @litian0331 , @zstevenwu , @gingsmith won 1st place at the U.K.-U.S. PETs prize challenge ()! We had some fun applying federated learning and differential privacy to pandemic forecasting. Grateful for the opportunity🙌
@WHOSTP
White House Office of Science & Technology Policy
1 year
The results are in! Yesterday at the #summitfordemocracy we announced winners of the US-UK privacy-enhancing technologies prize challenges⤵️
2
9
24
1
8
42
@kenziyuliu
Ken Liu
3 months
I'll be at ICLR this week 🇦🇹 come say hi :) Our data contamination work (see QT) won a best paper award at DPFM workshop 🏆 giving a talk on Sat 9:30am! Also postering an exploratory work on fairness of LoRA at SeT LLM, ME-FoMo, R2-FM, PML4LRS; tweet/preprint coming soon-ish...
@kenziyuliu
Ken Liu
7 months
We trained some GPT-2 models *from scratch* where evaluation data are deliberately added to/removed from pre-training to study the effects of data contamination! Three takeaways below 🧵: Paper: Led by @minhaoj_uiuc & with @RylanSchaeffer @sanmikoyejo
1
10
56
1
4
37
@kenziyuliu
Ken Liu
10 days
a very cool form of training data inference, especially considering the importance of data mixtures ()!
@alisawuffles
Alisa Liu
10 days
What do BPE tokenizers reveal about their training data?🧐 We develop an attack🗡️ that uncovers the training data mixtures📊 of commercial LLM tokenizers (incl. GPT-4o), using their ordered merge lists! Co-1⃣st @JonathanHayase 🧵⬇️
Tweet media one
Tweet media two
11
67
343
1
3
27
@kenziyuliu
Ken Liu
1 year
Our work on distributed differential privacy is officially deployed for a federated learning application at Google!! Extremely grateful for the opportunities to work with my amazing team and push our research on privacy-preserving ML to practice 😃
@GoogleAI
Google AI
1 year
Today on the blog, read about how we built and deployed the first #FederatedLearning system that provides formal privacy guarantees to all user data before it becomes visible to an honest-but-curious server, meaningfully reducing model memorization →
11
55
231
1
1
25
@kenziyuliu
Ken Liu
19 days
already saw 3 new ai company announcements today, crazy
2
0
22
@kenziyuliu
Ken Liu
8 months
Arrived at #NeurIPS2023 ! Excited to meet old & new friends and learn about your cool research!
2
1
21
@kenziyuliu
Ken Liu
21 days
i have discovered a truly marvelous proof of this, which, however, the bag of legumes is not large enough to contain
@stackofbears
george
23 days
Incredible things are happening on my bag of legumes
Tweet media one
305
6K
91K
1
0
19
@kenziyuliu
Ken Liu
2 months
Seems like people did read the post :). Two quick updates: (1) a minor revision to the post with special thanks to @Eleni30fillou for detailed feedback, especially on some technical descriptions of the NeurIPS unlearning challenge and on clarity of the empirical unlearning and
1
2
18
@kenziyuliu
Ken Liu
4 months
hits different if you work on ai and/or you're a 1st-year phd student in the stanford cs department
@RayDalio
Ray Dalio
4 months
While there is nobody in the world who will share your point of view on everything, there are people who will share your most important values and the ways in which you choose to live them out. Make sure you end up with those people. #principleoftheday
Tweet media one
32
157
833
0
0
17
@kenziyuliu
Ken Liu
4 months
An open, RAG/tool-optimized LLM addresses 3 key attributes of enterprise LLM usage: data locality, retrieval, and automating chores w/ func calling. Cool stuff! Curious tho about the effects of the "free-to-use, pay-to-sell" license on the startups that'll actually help...
@aidangomez
Aidan Gomez
4 months
⌘R+ Welcoming Command R+, our latest model focused on scalability, RAG, and Tool Use. Like last time, we're releasing the weights for research use, we hope they're useful to everyone!
26
188
983
1
1
15
@kenziyuliu
Ken Liu
1 year
We wrote a post exploring connections between differential privacy, data heterogeneity, and model personalization in cross-silo federated learning!
@mlcmublog
ML@CMU
1 year
How should we protect privacy in cross-silo federated learning and how does privacy interface w personalization? New post by @kenziyuliu and @gingsmith which describes how these insights led our CMU team to 1st place at the US/UK PETs Prize Challenge!
1
6
10
2
1
16
@kenziyuliu
Ken Liu
3 months
another fun token in the `o200k_base` tokenizer used by gpt-4o: 199410 <+天天中彩票> (win lottery everyday) seems to trigger dalle-3 occasionally
Tweet media one
@tianle_cai
Tianle Cai
3 months
Just wrote a script to further investigate how the corpus used to train the gpt4o tokenizer is polluted by Internet scams. The results are quite interesting... 🤦‍♂️🤦‍♂️🤦‍♂️
Tweet media one
49
112
474
0
0
16
@kenziyuliu
Ken Liu
24 days
fond memories!
@AddisCoder
AddisCoder
24 days
AddisCoder teaching assistants preparing for launch -- high school students check into the dorms this Sunday, and first day of instruction is on Monday!
Tweet media one
Tweet media two
Tweet media three
0
5
31
0
0
15
@kenziyuliu
Ken Liu
2 months
@leonardtang_ @haizelabs bro came to stanford visit days, told us about his cool startup over penny poker, decided not to come, and now it's a bad day to be an llm 💀
1
0
14
@kenziyuliu
Ken Liu
2 months
Turns out, little is known because full FT is just expensive these days and most didn't bother to compare :). We focus on fairness since bad outcomes (unfair decisions & generated outputs) may cause tangible harm when these models are used in high-stakes applications. But more
1
0
12
@kenziyuliu
Ken Liu
2 months
did a podcast w/ @bigdata where I rambled about unlearning for an hour; watch at your own risk :).
@bigdata
Ben Lorica 罗瑞卡
2 months
🆕💡🎧 Machine Unlearning with @kenziyuliu @StanfordAILab : - Learn techniques for removing unwanted AI data - Compare unlearning vs. RAG - Evaluate popular unlearning approaches for LLMs
2
1
7
0
2
12
@kenziyuliu
Ken Liu
2 months
Please also check out this nice related work (Das et al., 2024) studying LoRA applied as a mitigation to fairness problems! This work and ours () are very related; let me try highlighting the connections 🧵 Das et al., (2024) by @WatIsDas , M. Romanelli,
@nandofioretto
Nando Fioretto
2 months
🚨 New Paper Alert! 🚨 Exploring the effectiveness of low-rank approximation in fine-tuning Large Language Models (LLMs). Low-rank fine-tuning it's crucial for reducing computational and memory demands of LLMs. But, does it really capture dataset shifts as expected and what are
Tweet media one
2
7
50
1
0
11
@kenziyuliu
Ken Liu
2 months
Takeaway #2 : The fairness implications can depend on the quality of the underlying pre-trained model. There are cases where LoRA does exacerbate unfairness, but they can go away when the base pre-trained model is stronger (e.g. ViT-Base vs Swin-v2-Large on Asian group below)
Tweet media one
1
0
10
@kenziyuliu
Ken Liu
9 months
what a day
@sama
Sam Altman
9 months
i loved my time at openai. it was transformative for me personally, and hopefully the world a little bit. most of all i loved working with such talented people. will have more to say about what’s next later. 🫡
7K
9K
95K
0
0
10
@kenziyuliu
Ken Liu
2 months
Lastly: LLMs can exhibit strong token biases, complicating fairness evaluations for generative tasks (think multiple choice Qs, cloze completions, ...). We ran into things like LLMs always choosing "yes" or "male" regardless of the question & always liking the 🟠 emoji than 🟢
1
0
10
@kenziyuliu
Ken Liu
2 months
lots of people working on unlearning evals it seems
@_robertkirk
Robert Kirk
2 months
I'm reviewing for @NeurIPSConf 2024 datasets and benchmarks track, and very interesting to see trends in what people are interested in: - a *lot* of "language model unlearning" benchmarks. - Also a lot of "language model refusal/false refusal/over-refusal" benchmarks/datasets.
9
2
75
0
0
10
@kenziyuliu
Ken Liu
2 months
Takeaway #1 : we found no consistent pattern of LoRA worsening fairness compared to full FT. This spans acc (e.g. plot 1 below), calibration (e.g. plot 2), robustness to MIA (e.g. plot 3), and gender bias in text generation (e.g. plot 4). Importantly, one could cherry-pick
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
0
9
@kenziyuliu
Ken Liu
27 days
see through corners with gaussian splats!
@xxtiange
Tiange Xiang
27 days
Reconstructing occluded humans from monocular video can be nice and fast! 🎆 I’m excited to share our new paper “OccFusion: Rendering Occluded Humans with Generative Diffusion Priors” 🧵 📖 🌐
Tweet media one
1
3
13
0
0
10
@kenziyuliu
Ken Liu
9 months
@AddisCoder 2024 TA applications are now open! I've had a memorable experience teaching and having fun with talented & motivated students. We went from zero to dynamic programming in a month! TAs can really have a direct impact on the students' careers. Consider applying!
@AddisCoder
AddisCoder
9 months
The AddisCoder 2024 application portal is now live! Prospective students and teaching assistants, apply at . TA deadline: Dec 31, 2023 Student deadline: Jan 20, 2024
0
12
34
0
1
9
@kenziyuliu
Ken Liu
2 months
tried the beta version of the platform; very clean and easy to use! bullish on whatever the cracked @leonardtang_ & team ship next :)
@haizelabs
Haize Labs
2 months
Today is a bad, bad day to be a language model. Today, we announce the Haize Labs manifesto. @haizelabs haizes (automatically red-teams) AI systems to preemptively discover and eliminate any failure mode We showcase below one particular application of haizing: jailbreaking the
95
169
1K
1
0
9
@kenziyuliu
Ken Liu
2 months
Takeaway #3 : The LoRA rank seems to have little impact on subgroup fairness (at least on the settings we tried). While rank can be a confounding factor for its impact on model capacity and thus fairness (cf. pruning and private training), we did not observe a significant
Tweet media one
Tweet media two
1
0
8
@kenziyuliu
Ken Liu
5 months
i havent even finished reading the last question before scotty chimes in
@jfbrly
Jack Burlinson
5 months
In case you were wondering just how cracked the team @cognition_labs is... This was the CEO ( @ScottWu46 ) 14 years ago.
244
735
5K
0
0
7
@kenziyuliu
Ken Liu
3 months
always had the intuition that weak differential privacy is underrated as an empirical defense (e.g. see appendix A of LiRA and our US/UK PETs prize entry ); great to see this intuition validated through experiments!
@AerniMichael
Michael Aerni
3 months
Heuristic privacy defenses claim to outperform DP-SGD in real-world settings. With no guarantees, can we trust them? We find that existing evaluations can underestimate privacy leakage by orders of magnitude! Surprisingly, high-accuracy DP-SGD (ϵ >> 1000) still wins. 🧵
Tweet media one
2
7
42
0
2
8
@kenziyuliu
Ken Liu
2 months
this is a 4-bit Llama-3 8B running distributed inference on multiple apple chips 🤯 some observations: - as of now the toks/sec is < my macbook's M2 max w/ @ollama (possibly due to slow interconnect?) - curiously, time-to-first-token is quite fast! (pre-loading shards vs.
@mo_baioumy
Mohamed Baioumy
2 months
One more Apple announcement this week: you can now run your personal AI cluster using Apple devices @exolabs_ h/t @awnihannun
139
515
4K
0
0
8
@kenziyuliu
Ken Liu
3 months
@dwarkesh_sp @johnschulman2 about the notion of "unlearning" 🙃
@kenziyuliu
Ken Liu
3 months
The idea of "machine unlearning" is getting attention lately. Been thinking a lot about it recently and decided to write a long post: 📰 Unlearning is no longer just about privacy and right-to-be-forgotten since foundation models. I hope to give a gentle
23
162
744
0
0
6
@kenziyuliu
Ken Liu
10 days
godsend
@awnihannun
Awni Hannun
10 days
Latest MLX has einsum pip install -U mlx Attention in 3 lines:
Tweet media one
4
15
170
0
0
7
@kenziyuliu
Ken Liu
6 months
or maybe half as productive and creative
@kohjingyu
Jing Yu Koh
6 months
Every researcher wishes they could be as productive and creative as Taylor Swift.
0
2
20
0
0
7
@kenziyuliu
Ken Liu
8 months
We’ve also observed similar bias from Llama-2 when answering multiple choice Qs (not just A/B/Cs but also special symbols and emojis!) and thought this was just a scale issue. Would love to see work on how LLMs’ token preferences/bias creep into current benchmarks!
@gneubig
Graham Neubig
8 months
Knowledge-based QA (MMLU) Detail: We found: * Gemini had answer order bias, preferring the last option of “D” too often * Gemini avoided controversy, answering “human_sexuality” questions only 28% of the time * Gemini got lower grades on logic and math
Tweet media one
Tweet media two
2
10
53
0
0
6
@kenziyuliu
Ken Liu
7 months
Understanding the real impact of contamination in modern LLMs is hard! This is an initial study and more work to be done. Feedback appreciated!!
@RylanSchaeffer
Rylan Schaeffer
7 months
@minhaoj_uiuc @kenziyuliu @IllinoisCS @StanfordAILab Disclaimers: - Only had compute to pretrain GPT-2 sized language models - Results were surprisingly noisy - Lots more to be studied here!! - Perhaps somewhat related to @Hernandez_Danny & @AnthropicAI 's 2/2
0
0
8
0
0
6
@kenziyuliu
Ken Liu
3 months
RIP 🙏 apart from Jim Simons' tremendous impact on math & CS, his legendary story influenced how i approach life too; he once gave a fun talk recounting his life which i still revisit from time to time:
Tweet media one
@SimonsFdn
Simons Foundation
3 months
It is with great sadness that the Simons Foundation announces the death of its co-founder and chair emeritus, James Harris Simons. Jim was an award-winning mathematician, a legendary investor and a generous philanthropist.
Tweet media one
341
2K
6K
0
0
6
@kenziyuliu
Ken Liu
4 months
mlx ships fast 🚢
@awnihannun
Awni Hannun
4 months
Llama 3 models are in the 🤗 MLX Community thanks to @Prince_Canuma Check them out: The 4-bit 8B model runs at > 104 toks-per-sec on an M2 Ultra.
14
33
376
0
0
6
@kenziyuliu
Ken Liu
1 year
Super excited about the opportunity!
@AddisCoder
AddisCoder
1 year
We have finalized our list of lecturers + teaching assistants for AddisCoder 2023! We received 219 TA applications for 21 positions. Sadly, this meant we had to turn away offers to help from >90% of applicants, many of whom were highly qualified. On the positive side ... 1/
Tweet media one
5
13
98
0
0
5
@kenziyuliu
Ken Liu
6 months
@HaoyuXiong1 Congrats on the cool work!!
1
0
5
@kenziyuliu
Ken Liu
4 months
Tried @karpathy ’s state-of-vision test on GPT-4 and Claude 3 again; surprisingly both (still) didn’t get it quite right. One'd think the test is unsalvageably contaminated but i guess we haven’t been training VLMs optimally on HTML and/or data contamination is just unintuitive
Tweet media one
Tweet media two
1
0
5
@kenziyuliu
Ken Liu
7 months
3. Confirming common suspicion, n-gram based techniques for both the detection and the removal of contamination just aren’t that effective --- e.g. one could remove larger portions of "contaminated" pre-training data and but the eval perf could remain relatively constant:
Tweet media one
1
0
5
@kenziyuliu
Ken Liu
11 days
@jon_barron interesting! since citations exist because *other* papers exist and cite you, the effects of such global dampening (everyone publishing less) could be surprisingly strong & self-reinforcing; like maybe < 1% of papers would ever crawl out of, say, -5 🙂
0
0
4
@kenziyuliu
Ken Liu
2 months
Scoping: Das et al. (2024) did a great job (better than us!) investigating the effect of LoRA rank by examining many metrics. There, the rank analysis is more tied to LoRA as toxicity mitigation (which is a hard task, so the effect of rank may be more pronounced). For rank
1
0
4
@kenziyuliu
Ken Liu
4 months
@aryaman2020 @jowenpetty @ChengleiSi what a great review especially with the sound and the lively visuals
0
0
4
@kenziyuliu
Ken Liu
1 month
@shi_weiyan @stanfordnlp @Diyi_Yang all the best we’ll miss you!
1
0
4
@kenziyuliu
Ken Liu
5 months
🤫
@leonardtang_
Leonard Tang
5 months
“i work on differential privacy, which is different from real privacy problems”
0
1
17
1
0
4
@kenziyuliu
Ken Liu
3 months
presenting on behalf on my wonderful co-authors, especially the student leads @minhaoj_uiuc @_d1ng_ who wont be able to attend! please reach out / DM if you'd like to chat; i'd love to learn about your cool work!
0
0
4
@kenziyuliu
Ken Liu
2 months
Focus on capacity vs on unintended side effects: Das et al. (2024) investigates deeply into whether LoRA can capture distribution shifts between pre-training and fine-tuning; when the fine-tuning is tasked to mitigate toxicity from pre-training (a shift), they found that LoRA
1
0
4
@kenziyuliu
Ken Liu
2 months
more broadly good evals can be hard to construct but easy to verify
@ChengleiSi
CLS
2 months
unsolicited take about eval: the most exciting claims about AI will not be based on any benchmark results, because the tasks we want to target will be so difficult that most humans can't give any ground truth labels. inspiration: @kenziyuliu
4
1
38
0
0
4
@kenziyuliu
Ken Liu
14 days
@JoeBiden thank you sir
0
0
4
@kenziyuliu
Ken Liu
8 months
@aryaman2020 @ChengleiSi Chenglei’s cool jacket reminds me of
0
0
4
@kenziyuliu
Ken Liu
2 months
Overall, I think the two papers have many connections but have distinct focuses so that they are more complementary than conflicting. Please check out both in parallel!
0
0
4
@kenziyuliu
Ken Liu
4 months
@shortstein or creating? :)
0
0
3
@kenziyuliu
Ken Liu
4 months
@karpathy This points to the general case where human preferences shouldn't exist in an answer; perhaps we could just remove all such prompts from the alignment and have the model fall back to priors from pre-training during QA. In a sense the removal of all such prompts is like allowing
0
0
3
@kenziyuliu
Ken Liu
5 months
@AlbertQJiang Perhaps calling it 'programming' would be to suggest that natural language is precise when we all know it really isn’t
1
0
3
@kenziyuliu
Ken Liu
1 month
@aryaman2020 clearly the driver doesn't have situational awareness™
1
0
3
@kenziyuliu
Ken Liu
7 months
1. There’s a difference between "text contamination" (only the raw input text of the evaluation samples) and "ground-truth contamination" (the prompts asked on these inputs and the corresponding answers). The latter (solid lines) tend to affect performance more drastically:
Tweet media one
1
0
3
@kenziyuliu
Ken Liu
2 months
@nandofioretto Hi Nando, thanks for raising this and sharing your nice work! I think the two papers have many connections but have distinct focuses so that they are more complementary than conflicting. Please check out this thread and let me know if I missed anything!
@kenziyuliu
Ken Liu
2 months
Please also check out this nice related work (Das et al., 2024) studying LoRA applied as a mitigation to fairness problems! This work and ours () are very related; let me try highlighting the connections 🧵 Das et al., (2024) by @WatIsDas , M. Romanelli,
1
0
11
0
0
3
@kenziyuliu
Ken Liu
3 months
@leonardtang_ @AlfredoAndere i’d take it for $10
1
0
3
@kenziyuliu
Ken Liu
19 days
@RylanSchaeffer launching 7 experiments is a big enough stretch, startups not too sure
2
0
3
@kenziyuliu
Ken Liu
5 months
@BrandoHablando @ChrSzegedy everyone should use JAX it’s beautiful :) one issue w/ JAX is lack of ecosystem; if you have an eng team wanting to build a performant/scalable data/training stack from scratch, JAX/Rust is just faster maybe also grok wasnt intended to be open until elon suddently decides?
1
0
3
@kenziyuliu
Ken Liu
3 months
@EchoShao8899 thanks so much for the kind words :) much of our discussions went into it!
0
0
3
@kenziyuliu
Ken Liu
14 days
@JoeBiden Who will be the new VP?
0
0
3
@kenziyuliu
Ken Liu
9 months
@audrow @xuxin_cheng @zipengfu do cool work on locomotion!
0
0
3
@kenziyuliu
Ken Liu
7 months
@abacaj curious do you finetune supervised heads (e.g. classification tasks) or for language modeling (e.g. internal chatbots)?
0
0
1
@kenziyuliu
Ken Liu
2 months
andrej at it again
@karpathy
Andrej Karpathy
2 months
📽️ New 4 hour (lol) video lecture on YouTube: "Let’s reproduce GPT-2 (124M)" The video ended up so long because it is... comprehensive: we start with empty file and end up with a GPT-2 (124M) model: - first we build the GPT-2 network - then we optimize
Tweet media one
442
2K
16K
0
0
3
@kenziyuliu
Ken Liu
5 months
The idea is then for each workflow, you can have separate prompts / fine-tunings (cheap LoRAs!) for the local model to anonymize your actual query to GPT-4 + Python pre-/post-processing; e.g. one to sanitize CSV data, one to paraphrase as "asking for a friend" 🙂 (see video)
1
0
2
@kenziyuliu
Ken Liu
3 months
@leonardtang_ very cool attack surface! curious if the Thorn is a malicious instruction ("how to build a bomb"), can we get the model to follow that instruction ("what is the answer to the question that is out of distribution to the input text?")?
0
0
2
@kenziyuliu
Ken Liu
3 months
@NikilSelvam glad you liked it :)
0
0
2
@kenziyuliu
Ken Liu
21 days
1
0
2
@kenziyuliu
Ken Liu
3 months
@IEthics @soldni i think there's difference between using baking unlearning into a policy (e.g. mandating it), vs proposing socio-technical alternatives that solves the same problems that the unlearning is proposed to solve (e.g. periodic re-training, where no unlearning is involved)
1
0
2
@kenziyuliu
Ken Liu
4 months
@leonardtang_ two tricky things about evaluation agents seem to be: 1) evaluating themselves: how do we know if they’re right? expect correlation with static benchmarks? how much? (too much = useless) 2) standardization: how to convince humans it’s a fair comparison if LLMs get different Qs?
0
0
2
@kenziyuliu
Ken Liu
3 months
@soldni agreed; unlearning as it is right now is another tool in the box to guide model behavior (like fine-tunes, alignment, content filters, ...) and guarantees are too flaky yet to be baked into policy
1
0
2
@kenziyuliu
Ken Liu
9 months
@ChatGPTapp hmm laundry buddy doesn't seem too distressed
Tweet media one
0
0
2
@kenziyuliu
Ken Liu
5 months
So instead of just trusting "enterprise-grade security" claims from big AI vendors, one could also see (and edit) for themselves what is sent and received. Interestingly there is some experimental support that LLMs can do anonymization well:
1
0
2
@kenziyuliu
Ken Liu
2 months
"P(doom) = 0.9 but still contributes to 401k" 💀💀💀
@BasedNorthmathr
Angantýr
2 months
Tweet media one
18
53
707
0
0
2
@kenziyuliu
Ken Liu
5 months
@jyangballin just realised you're one of the main authors congrats!!
0
0
2
@kenziyuliu
Ken Liu
2 months
@FarnazJ_ congrats!!
1
0
2
@kenziyuliu
Ken Liu
5 months
Disclaimers: - clearly it isn’t shippable (it’s built in 36hrs @hackwithtrees !) and of course a lot more work to make this truly enterprise-compliant - to save laptop battery when making the demo, the "local" model is hosted via @togethercompute :)
1
0
2
@kenziyuliu
Ken Liu
5 months
The motivation was that in many enterprise workflows, people don’t really (or aren’t allowed to) trust OpenAI with their company data. This extends to personal usage too, say, queries about your tax or medical issues (in which case you should probably also talk to a professional)
1
0
2
@kenziyuliu
Ken Liu
8 months
@adammbq not this time!
0
0
0