Weijie Su Profile Banner
Weijie Su Profile
Weijie Su

@weijie444

Followers
5,061
Following
456
Media
48
Statuses
775
Explore trending content on Musk Viewer
@weijie444
Weijie Su
25 days
10 years ago, ML papers were math-heavy. Advice I got: less math, more empirics. Today, many ML/AI papers lack even a single math formula, let alone math thinking. My advice to young LLM researchers: do a little math if possible. It'll distinguish yours from the sea of LLM
50
175
2K
@weijie444
Weijie Su
2 years
I receive large volumes of emails from Chinese students asking for summer internships. The English was often broken, but it has recently become native sounding😅
48
28
1K
@weijie444
Weijie Su
2 months
New Research (w/ amazing @hangfeng_he ) "A Law of Next-Token Prediction in Large Language Models" LLMs rely on NTP, but their internal mechanisms seem chaotic. It's difficult to discern how each layer processes data for NTP. Surprisingly, we discover a physics-like law on NTP:
Tweet media one
8
91
426
@weijie444
Weijie Su
5 years
While learning French affects my English, it doesn’t affect math. So human brains are “locally elastic”: interaction depends on similarities! Our paper () shows deep learning is also locally elastic, with implications on memorization, generalization, etc.
4
31
248
@weijie444
Weijie Su
3 years
Are you fed up with VERY noisy reviews from NeurIPS, ICML, ICLR...? Do you have your best papers rejected but mediocre papers accepted? If so, check out this NeurIPS 2021 paper:
8
39
240
@weijie444
Weijie Su
3 years
It feels so good to write single-author papers! In , we introduce a phenomenological model toward understanding why #deeplearning is so effective. It is called *neurashed* because it was inspired by watershed. 8 pages and ~15mins read!
Tweet media one
6
38
235
@weijie444
Weijie Su
5 years
How does training time in #deeplearning depend on the learning rate? A new paper () uncovers a fundamental *distinction* between #nonconvex and convex problems from this viewpoint, showing why learning rate decay is *more* powerful in nonconvex settings.
Tweet media one
3
57
236
@weijie444
Weijie Su
3 years
Our deep learning theory paper () got accepted to PNAS! Introduced Layer-Peeled Model that can: 1. predict a hitherto unknown phenomenon that we term Minority Collapse in imbalanced training; 2. explain neural collapse discovered by Papyan, Han and Donoho.
Tweet media one
3
29
193
@weijie444
Weijie Su
7 months
My group has a postdoc opening in the "theoretical" aspects of LLMs (meaning doing research on LLMs without much compute). Send me an email if you're interested!
4
18
144
@weijie444
Weijie Su
2 years
A postdoc position is available. Pls shoot me an email if you're interested in any of the following three: the theoretical foundation of #deeplearning , #privacy -preserving machine learning, and game-theoretic approaches to estimation. Appreciate it very much for RTing🙏 1/4
3
59
125
@weijie444
Weijie Su
1 year
Good news while attending #ICML2023 : Our deep learning "theory" paper *A Law of Data Separation in Deep Learning* got accepted by PNAS! w/ amazing @hangfeng_he
7
9
125
@weijie444
Weijie Su
3 years
I'm genuinely honored and excited to receive the inaugural SIAM Activity Group on Data Science Early Career Prize in 2022 (). Very grateful to my amazing collaborators and mentors who helped make it possible for me.
12
9
120
@weijie444
Weijie Su
2 years
How does #deeplearning separate data throughout *all* layers? w/ @hangfeng_he we discovered a precise LAW emerging in AlexNet, VGG, ResNet: The law of equi-separation What is this law about? Can it guide design, training, and model interpretation? 1/n
Tweet media one
Tweet media two
3
25
111
@weijie444
Weijie Su
5 years
Very honored to receive 2020 Sloan Research Fellowship #sloanfellow . Many thanks to collaborators, colleagues @Wharton @Penn and students for help & support along the way. Excited to use the funding by @SloanFoundation to further my research in #MachineLearning and #DataScience .
Tweet media one
8
14
112
@weijie444
Weijie Su
5 months
Introducing Preference Matching RLHF for aligning LLMs with human preferences : 1. Standard RLHF is biased. Its algorithmic bias inherits from the reference model. 2. PM RLHF precisely aligns w/ RM preferences. This is mathematically provable and experimentally corroborated.
Tweet media one
1
20
111
@weijie444
Weijie Su
4 years
New interpretation of the *double descent* phenomenon: noise in features is ubiquitous, and we show using a random feature model that noise can lead to benign overfitting. Paper: . w/ Zhu Li and Dino Sejdinovic.
Tweet media one
3
17
100
@weijie444
Weijie Su
2 years
I'm very humbled to receive the 2022 Peter Gavin Hall IMS Early Career Prize. Thanks to my advisor, mentors, colleagues, and students who have helped me make it possible.
@InstMathStat
IMS
2 years
Weijie Su @weijie444 Wins Peter Gavin Hall IMS Early Career Prize "for fundamental contributions to the development of privacy-preserving data analysis methodologies; for groundbreaking theoretical advancements in understanding gradient-based optimization methods; 1/5
Tweet media one
20
8
124
5
1
103
@weijie444
Weijie Su
4 years
Why do we need a *phenomenological* approach for understanding deep learning? Two new papers that use *local elasticity* for understanding the effectiveness of neural networks: and (NeurIPS 2020), from a phenomenological perspective
4
16
99
@weijie444
Weijie Su
1 year
Can (very) simple math inform RLHF for large language models? New paper  says YES​! Problem: 'write a love story' has many good responses but 'what's the capital of Peru' doesn't. However, human preference rankings cannot tell the difference! Solution:⬇️
Tweet media one
2
12
99
@weijie444
Weijie Su
3 years
Is it worth 4 hours making a poster for #NeurIPS2021 ? Perhaps NOT for my paper "You Are the Best Reviewer of Your Own Papers: An Owner-Assisted Scoring Mechanism" cos from my experience there'll be ~5 ppl each spending 2mins skimming my poster in Gather Town on Dec 7 evening EST.
Tweet media one
3
8
97
@weijie444
Weijie Su
2 years
Sparsity is all I see at #icml2022
Tweet media one
2
2
85
@weijie444
Weijie Su
3 years
In a #NeurIPS2021 paper (), we introduced a *phenomenological* model toward understanding deep learning using SDEs, showing training is *successful* iff local elasticity exists---backpropagation has a larger impact on same-class examples.
Tweet media one
3
16
72
@weijie444
Weijie Su
6 months
Ongoing lawsuits against GenAI firms over possible use of #copyrighted data for training raise vital questions for our society. 🤖⚖️ How can we address the copyright challenges? New research proposes a solution: "An Economic Solution to Copyright Challenges of Generative AI"
Tweet media one
1
13
70
@weijie444
Weijie Su
2 years
Heading to the @TheSIAMNews Conference on Mathematics of Data Science #SIAMMDS22 . Looking forward to seeing many colleagues and attending great talks.
Tweet media one
2
9
68
@weijie444
Weijie Su
3 months
Very excited to give a short course on large language models at #JSM2024 in Portland! w/ Emily Getzen and @linjunz_stat AI for Stat and Stat for AI! @AmstatNews
Tweet media one
1
6
67
@weijie444
Weijie Su
1 year
Eager for #ICML2023 results? Let's improve peer review! #ICML2023 invited authors to rank their submissions - can it refine review scores? Our paper () extends isotonic mechanisms to exp family dist in using rankings to refine scores & boost estimation 🤖
Tweet media one
2
3
66
@weijie444
Weijie Su
7 months
📢 #ICML2024 authors! Help improve ML peer review! 🔬📝 Check your inbox for an email titled "[ICML 2024] Author Survey" and rank your submissions. 🏆📈 Your confidential input is crucial, and won't affect decisions. 🔒✅ Survey link in email or "Author Tasks" on OpenReview.
0
18
52
@weijie444
Weijie Su
5 years
The long road to #fair news...
Tweet media one
1
12
56
@weijie444
Weijie Su
7 months
New Research: Watermarking large language models is a principled approach to combatting misinformation, but how do we evaluate their statistical efficiency and design even more powerful detection methods? 🤔 Our new paper addresses these challenges using a new framework. 1/n
Tweet media one
2
9
57
@weijie444
Weijie Su
5 years
How to maintain the power of #DeepLearning while preserving #privacy ? We recently applied Gaussian differential privacy to training neural nets, obtaining improved utility. Check out the paper () and happy Thanksgiving!
2
8
54
@weijie444
Weijie Su
5 years
The *tightest* #privacy analysis of deep learning is given by Gaussian differential privacy, as shown by earlier work w/ @JinshuoD @Aaroth . Now the implementation is available in TensorFlow via its privacy package. Check out
2
19
55
@weijie444
Weijie Su
5 years
Learning rate is recognized as “the single most important hyper-parameter” in training deep learning. Inspired by statistical insights, our new paper () proposes a dynamic learning rate schedule by splitting SGD for stationarity detection.
0
7
47
@weijie444
Weijie Su
24 days
@HighFreqAsuka Agree, I would emphasize 'mathematical thinking' rather than math formulae, equations, inequalities...
2
1
47
@weijie444
Weijie Su
5 years
One-year postdoc position available in my group. Ideal for one on the job market but wishing to defer the tenure clock. Pls email me if you are interested in #privacy , #deeplearning theory, #optimization , high-dimensional statistics, and other areas of mutual interests.
0
9
46
@weijie444
Weijie Su
2 years
#NeurIPS reviews just came out. It confirmed that "I am the best reviewer of my own papers."☹️
@weijie444
Weijie Su
2 years
I’ll be giving a talk “When Will You Become the Best Reviewer of Your Own Papers?” at @iccopt2022 tomorrow starting at 4:05 in Rauch 137. #optimization #iccopt2022
0
2
31
1
2
44
@weijie444
Weijie Su
2 years
It feels fun to write single-author papers! In , a mechanism design (game theory) based framework of estimation asks: Alice knows the quality of her n papers. What questions can Bob ask about her papers to improve the accuracy of review ratings?
Tweet media one
2
6
44
@weijie444
Weijie Su
3 years
Our paper "Understanding the Acceleration Phenomenon via High-Resolution Differential Equations" () was accepted by Mathematical Programming. Thanks to my wonderful coauthors Bin, @SimonShaoleiDu , and Mike!
3
4
44
@weijie444
Weijie Su
3 years
Gradient clipping is used in training private deep learning models. Any *side effect*? New paper shows it changes the spectrum from an optimization viewpoint, resulting in slow convergence. New strategy is proposed. w my students Zhiqi, Hua, colleague Qi
1
2
39
@weijie444
Weijie Su
5 years
Our *local elasticity* paper () was accepted to #ICLR2020 ! Takeaway: DNNs learn locally and understand elastically, with slides (). First #DeepLearning paper, first paper w/o theorems. Excited to visit the land of Queen of Sheba 2020!
Tweet media one
0
3
38
@weijie444
Weijie Su
5 years
#NeurIPS2019 poster session. Reminds me of farmers’ markets in China when I was a child...
Tweet media one
1
0
34
@weijie444
Weijie Su
4 years
Too much competition leads to inefficiency, happening e.g. in the #gaokao exam taken by 10M students next week. That's the price we pay for involution. Surprisingly, our paper () shows the price of competition appears in high-dimensional linear models. 1/n
Tweet media one
1
4
34
@weijie444
Weijie Su
3 years
A new paper with Zhiqi, @JKlusowski , @CindyRush on type I and II errors tradeoff of SLOPE, asking: is there any fundamental difference between l1 and sorted l1 regularization? Our analysis leverages approximate message passing, developed by Donoho, Maleki, and @Andrea__M .
Tweet media one
1
3
35
@weijie444
Weijie Su
25 days
@qixing_huang Exactly! If everyone were at the Einstein level, doing mediocre research would win you a Nobel prize!
0
0
33
@weijie444
Weijie Su
4 years
Had an enormous pleasure to read the paper by Papyan, Han and Donoho. Highly recommend it to anyone who is interested in deep learning theory. Very elegant and mathematically concrete insights.
0
5
32
@weijie444
Weijie Su
5 years
I will be giving a talk at @Yale @yaledatascience this Monday introducing a new notion of differential privacy (). This was inspired by the hypothesis testing interpretation of privacy, joint work with Jinshuo and @Aaroth .
0
5
32
@weijie444
Weijie Su
2 years
I’ll be giving a talk “When Will You Become the Best Reviewer of Your Own Papers?” at @iccopt2022 tomorrow starting at 4:05 in Rauch 137. #optimization #iccopt2022
0
2
31
@weijie444
Weijie Su
10 months
Attending #NeurIPS2023 and find it more exciting than in past years. But each work/poster/oral talk individually becomes less interesting. Anyone has the same feeling?
1
0
30
@weijie444
Weijie Su
2 years
I'll be attending #neurips2022 from Nov 29-Dec 1. Looking forward to seeing old and new friends out there, as well as my former adviser Emmanuel Candes' plenary talk on conformal prediction!
0
3
29
@weijie444
Weijie Su
3 years
This new paper () proposes a new weighted training algorithm to improve the sample efficiency of learning from cross-task signals. To the best of our knowledge, it is the first weighted algorithm for cross-task learning with theoretical guarantees.
Tweet media one
2
4
30
@weijie444
Weijie Su
5 months
Heading to #ICLR2024
2
1
29
@weijie444
Weijie Su
7 months
ICML 2024 authors: Please participate in our study on improving peer review in ML! Rank your submissions confidentially under "Author Tasks" on OpenReview. More details at . Thank you! @ENAR_ibs
@icmlconf
ICML Conference
7 months
You have received an email titled "[ICML 2024] Author Survey" with a link to confidentially rank your submissions based on their relative quality. The survey can also be accessed under "Author Tasks" on OpenReview.
2
3
31
0
5
28
@weijie444
Weijie Su
5 years
Hi Boston friends, I will be giving two talks this Thu and Fri at Harvard and MIT on differential #privacy : 1:30-2:45 Thu at Pierce Hall 213 Harvard; 11-12 Fri at E18-304 MIT. Hope to see many of you if you are free of the #icml2020 #virus .
1
3
28
@weijie444
Weijie Su
1 year
#icml2023 is out! Are your expectations consistent with the outcomes? In any case, we're sending out a review feedback email via OpenReview to all #icml2023 authors. We appreciate your input that will help us design better peer review mechanisms for ml conferences.
1
3
27
@weijie444
Weijie Su
1 year
Just landed in Hawaii for @icml2023 . Happy to chat if you’re also around
1
0
27
@weijie444
Weijie Su
2 years
Again, my favorite papers got rejected. Really hope that I can "review" my own papers myself:
@ylecun
Yann LeCun
2 years
Verdict from @icmlconf : 3 out of 3 ..... rejected. If I go by tweet statistics, ICML has rejeted every single paper this year 🤣
27
46
1K
2
2
26
@weijie444
Weijie Su
1 year
The Isotonic Mechanism was experimented in @icmlconf 2023, requiring authors to provide rankings of their submissions to compute rank-calibrated review scores. A challenge is how to deal with multiple coauthors? A (not perfect but performant) solution:
Tweet media one
1
5
25
@weijie444
Weijie Su
5 years
Very pleasant visit at Yale; will give the same talk tomorrow @Wharton (). BTW, got best advice yesterday that it would double attendance with deep learning in the title. So the title now is Gaussian differential #privacy , w/ application to deep learning.
@weijie444
Weijie Su
5 years
I will be giving a talk at @Yale @yaledatascience this Monday introducing a new notion of differential privacy (). This was inspired by the hypothesis testing interpretation of privacy, joint work with Jinshuo and @Aaroth .
0
5
32
2
1
23
@weijie444
Weijie Su
5 years
Happening soon. Our poster ( #40 ) at #neurips19 will be presented starting 5pm: how can we efficiently do variable selection with confidence? It's a marriage between a #statis method and a signal processing algorithm. Paper , with @CindyRush , Jason, Zhiqi.
Tweet media one
0
5
23
@weijie444
Weijie Su
11 months
As ICML 2023, we will again collect rankings of submissions from authors at ICML 2024. Rankings will be used to assess the modified ratings from the Isotonic Mechanism. It won’t affect decision making, and pls provide rankings for improving peer review in the future!
@zicokolter
Zico Kolter
11 months
Now that at NeurIPS is upon us shortly ... it's time to start planning for ICML😀! Thrilled to serve with @kat_heller @adrian_weller @nuriaoliver as PCs, and @rsalakhu as general chair. Call for papers is here: Intro blog post:
4
24
245
0
8
24
@weijie444
Weijie Su
2 years
Surprised that it led to confusion, which was certainly not what I meant. As a non-Indo-European language speaker, I know how difficult it is to speak and write in English. I speak English with an accent, and I also speak Mandarin with a Wu accent. 1/2
@weijie444
Weijie Su
2 years
I receive large volumes of emails from Chinese students asking for summer internships. The English was often broken, but it has recently become native sounding😅
48
28
1K
2
0
23
@weijie444
Weijie Su
4 years
Can anyone help with this trivial (theoretical) deep learning question? The two images are cropped from the same image; we clearly know they're both cats. But, is there any deep learning THEORY that ensures the prediction is invariant under (proper) cropping? 1/n
Tweet media one
Tweet media two
6
2
22
@weijie444
Weijie Su
2 years
Will be attending #ICML starting tomorrow. Looking forward to seeing many friends for the first time in 3 years!
1
0
23
@weijie444
Weijie Su
5 years
Our Penn Institute for Foundations of Data Science received an NSF award (). Thanks to amazing PI Shivani, co-PIs Hamed, Sanjeev and @Aaroth . Looking forward to advancing data science at @Penn by integrating strength from @PennEngineers , @Wharton and beyond!
1
2
23
@weijie444
Weijie Su
10 months
Heading to Nola for #NeurIPS2023 . Looking forward to seeing many old and new friends.
0
0
22
@weijie444
Weijie Su
5 years
How to accurately track the overall #privacy cost under composition of many private algorithms? Our new paper () offers a new approach using the Edgeworth expansion in the f- #differentialprivacy framework. w/ Qinqing Zheng, @JinshuoD , and Qi Long.
1
4
21
@weijie444
Weijie Su
4 years
I thought today was the weekend but it's still a workday! So, a simple yet neat result (), which completely delineates the range of precision and recall of Lasso, with some counterintuitive outcomes related to the *Donoho-Tanner* phase transition.
2
2
20
@weijie444
Weijie Su
12 days
In 2021, Abel went to Wigderson (TCS); In 2024, Wolf went to Shamir (crypto)... I won't be surprised if someday it becomes a norm that Fields often goes to applied mathematicians.
@NobelPrize
The Nobel Prize
12 days
BREAKING NEWS The Royal Swedish Academy of Sciences has decided to award the 2024 #NobelPrize in Physics to John J. Hopfield and Geoffrey E. Hinton “for foundational discoveries and inventions that enable machine learning with artificial neural networks.”
Tweet media one
1K
14K
34K
0
1
21
@weijie444
Weijie Su
4 years
Many models can explain phenomena in deep learning. OK, but do you see one predicting a *new* surprising phenomenon? Super excited to share a paper "Layer-Peeled Model: Toward Understanding Well-Trained Deep Neural Networks" (). w/ Fang, @HangfengH , Long.
1
4
21
@weijie444
Weijie Su
3 years
That's the issue with virtual conferences! Nevertheless, I spent my entire afternoon yesterday making the poster since NeurIPS asked so. The bottom line is that I can post it on Twitter, which renders my effort slightly more meaningful.
1
0
21
@weijie444
Weijie Su
4 years
New paper: In *Federated f-Differential Privacy* (), we proposed a new privacy notion tailored to the setting where the clients ally in the attack. This privacy concept is adapted from f-differential privacy. w/ Qinqing Zheng, @ShuxiaoC , and Qi Long.
0
4
20
@weijie444
Weijie Su
5 years
Jinshuo will present our work Gaussian Differential #Privacy at #NeurIPS2019 9:05am East Meeting Rooms 8+15. Please stop by!
Tweet media one
@thegautamkamath
Gautam Kamath
5 years
Submissions are open for the NeurIPS 2019 Workshop on Privacy in Machine Learning (PriML 2019)! Anything at the intersection of privacy and machine learning is welcome! #privacy #NeurIPS2019
0
2
13
1
1
19
@weijie444
Weijie Su
5 years
Have lived in the States for 9 years. My first interview in English, however, reveals that I still maintain a Confucian way of thinking.
@dailypenn
The Daily Pennsylvanian
5 years
Three Penn professors have received the Sloan Fellowship which recognizes young scholars for their "unique potential to make substantial contributions to their field."
1
1
7
0
1
20
@weijie444
Weijie Su
9 months
Exciting opportunity for aspiring young data enthusiasts! 🌟 Join the 3-week Data Science Academy at Wharton, directed by Linda Zhao. Dive into the world of data science with hands-on learning on campus. Perfect for high school students. Link:
0
4
17
@weijie444
Weijie Su
6 months
Thanks, @emollick ! Yes, it's fine to train your model on my data, but please pay me accordingly!
@emollick
Ethan Mollick
6 months
There has been a lot of debate on how to deal with copyright issues for AI art, but this interesting paper is one of the first to offer a solution as to how compensation of copyright holders could technically work, given the explainability problem in AI.
Tweet media one
Tweet media two
16
69
291
1
2
20
@weijie444
Weijie Su
4 months
So sad, I took Luca's course on expanders back to my Stanford days
@Pooyahat
Pooya Hatami
4 months
I don't have any words.
0
7
58
0
0
19
@weijie444
Weijie Su
24 days
@a11enhw Yes, the ultimate goal is to understand nature. But we cannot expect any individual to have such a high standard. Being pragmatic is what an average person cares about
1
0
18
@weijie444
Weijie Su
2 months
Very short paper (5 mins read): Our extensive experiments show LLMs improve NTP capability following an exponential law, throughout all layers! It emerges for Transformer, Mamba, RWKV-based LLMs. It's universal!
Tweet media one
2
2
18
@weijie444
Weijie Su
4 years
Congratulations to Emmanuel!
@euromathsoc
European Mathematical Society
4 years
Congratulations to Yves Meyer, Ingrid Daubechies, Terence Tao and Emmanuel Candès, winners of the 2020 Princess of Asturias Award for Technical and Scientific Research!
0
10
25
0
0
16
@weijie444
Weijie Su
5 years
I would feel the same if my least favorable student happened to be the loudest one.
@justthefactsdoc
I Choose Freedom, PhD 🇺🇸🇺🇦✡️♥️🤍💙
5 years
How we all feel watching 45’s recent performances #DrFauci #TrumpIsAnIdiot #TrumpMeltdown
213
2K
8K
0
0
16
@weijie444
Weijie Su
5 years
Will attend #NeurIPS2019 in Vancouver 12/10-12/14. Very much looking forward to meeting old and new friends!
0
0
14
@weijie444
Weijie Su
5 years
Anyone who has a memory of the millennium bug, exactly 20 years ago? For me, I was so nervous that I couldn’t fall asleep that night.
2
1
15
@weijie444
Weijie Su
5 years
Congrats, Dr. Dong!
@Aaroth
Aaron Roth
5 years
Congratulations to Dr. Dong @JinshuoD for defending his PhD "Gaussian Differential Privacy and Related Techniques" earlier today!
Tweet media one
2
4
65
0
0
15
@weijie444
Weijie Su
5 years
@UofTStatSci
U of T Statistical Sciences
5 years
Congratulations to Monica Alexander ( @monjalexander ), @MuratAErdogdu & Stanislav Volgushev for being among this year’s winners of the @UofT Connaught New Researcher Award. 👏👏👏
Tweet media one
0
6
59
1
1
15
@weijie444
Weijie Su
5 years
@tariqnasheed Did anyone call a virus the black virus?
14
0
12
@weijie444
Weijie Su
4 years
Around #JSM2020 ? Want to know more about #privacy research from ml and stats perspectives? Want to enjoy the humor of @XiaoLiMeng1 , please join the session () tmrw at 1pm EST. Speakers @thegautamkamath , Jordan Awan, Feng Ruan and @JinshuoD .
1
3
13
@weijie444
Weijie Su
5 years
Nowadays it’s so common to rediscover the wheels, even for the best minds in our time. The amazing result #eigenvectorsfromeigenvalues () appeared long ago: Principal submatrices II: The upper and lower quadratic inequalities by Thompson and McEnteggert.
1
1
12
@weijie444
Weijie Su
5 years
The @Wharton Stats Dept is seeking applicants for tenure track professor positions ( ). Future colleagues please apply!
0
3
12
@weijie444
Weijie Su
4 years
The poem of privacy
@thegautamkamath
Gautam Kamath
4 years
Our discussant @XiaoLiMeng1 with a poem to conclude the session on Differential Privacy (hosted by @weijie444 ). He gives a shoutout to @TheHDSR , featuring some nice articles on DP, by @DanielOberski and @fraukolos , and by @mbhawes .
Tweet media one
0
3
14
0
0
12
@weijie444
Weijie Su
2 years
The law of equi-separation says that each layer roughly improves a measure of data separation by an equal multiplicative factor. The measure is called separation fuzziness, which is basically the inverse signal-to-noise ratio for classification.
Tweet media one
1
0
12
@weijie444
Weijie Su
2 years
Fairness @ #JSM2022
@zhun_deng
Zhun Deng
2 years
How to make classical fair learning algorithms work for deep learning? How to deal with severe class imbalance and subgroup imbalance? Excited to talk about our new work at JSM on Aug. 8th.
Tweet media one
2
13
64
0
1
12
@weijie444
Weijie Su
2 months
We call it the Equi-Learning Law. Intuitively, one might expect token embeddings to differentiate across LLM layers. Remarkably, a universal geometric law governing tens of thousands of tokens emerges, even in models of immense complexity.
2
1
12
@weijie444
Weijie Su
5 years
Congrats to Lin Xiao for the TOT award #NeurIPS2019 ! Very well deserved. Still miss a bit the summer of 2013 working with Lin at Redmond.
@SebastienBubeck
Sebastien Bubeck
5 years
Congratulations to our colleague Lin Xiao @MSFTResearch for the #NeurIPS2019 test of time award!!! Online convex optimization and mirror descent for the win!! (As always? :-).)
1
45
248
1
0
12
@weijie444
Weijie Su
5 years
The most fundamental recipe in #privacy research is perhaps to understand “how private are private algorithms.” Here comes a nice read: . BTW, the author @JinshuoD is on the job market!
@JinshuoD
Jinshuo Dong
5 years
My very first tweet about my very first blog post: How Private Are Private Algorithms? In fact, this question is answered 87 years ago... #OldiesButGoodies #differentialprivacy
2
0
12
0
2
12
@weijie444
Weijie Su
2 years
Congratulations to @daniela_witten on winning the COPSS Presidents’ Award! #JSM2022
0
2
12
@weijie444
Weijie Su
2 years
What a great year for Penn CIS!
@vijay_r_kumar
Vijay Kumar
2 years
Delighted to welcome new faculty @SurbhiGoel_ , @LingjieLiu1 , @RyanMarcus , @zkproofs , Erik Waingarten and Eric Wong to @PennCIS @pennengineers . We are on a roll...
2
17
173
0
0
11
@weijie444
Weijie Su
3 months
Does anyone have a photo of the Sequoia lounge of Stanford’s statistics department showing Pao-Lu Hsu in the hall of fame? It must be taken several years ago. Thanks!!
0
3
11