KuninDaniel Profile Banner
Daniel Kunin Profile
Daniel Kunin

@KuninDaniel

Followers
728
Following
158
Statuses
67

PhD student @ICMEStanford Creator @SeeingTheory

Stanford University
Joined December 2020
Don't wanna be here? Send us removal request.
@KuninDaniel
Daniel Kunin
5 months
🌟Announcing NeurIPS spotlight paper on the transition from lazy to rich🔦 We reveal through exact gradient flow dynamics how unbalanced initializations promote rapid feature learning co-led @AllanRaventos and @ClementineDomi6 @FCHEN_AI @klindt_david @SaxeLab @SuryaGanguli
5
41
248
@KuninDaniel
Daniel Kunin
1 month
RT @MasonKamb: Excited to finally share this work w/ @SuryaGanguli. Tl;dr: we find the first closed-form analytical theory that replicates…
0
152
0
@KuninDaniel
Daniel Kunin
2 months
Come check out our #NeurIPS2024 spotlight poster on feature learning tomorrow! 📍East Exhibit Hall A-C #2102 📅Thu 12 Dec 4:30 p.m. — 7:30 p.m. PST
@KuninDaniel
Daniel Kunin
5 months
🌟Announcing NeurIPS spotlight paper on the transition from lazy to rich🔦 We reveal through exact gradient flow dynamics how unbalanced initializations promote rapid feature learning co-led @AllanRaventos and @ClementineDomi6 @FCHEN_AI @klindt_david @SaxeLab @SuryaGanguli
0
7
50
@KuninDaniel
Daniel Kunin
4 months
@ziv_ravid @GalantiTomer @NiketPatel91154 Another related work that studies the relationship between SGD and the rank & sparsity of weights (although seems your analysis of local rank decreasing doesn't depend on noisy gradients?)
0
0
4
@KuninDaniel
Daniel Kunin
5 months
RT @klindt_david: Great job, it was an honor being part of this amazing project! Congrats to the team 💪
0
2
0
@KuninDaniel
Daniel Kunin
5 months
Also, big shoutout to @yasamanbb, @CPehlevan, and @HSompolinsky for coordinating last year's 'Deep Learning from Physics and Neuroscience' program @KITP_UCSB. Our amazing team met there, and this project is a direct result of the conversations we had!
0
1
6
@KuninDaniel
Daniel Kunin
5 months
Applying our function space analysis to shallow ReLU networks, we find that rapid feature learning occurs from unbalanced initializations that promote faster learning in early layers, driving a large change in activation patterns, but a small change in parameter space
Tweet media one
1
1
7
@KuninDaniel
Daniel Kunin
5 months
We find three regimes in function space: 1. lazy akin to linear regression, 2. rich akin to silent alignment (Atanasov et al. 2021), 3. delayed-rich initially lazy followed by rich We extend this analysis (with mirror flows and implicit biases) to wide & deep linear networks
Tweet media one
1
1
4
@KuninDaniel
Daniel Kunin
5 months
We derive exact gradient flow solutions for a minimal two-layer linear model displaying lazy and rich learning, which reveals that the relative scale between layers influence feature learning through conserved quantities that constrain the geometry of learning trajectories
Tweet media one
1
1
7
@KuninDaniel
Daniel Kunin
5 months
Reproducing Fig.1 in Chizat et al. 2019 we find that even at small overall scale, the relative scale between layers can transition the network between rich and lazy learning and the best generalization occurs at small scale and large relative scale!
Tweet media one
1
1
9
@KuninDaniel
Daniel Kunin
7 months
Really cool theory project on feature learning. If you are at the HiLD workshop @icmlconf check it out!
@AllanRaventos
Allan Raventós
7 months
Interested in exactly solvable models of learning dynamics and implicit bias? Come check out our "Get Rich Quick" poster at the HiLD Workshop @icmlconf at 10am! With @KuninDaniel, myself, @ClementineDomi6, @FCHEN_AI, @klindt_david, @SaxeLab, and @SuryaGanguli.
Tweet media one
1
1
15
@KuninDaniel
Daniel Kunin
7 months
RT @AllanRaventos: Interested in exactly solvable models of learning dynamics and implicit bias? Come check out our "Get Rich Quick" poster…
0
13
0
@KuninDaniel
Daniel Kunin
1 year
RT @DebOishi: Reminder! Happening Tomorrow! @ELLISforEurope
0
1
0
@KuninDaniel
Daniel Kunin
1 year
To get the zoom link and get notified about other interesting talks check out
0
1
2
@KuninDaniel
Daniel Kunin
1 year
RT @DebOishi: We are delighted to announce our next speakers for the @ELLISforEurope RG are fr @Stanford, the authors of the NeurIPS2023 pa…
0
4
0
@KuninDaniel
Daniel Kunin
1 year
@ItsNeuronal Neat! In part of our work we track empirically the degree of redundancy in networks through training. Maybe your metric would help. Let's chat about it tomorrow at the poster session!
0
0
2