Sanae Lotfi Profile
Sanae Lotfi

@LotfiSanae

Followers
2,163
Following
318
Media
41
Statuses
373

PhD candidate at NYU, research intern @MSFTResearch | @GoogleDeepMind and @MSFTResearch Fellow | Prev. @MetaAI (FAIR) and @AmazonScience

New York City, NY
Joined August 2020
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@LotfiSanae
Sanae Lotfi
2 years
This Moroccan Arab Muslim first-generation woman just gave her first long talk for her award-winning paper at #ICML2022 ! I dedicate this achievement to all the underrepresented groups that I proudly represent! So overwhelmed by all the support that I received! Many thanks! 1/2
Tweet media one
Tweet media two
61
67
1K
@LotfiSanae
Sanae Lotfi
2 years
I'm so grateful and honored to receive the Microsoft Research PhD Fellowship @MSFTResearch !!🥳 This fellowship means that my research and that of my group led by @andrewgwils is recognized as meaningful and impactful for the machine learning community! 1/3
Tweet media one
38
31
745
@LotfiSanae
Sanae Lotfi
2 years
I'm so proud that our paper on the marginal likelihood won the Outstanding Paper Award at #ICML2022 !!! Congratulations to my amazing co-authors @Pavel_Izmailov , @g_benton_ , @micahgoldblum , @andrewgwils 🎉 Talk on Thursday, 2:10 pm, room 310 Poster 828 on Thursday, 6-8 pm, hall E
Tweet media one
@andrewgwils
Andrew Gordon Wilson
2 years
I'm happy that this paper will appear as a long oral at #ICML2022 ! It's the culmination of more than a decade of thinking about when the marginal likelihood does and doesn't make sense for model selection and hyper learning, and why. It was also a great collaborative effort.
4
19
219
13
33
322
@LotfiSanae
Sanae Lotfi
1 year
We talk a lot about LLMs but you know what we don’t talk enough about? PhD student mental health, community building, and effective research practices! That’s why Yaniv Yacoby and I are organizing a workshop at NYU entitle “Everything I wish I knew as I navigated my PhD”. 1/3 👇
Tweet media one
5
24
235
@LotfiSanae
Sanae Lotfi
2 years
I appreciate this apology, but I am saddened that you are still implying a lack of scientific integrity. 1/N
@BlackHC
Andreas Kirsch 🇺🇦
2 years
I'm sorry for the negative attention I brought on @LotfiSanae in what should have been a joyous moment and who has authored an amazing & exciting paper on a complex topic. My intention was scientific integrity, but I lost sight of the broader context of this academic debate &
1
0
56
7
11
187
@LotfiSanae
Sanae Lotfi
9 months
It was a great honor to be distinguished as a Rising Star in ML by @ml_umd 🥳 Many thanks to the organizers and professors with whom I had great discussions! Next stop: #NeurIPS2023 ; so excited to share our work on generalization bounds for LLMs and host @MuslimsinML there!
Tweet media one
Tweet media two
Tweet media three
6
6
143
@LotfiSanae
Sanae Lotfi
1 year
Went to Harvard to give a talk about our work on the marginal likelihood and PAC-Bayes bounds, ended up meeting @ConanOBrien there. It turns out he’s a big Bayesian!
Tweet media one
2
2
143
@LotfiSanae
Sanae Lotfi
8 months
Can LLMs generalize meaningfully beyond their training data? We answer this question by computing the first non-vacuous generalization bounds for LLMs. Wanna learn more? Find us the M3L and SSL workshops 😉 w/ @m_finzi @KuangYilun @timrudner @micahgoldblum @andrewgwils
Tweet media one
1
13
135
@LotfiSanae
Sanae Lotfi
2 years
Had a great time presenting our work on PAC-Bayes bounds. Thanks everyone who stopped by; many insightful discussions! I’m at #NeurIPS2022 this week, ping me if you want to discuss generalization in deep learning or Taylor Swift concert tickets! Paper:
Tweet media one
5
7
122
@LotfiSanae
Sanae Lotfi
2 years
🚨 #NeurIPS2022 poster today: 4-6pm, Hall J #306 🚨 Why do CNNs generalize so much better than MLPs? Why can neural networks fit random labels and still generalize? What is the value of encoding invariances in our models? 1/N
2
16
104
@LotfiSanae
Sanae Lotfi
1 year
We’re organizing Muslims in ML @NeurIPSConf this year 🥳 Submit a two-page abstract if you would like to showcase and present your work at the workshop by September 22 ⏰
@MuslimsinML
Muslims in ML Workshop
1 year
Excited to announce that the Muslims in Machine Learning (MusIML) workshop is back @NeurIPSConf ! If you self-identify as Muslim, or work on research that address challenges faced by Muslims, we'd love to showcase your work. Submit an abstract by Sep 22:
1
5
31
2
17
98
@LotfiSanae
Sanae Lotfi
3 years
This competition resulted in hundreds of submissions and many interesting solutions that will be presented by the winners 🏅 tomorrow starting 1pm ET 🎉 Join us to discover these scalable approximate inference methods for Bayesian Deep Learning:
@andrewgwils
Andrew Gordon Wilson
3 years
I heard a rumour there is this amazing Approximate Inference in Bayesian Deep Learning competition at #NeurIPS2021 tomorrow, starting at 1 pm ET. From what I understand, the winners will be revealing their solutions, and the link to join is . 🤫
Tweet media one
0
43
193
3
17
90
@LotfiSanae
Sanae Lotfi
4 years
Very excited to announce that I will be joining the Center for Data Science at New York University as a PhD student. I am also very honored to receive the DeepMind Fellowship. Thank you @DeepMind and @NYUDataScience !
@NYUDataScience
NYU Data Science
4 years
Meet Sanae Lotfi, a PhD CDS student who is one of three DeepMind Fellows joining us this academic year. We’re also proud to note that Sanae has recently won the McKinsey First Generation Achievement Award earlier this month. Learn more on our blog:
3
23
149
8
4
87
@LotfiSanae
Sanae Lotfi
8 months
So happy we ended up having a last minute talk by @MarzyehGhassemi on Ethical AI in Health @MuslimsinML ! I am a big fan of @MarzyehGhassemi ’s work and I also like her so much as a female researcher, mentor, and role model. Very thankful I got a front seat to her talk! 🥳
Tweet media one
Tweet media two
Tweet media three
2
8
82
@LotfiSanae
Sanae Lotfi
11 months
Being a Visiting Researcher at Meta has been amazing, but it’s coming to an end! I’m looking for internship opportunities for Summer 2024, possibly my last PhD internship. I’m interested in the foundations of deep learning with a recent focus on LLMs. Reach out if you see a fit!
1
5
74
@LotfiSanae
Sanae Lotfi
2 years
That’s not true. We re-ran any experiments affected by a minor bug (one panel only in the main text) in the camera ready, and the qualitative results remain unchanged. We engaged extensively with all your questions via email, up through you publicizing your review.
@BlackHC
Andreas Kirsch 🇺🇦
2 years
Not sure if this paper deserves such an award... After pointing out a bug, the authors reran some experiments. But the results in the revisions contradict each other and the authors never replied to my Qs in the email thread again, so yay🔥 PC review:
3
1
28
2
1
58
@LotfiSanae
Sanae Lotfi
1 year
Wrapped up my Boston visit with a talk at @MIT_CSAIL ! Many thanks to all the brilliant researchers who invited me or made time to chat! Thank you @MarzyehGhassemi , @irenetrampoline and Yaniv for being amazing hosts as well! Bonus: I discovered more Taylor Swift fans in academia!
Tweet media one
0
0
56
@LotfiSanae
Sanae Lotfi
2 years
It has been extremely stressful not to be able to defend our work on Twitter, just to avoid an unproductive discourse. We wanted to enjoy ICML but were instead stressed out all week glued to Twitter as our integrity was called into question. 11/N
1
2
49
@LotfiSanae
Sanae Lotfi
2 years
I am at ICML this week and would love to meet and chat with some of you there! If you would like to talk about generalization in deep learning and other topics, ping me here or find me somewhere in the Baltimore Convention Center! #ICML2022
0
2
48
@LotfiSanae
Sanae Lotfi
1 year
Our extended paper on the marginal likelihood was accepted to JMLR 🎉 In this version, we expand on the marginal likelihood’s connection to PAC-Bayes bounds, its approximations, and its use for architecture search. Check the thread👇for more details and stay tuned for more!
@andrewgwils
Andrew Gordon Wilson
1 year
Last year at ICML, we presented marginal likelihood pathologies in model selection and hyper learning. We now have a 60 page JMLR extension featuring: 1) should we be comforted by connections with PAC-Bayes? 2) approximations; 3) architecture search. 1/16
3
18
95
3
4
44
@LotfiSanae
Sanae Lotfi
2 years
Thrilled that this work with @Pavel_Izmailov , @g_benton_ , @micahgoldblum , and @andrewgwils got accepted for a long oral presentation at ICML 2022. It was fun to investigate and articulate the subtle ways in which the marginal likelihood is different from generalization!
Tweet media one
2
3
44
@LotfiSanae
Sanae Lotfi
2 years
Come join me tomorrow, alongside @polkirichenko and @shiorisagawa , to discuss methods and challenges in tackling distribution shift in deep learning. No matter what you are working on, come share with us how distribution shift affects your work and/or its applications! #ICML2022
@polkirichenko
Polina Kirichenko
2 years
We are excited to present our breakout session on robustness to distribution shift at @WiMLworkshop @icmlconf together with @shiorisagawa @LotfiSanae ! Join our session & discussion tomorrow, Monday July 18, at 11am at the Exhibit Hall G at Level 100 Exhibition Halls 🙂 #ICML2022
Tweet media one
1
25
269
2
2
32
@LotfiSanae
Sanae Lotfi
2 years
Just know Andreas that none of us hold a grudge against you and hope we can interact in a friendly way going forward. Thanks for extending an olive branch! 12/N, N=12.
2
0
31
@LotfiSanae
Sanae Lotfi
2 years
🔥Talk tomorrow, Dec. 5, at the @NorthAfricansML workshop, 10-11 AM GMT+1 🔥 "Are the Marginal Likelihood and PAC-Bayes Bounds the right proxies for Generalization?" 1/11
Tweet media one
1
5
30
@LotfiSanae
Sanae Lotfi
2 years
Did you engage with the authors of the papers that you review today? Channeling my inner @kchonyc 😅
0
3
29
@LotfiSanae
Sanae Lotfi
2 years
Big thanks to my amazing co-authors @Pavel_Izmailov , @g_benton_ , @micahgoldblum , and my incredible advisor @andrewgwils who put a lot of effort into this work and supported me all along! Want to learn more about this work? check our poster 828, hall E between 6 and 8 pm EST!
0
1
26
@LotfiSanae
Sanae Lotfi
8 months
When is it responsible to deploy a general purpose AI system? @OpenAI ’s @_lamaahmad is answering this question right now @MuslimsinML ! Very exciting talk about identifying the potential risks of @OpenAI ’s systems through Red Teaming!
Tweet media one
1
4
20
@LotfiSanae
Sanae Lotfi
2 years
This work was a team effort. We all care about scientific rigor and stand by the work. I do not think it is acceptable to single out any other author and direct attacks onto them. We all take responsibility and are committed to addressing questions about it. 2/N
1
1
19
@LotfiSanae
Sanae Lotfi
2 years
We will release this exciting work soon w/ @m_finzi , @snymkpr , Andres Potapczynski, @micahgoldblum and @andrewgwils ! 🔥
@DimitrisPapail
Dimitris Papailiopoulos
2 years
@thegautamkamath Best title so far "PAC-Bayes Compression Bounds So Tight That They Can Explain Generalization" I'm intrigued..
2
0
11
0
0
19
@LotfiSanae
Sanae Lotfi
2 years
@biotin10000mcg "A formal definition of a first-generation college student is a student whose parent(s) did not complete a four-year college or university degree." My parents actually never went to school. So proud of them for raising us and educating us to appreciate science this much!
1
0
18
@LotfiSanae
Sanae Lotfi
2 years
I'd like to thank my incredible advisor @andrewgwils for his continuous support, my supportive colleagues @Pavel_Izmailov and @micahgoldblum for their feedback on my proposal, and my amazing masters advisor @69alodi and Prof. Julia Kempe for supporting my application! 2/3
2
0
16
@LotfiSanae
Sanae Lotfi
9 months
Very cool work! I was just yesterday mentioning during the lab I TA that we have more intuition about the function space than the parameter space, and that it would be cool to thoroughly investigate function space MAP estimation — we don’t have to wonder anymore! 🥳
@timrudner
Tim G. J. Rudner
9 months
When training machine learning models, should we learn most likely parameters—or most likely functions? We investigate this question in our #NeurIPS2023 paper and made some fascinating observations!🚀 Paper: w/ @ShikaiQiu @psiyumm @andrewgwils 🧵1/10
Tweet media one
3
23
139
0
4
17
@LotfiSanae
Sanae Lotfi
10 months
Some of the PhD struggle is necessary and a byproduct of growth, but some of it is not and needs to be dealt with effectively and compassionately! We’ll try to figure which is which with amazing peers @SCSatCMU 👇 Let’s keep discussing PhD student mental health as a community!
@narijohnson
Nari Johnson
10 months
📣📣 @SCSatCMU PhD students: I'm so excited to share that @charvvvv_ , @LotfiSanae , Yaniv Yacoby, and I are bringing this amazing workshop to CMU! Join us to reflect on PhD student mental health, community-building🫂, and academic culture. (1/n) 👇
1
16
89
0
2
15
@LotfiSanae
Sanae Lotfi
2 years
I just donated too! Thanks @bneyshabur for bringing this into my feed! So grateful and proud of @ml_collective and @DeepIndaba for their effort to support African researchers! P.S: you can also help with your time through the Indaba Mentorship Programme:
@ml_collective
ML Collective
2 years
8 researchers from Nigeria got accepted into Deep Learning Indaba @DeepIndaba but couldn't afford the trip. Access to research, mentorship, and networking opportunities like this is vital for early-stage researchers. You can help make their trips possible!
7
199
245
1
4
15
@LotfiSanae
Sanae Lotfi
2 years
@shortstein The reviewer in me likes it, the author in me does not. What a dilemma! In general, I believe we should treat the papers we review the same way we want our papers to be treated, so I'll probably pass on this option whenever possible😅
1
0
14
@LotfiSanae
Sanae Lotfi
2 years
I also don’t appreciate your suggestion that we did not deserve the award or that people should complain to ICML. We are proud of this work, and feel it is deserving of such support. 3/N
1
0
13
@LotfiSanae
Sanae Lotfi
2 years
Big thanks to all of my collaborators, labmates and mentors who inspire every day and to my amazing partner for his ample feedback and help with my research proposal! 3/3
3
1
11
@LotfiSanae
Sanae Lotfi
2 years
We explore many ways in which the question ``how likely is my prior to generate the training data?’’ differs from the question we care about ``how likely is my posterior to model withheld points from the same distribution?’’. We also distinguish this from hypothesis testing. 5/N
1
1
10
@LotfiSanae
Sanae Lotfi
1 year
The workshop is happening on Monday April 3rd at NYU (60 5th Ave) from 12-2pm and It’s open to all NYU PhD students. You can find the link to register in your inbox or DM me to send it to you! 2/3
1
0
10
@LotfiSanae
Sanae Lotfi
2 years
@BlackHC That’s also incorrect. We did not cherry-pick the experimental setting as I used exactly the same models and checkpoints in both arxiv versions. I did not do any re-training after we fixed the bug. I’m sorry but we don’t think you’re engaging in good faith. Thanks for the review.
1
0
10
@LotfiSanae
Sanae Lotfi
2 years
We are agnostic as to whether the CLML is better or not than validation, and this is not at all the focus of our work. Sometimes it will be, others it won’t. If one can do something simpler and achieve better performance, we are supportive of that. There is nothing to hide. 9/N
1
1
10
@LotfiSanae
Sanae Lotfi
2 years
The marginal likelihood is regularly used as a proxy for generalization, in seminal early work, and in recent papers, for hyperparameter learning, and model selection. While it can sometimes be practical for these purposes… 4/N
1
1
9
@LotfiSanae
Sanae Lotfi
2 years
So excited to be presenting our work on PAC-Bayes bounds 🔥tonight🔥 alongside an amazing lineup of female speakers! Join us if you want to hear more about this work w/ @m_finzi , @snymkpr , A. Potapczynski, @micahgoldblum , @andrewgwils ; or to chat! Paper:
@mahtabm81
Mahtab Mirmomeni, PhD
2 years
Join us at the Women in AI Ignite today 6:00 pm CT @NeurIPSConf Great lineup #womeninai A room full of inspiring female figures reminds me of how important it is to support women around the world to live freely and reach their true potential. #WomenLifeFreedom @anoushnajarian
Tweet media one
0
7
24
0
0
8
@LotfiSanae
Sanae Lotfi
9 months
Been using CoLA pretty much since it came out and I like it a lot! It makes large matrix inversion, eigenspectrum decomposition and accessing other quantities I need for my research much more memory-efficient and scalable. Works with PyTorch, JAX and supports GPUs. Try it out!
@m_finzi
Marc Finzi
9 months
Like differentiation, numerical linear algebra is a notorious bottleneck that is ripe for automation and acceleration. At #NeurIPS2023 we introduce Compositional Linear Algebra (CoLA)!🥤 w/ A. Potap, G. Pleiss, @andrewgwils . 🧵 [1/7]
Tweet media one
4
44
270
0
0
7
@LotfiSanae
Sanae Lotfi
1 year
I’m so excited to be part of this workshop and host Yaniv Yacoby who created several successful initiatives around improving PhD student mental health and community at Harvard. I hope this workshop results in follow-up initiatives to strengthen our graduate community at NYU! 3/3
1
0
8
@LotfiSanae
Sanae Lotfi
2 years
Big thanks to the @NorthAfricansML workshop organizers as well. It's great to see a NeurIPS workshop that is fully accessible in the North African timezone. Hopefully we get to see NeurIPS in Africa soon! 11/11
0
2
7
@LotfiSanae
Sanae Lotfi
2 years
@FelixHill84 @BlackHC I have no doubt! I realize how Twitter can make things escalate quickly and amplify misunderstandings, especially when it comes to technical and nuanced discussions!
0
0
7
@LotfiSanae
Sanae Lotfi
3 years
The final phase of our #NeurIPS 2021 competition "Approximate Inference in Bayesian Deep Learning" has officially started! You can test the fidelity of your approximate inference procedure with the opportunity to present it at the #NeurIPS Bayesian deep learning workshop (+$)!
@bdl_competition
NeurIPS Approximate Inference in BDL Competition
3 years
We are officially entering the final stage of the competition! Now is a great time to get involved :) Competition website: Submission system:
1
5
23
0
2
7
@LotfiSanae
Sanae Lotfi
2 years
In short, the LML has already been compared extensively to various types of validation (Minka on automatic PCA, Rasmussen for GPs..), and in some cases is preferable, in others it is not. As we show, the CLML will typically be more aligned with generalization than the LML. 8/N
1
0
7
@LotfiSanae
Sanae Lotfi
2 years
We did not remove the comparison to non-BMA cross validation from the CR. We just never had it in arxiv v1 and we built the CR on v1 and not v2. There was no deliberate action. We have no issue with BMA validation doing better for architecture search & will highlight in v3. 10/N
1
0
6
@LotfiSanae
Sanae Lotfi
2 years
We are excited to see you at the poster session: Tuesday 4-6pm, Hall J #306 9/N, N=9
Tweet media one
0
0
6
@LotfiSanae
Sanae Lotfi
2 years
Your main concern at this point seems to be whether the CLML is competitive with validation loss. We have already largely responded to this point. See the attached reply we had in your blog. 7/N
Tweet media one
1
1
6
@LotfiSanae
Sanae Lotfi
2 years
Although it is not the focus of our paper, we show how trivially modifying the log marginal likelihood (LML), for the CLML, can lead to a better correlation with generalization for model comparison and hyper learning. We stand by this observation. 6/N
1
1
6
@LotfiSanae
Sanae Lotfi
2 years
Where do I send the $5M check?
@ZerzarBukhari
Zerzar Bukhari
2 years
OMG even start-up founders are getting automated
Tweet media one
0
0
3
0
0
5
@LotfiSanae
Sanae Lotfi
2 years
With these bounds we find that models which encode locality and equivariances are more compressible when these biases match the structure in the dataset, and hence have tighter generalization bounds. 6/N
Tweet media one
1
0
5
@LotfiSanae
Sanae Lotfi
2 years
We explore these questions in our new #NeurIPS2022 paper “PAC-Bayes Compression Bounds So Tight That They Can Explain Generalization” w/ @m_finzi , @snymkpr , Andres Potapczynski, @micahgoldblum and @andrewgwils ! 2/N
1
0
5
@LotfiSanae
Sanae Lotfi
2 years
@nsaphra @BlackHC @hewal_oscar Thanks for pointing that out Naomi. The camera ready has updated figures indeed and a discussion about how we never considered cross validation to be a competing metric (see appendix L).
1
0
5
@LotfiSanae
Sanae Lotfi
4 years
@69alodi @NYUDataScience @polymtl @DS4DM It was a pleasure to be supervised by you and Dominique. I am so grateful to you, @69alodi for all the freedom and guidance you offered. I also feel immense gratitude towards our team at @DS4DM . Congratulations on creating this positive, supportive, and collaborative environment.
0
1
5
@LotfiSanae
Sanae Lotfi
4 years
Join the Q&A session at 18:30 ET to learn more about our work "Stochastic Damped L-BFGS with Controlled Norm of the Hessian Approximation", with Tiphaine Bonniot, Dominique Orban and @69alodi OPT2020 Workshop schedule: Paper:
1
1
5
@LotfiSanae
Sanae Lotfi
2 years
Join us tomorrow at the workshop to discuss these results and open questions! Big thanks to collaborators: @Pavel_Izmailov , @g_benton_ , @m_finzi , @snymkpr , A. Potapczynski, @micahgoldblum , and @andrewgwils . 10/11
1
1
4
@LotfiSanae
Sanae Lotfi
8 months
@OpenAI @_lamaahmad @MuslimsinML Lots of great questions, impressed by how much work goes into alignment and risk assessment @OpenAI !
Tweet media one
Tweet media two
0
1
4
@LotfiSanae
Sanae Lotfi
2 years
@khademinori Thank you, Milad!
0
0
4
@LotfiSanae
Sanae Lotfi
4 years
Thank you very much @DS4DM . Very honoured to receive this award. Thank you @McKinsey for your trust.
@DS4DM
DS4DM
4 years
Congratulations to our master student Sanae Lotfi to receive the McKinsey First Generation Achievement Award!
1
3
12
0
0
4
@LotfiSanae
Sanae Lotfi
2 years
@zicokolter Thank you very much, Zico! I feel encouraged by your support!
1
0
3
@LotfiSanae
Sanae Lotfi
10 months
@andrewgwils @gruver_nate @m_finzi @ShikaiQiu Very happy to see the final paper and extensive exploration, well done 👏
0
0
3
@LotfiSanae
Sanae Lotfi
3 years
@Khalid_Montreal First place goes to @riken_en ( @tmoellenhoff , Y. Shen, @ShhhPeaceful , @PeterNickl_ , @EmtiyazKhan ) in both tracks! Second place goes to @niket096 and A. Thin are in the extended and tie with @ArnaudDelaunoy for second in the light track. We'll hear from them all tomorrow 🎉
0
0
3
@LotfiSanae
Sanae Lotfi
2 years
Our bounds also explain why transfer learning works; the difference between a pre-trained checkpoint and the fine-tuned model is far more compressible than the difference between a random initialization and the corresponding model trained from scratch. 8/N
1
0
3
@LotfiSanae
Sanae Lotfi
2 years
@andrewgwils Thanks a lot Andrew! I appreciate your support immensely!
0
0
3
@LotfiSanae
Sanae Lotfi
4 years
@akazachk Good luck for your next journey Aleks! I am sure you will be a great professor and a fantastic mentor to all your future students!
1
0
3
@LotfiSanae
Sanae Lotfi
2 years
@ZerzarBukhari Thank you for being as supportive as it gets!
0
0
3
@LotfiSanae
Sanae Lotfi
2 years
Very interesting work on how to properly do transfer learning by capturing much more than just the initialization from pre-trained models.
@micahgoldblum
Micah Goldblum
2 years
Typical transfer learning pipelines involve initializing at pre-trained weights and hoping that relevant learned information magically transfers even when the weights change during fine-tuning. But you can transfer so much more than just initialization! 1/4
1
31
177
0
0
3
@LotfiSanae
Sanae Lotfi
2 years
We achieve state-of-the-art bounds with data-independent priors by considerably compressing our models. 4/N
Tweet media one
1
0
2
@LotfiSanae
Sanae Lotfi
2 years
On the other hand, when these structures are broken, the models are less compressible and hence generalize worse even than generic MLPs. We also find that checkpoints which fit random labels are not compressible, explaining why such flexible models generalize well. 7/N
Tweet media one
1
0
2
@LotfiSanae
Sanae Lotfi
2 years
By training in a low-dimensional affine subspace and using aggressive quantization, we design a practical method to achieve the trade-off between the training error and the KL divergence as we change the subspace dimension. 5/N
Tweet media one
1
0
2
@LotfiSanae
Sanae Lotfi
2 years
PAC-Bayes bounds are another expression of Occam’s razor where simpler descriptions of the data generalize better, that can be used to understand generalization in deep learning. 5/11
@LotfiSanae
Sanae Lotfi
2 years
🚨 #NeurIPS2022 poster today: 4-6pm, Hall J #306 🚨 Why do CNNs generalize so much better than MLPs? Why can neural networks fit random labels and still generalize? What is the value of encoding invariances in our models? 1/N
2
16
104
1
0
2
@LotfiSanae
Sanae Lotfi
1 year
Finally got to watch this episode and learned a lot! Excited for the next one! 🔥🤗
@ZerzarBukhari
Zerzar Bukhari
1 year
Ep 3 is all about the chip wars 🔥
0
0
4
0
0
2
@LotfiSanae
Sanae Lotfi
2 years
@felix_led Thank you very much Felix!
0
0
2
@LotfiSanae
Sanae Lotfi
2 years
@SMRhadou aaaww that truly warms my heart! Likewise here, so proud of you here!
0
0
2
@LotfiSanae
Sanae Lotfi
2 years
@arpitabiswas777 @NeurIPSConf Great talk today! 🔥
0
0
2
@LotfiSanae
Sanae Lotfi
4 years
@Khalid_Montreal @DS4DM Thank you very much @Khalid_Montreal ! Your help and support helped a lot with that !
0
0
2
@LotfiSanae
Sanae Lotfi
2 years
Work by Germain et al., 2017 () has shown that minimizing the PAC-Bayes bound is equivalent to maximizing the marginal likelihood when choosing the loss function to be negative log-likelihood. 8/11
2
0
2
@LotfiSanae
Sanae Lotfi
2 years
@micahgoldblum You should at least add her in the acknowledgement paragraph for moral support and feedback (pretty sure she gives you feedback, doesn't she?)
1
0
2
@LotfiSanae
Sanae Lotfi
2 years
Previous works choose between loose data-independent bounds and data-dependent priors which poorly explain generalization. Instead, data-dependent bounds essentially say that neural nets generalize because they have good validation accuracy … 3/N
Tweet media one
1
0
2
@LotfiSanae
Sanae Lotfi
1 year
@FalaahArifKhan So happy you’ll be there!
0
0
1
@LotfiSanae
Sanae Lotfi
2 years
0
0
1
@LotfiSanae
Sanae Lotfi
3 years
@tarantulae We had twelve teams participate in each of the two phases of the competition
1
0
1
@LotfiSanae
Sanae Lotfi
3 years
@EmtiyazKhan Thank you for participating, we are excited to learn more about your solution tomorrow!
0
0
1
@LotfiSanae
Sanae Lotfi
2 years
@ShresthaRobik @polkirichenko @shiorisagawa I don't think so but we can try to make the slides available. @polkirichenko @shiorisagawa
1
0
1
@LotfiSanae
Sanae Lotfi
2 years
@FelixHill84 Thank you very much, Felix. I really appreciate your kind and supportive words!
0
0
1
@LotfiSanae
Sanae Lotfi
2 years
0
0
1
@LotfiSanae
Sanae Lotfi
2 years
@kasshout Thank you very much!
0
0
1
@LotfiSanae
Sanae Lotfi
2 years
@Jessenathank Thank you!
0
0
1
@LotfiSanae
Sanae Lotfi
2 years
0
0
1
@LotfiSanae
Sanae Lotfi
2 years
@69alodi I am very lucky that I have and had your support as my master's co-supervisor and my mentor to this day! The trust, support and mentorship you keep offering me are so much appreciated! Thank you very much Andrea!
1
0
1