Maithra Raghu Profile
Maithra Raghu

@maithra_raghu

Followers
17,379
Following
481
Media
76
Statuses
417

Cofounder and CEO @Samaya_AI . Formerly Research Scientist Google Brain ( @GoogleAI ), PhD in ML @Cornell .

Joined July 2017
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@maithra_raghu
Maithra Raghu
1 year
Does One Large Model Rule Them All? New post with @matei_zaharia and @ericschmidt on the future of the AI ecosystem. Our key question: does the rise of large, general AI models means the future AI ecosystem is dominated by a single general model? ⬇️
Tweet media one
27
85
292
@maithra_raghu
Maithra Raghu
4 years
A Survey of Deep Learning for Scientific Discovery To help facilitate using DL in science, we survey a broad range of deep learning methods, new research results, implementation tips & many links to code/tutorials Paper Work with @ericschmidt Thread⬇️
Tweet media one
15
432
1K
@maithra_raghu
Maithra Raghu
3 years
Do Vision Transformers See Like Convolutional Neural Networks? New paper The successes of Transformers in computer vision prompts a fundamental question: how are they solving these tasks? Do Transformers act like CNNs, or learn very different features?
Tweet media one
27
253
1K
@maithra_raghu
Maithra Raghu
4 years
Thrilled to share that I successfully defended my PhD today!! This milestone wouldn't have been possible without the support and guidance of my collaborators, mentors, friends and family -- thank you so much!!! Thanks also to everyone who attended my (virtual) defense!
Tweet media one
73
21
1K
@maithra_raghu
Maithra Raghu
4 years
New blog post: "Reflections on my (Machine Learning) PhD Journey" 2020 has marked the end of my six year PhD journey, filled with struggles, success and evolution of personal & research perspectives. In the post I share experiences and lessons learned ⬇️
Tweet media one
15
146
954
@maithra_raghu
Maithra Raghu
2 years
After almost 6.5 years, I left Google Brain earlier this month. It's been an incredible journey of gaining insights on many exciting areas of machine learning, and watching the field grow and evolve (so much, so quickly!)
12
10
597
@maithra_raghu
Maithra Raghu
2 years
A few months ago I left Google Brain to pursue my next adventure: building @samaya_AI ! We're excited to bring the latest AI advances to the Knowledge Discovery process!
17
21
464
@maithra_raghu
Maithra Raghu
10 months
Many of these trends don't hold. Last week we celebrated @geoffreyhinton 's retirement, and a few weeks earlier saw @kkariko receive the Nobel Prize. Their research took decades to come together, and had enormous impact at a world scale. We'd be much worse off if they'd pivoted!
@_jasonwei
Jason Wei
10 months
Enjoyed visiting UC Berkeley’s Machine Learning Club yesterday, where I gave a talk on doing AI research. Slides: In the past few years I’ve worked with and observed some extremely talented researchers, and these are the trends I’ve noticed: 1. When
45
275
2K
12
35
423
@maithra_raghu
Maithra Raghu
6 years
Excited to be one of the Forbes 30 under 30 in Science! #ForbesUnder30 Full list:
29
20
357
@maithra_raghu
Maithra Raghu
6 years
We're releasing tutorials on our work using CCA to compare and probe representations in deep neural networks: There are Jupyter notebooks overviewing the technique, descriptions of results, and discussions of open problems. We hope this is useful resource!
3
78
360
@maithra_raghu
Maithra Raghu
7 years
T-shirts at #NIPS2017 -- any interest @goodfellow_ian :)
Tweet media one
5
89
340
@maithra_raghu
Maithra Raghu
10 months
Congratulations @geoffreyhinton on your retirement and an incredible research career!
Tweet media one
2
6
309
@maithra_raghu
Maithra Raghu
5 years
Our paper on Understanding Transfer Learning for Medical Imaging has been accepted to #NeurIPS2019 !! Preprint: As a positive datapoint: we had a good reviewing experience, with detailed feedback and mostly useful comments. Thanks to the Program Chairs!
Tweet media one
8
57
281
@maithra_raghu
Maithra Raghu
4 years
Delighted our new paper "Anatomy of Catastrophic Forgetting: Hidden Representations and Task Semantics" just won Best Paper at the Continual Learning Workshop at #ICML2020 !! Paper: Oral *tomorrow*, details at: ⬇️ Paper thread
3
40
262
@maithra_raghu
Maithra Raghu
3 years
LaMDA: Language Models for Dialogue Applications Paper: Blogpost: Excited to see this paper come out! I enjoyed the weddell seal conversation with LaMDA in our 2021 research summary blogpost!
Tweet media one
6
44
248
@maithra_raghu
Maithra Raghu
3 years
NeurIPS poster presentation happening tomorrow, 8:30am - 10am PT. Hope to see you there! #NeurIPS2021 @NeurIPSConf
Tweet media one
@maithra_raghu
Maithra Raghu
3 years
Do Vision Transformers See Like Convolutional Neural Networks? New paper The successes of Transformers in computer vision prompts a fundamental question: how are they solving these tasks? Do Transformers act like CNNs, or learn very different features?
Tweet media one
27
253
1K
6
26
242
@maithra_raghu
Maithra Raghu
6 years
My favourite #NeurIPS2018 swag: an ML version of cards against humanity by @MSFTResearch
Tweet media one
2
31
224
@maithra_raghu
Maithra Raghu
3 years
Happy to share our paper on ViTs and CNNs was accepted to #NeurIPS2021 ! Our other two submissions this year were rejected. I still think they have some great results and am looking forward to improving the papers with the received feedback.
@maithra_raghu
Maithra Raghu
3 years
Do Vision Transformers See Like Convolutional Neural Networks? New paper The successes of Transformers in computer vision prompts a fundamental question: how are they solving these tasks? Do Transformers act like CNNs, or learn very different features?
Tweet media one
27
253
1K
3
19
219
@maithra_raghu
Maithra Raghu
3 years
Pointer Value Retrieval: A new benchmark for understanding the limits of neural network generalization We introduce a rich family of tasks with a pointer-value rule, to study mechanisms NN of generalization, from memorization to reasoning. Paper:
Tweet media one
5
47
197
@maithra_raghu
Maithra Raghu
4 years
Do Wide and Deep neural networks Learn the Same Things? Paper: We study representational properties of neural networks with different depths and widths on CIFAR/ImageNet, with insights on model capacity effects, feature similarity & characteristic errors
@thao_nguyen26
Thao Nguyen
4 years
Do wide and deep neural networks learn the same thing? In a new paper () with @maithra_raghu and @skornblith we study how width and depth affect learned representations within and across models trained on CIFAR and ImageNet. 1/6
2
33
148
4
18
202
@maithra_raghu
Maithra Raghu
6 years
"Transfusion: Understanding Transfer Learning with Applications to Medical Imaging" The benefits of transfer are nuanced. With *no* feature reuse and only the pretrained weight scaling, we can regain the effects of transfer. More findings in the paper!
Tweet media one
Tweet media two
3
38
196
@maithra_raghu
Maithra Raghu
2 years
Excited to share the @icmlconf 2022 Workshop on Knowledge Retrieval and Language Models Please consider submitting! We welcome work across topics including LM grounding, open-domain Q&A, bias in retrieval, analyses of scale, transfer and LM phenomena.
Tweet media one
3
26
189
@maithra_raghu
Maithra Raghu
5 years
New blogpost on citation trends in @NeurIPSConf and @icmlconf : I scraped paper citations and studied topic trends, citation distributions and academia/industry splits. Releasing scraper, data and a tutorial! Post: Code/Data:
Tweet media one
Tweet media two
5
33
167
@maithra_raghu
Maithra Raghu
30 days
Super exciting to see AI achieving a Silver Medal at IMO today, both as an AI researcher and more personally, as someone who spent many years competing in math Olympiads. Some quick (possibly controversial!) thoughts:
Tweet media one
1
12
159
@maithra_raghu
Maithra Raghu
4 years
ICLR Town: Pokemon-esque environment to wander around and bump into people, which syncs almost seamlessly with video-chatting capabilities. What a fun idea for virtual (research) conferences! Thanks @iclr_conf organizers!! #ICLR2020 #iclr (Uses )
Tweet media one
1
27
157
@maithra_raghu
Maithra Raghu
5 years
Rapid Learning or Feature Reuse? New paper: We analyze MAML (and meta-learning and meta learning more broadly) finding that feature reuse is the critical component in the efficient learning of new tasks -- leading to some algorithmic simplifications!
@OriolVinyalsML
Oriol Vinyals
5 years
Rapid Learning or Feature Reuse? Meta-learning algorithms on standard benchmarks have much more feature reuse than rapid learning! This also gives us a way to simplify MAML -- (Almost) No Inner Loop (A)NIL. With Aniruddh Raghu @maithra_raghu Samy Bengio.
Tweet media one
9
163
602
2
26
152
@maithra_raghu
Maithra Raghu
2 years
Headed to #ICML2022 for the first in-person conference since pre-covid! Looking forward to exciting ML discussions with old & new friends Our workshop on Knowledge Retrieval and Language Models is on Friday 22nd. Do stop by (or tune in online)! @icmlconf
Tweet media one
1
16
127
@maithra_raghu
Maithra Raghu
4 years
Excited to attend #NeurIPS2020 ! My amazing collaborator @thao_nguyen26 **who is applying to PhD programs this year** will be presenting Do Wide and Deep Neural Nets Learn the Same Things? at @WiMLworkshop posters *today* & in Inductive Biases Workshop
Tweet media one
2
18
125
@maithra_raghu
Maithra Raghu
1 year
Lost in the Middle: How Language Models Use Long Contexts Exciting work exploring the effectiveness of long context, led by @nelsonfliu and with Kevin Lin, Ashwin Paranajape, John Hewitt, @percyliang @Fabio_Petroni @MicheleBevila20
Tweet media one
4
24
125
@maithra_raghu
Maithra Raghu
6 years
Very excited about our latest preprint: , joint work with @arimorcos and Samy Bengio. We apply Canonical Correlation (CCA) to study the representational similarity between memorizing and generalizing networks, and also examine the training dynamics of RNNs.
@GoogleAI
Google AI
6 years
Do different networks learn similar representations to solve the same tasks? How do RNN representations evolve over training? What can representational similarity tell us about generalization? Using CCA, @arimorcos and @maithra_raghu try to find out!
5
97
299
0
41
117
@maithra_raghu
Maithra Raghu
6 years
My entry to #MachineLearning (from another field) wouldn't have happened without #NIPS2014 . But the reason I went, and found a welcoming community was due to #WiML2014 . Now #WiML2018 's organizer call is open . Apply by 25/03! The impact can't be overstated.
0
40
111
@maithra_raghu
Maithra Raghu
5 years
How do representations evolve as they go through the transformer? How does the Masked Language Model objective affect these compared to Language Models? How much do different tokens change and influence other tokens? Answers in the paper by @lena_voita : !
Tweet media one
0
22
103
@maithra_raghu
Maithra Raghu
7 years
Deep Learning: Bridging Theory and Practice happening tomorrow at #NIPS2017 ! Final Schedule: @lschmidt3 @OriolVinyalsML @rsalakhu We have an exciting program with talks by Yoshua Bengio @goodfellow_ian Peter Bartlett Doina Precup Percy Liang Sham Kakade!
2
25
95
@maithra_raghu
Maithra Raghu
5 years
How does transfer learning for medical imaging affect performance, representations and convergence? Check out the blogpost below and our #NeurIPS2019 paper for some of the surprising conclusions, new approaches and open questions!
@GoogleAI
Google AI
5 years
How does transfer learning for medical imaging affect performance, representations and convergence? In a new #NeurIPS2019 paper, we investigate this across different architectures and datasets, finding some surprising conclusions! Learn more below:
3
174
473
2
27
87
@maithra_raghu
Maithra Raghu
4 years
Presenting this at @iclr_conf *today*! Talk and Slides: Poster Sessions: (i) 10am - 12 Pacific Time, (ii) 1pm - 3pm Pacific Time Thanks to the organizers for a *fantastic* virtual conference, hope to see you there! #iclr #ICLR2020
@OriolVinyalsML
Oriol Vinyals
5 years
Rapid Learning or Feature Reuse? Meta-learning algorithms on standard benchmarks have much more feature reuse than rapid learning! This also gives us a way to simplify MAML -- (Almost) No Inner Loop (A)NIL. With Aniruddh Raghu @maithra_raghu Samy Bengio.
Tweet media one
9
163
602
1
15
87
@maithra_raghu
Maithra Raghu
6 years
Motivating the Rules of the Game for Adversarial Example Research: Fantastic and nuanced position paper by @jmgilmer @ryan_p_adams @goodfellow_ian on better bridging the gap between research on adversarial examples and realistic ML security challenges.
0
46
88
@maithra_raghu
Maithra Raghu
7 years
First foray into Deep RL We test on a game with continuously tuneable difficulty and *known* optimal policy. We study different RL algorithms, supervised learning, and multiagent play. @jacobandreas
Tweet media one
0
29
84
@maithra_raghu
Maithra Raghu
5 years
Our paper on using Machine Learning (Direct Uncertainty Prediction) for predicting doctor disagreements and medical second opinions will be at @icmlconf next week! Blog: Paper: #icml2019 #DeepLearning
Tweet media one
Tweet media two
1
18
82
@maithra_raghu
Maithra Raghu
10 months
Probably the best we can do is be master of our crafts (know the field well, write good code, collaborate, bring energy & challenge ourselves), and be *brave* --- take risks and try things, even if they're hard, they don't get external validation, and the outcomes are uncertain.
0
9
77
@maithra_raghu
Maithra Raghu
3 years
Really enjoyed this discussion with @jaygshah22 on our work on exploring neural network hidden representations, our recent paper on ViTs and CNNs, and PhD experiences + the ML research landscape! Video:
@jaygshah22
Jay Shah
3 years
In a chat with @maithra_raghu , Sr. Research Scientist at @GoogleAI about analyzing internal representations of #DeepLearning models, comparing vision transformers and CNNs, how she developed her interest in ML, and useful tips for researchers/PhD students!
Tweet media one
0
0
10
1
7
79
@maithra_raghu
Maithra Raghu
7 years
NIPS workshop on theory and practice in Deep Learning #nips2017 @NipsConference @OriolVinyalsML @lschmidt3 @rsalakhu
2
19
79
@maithra_raghu
Maithra Raghu
6 years
Very cool! Hundreds of ML tasks with links to papers and leaderboards.
1
22
68
@maithra_raghu
Maithra Raghu
5 years
Had a fantastic week learning about exciting research directions and meeting old and new friends at #NeurIPS2019 . Thanks to the organizers, volunteers and participants for a wonderful conference! My talk at #ML4H is at (~44 mins), and posters below!
Tweet media one
Tweet media two
1
8
70
@maithra_raghu
Maithra Raghu
3 years
Looking forward to attending #ICLR2021 next week! We're presenting three papers on questions exploring neural network representations, properties of training and algorithms for helping the learning process.
1
7
69
@maithra_raghu
Maithra Raghu
5 years
Excited to announce our @icmlconf workshop on understanding phenomena in deep neural networks! With fantastic speakers including @orussakovsky @ChrSzegedy @KordingLab @beenwrekt @AudeOliva Submission deadline May 5! #DeepLearning #AI #MachineLearning
1
12
67
@maithra_raghu
Maithra Raghu
2 years
Most grateful to my wonderful, supportive colleagues from whom I have learned so much. Hope to share more on next steps in the coming weeks!
1
2
57
@maithra_raghu
Maithra Raghu
2 years
And at last(!!) Google's response to ChatGPT Excited to see Google putting some of these advances out, especially after many years seeing first-hand the development of LaMDA and other AI technology.
0
5
55
@maithra_raghu
Maithra Raghu
4 years
Delighted to be named one of this year's #STATWunderkinds for our work on machine learning in medicine: Grateful to my collaborators and mentors for their advice and support throughout! @statnews
4
9
55
@maithra_raghu
Maithra Raghu
10 months
On AGI and Self-Improvement With @ericschmidt Questions on AGI are at heart of debate on AI capabilities & risks. To get there AI must learn "on the fly". We outline definitions of AGI, explore this gap, and examine the crucial role of *self-improvement*
Tweet media one
5
6
54
@maithra_raghu
Maithra Raghu
10 months
It is usually is very hard to predict *true* breakthroughs, which are often *novel* and have high impact. The novelty means that it's a slow process to be recognized as a breakthrough, and it can be a long and lonely road in the meantime
1
0
51
@maithra_raghu
Maithra Raghu
7 years
A blogpost I wrote on our paper SVCCA, at #nips2017 ! With Justin Gilmer, @jasonyo @jaschasd -- hoping many people will try it out on their networks with the open source code:
@GoogleAI
Google AI
7 years
In order to build better and more robust DNN-based systems, one must be able to effectively interpret the models. We introduce a simple and scalable method to both compare and interpret the representations learned by DNNs
3
201
497
0
16
46
@maithra_raghu
Maithra Raghu
10 months
Furthermore, a lot of best research practices are determined by the maturity and state of the field. Right, now, in LLM research, it's important to write good code and have good infra. That was hardly the case earlier in deep learning when we barely had libraries for autodiff!
1
0
41
@maithra_raghu
Maithra Raghu
4 years
Fantastic workshop on the theory of deep learning at Bellairs in Barbados! Five days of incredible talks from @ShamKakade6 @HazanPrinceton @ylecun @suriyagnskr @prfsanjeevarora and many others! Huge thanks to the organizers ( @prfsanjeevarora and Denis Therien)!
@ShamKakade6
Sham Kakade
4 years
Bellairs. Day 5 @HazanPrinceton and myself: double feature on controls+RL. +spotlights: @maithra_raghu : meta-learning as rapid feature learning. Raman Arora: dropout, capacity control, and matrix sensing . @HanieSedghi : module criticality and generalization! And that is a wrap!🙂
0
3
33
1
5
38
@maithra_raghu
Maithra Raghu
5 years
Exploring the AI Landscape: New blog by @bclyang and me! We'll be covering topics in AI from fundamental research to considerations for deployment. Our first post: is on Digital Health and AI for Health, a longstanding interest!
0
5
39
@maithra_raghu
Maithra Raghu
4 years
Excellent page by @Worldometers : has detailed statistics on the coronavirus --- number of cases, severity, breakdown by country, and many others.
Tweet media one
1
17
36
@maithra_raghu
Maithra Raghu
3 years
An analysis of self-attention reveals some reasons for this difference: very early ViT layers learn to incorporate local and *global* spatial information, unlike CNN early layers with their smaller receptive field size.
Tweet media one
1
2
38
@maithra_raghu
Maithra Raghu
10 months
@geoffreyhinton often wrote quick matlab code and even computed gradients by hand! (I was always inspired that even at that level of seniority, he could quickly prototype his own ideas!)
2
0
35
@maithra_raghu
Maithra Raghu
6 years
Heading to #NeurIPS2018 this week! Looking forward to meeting old friends and new! Let me know if you'll be around and want to chat. @arimorcos and I will be presenting our paper on the Wednesday poster session, hope to see you there!
0
4
37
@maithra_raghu
Maithra Raghu
1 year
It's a delight and privilege to work with such an amazing team at
@Fabio_Petroni
Fabio Petroni
1 year
🎉🌐 Big news from @samaya_AI . We have two shiny new offices in #London & #MountainView 🏢, staffed with an incredible team of brilliant minds💡🚀. Check out our freshly launched website at 🌟
Tweet media one
Tweet media two
3
10
97
1
0
33
@maithra_raghu
Maithra Raghu
1 year
This article, on the lack of an AI moat at Google and OpenAI has been making the rounds: While it's true that that there is exciting, fast-paced opensource activity in AI, and we may see many current LLMs commoditize, there are still *quality moats*
2
4
34
@maithra_raghu
Maithra Raghu
14 days
I'm deeply saddened to hear about the passing of @SusanWojcicki We met just a couple months back, and she offered sage advice on running a company, even giving feedback on our new product features. I was struck by her insight, her groundedness and her warmth. Sending her family
0
0
34
@maithra_raghu
Maithra Raghu
9 months
AI winning IMO gold would be impressive, but an AI coming up with IMO *questions* would be even more impressive to me. Can it understand and use different theorems intelligently to come up with hard, creative and truly new questions? Can it do this consistently?
@natfriedman
Nat Friedman
9 months
👀 Alex has launched a $10M challenge for the first AI to win IMO Gold.
18
43
574
6
1
29
@maithra_raghu
Maithra Raghu
6 years
Bots destroying humans at @OpenAI
Tweet media one
0
7
31
@maithra_raghu
Maithra Raghu
6 years
Excited to be speaking at @reworkdl deep learning summit today , and Stanford's HealthAI @ai4healthcare hackathon tomorrow! What with the ICML deadline just wrapping up, it's been a busy week 😅
1
2
31
@maithra_raghu
Maithra Raghu
4 years
Very interesting work on identifying, understanding and reconstructing the representations learned by neural networks! (I've also enjoyed @distillpub 's "Building Blocks of Interpretability" and "Zoom In" which this work builds on)
@nickcammarata
Nick
4 years
Excited to share a new paper, Curve Circuits We reverse engineer a non-trivial 50k+ parameter learned algorithm from the weights of a neural network and use its core ideas to craft an artificial artificial neural network from scratch that reimplements it
Tweet media one
49
334
2K
0
3
32
@maithra_raghu
Maithra Raghu
1 year
Very exciting work by @matei_zaharia @alighodsi and quite literally all of @databricks (who created the dataset!) Lots of interesting followup questions from this --- how well can we use this to bootstrap synthetic data, etc.
@alighodsi
Ali Ghodsi
1 year
Free Dolly! Introducing the first *commercially viable*, open source, instruction-following LLM. Dolly 2.0 is available for commercial applications without having to pay for API access or sharing data with 3rd parties.
55
448
2K
2
2
31
@maithra_raghu
Maithra Raghu
6 years
I've been enjoying reading @beenwrekt 's posts on #ReinforcementLearning : (new post today!), and it's great to see these insights come together in paper format!
@hardmaru
hardmaru
6 years
"Simple random search provides a competitive approach to reinforcement learning", by Mania, Guy and @beenwrekt Paper: Code: Blog:
6
128
447
0
12
30
@maithra_raghu
Maithra Raghu
2 years
Thrilled to be working with my amazing co-founder @Fabio_Petroni , and a growing team of incredible researchers & engineers!
1
0
30
@maithra_raghu
Maithra Raghu
3 years
Transformer representations can generalize across data modalities! Very interesting result, lots of promise for more progress in multi-modal learning!
@IMordatch
Igor Mordatch
3 years
What are the limits to the generalization of large pretrained transformer models? We find minimal fine-tuning (~0.1% of params) performs as well as training from scratch on a completely new modality! with @_kevinlu , @adityagrover_ , @pabbeel paper: 1/8
4
71
353
0
5
29
@maithra_raghu
Maithra Raghu
4 years
Although there were ups and downs, I'm deeply grateful to the many rich experiences during my PhD, and hope this blogpost might be helpful to others on the journey. Wishing everyone a happy new year!!
0
0
29
@maithra_raghu
Maithra Raghu
4 years
With the rapid pace of progress in Machine Learning, it's hard not to feel publication pressure during the PhD. But while writing papers is important, the main research goal of the PhD (to me at least!) is to make you an independent researcher, with a rich research vision
1
0
30
@maithra_raghu
Maithra Raghu
3 years
But attending locally is also very important! It is automatically encoded in CNNs, but larger ViTs only learn to do this with enough data (which is needed for their strong performance also.)
Tweet media one
1
0
29
@maithra_raghu
Maithra Raghu
7 years
Another research update: Final version of our #nips2017 @NipsConference paper SVCCA: with accompanying code:(!!) We look at deep learning dynamics and interpret the latent representations. With Justin Gilmer, @jasonyo , @jaschasd
Tweet media one
Tweet media two
1
7
28
@maithra_raghu
Maithra Raghu
4 years
Looking forward to speaking at @RAAISorg this Friday! Many exciting ML research areas, from health to privacy to bioengineering. Details on the talks, research and speakers at:
@raais
RAAIS
4 years
Tweet media one
0
5
8
0
10
27
@maithra_raghu
Maithra Raghu
1 year
It was awesome having @samaya_AI as part of the first batch of AI Grant companies! Grateful to @natfriedman and @danielgross for creating an energizing community for AI-native products. Consider applying!
@natfriedman
Nat Friedman
1 year
AI Grant's second batch is now accepting applications!
42
87
560
0
2
20
@maithra_raghu
Maithra Raghu
3 years
Using local and global info allows ViT earlier layers to learn better representations, which are strongly propagated through residual connections. Surprisingly ViT has stronger residual connections than ResNet! These help explain the uniform structure of ViT representations
Tweet media one
Tweet media two
2
1
28
@maithra_raghu
Maithra Raghu
9 months
Looking forward to heading to #NeurIPS2023 next week! This year marks a decade(!) of attending NeurIPS! It's remarkable to see how much the field has advanced in 10 years! These past 2 years of building @samaya_AI has been incredible, and we are continuing to grow!
1
1
28
@maithra_raghu
Maithra Raghu
4 years
Thanks to @atJustinChen and @statnews for an in-depth followup discussion on our research work and motivations. We talk about neural networks, techniques to better understand them, and ways this can inform their design and usage as assistive tools.
0
2
27
@maithra_raghu
Maithra Raghu
1 year
We believe not! The future ecosystem will be rich, with set of *Specialized AI Systems* and a few *General AI Models*, with many entities participating. Specialized AI Systems develop for well-defined, high-value workflows, while General AI models tackle a heavy tail of uses
3
1
24
@maithra_raghu
Maithra Raghu
9 months
Sending best wishes to friends, former colleagues and the team at @OpenAI . You've made incredible, world changing contributions to AI, and it was sad to see the developments of the past few days. Wishing you the best in navigating these transitions.
0
0
21
@maithra_raghu
Maithra Raghu
3 years
Using representational similarity measures, we investigate the internal structure of the two architectures, finding striking differences, with ViT to having a much more uniform representation across all layers
Tweet media one
2
0
25
@maithra_raghu
Maithra Raghu
4 years
Some good news: the recent ruling forcing international students to choose between leaving the country and safety (taking online classes) has been rescinded:
0
1
25
@maithra_raghu
Maithra Raghu
4 years
To me, a big surprise of the PhD was how much it really is a journey, with evolving perspectives (both personal and research) affecting interest in specific problems, research directions and broader subfields. Importantly, it's hard to predict this evolution going in!
1
0
24
@maithra_raghu
Maithra Raghu
6 years
Thanks so much to the organizers and @MITEECS for hosting the EECS Rising Stars 2018! Entertaining, insightful and inspiring discussion by panelists and speakers on research and academia, and a truly unique opportunity to meet my fantastic peers across all of EECS!
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
0
22
@maithra_raghu
Maithra Raghu
1 year
It takes a lot of technical knowledge, effort & iteration to build high quality AI systems for specific, valuable uses. So while "base models" may commoditize (also discussed in ), there are plenty of chances of moats for focused, high-value AI products.
3
0
22
@maithra_raghu
Maithra Raghu
2 years
Totally agree. Public criticism disproportionally impacts the graduate student leading the project, and ML publishing is already very high pressure. Twitter also isn't the right place for a nuanced scientific discussion.
@zicokolter
Zico Kolter
2 years
I realize this is seemingly an unpopular opinion, but I can't get onboard with these Twitter criticisms of some of the recent #ICML2022 best paper awardees. I've been thinking about this all day. A thread... 🧵 1/N
20
86
911
0
0
23
@maithra_raghu
Maithra Raghu
2 years
So excited to be working together!!
@Fabio_Petroni
Fabio Petroni
2 years
Today is my first day as a CTO (and co-founder) of @samaya_AI . The last 4 years at FAIR have been incredible. Now I'm looking forward to bringing the latest knowledge discovery technologies to market!
17
10
202
0
0
23
@maithra_raghu
Maithra Raghu
4 years
We provide links to incredible resources developed by the community: software packages & high level APIs, freely available DL tutorials, sites with summaries/discussions/code of new research, repositories of DL pipelines & pretrained models, data curation & analysis packages
Tweet media one
Tweet media two
1
1
21
@maithra_raghu
Maithra Raghu
6 years
I've gained a lot from the interesting paper links, tutorials and code releases posted on Twitter. However it's important to recognise the drawbacks of the filter bubble: Particularly poignant: "Algorithms know what you've been, not what you want to be."
1
2
21
@maithra_raghu
Maithra Raghu
4 years
New paper Teaching with Commentaries We introduce commentaries, metalearned information to help neural net training & give insights on learning process, dataset & model representations Led by @RaghuAniruddh & w/ @skornblith @DavidDuvenaud @geoffreyhinton
@RaghuAniruddh
Aniruddh Raghu
4 years
Teaching with Commentaries: We study the use of commentaries, metalearned auxiliary information, to improve neural network training and provide insights. With @maithra_raghu , @skornblith , @DavidDuvenaud , @geoffreyhinton Thread⬇️
Tweet media one
2
6
52
0
3
22
@maithra_raghu
Maithra Raghu
5 years
Intriguing invited talk at #DeepPhenomena from Chiyuan Zhang on the effect of resetting different layers: Are all layers created equal? #ICML2019 @icmlconf
Tweet media one
1
3
22
@maithra_raghu
Maithra Raghu
4 years
Enjoyed speaking at @RealAAAI workshop on Learning Network Architectures During Training: I overviewed our work on techniques to gain insights from neural representations for model & algorithm design. All talk videos are on the workshop page above! ⬆️
0
3
21
@maithra_raghu
Maithra Raghu
4 years
@karpathy I usually mute all notifications and put on do not disturb. Sometimes takes me a little longer to respond to things, but the mental space is worth it :)
2
0
22