![Arslan Chaudhry Profile](https://pbs.twimg.com/profile_images/1730327160391663616/JfEJYjCT_x96.jpg)
Arslan Chaudhry
@arslan_mac
Followers
610
Following
76
Statuses
64
Research Scientist @GoogleDeepMind | Ex @UniofOxford, @AIatMeta, @GoogleAI
Mountain View, CA
Joined September 2011
RT @Olympics: Believe it, Nadeem. You just became Pakistan's first ever athletics Olympic champion. šµš°š„¹
0
21K
0
I will be at the ICML in Vienna. If you are interested in continual learning in foundation models or other forgetting or adaptation related challenges in (V)LMs, come chat to me. I will be at/ around the @GoogleDeepMind booth.
2
0
16
RT @yeewhye: We're looking for an exceptional junior researcher in AI/ML with strong interests in diversity, equity and inclusion to fill tā¦
0
22
0
Very well done @usmananwar391 and coauthors.
Iām super excited to release our 100+ page collaborative agenda - led by @usmananwar391 - on āFoundational Challenges In Assuring Alignment and Safety of LLMsā alongside 35+ co-authors from NLP, ML, and AI Safety communities! Some highlights below...
1
0
3
RT @schwarzjn_: @thegautamkamath I have no doubt about the positive intentions behind this, but my fear is that it's yet another policy thaā¦
0
3
0
@vfsglobalcare Could you please respond to the DM? Itās been more than 6 hours and we are running short on time here.
1
0
0
Gemini has achieved a classification accuracy of 90% on MMLU (with a combo of greedy sampling and a majority vote of 32 COT samples). This is remarkable! For reference, expert-level perf, defined at 95th percentile human test-taker accuracy in *each subject*, would be ~89.8%.
Exciting times, welcome Gemini (and MMLU>90)! State-of-the-art on 30 out of 32 benchmarks across text, coding, audio, images, and video, with a single model š¤Æ Co-leading Gemini has been my most exciting endeavor, fueled by a very ambitious goal. And that is just the beginning! A long š post about our Gemini journey & state of the field. The biggest challenges in LLMs are far from trivial or obvious. Evaluation and data stand out to me. We've moved beyond the simpler "Have we won in Go/Chess/StarCraft?" to āIs this answer accurate and fair? Is this conversation good? Does this complex piece of text prove the theorem?ā Exciting potential coupled with monumental challenges. The field is less ripe further down the model pipeline. Pretraining is relatively well understood. Instruction tuning and RLHF, less so. In AlphaGo and AlphaStar we spent 5% of compute in pre-training and the rest in the very important RL phase, where the model learns from its successes or failures. In LLMs, we spend most of our time on pretraining. I believe thereās huge potential to be untapped. Cakes with lots of cherries, please š @Google has demonstrated its ability to move fast. It has been an absolute blast to see the energy from my colleagues and the support received. A ārandomā highlight is coauthoring our tech report with a co-founder. Another is coleading with @JeffDean. But beyond individuals, Gemini is about teamwork: it is important to recognize the collective effort behind such achievements. Picture a room full of brilliant people, and avoid attributing success solely to one person. On a personal note, recently I celebrated my 10 year anniversary at Google, and itās been 8 years since @quocleix and I co-authored āA Neural Conversational Modelā, which gave us a glimpse of what was, has, and is yet to come. Back then, that line of work received a lot of skepticism. Lessons learned: whatever your passion is, push for it! Zooming back out, thereās lots of change in our field, and the stakes couldnāt be higher. Excited for whatās to come from Gemini, but humbled by the responsibility to āget it rightā. 2024 will be drastic. Welcome Gemini!
2
0
4
Our team in @GoogleDeepMind Mountain View is looking for a Research Scientist/ Research Engineer. If you are interested in working on multimodal foundation models and their adaptability over time, come chat with me. I will be attending ICML and my schedule is quite flexible.
6
18
245
RT @bschoelkopf: Another AI paradox: people are excited about LLMs, some even think that AGI is just around the corner. But some students aā¦
0
309
0
Our work on a new large-scale benchmark composed of 30 years of computer vision research to explore knowledge accrual in continual learning systems.
Introducing NEVISā22, a new benchmark developed using 30 years of computer vision research. This provides an opportunity to explore how AI models can continually build on their knowledge to learn future tasks more efficiently. ā”ļø
0
0
3
@timrudner @NYUDataScience @andrewgwils @yaringal @yeewhye This is such a good news! Congratulations, Tim!
1
0
1
About time!
Doina Precup, @rpascanu and I are very excited to announce the first International Conference on Lifelong Learning Agents (CoLLA)! If you are working in lifelong/continual learning, meta-learning, multi-task learning, or OOD generalization, consider submitting your work to CoLLA!
0
0
0