Lisa Lee Profile Banner
Lisa Lee Profile
Lisa Lee

@rl_agent

Followers
5,942
Following
0
Media
30
Statuses
128

Research Scientist at Google DeepMind. Core contributor to Gemini post-training & multimodal pre-training.

Joined February 2016
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
@rl_agent
Lisa Lee
3 years
Five years ago, I left Google to pursue a PhD in Machine Learning. Tomorrow, I'm very excited to join Google Brain as a Research Scientist! Looking forward to meeting everyone :) My PhD thesis can be viewed at: Thesis defense:
56
200
3K
@rl_agent
Lisa Lee
3 years
Update: I will be a Research Scientist at Google Brain starting this fall! Super grateful to all of the fantastic researchers I met through the interview process. Thanks for valuable career advice & research chats – I'll take these lessons with me in the next stage of my career!
30
8
1K
@rl_agent
Lisa Lee
7 months
Gemini is the most fun project I've worked on in my career. I feel so lucky to work with incredible teammates in Gemini. We all worked really hard for this public release. Looking forward to more learnings in 2024. Chat with Bard (running Gemini Pro):
12
34
466
@rl_agent
Lisa Lee
6 years
Just released our #ICML2018 paper "Gated Path Planning Networks" as well as a @PyTorch implementation for replicating our experiments: - with Emilio Parisotto, @dchaplot , Eric Xing, @rsalakhu See you @icmlconf in Stockholm!
Tweet media one
0
67
242
@rl_agent
Lisa Lee
5 years
I wrote a Colab tutorial on MaxEnt RL: It implements the graphical model from @svlevine 's "RL as Inference" tutorial for a simple chain environment. Play around with the reward function to learn different policies using the forward-backward algorithm!
0
54
228
@rl_agent
Lisa Lee
1 year
My dog Eevee 🐶 is featured in our recent research paper from Google @DeepMind : Barkour: Benchmarking animal-level agility with quadruped robots Inspired by dog agility competitions, we introduce a diverse and challenging obstacle course for robotic locomotion. (1/n)
3
22
175
@rl_agent
Lisa Lee
6 years
I was a TA for @rsalakhu 's 10-703 Deep Reinforcement Learning, with 500+ enrolled students. It was a little scary to hold office hours at first, but it got easier as I got to know more students. I was so touched to read these words from students. Thank you!
Tweet media one
3
5
165
@rl_agent
Lisa Lee
5 years
Excited to give two Contributed Talks @ #iclr2019 on Monday w/ Ben Eysenbach on our new work: Exploration & Meta-RL via State Marginal Matching w/ Ben, Emilio @rsalakhu @svlevine 12:15 @ TARL 15:50 @ SPiRL
1
26
147
@rl_agent
Lisa Lee
1 year
What makes it hard for robots to generalize to new environments? In our study, we broke down the notion of an “environment” into smaller, more manageable factors of variation, such as lighting or camera placement. 1/n
3
34
141
@rl_agent
Lisa Lee
6 years
I'm excited to present our Gated Path Planning Networks paper at #ICML2018 today. Come say hi! Oral Talk: 14:10 @ A1 Poster: 18:15 - 21:00 @ Hall B #134 Paper: Code: - with Emilio, @dchaplot , Eric, @rsalakhu
Tweet media one
1
32
112
@rl_agent
Lisa Lee
11 months
Our new work is featured in the @NYTimes : RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control tl;dr We fine-tune a VLM to predict robot actions directly as text, and see emergent capabilities in the embodied agent. (1/n)
6
17
105
@rl_agent
Lisa Lee
11 months
Visited @McGillU in Montreal. Such a beautiful campus in a beautiful city 🇨🇦🍁
Tweet media one
Tweet media two
3
6
97
@rl_agent
Lisa Lee
5 years
Video & slides for LIRE workshop @ #NeurIPS2019 are now up: Check out the Talks and Panel by @RaiaHadsell @tommmitchell Jeff Bilmes @pabbeel @YejinChoinka Tom Griffiths & more. Thanks to all speakers & presenters for making the workshop a success!
Tweet media one
Tweet media two
Tweet media three
0
28
86
@rl_agent
Lisa Lee
5 years
I just started at Stanford this week as a visiting researcher in @chelseabfinn 's lab, and I'm also still part-time at Google Brain Robotics. If you're around in the area and would like to chat about research, please feel free to reach out anytime! (My office is in Gates)
0
0
75
@rl_agent
Lisa Lee
2 years
Our team is hiring! I only recently joined Dale's team after PhD, but am grateful to be part of a team of super talented & wonderful researchers, with diverse areas of expertise in core ML, reinforcement learning, representation learning, planning, discrete optimization, etc.
@hanjundai
Hanjun Dai
2 years
Our team at Google Brain (w/ Dale Schuurmans, @daibond_alpha , @rl_agent , @mengjiao_yang and many others) is hiring a SWE to work on representation learning for reasoning, search & decision making. Apply below if you are interested!
3
14
93
3
9
69
@rl_agent
Lisa Lee
3 years
There are many ways to parameterize an RL problem (state/action space, rewards, time freq. & horizon, etc). Researchers often assume the problem formulation to be fixed, & iterate on algorithm design. But conversely: How does the design of envs & tasks affect the RL algorithm?
1
5
65
@rl_agent
Lisa Lee
6 years
Yo-Yo Ma ( @YoYo_Ma ) playing Bach Cello Suite No. 3 Prelude at the AI for Social Good ( #ai4good ) workshop in Montreal! #NeurIPS2018
1
10
65
@rl_agent
Lisa Lee
6 years
Slides & video from my talk at #icml2018 on Gated Path Planning Networks: - with Emilio, @dchaplot , Eric, and @rsalakhu
Tweet media one
0
21
64
@rl_agent
Lisa Lee
4 years
Instead of rewards or demos, can we use weak supervision to accelerate goal-conditioned RL & learn interpretable latent policies? Excited to share our work from during my visit @Stanford & Google, with my amazing collaborators Ben Eysenbach, @rsalakhu @shaneguML & @chelseabfinn !
@chelseabfinn
Chelsea Finn
4 years
Supervising RL is hard, especially if you want to learn many tasks. To address this, we present: Weakly-Supervised Reinforcement Learning for Controllable Behavior with @rl_agent , Eysenbach, @rsalakhu , @shaneguML @GoogleAI thread 👇
Tweet media one
2
59
276
0
11
58
@rl_agent
Lisa Lee
5 years
Congratulations & thanks Kamalika ( @kamalikac ) and Russ ( @rsalakhu ) for organizing #ICML2019 ! And thanks for having us workflow chairs onboard! It was an amazing experience to see the conference come together from start to finish. @dchaplot @pliang279
@rsalakhu
Russ Salakhutdinov
5 years
From your ICML2019 Program Chairs with @kamalikac . We are done my friends! We hope you enjoyed ICML this year! And big thanks to all members of Organizing Committee and our workflow chairs for making it a successful conference!
Tweet media one
11
17
324
0
6
50
@rl_agent
Lisa Lee
6 years
Pics from our #ICML2018 Workshop on Theoretical Foundations & Applications of Deep Generative Models Huge thanks to @ZhitingHu who put a lot of work into organizing & thanks to speakers for their amazing talks! with Zhiting, @andrewgwils @rsalakhu & Eric
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
8
50
@rl_agent
Lisa Lee
3 years
I'm always grateful for my collaborators & mentors from my PhD, including my thesis committee @rsalakhu @ericxing @chelseabfinn @svlevine ; intern hosts N.Heess (DeepMind), @shaneguML (Google Robotics); friends @ben_eysenbach @alshedivat @dchaplot E.Parisotto C.Huang &many others.
3
2
50
@rl_agent
Lisa Lee
5 years
I'm Workflow Chair for #ICML2019 with @dchaplot . It's been eye-opening to see how much work goes into organizing a major ML conference, and a great honor to work with Program Chairs @kamalikac & @rsalakhu !
@dchaplot
Devendra Chaplot
5 years
Excited to be a workflow chair for #icml2019 with Lisa Lee ( @rl_agent ). Reminder ICML deadline is early this year. Submissions open now. Abstract deadline in one week on Jan 18, 4pm PT. @icmlconf
0
7
27
0
3
47
@rl_agent
Lisa Lee
6 years
1/ Excited to present two workshop papers at #NeurIPS on RL + Navigation, including this paper at the Deep RL workshop (Fri 15:00). tl;dr We introduce a series of attention operators to disentangle text/visual representations & enable cross-task transfer for embodied navigation.
@dchaplot
Devendra Chaplot
6 years
Check out our poster on 'Cross-Task Knowledge Transfer for Visually-Grounded Navigation' at the #neurips2018 DeepRL Workshop (Friday, 3pm, Room 220). - with @rl_agent @rsalakhu @deviparikh @DhruvBatraDB
Tweet media one
0
13
57
1
8
45
@rl_agent
Lisa Lee
4 years
I'm excited to present Weakly-Supervised RL for Controllable Behavior at #NeurIPS2020 today! Stop by our poster @ 9-11 pm PST in Session 7 (Deep Learning) D0-A0. w/ B.Eysenbach @rsalakhu @shaneguML @chelseabfinn Paper: Talk:
Tweet media one
@chelseabfinn
Chelsea Finn
4 years
Supervising RL is hard, especially if you want to learn many tasks. To address this, we present: Weakly-Supervised Reinforcement Learning for Controllable Behavior with @rl_agent , Eysenbach, @rsalakhu , @shaneguML @GoogleAI thread 👇
Tweet media one
2
59
276
0
14
43
@rl_agent
Lisa Lee
5 years
Check out Emilio's new paper: Concurrent Meta Reinforcement Learning (w/ @Yuhu_ai_ , @rsalakhu , and others) tl;dr CMRL learns a multi-agent communication protocol to coordinate exploration between parallel rollout agents.
0
10
39
@rl_agent
Lisa Lee
11 months
RT-2 paper is now on ArXiv: Videos: Blog post by @YevgenChebotar & @TianheYu : It also appears on the @NYTimes podcast Hard Fork hosted by @KevinRoose & @CaseyNewton (gifted link):
@rl_agent
Lisa Lee
11 months
Our new work is featured in the @NYTimes : RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control tl;dr We fine-tune a VLM to predict robot actions directly as text, and see emergent capabilities in the embodied agent. (1/n)
6
17
105
0
9
40
@rl_agent
Lisa Lee
6 years
And check out Emilio's other paper titled "Neural Map: Structured Memory for Deep Reinforcement Learning":
@dchaplot
Devendra Chaplot
6 years
Also, check out our poster on LSTM Iteration Networks: An Exploration of Differentiable Path Finding in #ICLR2018 Workshop at 11am on Monday. (with @binarymatroid , Emilio and @rsalakhu ) PDF:
0
5
15
1
1
35
@rl_agent
Lisa Lee
6 years
Come say hi! #ICLR2018
Tweet media one
1
4
33
@rl_agent
Lisa Lee
2 years
Check out our blog post on the Multi-Game Decision Transformer, one of my first projects since starting at Google Brain :) We study the generalization capabilities of a single Transformer agent trained on many Atari games. Blog post authored by @winniethexu @kuanghueilee
@GoogleAI
Google AI
2 years
Introducing the Multi-Game Decision Transformer: Learn how it trains an agent that can play 41 Atari games, can be quickly adapted to new games via fine-tuning, and significantly improves upon the few alternatives for training multi-game agents →
8
156
527
0
5
31
@rl_agent
Lisa Lee
6 years
My friend @alshedivat got a Best Paper award at ICLR 2018! Check out his Oral presentation this Thursday at 10:15am :)
@OpenAI
OpenAI
6 years
Continuous Adaptation via Meta-Learning in Nonstationary and Competitive Environments (, tl;dr: agents that meta-learn outcompete agents that don't, ). Best paper award, Thursday May 3rd, 10:15-10:30am, Exhibition Hall A
3
19
61
0
4
30
@rl_agent
Lisa Lee
6 years
2/ On the Complexity of Exploration in Goal-Driven Navigation w/ @alshedivat @rsalakhu Eric Xing R2Learning workshop (Sat 10:15) #NeurIPS2018 tl;dr We measure the complexity of RL environments by computing expected hitting times in goal dependency graphs.
Tweet media one
0
11
31
@rl_agent
Lisa Lee
10 months
Check out our #NeurIPS2023 paper, led by KAIST PhD student Changyeon Kim ( @cykim1006 ): Guide Your Agent with Adaptive Multimodal Rewards Website: Paper: Code:
@cykim1006
Changyeon Kim
10 months
Excited to share Adaptive Return-conditioned Policy (ARP): a return-conditioned policy utilizing adaptive multimodal reward from pre-trained CLIP encoders! ARP can mitigate goal misgeneralization and execute unseen text instructions! 🧵👇 1/N
2
9
57
0
3
30
@rl_agent
Lisa Lee
2 years
Honored to be featured in Sean Welleck's "The Thesis Review" podcast! Thanks @wellecks for the fun conversation & great questions. Check out other interesting episodes on Deep Learning, NLP, Active Learning, Graphical Models, RL & more at:
@thesisreview
The Thesis Review Podcast
2 years
Episode 40 of The Thesis Review: Lisa Lee ( @rl_agent ), "Learning Embodied Agents with Scalably-Supervised RL" We discuss her thesis work on reinforcement learning, including exploration, weak supervision, and embodied agents, along with trends in RL.
Tweet media one
1
6
20
0
2
23
@rl_agent
Lisa Lee
2 years
Forgetful Causal Masking makes language models better zero-shot learners Led by @haoliuhl and @YoungGeng ( @UCBerkeley & @GoogleAI ) A simple technique can greatly improve zero- & few-shot performance of LLMs on downstream language-understanding tasks. 🧵
@haoliuhl
Hao Liu
2 years
Can language model pretraining be even better? Our paper shows that by randomly masking input tokens during pretraining, the zero-shot, few-shot, and fine-tuning performance can be significantly improved. 🧵
Tweet media one
2
41
303
0
4
23
@rl_agent
Lisa Lee
1 year
I loved my time at DeepMind in Nicolas Heess's team. Met so many talented & kind researchers and engineers there. Now, looking forward to exciting collaboration opportunities with old and new colleagues under the same umbrella. :)
@demishassabis
Demis Hassabis
1 year
The phenomenal teams from Google Research’s Brain and @DeepMind have made many of the seminal research advances that underpin modern AI, from Deep RL to Transformers. Now we’re joining forces as a single unit, Google DeepMind, which I’m thrilled to lead!
159
651
4K
0
1
20
@rl_agent
Lisa Lee
6 months
Submissions are now open for the ICLR 2024 workshop on Generative Models for Decision Making: How can we combine generative models and decision making for better exploration, planning, sample efficiency, transfer learning, and alignment with human priors?
1
2
18
@rl_agent
Lisa Lee
3 years
EcoRL workshop @NeurIPSConf brings together exciting talks & papers to discuss: "How does task design influence agent learning?" I'm personally excited about the step towards data-centric understanding of RL, complementing today's algorithm-centered view.
@EcoTheoryRL
Ecological Theory of Reinforcement Learning
3 years
Starting 8:10 ET, we will have @ShaneLegg Co-Founder & Chief Scientist @DeepMind with a thought-provoking talk titled “Artificial What?” : (). Submit questions for @ShaneLegg at:  for the Live Q&A following the talk.
1
0
10
1
1
18
@rl_agent
Lisa Lee
1 year
This work was done with wonderful collaborators at Google @DeepMind . Check out our blog post, co-authored by Ken Caluwaerts and Atil Iscen: Blog Post: Paper: Website: (6/6)
1
3
17
@rl_agent
Lisa Lee
6 years
Come check out our ICML Workshop next week!
@rsalakhu
Russ Salakhutdinov
6 years
Excited about our 2-day @icmlconf Workshop on Theoretical Foundations and Applications of Deep Generative Models with Zhiting Hu, @andrewgwils , @rl_agent & Eric Xing, on Sat Jul 14th, 8am - 6pm @ A5 Check out the program, posters, & terrific speakers!
0
23
86
0
3
16
@rl_agent
Lisa Lee
3 years
Thanks Russ for being an amazing advisor to all of us!
@rsalakhu
Russ Salakhutdinov
3 years
1/3: 2021 has been an incredible year: 5 of my PhD students (some co-advised) have graduated this year. It's a privilege to have worked with such talented people! Devendra Chaplot, Lisa Lee, Emilio Parisotto, Shrimai Prabhumoye & Hubert Tsai Check out their theses/papers here:
2
9
177
1
0
15
@rl_agent
Lisa Lee
2 years
Instruction-Following Agents with Jointly Pre-Trained Vision-Language Models Paper: Code: w/ @haoliuhl * @kimin_le2 @pabbeel from @berkeley_ai @GoogleAI We showcase the importance of a jointly pre-trained encoder for grounded RL:
@haoliuhl
Hao Liu
2 years
How to pretrain large language-vision models to help seeing, acting, and following instructions? We found that using models jointly pretrained on image-text pairs and text-only corpus significantly outperforms baselines. A 🧵 on the paper InstructRL
3
13
88
1
2
15
@rl_agent
Lisa Lee
2 years
サンフランシスコからこんばんは!日本語で初めてのツイートです。 リサと申します。子供の頃、大阪府豊中市で中豊島小学校の生徒でした。今はGoogle Brainで人工知能の研究をやってます。日本にいる大切な友達と連絡を取り続けるように、ずっと日本語を練習続けてきました。 よろしくお願い致します。
0
0
12
@rl_agent
Lisa Lee
2 years
@zacharylipton So sorry to hear that. MLD is lucky to have you :)
0
0
9
@rl_agent
Lisa Lee
4 years
Learning a stationary reward function via gradient descent to match the expert state density. This reward can be reused for solving downstream tasks & behavior transfer across dynamics. #CoRL2020 Project led by Tianwei Ni, Harshit Sikchi, Yufei Wang & @the_tejus from CMU!
@harshit_sikchi
Harshit Sikchi
4 years
Can we extract reward functions when only expert state density/samples are given? Our #CORL_2020 paper derives an analytical gradient to match general f-divergence! Equal contribution work with coauthors T. Ni, Y. Wang, @the_tejus , @rl_agent , B. Eysenbach
1
2
14
0
0
10
@rl_agent
Lisa Lee
1 year
These results suggest that tackling harder factors (e.g. new camera poses) may be more important to closing the generalization gap than other factors (e.g. new backgrounds). We hope these results serve as a guide for future data collection and model design in robotics. 4/n
Tweet media one
1
1
8
@rl_agent
Lisa Lee
1 year
Check out our full paper at: This work was led by Annie Xie (PhD student at Stanford), and w/ Ted Xiao ( @xiao_ted ) & Chelsea Finn ( @chelseabfinn ). 6/6
Tweet media one
0
1
7
@rl_agent
Lisa Lee
1 year
Using a custom-built quadruped robot, we train two strong baselines: 1. Individual specialist skill policies trained via on-policy RL and large-scale parallel simulation. 2. A single, generalist Transformer-based policy trained by distilling the specialist policies. (3/n)
1
1
6
@rl_agent
Lisa Lee
1 year
The generalist Locomotion Transformer policy automatically infers which skill to use, such as walking or slope climbing, from its terrain and proprioceptive observations. It predicts low-level locomotion actions, conditioned on velocity command & environment observations. (4/n)
1
0
6
@rl_agent
Lisa Lee
11 months
Also visited @Mila_Quebec . It's a nice collaborative space for research in machine learning and AI. Thanks @zhu_zhaocheng and @zdhnarsil for the tour!
Tweet media one
1
0
6
@rl_agent
Lisa Lee
2 years
My PhD advisor's origin story :-)
@rsalakhu
Russ Salakhutdinov
2 years
1/4 I was watching @geoffreyhinton interview w/t @pabbeel - it reminded me of a couple fun stories about Geoff and how he got me into deep learning. Back in 2005, I was not thinking about PhD & had a good career as ML engineer. One morning, on my way to work I bumped into Geoff.
13
43
605
0
0
6
@rl_agent
Lisa Lee
4 months
7B and 2B models based on Gemini, both pre-trained & fine-tuned checkpoints, are now publicly available. Technical report: Opensource codebase for inference and serving: HuggingFace:
@GoogleDeepMind
Google DeepMind
4 months
Introducing Gemma: a family of lightweight, state-of-the-art open models for developers and researchers to build with AI. 🌐 We’re also releasing tools to support innovation and collaboration - as well as to guide responsible use. Get started now. →
Tweet media one
134
535
2K
1
0
6
@rl_agent
Lisa Lee
1 year
The Barkour course consists of four obstacles in a 5m x 5m area, testing a diverse set of skills while keeping the setup within a small footprint. The scoring system is inspired by dog competition rules, with penalties for skipping/failing obstacles, or moving too slowly. (2/n)
1
0
5
@rl_agent
Lisa Lee
3 years
EcoRL Workshop will resume at 10 am PT for Session 2! Link: Submit questions for Live Q&A w/ our speakers: - Benjamin Van Roy @Stanford - @WarrenBPowell @Princeton - @yayitsamyzhang @UCBerkeley FAIR @UTAustin - Tom Griffiths @Princeton - @Mvandepanne @UBC
@EcoTheoryRL
Ecological Theory of Reinforcement Learning
3 years
Session II starts at 13:00 ET with Professor Benjamin van Roy from @Stanford @DeepMind and a live talk on “Environment Capacity.”  If you have any questions during the talk, please submit them at:
1
0
3
1
1
5
@rl_agent
Lisa Lee
1 year
We report the performance of the learned policies, which can complete the Barkour course at roughly half the speed of a dog. Additionally, state-of-the-art RL methods fail to learn a single policy to complete the course, underscoring the complexity and value of Barkour. (5/n)
Tweet media one
1
0
4
@rl_agent
Lisa Lee
1 year
Surprisingly, most combinations of factors do not have a compounding effect on generalization performance. For example, generalizing to the combination of new table textures & new distractors is no harder than new table textures alone (which is the harder of the two factors). 3/n
Tweet media one
1
1
4
@rl_agent
Lisa Lee
1 year
We also provide a simulated robot environment with 19 tasks and 11 factors of variation, to facilitate more controlled evaluations of generalization: 5/n
1
1
4
@rl_agent
Lisa Lee
1 year
We then independently evaluated generalization to each factor and found a pretty consistent ordering of difficulty across different datasets, models, and even between sim vs. real setups: 2/n
Tweet media one
1
1
4
@rl_agent
Lisa Lee
1 year
@ikostrikov @OpenAI Congrats Ilya! :)
0
0
2
@rl_agent
Lisa Lee
3 years
Then at 1:10 pm PT, join us for a Live Panel Discussion with: - Joelle Pineau @Mila_Quebec - Tom Griffiths @Princeton - @pyoudeyer @MSFTResearch @FlowersINRIA - @jeffclune @UBC moderated by @shaneguML . EcoRL Workshop:
0
0
2
@rl_agent
Lisa Lee
4 years
@rsalakhu Congratulations Russ, and thanks for being an amazing advisor all these years :)
0
0
2
@rl_agent
Lisa Lee
11 months
Emergent capabilities include: - Generalization to novel objects - Symbol understanding ("move apple to Google") - Human recognition ("move apple to Taylor Swift") - Semantic reasoning ("pick land animal", "pick object that is different", "move banana to the sum of 2+1") (2/n)
1
0
2
@rl_agent
Lisa Lee
2 years
@SimonShaoleiDu @NSF Congrats Simon! Well-deserved!
1
0
2
@rl_agent
Lisa Lee
2 years
@PangWeiKoh Congrats Pang-Wei, very well-deserved! 😊
0
0
1
@rl_agent
Lisa Lee
2 years
0
0
1
@rl_agent
Lisa Lee
3 years
@annadgoldie Congrats Anna!
0
0
1
@rl_agent
Lisa Lee
2 years
InstructRL is much simpler yet outperforms prior SOTA across a wide range of visual manipulation tasks. We found that InstructRL can generalize to unseen instructions, and can even understand longer, detailed step-by-step instructions to solve the task better.
Tweet media one
0
0
1
@rl_agent
Lisa Lee
1 year
@shaneguML @GoogleAI @OpenAI @johnschulman2 Congrats Shane! We'll miss you at Brain, but looking forward to your future endeavors.
0
0
1
@rl_agent
Lisa Lee
3 years
@pabbeel @IEEEorg Congrats Pieter!
0
0
1
@rl_agent
Lisa Lee
1 year
@vdean314 Break a leg, Victoria! 🥳
0
0
1