![Violet X. Profile](https://pbs.twimg.com/profile_images/1505458917312118789/KasJtjd8_x96.jpg)
Violet X.
@ZiyuX
Followers
158
Following
599
Statuses
106
PhD student @Stanford. Working on LLM-based agents
United States
Joined October 2011
RT @AndrewYNg: Introducing Agentic Object Detection! Given a text prompt like “unripe strawberries” or “Kellogg’s branded cereal” and an i…
0
731
0
RT @jiayi_pirate: We reproduced DeepSeek R1-Zero in the CountDown game, and it just works Through RL, the 3B base LM develops self-verifi…
0
1K
0
RT @synth_labs: Ever watched someone solve a hard math problem? Their first attempt is rarely perfect. They sketch ideas, cross things out…
0
51
0
RT @rm_rafailov: "Superintelligence isn't about discovering new things; it's about discovering new ways to discover" -> Meta RL
0
48
0
RT @rm_rafailov: We have a new position paper on "inference time compute" and what we have been working on in the last few months! We prese…
0
236
0
RT @sunfanyun: Training RL/robot policies requires extensive experience in the target environment, which is often difficult to obtain. How…
0
44
0
Excited about our new paper - Hypothetical Minds! The hypothesis-search-based approach shows a lot of promise in adapting to diverse agents in multi-agent settings. Check out the full paper for more!
Very excited to release a new paper introducing Hypothetical Minds! A LLM agent for multi-agent settings that generates hypotheses about other agents' latent states in natural language, adapting to diverse agents across collaborative, competitive, and mixed-motive domains🧵
0
0
6
RT @jphilippfranken: Constitutional AI showed LMs can learn to follow constitutions by labeling their own outputs. But why can't we just te…
0
34
0
RT @rm_rafailov: We have a new preprint out - your language model is not a reward, it’s a Q function! 1. The likelihood of the preferred an…
0
155
0
RT @gandhikanishk: Language models struggle to search, not due to an architecture problem, but a data one! They rarely see how to search or…
0
110
0
RT @jphilippfranken: When prompting language models to complete a task, users often leave important things unsaid. Can language models tea…
0
28
0
RT @ericzelikman: Language models today are trained to reason either 1) generally, imitating online reasoning data or 2) narrowly, self-tea…
0
187
0
RT @sunfanyun: Our research introduces a system that enables you to generate 3D environments from text prompts and train embodied AI agents…
0
60
0
RT @Saboo_Shubham_: OpenSource Multimodal AI Agent for Self-operating computers 🔥 Imagine an AI that doesn't just learn and adapt but int…
0
53
0
RT @jphilippfranken: Social Contract AI (SCAI): Aligning AI Assistants with Implicit Group Norms. We explore whether we can use an AI assi…
0
15
0
RT @jphilippfranken: 1/n Excited to share our work on Naive Information Aggregation in Human Social Learning (just out in Cognition)! Work…
0
10
0