![Cristina Scheau Profile](https://pbs.twimg.com/profile_images/1241604609149300736/sR5vlj6D_x96.jpg)
Cristina Scheau
@cristina_scheau
Followers
1K
Following
573
Statuses
1K
head of ChatgptSearch@OpenAI, prev 🚗, Meta
San Francisco, CA
Joined June 2008
Congrats @EdwardSun0909 , @isafulf and the rest of the team for building an incredible model and product. It has been one of my favorite models so far!
Excited to finally share what I’ve been working on since joining OpenAI last June! The goal of deep-research is to enable reasoning models with tools to tackle long-horizon tasks in the real world and discover new knowledge. It’s a highly autonomous agent—hand it a hard problem, grab a coffee, and come back to a well-researched solution in 10–30 minutes. Trained end-to-end with reinforcement learning in a tool-enabled environment, deep-research is built to seek truth and understand the universe. A key milestone is its performance on humanity’s "last exam," demonstrating the true power of an end-to-end trained agent. 2025 is the year of agents. Looking forward to what’s ahead!
0
0
8
RT @OpenAI: OpenAI o1 is now out of preview in ChatGPT. What’s changed since the preview? A faster, more powerful reasoning model that’s b…
0
2K
0
AGI will need access to realtime information and be capable of reasoning with fresh data. It’s been a fun journey building something new and making ChatGPT hopefully even better than today. This is just the beginning. I believe agentic browsing will revolutionize our daily lives.
🌐 Introducing ChatGPT search 🌐 ChatGPT can now search the web in a much better way than before so you get fast, timely answers with links to relevant web sources.
3
2
138
Before Phi-3 takes over the news cycle 🙂, sharing my public congratulations to my former colleagues @AIAtMeta for llama3 & launch. I’ve been busy trying it out and I’m very impressed.🔥A huge week for the open source AI. Andrej has excellent notes
Congrats to @AIatMeta on Llama 3 release!! 🎉 Notes: Releasing 8B and 70B (both base and finetuned) models, strong-performing in their model class (but we'll see when the rankings come in @ @lmsysorg :)) 400B is still training, but already encroaching GPT-4 territory (e.g. 84.8 MMLU vs. 86.5 4Turbo). Tokenizer: number of tokens was 4X'd from 32K (Llama 2) -> 128K (Llama 3). With more tokens you can compress sequences more in length, cites 15% fewer tokens, and see better downstream performance. Architecture: no major changes from the Llama 2. In Llama 2 only the bigger models used Grouped Query Attention (GQA), but now all models do, including the smallest 8B model. This is a parameter sharing scheme for the keys/values in the Attention, which reduces the size of the KV cache during inference. This is a good, welcome, complexity reducing fix and optimization. Sequence length: the maximum number of tokens in the context window was bumped up to 8192 from 4096 (Llama 2) and 2048 (Llama 1). This bump is welcome, but quite small w.r.t. modern standards (e.g. GPT-4 is 128K) and I think many people were hoping for more on this axis. May come as a finetune later (?). Training data. Llama 2 was trained on 2 trillion tokens, Llama 3 was bumped to 15T training dataset, including a lot of attention that went to quality, 4X more code tokens, and 5% non-en tokens over 30 languages. (5% is fairly low w.r.t. non-en:en mix, so certainly this is a mostly English model, but it's quite nice that it is > 0). Scaling laws. Very notably, 15T is a very very large dataset to train with for a model as "small" as 8B parameters, and this is not normally done and is new and very welcome. The Chinchilla "compute optimal" point for an 8B model would be train it for ~200B tokens. (if you were only interested to get the most "bang-for-the-buck" w.r.t. model performance at that size). So this is training ~75X beyond that point, which is unusual but personally, I think extremely welcome. Because we all get a very capable model that is very small, easy to work with and inference. Meta mentions that even at this point, the model doesn't seem to be "converging" in a standard sense. In other words, the LLMs we work with all the time are significantly undertrained by a factor of maybe 100-1000X or more, nowhere near their point of convergence. Actually, I really hope people carry forward the trend and start training and releasing even more long-trained, even smaller models. Systems. Llama 3 is cited as trained with 16K GPUs at observed throughput of 400 TFLOPS. It's not mentioned but I'm assuming these are H100s at fp16, which clock in at 1,979 TFLOPS in NVIDIA marketing materials. But we all know their tiny asterisk (*with sparsity) is doing a lot of work, and really you want to divide this number by 2 to get the real TFLOPS of ~990. Why is sparsity counting as FLOPS? Anyway, focus Andrej. So 400/990 ~= 40% utilization, not too bad at all across that many GPUs! A lot of really solid engineering is required to get here at that scale. TLDR: Super welcome, Llama 3 is a very capable looking model release from Meta. Sticking to fundamentals, spending a lot of quality time on solid systems and data work, exploring the limits of long-training models. Also very excited for the 400B model, which could be the first GPT-4 grade open source release. I think many people will ask for more context length. Personal ask: I think I'm not alone to say that I'd also love much smaller models than 8B, for educational work, and for (unit) testing, and maybe for embedded applications etc. Ideally at ~100M and ~1B scale. Talk to it at Integration with
0
0
6
RT @Cruise: Sometimes, vision is not enough (especially at night). On this fully driverless ride, our perception stack picks up this cyclis…
0
19
0
Last night, I had my first driverless ride for 1 hr. The ride was smooth and flawless. It’s been a surreal and very rewarding experience. Hard to describe it, but i am still speechless. It is deeply gratifying to be part of such an incredible journey. #lifeatcruise
2
0
40
RT @olivercameron: The future really is here. I was in a @Cruise AV for 70+ minutes today—moving all around SF—and it was dramatically saf…
0
20
0
RT @Cruise: Meet Poppy, one of our favorite self-driving vehicles who’s making friends across #SanFrancisco. Follow her adventures @PoppyTh…
0
51
0