![Kelly Buchanan Profile](https://pbs.twimg.com/profile_images/1544686923536601090/7qXuYC_W_x96.jpg)
Kelly Buchanan
@ekellbuch
Followers
925
Following
12K
Statuses
527
Postdoctoral Fellow @Stanford building reliable AI for science. PhD from @cu_neurotheory @ZuckermanBrain. Industry: Research @GoogleAI
Palo Alto, CA
Joined July 2011
RT @allen_ai: We took our most efficient model and made an open-source iOS app📱but why? As phones get faster, more AI will happen on devic…
0
91
0
RT @deedydas: New research shows that LLMs don't perform well on long context. Perfect needle-in-the-haystack scores are easy—attention me…
0
263
0
RT @SuryaGanguli: My @TEDAI2024 talk is out! I discuss our work, spanning AI, physics, math & neuroscience, to deve…
0
42
0
RT @iScienceLuvr: Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach We study a novel language model architect…
0
180
0
RT @jenzhuscott: The founders of 3 new tech stars from China (DeepSeek, UniTree, Game Science) have in common: - post 80’s gen - born in…
0
576
0
RT @jacobaustin132: Making LLMs run efficiently can feel scary, but scaling isn’t magic, it’s math! We wanted to demystify the “systems vie…
0
364
0
RT @simran_s_arora: Officially rebranding "Ask Me Anything: A simple strategy for prompting language models" to "Ask Me Anything: A simple…
0
5
0
RT @GoogleAI: Today we introduce ZeroBAS, a neural method to synthesize binaural audio from mono audio recordings paired with positional in…
0
124
0
RT @GoogleAI: Large language models are often limited by restrictions on the length of their inputs. To address this, we propose Chain-of-A…
0
196
0
RT @karpathy: TinyZero reproduction of R1-Zero "experience the Ahah moment yourself for < $30" Given a base model, the RL finetuning can b…
0
428
0
RT @brad19brown: My fellow code monkeys (@jordanjuravsky @ryansehrlich) and I are excited to release CodeMonkeys: a system for solving SWE-…
0
40
0
RT @bfspector: We got early access to some of the very first Nvidia B200’s. We share initial benchmark results and wrote the fastest (publi…
0
26
0
RT @heyyalexwang: did you know you've been doing test-time learning this whole time? transformers, SSMs, RNNs, are all test-time regressor…
0
108
0
RT @behrouz_ali: Attention has been the key component for most advances in LLMs, but it can’t scale to long context. Does this mean we need…
0
608
0
RT @yifan_zhang_: 1/ Introducing “Tensor Product Attention Is All You Need” (TPA) and Tensor ProducT ATTenTion Transformer (T6)! 🚀 Ever wo…
0
66
0
RT @tamaybes: We're hiring a Technical Lead to develop a next-generation computer-use benchmark at Epoch AI. This will be for evaluating re…
0
17
0
RT @AnjneyMidha: hi infra ppl - some updates on that @mabb0tt and i are teaching this quarter: 1. lots of people a…
0
124
0
RT @MichaelWornow: 1/ 🏥 Does your New Year’s resolution involve more rigorous evaluation of ML models in healthcare? 📝 If so, our latest wo…
0
9
0
RT @karpathy: DeepSeek (Chinese AI co) making it look easy today with an open weights release of a frontier-grade LLM trained on a joke of…
0
3K
0