Shalev_lif Profile Banner
Shalev Lifshitz Profile
Shalev Lifshitz

@Shalev_lif

Followers
1K
Following
1K
Media
66
Statuses
1K

a non-artificial intelligence working on scaling artificial reasoning

Toronto
Joined September 2017
Don't wanna be here? Send us removal request.
@Shalev_lif
Shalev Lifshitz
6 days
As we learn to use compute more efficiently, we will use more compute.
1
2
5
@Shalev_lif
Shalev Lifshitz
2 months
Best poster moment at #NeurIPS2024
Tweet media one
29
773
11K
@Shalev_lif
Shalev Lifshitz
7 days
A new replication of DeepSeek's RL results! Here are my notes and some quick thoughts:. Method:.- Uses PPO instead of GRPO (DeepSeek-R1), still works.- Data is 8K (query, final answer) examples from MATH.- Rule-based reward modelling (no neural reward).- Initialize model to
Tweet media one
@junxian_he
Junxian He
7 days
We replicated the DeepSeek-R1-Zero and DeepSeek-R1 training on 7B model with only 8K examples, the results are surprisingly strong. šŸš€ Starting from Qwen2.5-Math-7B (base model), we perform RL on it directly. No SFT, no reward model, just 8K MATH examples for verification, the
Tweet media one
Tweet media two
7
50
357
@Shalev_lif
Shalev Lifshitz
3 months
Great to see @geoffreyhinton at the @VectorInst office today. Here @michaelrzhang is presenting his work on qualitative eval of LLMs!. Very cool to have a Nobel Laureate + Turing Award winner around the campus.
Tweet media one
4
14
330
@Shalev_lif
Shalev Lifshitz
2 months
My question to @ilyasut at NeurIPS 2024: Do LLMs generalize multi-hop reasoning out-of-distribution?
7
18
293
@Shalev_lif
Shalev Lifshitz
2 months
Absolutely stacked panel at the System-2 Reasoning at Scale workshop at NeurIPS with Josh Tenenbaum, @MelMitchell1, @fchollet, @jaseweston, @DBahdanau, @dawnsongtweets, and. @Yoshua_Bengio (with @nouhadziri moderating). An amazing end to the conference. Will add notes below.
Tweet media one
11
24
241
@Shalev_lif
Shalev Lifshitz
1 year
@karpathy This reminds me of a meme @_jasonwei posted a while back! That is, once you play with these models so much you kind of develop your own mini test suite to gain intuition of its performance.
Tweet media one
3
9
198
@Shalev_lif
Shalev Lifshitz
2 months
@ilyasut giving a talk at the NeurIPS 2024 Test of Time awards! . Will add more photos below, throughout the talk.
Tweet media one
1
7
106
@Shalev_lif
Shalev Lifshitz
2 months
See the full paper by @shreyaskapur, @jenner_erik, and Stuart Russel! .
2
5
103
@Shalev_lif
Shalev Lifshitz
1 month
In a few years PhDs wonā€™t be coding much. Theyā€™ll have a fleet of agents coding up, running, and tuning their experiments. At that time, the most valuable skill will be deep expertise, as suggested by @RogerGrosse.
@roydanroy
Dan Roy šŸ‡ØšŸ‡¦šŸ‡©šŸ‡°šŸ‡±šŸ‡¹
1 month
@tunguz In all seriousness, PhDs today will have tools so powerful that previous generations wonā€™t know what to think of them. I think it is the most exciting time to be working. Just donā€™t work in an old way.
8
4
102
@Shalev_lif
Shalev Lifshitz
1 year
šŸ„³ Great news! Our paper STEVE-1 has been accepted at #NeurIPS 2023 as a spotlight!. I'm so proud to have worked on this project with my amazing collaborators @keirp1 @SirrahChan @jimmybajimmyba @SheilaMcIlraith. āœˆļø Very excited to present our work in New Orleans! āœˆļø. Project
7
6
97
@Shalev_lif
Shalev Lifshitz
2 months
The reason I initially got into RL is because the problem setting just seems much more like what AGI would be based on: optimizing for a goal, without basing on human-provided answers. Glad to see RL is making a comeback.
@_jasonwei
Jason Wei
2 months
2022: I never wrote a RL paper or worked with a RL researcher. I didnā€™t think RL was crucial for AGI. Now: I think about RL every day. My code is optimized for RL. The data I create is designed just for RL. I even view life through the lens of RL. Crazy how quickly life changes.
4
1
89
@Shalev_lif
Shalev Lifshitz
5 months
@jxmnop The ā€˜fastā€™ operations area of the GPU has lower memory. I recommend reading the first few pages of the Flash Attention paper, it goes into this and explains it super well!.
1
2
77
@Shalev_lif
Shalev Lifshitz
2 months
Genie 2: open-ended *playable* world modelling. Amazing work from the team @GoogleDeepMind! They even show the generalist SIMA agent playing inside of Genie 2. Weā€™re one step closer to the near-infinite training data regime for embodied agents. This is big!.
@jparkerholder
Jack Parker-Holder
2 months
Introducing šŸ§žGenie 2 šŸ§ž - our most capable large-scale foundation world model, which can generate a diverse array of consistent worlds, playable for up to a minute. We believe Genie 2 could unlock the next wave of capabilities for embodied agents šŸ§ .
0
7
39
@Shalev_lif
Shalev Lifshitz
1 month
@xuanalogue One good thing about o1/3 from a safety perspective is that the long CoTs are at least in natural language. If we make the models small enough so that they canā€™t do much multi-hop reasoning implicitly, this may be helpful from a safety sense.
5
0
33
@Shalev_lif
Shalev Lifshitz
1 year
@DrJimFan Agreed about the GPT-3 comparison, we need to be thinking about a few papers down the line. And regarding world models: Iā€™m still on the fence for LLMs but with video-gen it seems much more clear. Though good science is still needed. @yudapearl had a great question about this a.
1
3
25
@Shalev_lif
Shalev Lifshitz
3 months
@hyhieu226 Asked an undisclosed language model, got: "BERT isn't a person named Bert". Lol.
1
0
26
@Shalev_lif
Shalev Lifshitz
6 days
Itā€™s crazy watching mainstream media try to explain the DeepSeek results. I just watched CNBC compare the cost of a single training run for DeepSeek-V3 ($5.6M) to OpenAIā€™s total spend in a given year ($5.4B). Come onā€¦ seriously?.
5
2
87
@Shalev_lif
Shalev Lifshitz
2 months
We are in a period of time where there is too little diversity in research, almost everyone is toying around with the same architectures and ideas. We need much more diversity, and academia is much better at this than industry (the freedom to explore much more broadly).
2
4
26
@Shalev_lif
Shalev Lifshitz
5 months
@karpathy @levelsio Future startups: a few people + AI copilot = billion dollar valuation.
1
2
25
@Shalev_lif
Shalev Lifshitz
2 years
Happy you found our work interesting @_akhaliq, it was great chatting about the impact and future directions of STEVE-1 at ICML today!.
@_akhaliq
AK
2 years
STEVE-1: A Generative Model for Text-to-Behavior in Minecraft. paper page: Constructing AI models that respond to text instructions is challenging, especially for sequential decision-making tasks. This work introduces an instruction-tuned Video
1
6
24
@Shalev_lif
Shalev Lifshitz
7 days
@denny_zhou Agree about RL democratizing reasoning. But I feel that it's not quite RLHF, and just "RL with good SFT warmup", since we're using even less human feedback now (no neural reward model).
3
0
22
@Shalev_lif
Shalev Lifshitz
2 months
Very fortunate to have a talk from Ilya at NeurIPS this year. He sees what others cannot.
0
0
22
@Shalev_lif
Shalev Lifshitz
4 months
I'm in awe at the fact that "openai", "gemini", and "claude" are all the same length. Thank you Deep Learning god for keeping my code aligned and beautiful. šŸ„¹šŸ™. P.S. Yes, 'gemini' and 'claude' should really be 'gdm' and 'anthropic' or something, but. oh well.
Tweet media one
1
0
21
@Shalev_lif
Shalev Lifshitz
7 days
@rm_rafailov I think it is likely to have curated data and synthetic data in pre/mid-training, but I still find it quite impressive that such simple RL works on top. That being said, it does seem the general community on Twitter is starting to think that you only need RL and pre-training on.
1
2
20
@Shalev_lif
Shalev Lifshitz
7 months
@ClementDelangue @huggingface Amazing news! Congrats and well-deserved!.
0
0
15
@Shalev_lif
Shalev Lifshitz
7 days
@Swarooprm7 @johnschulman2 Unlikely since full o1 CoTs are hidden.
2
0
17
@Shalev_lif
Shalev Lifshitz
2 months
@pcastr What about this line of work? (This particular paper is a great one by @akyurekekin).
0
3
18
@Shalev_lif
Shalev Lifshitz
4 months
@karpathy Great work from Google. In a few years, youā€™ll have AI-generated lessons that can be either listened to or interacted with - and personalized per student. I also wonder what post-podcast interfaces AI might unlockā€¦.
1
1
18
@Shalev_lif
Shalev Lifshitz
6 months
Itā€™s hard to overstate the significance of the ā€˜AI scientistā€™ line of work to the development of general intelligence. The frontier labs have closed teams working on automating AI research, but this is one of the first major *open* steps in the same direction.
@SakanaAILabs
Sakana AI
6 months
Introducing The AI Scientist: The worldā€™s first AI system for automating scientific research and open-ended discovery!. From ideation, writing code, running experiments and summarizing results, to writing entire papers and conducting peer-review, The AI
Tweet media one
Tweet media two
Tweet media three
Tweet media four
2
0
15
@Shalev_lif
Shalev Lifshitz
2 months
Thought I entered the matrix today at the @GoogleDeepMind booth when I saw 100 people all wearing headphones and looking at a screen.
Tweet media one
2
0
17
@Shalev_lif
Shalev Lifshitz
2 years
STEVE-1 follows text and visual goals in Minecraft while using only raw pixels and keyboard/mouse controls! Very excited to finally share this work!. w/ @keirp1, @SirrahChan, @jimmybajimmyba, @SheilaMcIlraith.
@keirp1
Keiran Paster
2 years
Meet STEVE-1, an instructable generative model for Minecraft. STEVE-1 follows both text and visual instructions and acts on raw pixel inputs with keyboard and mouse controls. Best of all - it only cost $60 to train!. w/ @Shalev_lif @SirrahChan @jimmybajimmyba @SheilaMcIlraith
1
7
14
@Shalev_lif
Shalev Lifshitz
2 months
@jparkerholder Really amazing work! We can finally combine open-ended agents and open-ended world models. Weā€™re getting closer to the near-infinite training data regime.
0
2
16
@Shalev_lif
Shalev Lifshitz
2 months
Superintelligence is next, the long term:.- agentic (current models are barely agentic).- reasoning. seeing some early signs, unpredictable, chess AIs are unpredictable to best human chess players. Weā€™ll have to deal with LLMs that are unpredictable. - self-awareness.
0
2
14
@Shalev_lif
Shalev Lifshitz
7 days
@johnschulman2 Seems like this comes from the long CoT SFT warmup that happens before RL (including the synthetic reasoning data that is a part of pre/mid-training). Unless there's something about these words or the way they are used in natural text that causes them to emerge through pure RL,.
1
0
14
@Shalev_lif
Shalev Lifshitz
6 months
I loved this quote from @cong_ml (co-first-author of The AI Scientist) in todayā€™s AI papers of the week space hosted by @iScienceLuvr and @arankomatsuzaki: . ā€œThis is the ā€˜Will Smith eating spaghettiā€™ moment of AI science. This is the worst it will ever be.ā€. Well said @cong_ml,
1
0
12
@Shalev_lif
Shalev Lifshitz
1 year
@DrJimFan Itā€™s being used as a no-op (or one-op) in the current demonstrations, but it can also be used as an action taking agent, and I have to assume they are testing this internally. Formula to convert Sora to an action-taking agent: Condition on past frames from an embodied agentā€™s.
4
2
12
@Shalev_lif
Shalev Lifshitz
2 years
@DrJimFan Easy verification is key!.
1
0
12
@Shalev_lif
Shalev Lifshitz
2 months
ā€œPre-training as we know it will endā€
Tweet media one
1
0
12
@Shalev_lif
Shalev Lifshitz
2 months
Hominids have a different slope (ie, scaling law) between body mass and brain mass (note, axes are log scale!), compared to other animals. An example from nature
Tweet media one
1
0
12
@Shalev_lif
Shalev Lifshitz
1 year
@RogerGrosse Thanks for sharing, fascinating read. Truly ahead of his time. Great quote from the paper: "It is sometimes worthwhile to take science fiction seriously.".
0
0
11
@Shalev_lif
Shalev Lifshitz
6 months
Looking forward to the future of post-training at Anthropic now that John has joined the team!.
@johnschulman2
John Schulman
6 months
I shared the following note with my OpenAI colleagues today:. I've made the difficult decision to leave OpenAI. This choice stems from my desire to deepen my focus on AI alignment, and to start a new chapter of my career where I can return to hands-on technical work. I've decided.
0
1
11
@Shalev_lif
Shalev Lifshitz
1 month
Its been way too long since I saw my two-legged thumb friends and 4-armed spider friends. Brings me back.
@carlo_sferrazza
Carlo Sferrazza
2 months
šŸšØ New reinforcement learning algorithms šŸšØ. Excited to announce MaxInfoRL, a class of model-free RL algorithms that solves complex continuous control tasks (including vision-based!) by steering exploration towards informative transitions. Details in the thread šŸ‘‡
0
1
11
@Shalev_lif
Shalev Lifshitz
7 days
Why do similar words like "wait" and "alternatively" appear in both o1 and r1 CoTs?. I think either:.1. The pre-RL SFT data is coming from similar sources. 2. Different sources are generating similar pre-RL SFT data (for example, maybe the deepseek team saw the few published o1.
@johnschulman2
John Schulman
7 days
There are some intriguing similarities between the r1 chains of thought and the o1-preview CoTs shared in papers and blog posts (eg . In particular, note the heavy use of the words "wait" and "alternatively" as a transition words for error correction and.
5
0
11
@Shalev_lif
Shalev Lifshitz
1 month
Remember when 5-6 years were bullish timelines?.
@tsarnick
Tsarathustra
1 month
Yann LeCun says AGI will not happen in the next 2 years - it will take 5-6 years if everything goes well because the history of AI shows that people keep underestimating how hard it is
0
0
11
@Shalev_lif
Shalev Lifshitz
2 years
Excited to be in Hawaii this week to present our recent work STEVE-1 at ICML 2023! STEVE-1 is an instructable agent that can accept any text instruction and act using raw keyboard/mouse controls. By treating policy learning as a generative task, we show that itā€™s possible to
Tweet media one
2
0
11
@Shalev_lif
Shalev Lifshitz
1 year
100% recall for needle-in-a-haystack retrieval over 11 HOURS of audio? This just gets better and betterā€¦.
@JeffDean
Jeff Dean
1 year
Audio haystack. For audio, Gemini 1.5 Pro achieves 100% recall when looking for different audio needles hidden in ~11 hours of audio.
Tweet media one
0
0
10
@Shalev_lif
Shalev Lifshitz
11 months
This is such an important point and a huge question mark. How many jobs can be automated with a 1% error rate? If AGI = automating most economically valuable work, the this question is imperative.
@EugeneVinitsky
Eugene Vinitsky šŸ’
11 months
@Shalev_lif There are not many problems where a 1% error rate will be acceptable. Problems that have that property (for example reading long docs where compliance rates are already low) will crack.
4
1
9
@Shalev_lif
Shalev Lifshitz
2 months
@fchollet @polynoamial Search as an idea is fully general, but most implementations are specialized. The interesting thing about o1 is that it discovers truly general search techniques (ie, backtracking) via train-time RL. @polynoamial mentioned this recently.
1
1
8
@Shalev_lif
Shalev Lifshitz
8 months
@Sumnasca @ilyasut Cracked is a common term I've heard in startup/bay culture.
0
0
7
@Shalev_lif
Shalev Lifshitz
4 months
Wow! Congratulations to @HopfieldJohn and @geoffreyhinton who have just been awarded the Nobel Prize in Physics for their foundational work re: neural networks!. @geoffreyhinton now holds both an ACM Turing Award *and* a Nobel Prize in Physics. What an amazing achievement! I.
@NobelPrize
The Nobel Prize
4 months
BREAKING NEWS.The Royal Swedish Academy of Sciences has decided to award the 2024 #NobelPrize in Physics to John J. Hopfield and Geoffrey E. Hinton ā€œfor foundational discoveries and inventions that enable machine learning with artificial neural networks.ā€
Tweet media one
5
0
7
@Shalev_lif
Shalev Lifshitz
2 months
The trend these days is doing more inference during training and more training during inference.
1
0
8
@Shalev_lif
Shalev Lifshitz
8 months
Ironically this probably gives the work even more attention. New strategy unlocked?.
@CVPR
#CVPR2025
8 months
Poster missing? Improvise! Sorry about that.
Tweet media one
1
0
7
@Shalev_lif
Shalev Lifshitz
2 months
AGI itself is a flawed term that we should move away from. More concrete notions like ā€œautomating x% of remote workersā€, ā€œsaving x-type workers y hours each weekā€, ā€œincreasing productivity by x%ā€, and other economically driven definitions are much more meaningful and useful. AGI.
@aidan_mclau
Aidan McLaughlin
2 months
o1 is "better than most humans at most tasks" because, yes, humans exist exclusively in amnesic disembodied multi-turn chat interfaces.
1
0
7
@Shalev_lif
Shalev Lifshitz
2 years
Exciting advancement by Google DeepMind with the newly introduced Robotic Transformer 2 (RT-2)!. Main idea: train a single agent on 1) web-scale text/image data and 2) robot data, to create a single Vision-Language-Action (VLA) model. This VLA model can directly output actions to
Tweet media one
0
1
6
@Shalev_lif
Shalev Lifshitz
2 months
As you scale things up, tiny hacks matter less, what matters more is matrix multiplication and ReLU.
1
0
7
@Shalev_lif
Shalev Lifshitz
1 month
@denny_zhou Humans generate one action at a time but we perform search when solving problems.
1
0
7
@Shalev_lif
Shalev Lifshitz
4 months
At the announcement of the 2024 Nobel Prize in Physics this morning, @geoffreyhinton was asked about the future impact of AI on our civilizations. Here's what he said:. "I think it will have a huge influence. It'll be comparable with the industrial revolution, but instead of.
1
1
7
@Shalev_lif
Shalev Lifshitz
2 months
Abstraction is difficult to see inside an LLM, but neural nets can speak out about their abstractions.
1
0
7
@Shalev_lif
Shalev Lifshitz
7 months
@nathanbenaich @twentybn @muellerfreitag Yup, it's timing but also now the tech is really getting there and making it possible. Really excited for the future of Fei-Fei's company, she's awesome!.
0
0
7
@Shalev_lif
Shalev Lifshitz
9 days
@XingyouSong I can hear this picture šŸ˜†.
1
0
7
@Shalev_lif
Shalev Lifshitz
5 months
@elonmusk Itā€™s been interesting to see the distribution of opinions on 1047 among researchers. For those unaware: lots of AI researchers are in favour of regulation, but thereā€™s been criticism that 1047 hurts open-source too much (though a recent change in 1047 tries to address this).
1
1
7
@Shalev_lif
Shalev Lifshitz
3 months
@AnthropicAI @github Nice! Claude has been my go-to coding assistant (other than copilot). Happy to finally be able to use it with copilot. Thanks for the hard work!.
0
0
7
@Shalev_lif
Shalev Lifshitz
3 months
@DannyDriess @physical_int Didnā€™t know you were at @physical_int Danny! Amazing work!.
0
0
7
@Shalev_lif
Shalev Lifshitz
2 months
Future benchmarks may need to have a compute price, to account for the efficiency of inference-time algorithms.
0
0
7
@Shalev_lif
Shalev Lifshitz
2 months
@ericzelikman I love this šŸ˜†.
0
0
8
@Shalev_lif
Shalev Lifshitz
1 year
@DrJimFan Ah oops, I misinterpreted that too. Thanks for the clarification! Still extremely impressive and hopefully we get this level of truly autonomous capability in 2024.
0
0
1
@Shalev_lif
Shalev Lifshitz
1 year
@Deepneuron *If* we think AI will eventually be able to do almost all economically valuable work, this I guess we need an economy-scale data centre to power it!.
0
0
6
@Shalev_lif
Shalev Lifshitz
1 month
The relief when your method successfully transfers to a new benchmark.
Tweet media one
0
0
6
@Shalev_lif
Shalev Lifshitz
3 months
@gadgetKarens @EugeneVinitsky Hahaha this made me laugh. @EugeneVinitsky seems like you have an AI friend.
0
0
6
@Shalev_lif
Shalev Lifshitz
7 days
@inductionheads @rm_rafailov It's more that the important moves must have a high-enough probability as to be sampled enough to be reinforced during RL. In other words, the good moves need to be common enough so that they happen during training.
0
0
6
@Shalev_lif
Shalev Lifshitz
6 months
@aheze0 throwback to multivar calc, good old stokes theorem.
0
0
4
@Shalev_lif
Shalev Lifshitz
7 days
Extra important question:.Does Qwen-2.5-Math-Base have synthetic reasoning / long CoT data in pre/mid-training ? If so, should we still call the Zero model a ā€œcold startā€? This is all very new so the community still has to figure out these details. But this is an important.
@Shalev_lif
Shalev Lifshitz
9 days
Assuming a base model was trained with synthetic long reasoning during pre/mid-training, and then RL is applied, is it right to view this as RL with no SFT?. In other words, is this a cold start?.
1
0
6
@Shalev_lif
Shalev Lifshitz
1 month
The general consensus among top folks has been that math will be solved first, then code. Seems right. We'll get ASI in math and code before we get AGI for everything else.
1
0
6
@Shalev_lif
Shalev Lifshitz
4 months
Interesting theory paper on hallucinations. "By establishing the mathematical certainty of hallucinations, we challenge the prevailing notion that they can be fully mitigated.". [2409.05746] LLMs Will Always Hallucinate, and We Need to Live With This (.
0
0
6
@Shalev_lif
Shalev Lifshitz
5 months
Current AIs aren't curious! They seldom ask questions to learn more. Why is it important for AI to ask questions?. For one, it can help prevent hallucinations: ie, by asking follow-up questions instead of generating responses with high uncertainty. But more importantly: how do
Tweet media one
2
0
6
@Shalev_lif
Shalev Lifshitz
4 months
@tszzl Seems like the global banking system will run on Cobalt until the end of timeā€¦.
0
0
5
@Shalev_lif
Shalev Lifshitz
1 year
A major issue with current LLMs is that they often donā€™t ask questions when they are unsure about something, whether that be facts, user intent, etc. We must train these systems to ask questions, not just answer them.
1
0
5
@Shalev_lif
Shalev Lifshitz
1 year
@DrJimFan Agreed. We canā€™t just train a domain-specific IDM on every domain. Weā€™re moving down the hierarchy of core pieces needed for generalist embodied agents:.ā€” Semantic understanding and high level planning with text actions: a general VLM. ā€” Low level dynamic planning: a general.
0
1
4
@Shalev_lif
Shalev Lifshitz
6 months
News from ICML: Llama 3 405b might drop tomorrow.
0
0
5
@Shalev_lif
Shalev Lifshitz
11 months
@AndrewLampinen Really awesome, canā€™t wait to delve deeper into the paper this week!.
1
0
4
@Shalev_lif
Shalev Lifshitz
11 months
Huge developments, especially with the chat interface for everyday users. Mistral is shaping up to be a major competitor!.
@GuillaumeLample
Guillaume Lample @ NeurIPS 2024
11 months
Today, we are releasing Mistral Large, our latest model. Mistral Large is vastly superior to Mistral Medium, handles 32k tokens of context, and is natively fluent in English, French, Spanish, German, and Italian. We have also updated Mistral Small on our API to a model that is
Tweet media one
Tweet media two
Tweet media three
0
0
5
@Shalev_lif
Shalev Lifshitz
6 months
Very excited to have Zhijing join @UofTCompSci and @VectorInst!.
@ZhijingJin
Zhijing Jin
6 months
Happy to announce that I'm joining as an Asst. Prof. in CS at UToronto @UofTCompSci+@VectorInst in Fall '25, working on #NLProc, Causality, and AI Safety! I want to sincerely thank my dear mentors, friends, collabs & many who mean a lot to me. Welcome #PhDs/Research MSc to apply!.
1
0
5
@Shalev_lif
Shalev Lifshitz
2 months
@iScienceLuvr Interesting that this and coconut are released in the same week.
1
0
5
@Shalev_lif
Shalev Lifshitz
1 year
2024 is going to be exciting.
@RWKV_AI
RWKV
1 year
Introducing Eagle-7B. Based on the RWKV-v5 architecture, bringing into opensource space, the strongest.- multi-lingual model . (beating even mistral).- attention-free transformer today . (10-100x+ lower inference). With comparable English performance with the best 1T 7B models
Tweet media one
1
1
5
@Shalev_lif
Shalev Lifshitz
1 month
We are going to have ā€œsuperintelligentā€ AI that can make scientific breakthroughs in certain domains (ie, math) before we have ā€œAGIā€ in all domains.
0
0
5
@Shalev_lif
Shalev Lifshitz
5 months
Ilyaā€™s new company just raised $1B. Iā€™m excited. Itā€™s hard to overstate Ilyaā€™s brilliance.
@ssi
SSI Inc.
5 months
SSI is building a straight shot to safe superintelligence. Weā€™ve raised $1B from NFDG, a16z, Sequoia, DST Global, and SV Angel. Weā€™re hiring:
0
0
5
@Shalev_lif
Shalev Lifshitz
1 year
@DrJimFan Weā€™ve exhausted a lot of the text on the internet at this point, so the 10s if not 100s of TRILLIONS of tokens on YouTube are key for future LLM/VLM training. And merging with something like Sora (maybe directly, maybe through a dynamic prompting layer) is a key next step.
1
1
5
@Shalev_lif
Shalev Lifshitz
1 month
Happy New Year!.
0
0
5
@Shalev_lif
Shalev Lifshitz
20 days
@OpenAI should make it possible to upload .py files to o1! GPT-4o accepts .py files, but o1 doesn't. Why?
Tweet media one
1
0
5
@Shalev_lif
Shalev Lifshitz
1 year
@DrJimFan Very exciting times indeed. 2023 was an AI culture shock. As the dust settles, 2024 will reveal the practical path towards truly generalist agents.
0
0
4
@Shalev_lif
Shalev Lifshitz
1 year
@DrJimFan Yup, and I've been hearing that some companies are training code models on commit histories and their changes. Will usher a new way to code for sure.
0
0
5
@Shalev_lif
Shalev Lifshitz
22 days
another day, another 6am coding grind.
0
0
5
@Shalev_lif
Shalev Lifshitz
5 months
@AnthropicAI I was literally just wishing for this yesterday :).
0
0
5
@Shalev_lif
Shalev Lifshitz
24 days
This is me leading up to conference deadlines. I get out of bed, into my chair, and keep coding where I left off, trying to solve some problem I couldnā€™t figure out before I went to sleep. Not very sustainable though šŸ˜… (turns out breakfast is important).
@byraied
raied.
24 days
I read this often.
Tweet media one
1
0
5
@Shalev_lif
Shalev Lifshitz
3 months
@AnthropicAI Thanks! Been waiting to add it as a benchmark!.
0
0
5
@Shalev_lif
Shalev Lifshitz
1 year
@arankomatsuzaki Personalized AI that knows your paper preferences, what your working on, is a semi expert in your field, and can even suggest ideas related to the papers and your current research. Itā€™s comingā€¦.
1
0
5
@Shalev_lif
Shalev Lifshitz
1 year
@karpathy Now this is what weā€™ve been waiting for!.
0
1
5
@Shalev_lif
Shalev Lifshitz
1 year
@deliprao We did the same thing with AlexNet for years in computer vision, I donā€™t see a real problem as itā€™s always good for newcomers, and to pay respects to those who built the foundations of our work.
0
0
5
@Shalev_lif
Shalev Lifshitz
2 months
@slow_developer Importantly missing: Ilya said pretraining ā€œas we know itā€ is over.
0
0
5
@Shalev_lif
Shalev Lifshitz
23 days
@xai Awesome. Congrats to the team!.
0
1
4