rezendi Profile Banner
Jon Evans Profile
Jon Evans

@rezendi

Followers
10K
Following
17K
Statuses
50K

engineer / novelist, occasional journalist / CTO / co-founder / archivist / peripatetist; see https://t.co/ozL3Xbg4RT

Latent space, these days
Joined March 2008
Don't wanna be here? Send us removal request.
@rezendi
Jon Evans
15 days
Some personal professional news: I'm joining to help use AI to answer hard questions accurately, and I am very excited to do so.
0
0
13
@rezendi
Jon Evans
3 hours
RT @MadelnCanada: Uptown, midtown, downtown of Toronto
Tweet media one
0
576
0
@rezendi
Jon Evans
1 day
RT @KelseyTuoc: I think it's completely fine that DOGE's staffers are young and the hysteria over their ages was inappropriate but you can'…
0
99
0
@rezendi
Jon Evans
2 days
RT @DanielleFong: follower dropped this interesting take in my DMs. truth computes, deception doesn’t. reminds me of @paulg adopting a stat…
0
4
0
@rezendi
Jon Evans
3 days
RT @aelfred_D: BREAKING: DOGE has discovered that USAID delved too greedily and too deep in the mines of Moria, awakening in the darkness o…
0
2K
0
@rezendi
Jon Evans
3 days
RT @emollick: Economist @joshgans uses o1-pro to generate a (minor, fun) paper in an hour based on an idea of his, and it gets published in…
0
274
0
@rezendi
Jon Evans
3 days
“Without the proper drone and electronic warfare support, an infantry unit will survive only a few hours on the battlefield.”
0
0
2
@rezendi
Jon Evans
3 days
This is amazing
0
0
0
@rezendi
Jon Evans
4 days
RT @meaning_enjoyer: the endless struggle between people loyal to coalitions and people loyal to principles
0
11
0
@rezendi
Jon Evans
5 days
@DavidSHolz berkeley, but yo
0
0
3
@rezendi
Jon Evans
7 days
RT @ryrzny: For the many people in tech cheering for these tariffs, here is a little supply chain 101. If you are manufacturing something…
0
10
0
@rezendi
Jon Evans
7 days
RT @ATabarrok: Now that you mention it, I am in fact tired of all this winning.
0
43
0
@rezendi
Jon Evans
8 days
RT @jasminewsun: where did the tech right come from? @pmarca's recent interview with @DouthatNYC is very illuminating here's a 4-part gra…
0
460
0
@rezendi
Jon Evans
9 days
I wrote about the unlooked-for return of the Zizians:
0
1
2
@rezendi
Jon Evans
10 days
RT @DavidSHolz: i worry a unsustainable need to raise billions of dollars has broken the discourse around agi/asi timelines. theres too man…
0
48
0
@rezendi
Jon Evans
10 days
RT @natfriedman: I am hiring a Synchrotron Tomography Reconstruction Expert at Vesuvius Challenge. If you know anyone who can do this job,…
0
29
0
@rezendi
Jon Evans
11 days
RT @janrosenow: Batteries start playing a huge role in energy systems around the world. Look at California - not long ago you wouldn’t eve…
0
278
0
@rezendi
Jon Evans
11 days
@AndrewCurran_ This question was asked less than a year ago
1
0
3
@rezendi
Jon Evans
11 days
RT @zerohedge: "Sir, China just released another model"
Tweet media one
0
5K
0
@rezendi
Jon Evans
12 days
"These thoughts are *emergent* (!!!) and this is actually seriously incredible, impressive and new"
@karpathy
Andrej Karpathy
12 days
I don't have too too much to add on top of this earlier post on V3 and I think it applies to R1 too (which is the more recent, thinking equivalent). I will say that Deep Learning has a legendary ravenous appetite for compute, like no other algorithm that has ever been developed in AI. You may not always be utilizing it fully but I would never bet against compute as the upper bound for achievable intelligence in the long run. Not just for an individual final training run, but also for the entire innovation / experimentation engine that silently underlies all the algorithmic innovations. Data has historically been seen as a separate category from compute, but even data is downstream of compute to a large extent - you can spend compute to create data. Tons of it. You've heard this called synthetic data generation, but less obviously, there is a very deep connection (equivalence even) between "synthetic data generation" and "reinforcement learning". In the trial-and-error learning process in RL, the "trial" is model generating (synthetic) data, which it then learns from based on the "error" (/reward). Conversely, when you generate synthetic data and then rank or filter it in any way, your filter is straight up equivalent to a 0-1 advantage function - congrats you're doing crappy RL. Last thought. Not sure if this is obvious. There are two major types of learning, in both children and in deep learning. There is 1) imitation learning (watch and repeat, i.e. pretraining, supervised finetuning), and 2) trial-and-error learning (reinforcement learning). My favorite simple example is AlphaGo - 1) is learning by imitating expert players, 2) is reinforcement learning to win the game. Almost every single shocking result of deep learning, and the source of all *magic* is always 2. 2 is significantly significantly more powerful. 2 is what surprises you. 2 is when the paddle learns to hit the ball behind the blocks in Breakout. 2 is when AlphaGo beats even Lee Sedol. And 2 is the "aha moment" when the DeepSeek (or o1 etc.) discovers that it works well to re-evaluate your assumptions, backtrack, try something else, etc. It's the solving strategies you see this model use in its chain of thought. It's how it goes back and forth thinking to itself. These thoughts are *emergent* (!!!) and this is actually seriously incredible, impressive and new (as in publicly available and documented etc.). The model could never learn this with 1 (by imitation), because the cognition of the model and the cognition of the human labeler is different. The human would never know to correctly annotate these kinds of solving strategies and what they should even look like. They have to be discovered during reinforcement learning as empirically and statistically useful towards a final outcome. (Last last thought/reference this time for real is that RL is powerful but RLHF is not. RLHF is not RL. I have a separate rant on that in an earlier tweet
0
0
1
@rezendi
Jon Evans
12 days
RT @jackclarkSF: I have sympathy for people studying DeepSeek cold. Reads like: Cyberbla's new "Pastrami" technique has increased throughp…
0
44
0