![Joan Velja Profile](https://pbs.twimg.com/profile_images/1886830479384387584/TeMCVm30_x96.jpg)
Joan Velja
@Joanvelja
Followers
96
Following
644
Statuses
222
Bump up the epsilon! @UniofOxford | @LondonSafeAI | MS AI @UvA_Amsterdam
Joined December 2020
@johnhewtt In other words, if we assume, prior to seeing the machine's behavior, that the NL term captures the machine's internal representation, does the delta represent how much that assumption needs to be "corrected" to align with the machine's actual behavior (or a desired behavior)?
0
0
0
@geoffreyirving @krishnanrohit It’s the massive hiring during Covid (and subsequent layoffs) that make the chart, not GPT
0
0
2
@YouJiacheng @stalkermustang @__nmca__ @ahelkky Yep but seems unclear how (“learned scoring fn”); also selection is top 50
1
0
0
@__nmca__ @stalkermustang @ahelkky So what you are saying is that o3 has learned to become its own critic...?
0
0
1
@willccbb I reckon this is pretty much what’s behind Pro version and ARC score, which provides them yet another dimension for scaling
0
0
1
RT @natolambert: The fact that we seem to be marching straight towards another cold war, where AI is the defining technology, is hard to em…
0
18
0
@xuanalogue To be fair there seems to be discourse going on around this backtracking being part of the pretraining corpora (in subreddit-esque corpus), which the models learn to exploit during the outcome based RL training they go thru. Someone had found some evidence of this, ill follow up.
1
0
1
@willccbb @aryaman2020 I think a foundational piece of research most forget to mention is Belkin et al: Which came before all the OAI work on DD/Grokking
0
0
1
RT @mpshanahan: I agree with @AmandaAskell. It is a fallacy. It can be appropriate to say both a) that LLMs are next-token predictors, at a…
0
36
0