Joanvelja Profile Banner
Joan Velja Profile
Joan Velja

@Joanvelja

Followers
96
Following
644
Statuses
222

Bump up the epsilon! @UniofOxford | @LondonSafeAI | MS AI @UvA_Amsterdam

Joined December 2020
Don't wanna be here? Send us removal request.
@Joanvelja
Joan Velja
11 days
@CFGeek Even better, if you will
0
3
25
@Joanvelja
Joan Velja
32 minutes
@johnhewtt In other words, if we assume, prior to seeing the machine's behavior, that the NL term captures the machine's internal representation, does the delta represent how much that assumption needs to be "corrected" to align with the machine's actual behavior (or a desired behavior)?
0
0
0
@Joanvelja
Joan Velja
13 hours
@geoffreyirving @krishnanrohit It’s the massive hiring during Covid (and subsequent layoffs) that make the chart, not GPT
0
0
2
@Joanvelja
Joan Velja
15 hours
@lateinteraction @DimitrisPapail Much much appreciated!
0
0
1
@Joanvelja
Joan Velja
16 hours
@YouJiacheng @stalkermustang @__nmca__ @ahelkky Yep but seems unclear how (“learned scoring fn”); also selection is top 50
1
0
0
@Joanvelja
Joan Velja
18 hours
@__nmca__ @stalkermustang @ahelkky So what you are saying is that o3 has learned to become its own critic...?
0
0
1
@Joanvelja
Joan Velja
20 hours
@PreethiLahoti Pinged in DMs :)
0
0
0
@Joanvelja
Joan Velja
21 hours
@willccbb The exact amount we will never know :) time will tell
0
0
1
@Joanvelja
Joan Velja
21 hours
@willccbb I reckon this is pretty much what’s behind Pro version and ARC score, which provides them yet another dimension for scaling
0
0
1
@Joanvelja
Joan Velja
2 days
RT @natolambert: The fact that we seem to be marching straight towards another cold war, where AI is the defining technology, is hard to em…
0
18
0
@Joanvelja
Joan Velja
2 days
@xuanalogue To be fair there seems to be discourse going on around this backtracking being part of the pretraining corpora (in subreddit-esque corpus), which the models learn to exploit during the outcome based RL training they go thru. Someone had found some evidence of this, ill follow up.
1
0
1
@Joanvelja
Joan Velja
2 days
@willccbb @aryaman2020 I think a foundational piece of research most forget to mention is Belkin et al: Which came before all the OAI work on DD/Grokking
0
0
1
@Joanvelja
Joan Velja
2 days
From
0
0
0
@Joanvelja
Joan Velja
2 days
RT @mpshanahan: I agree with @AmandaAskell. It is a fallacy. It can be appropriate to say both a) that LLMs are next-token predictors, at a…
0
36
0
@Joanvelja
Joan Velja
3 days
1
0
0