![Jan Brauner Profile](https://pbs.twimg.com/profile_images/1336728824709591040/jACOeaZe_x96.jpg)
Jan Brauner
@JanMBrauner
Followers
1K
Following
4K
Statuses
466
Technical staff member at EU AI Office, Previously: RAND, ML PhD at Oxford (@OATML_Oxford), and, once upon a time, medical doctor.
Joined December 2020
@AdriGarriga I'm asking why it took so long to get from LLM (say gpt-3) to the first reasoning model (say o1), given that the benchmark gains are so big and the technical challenge seems smaller than rlhf.
0
0
1
Another amazing contribution from Owain et al!
New paper: We train LLMs on a particular behavior, e.g. always choosing risky options in economic decisions. They can *describe* their new behavior, despite no explicit mentions in the training data. So LLMs have a form of intuitive self-awareness 🧵
0
0
6
@jankulveit I just click on 1-5 posts that I don't like and say that I don't like it, and that usually fixes it as well.
0
0
2
@Altimor On the other hand, the human brain shows that huge efficiency gains (compared to current NNs) are possible. A legion of human-expert-level AIs working on AI R&D may well make a lot of progress?
1
0
1
@tamaybes ML conference review is incredibly noisy. It's common to get terrible reviews one time and great ones the next time, even without changing the paper. I'd encourage you to submit again to another conference (if you think the paper is worth it).
0
0
5
Pure fire coming from Owain et al, as usual!
New paper: We measure *situational awareness* in LLMs, i.e. a) Do LLMs know they are LLMs and act as such? b) Are LLMs aware when they’re deployed publicly vs. tested in-house? If so, this undermines the validity of the tests! We evaluate 19 LLMs on 16 new tasks 🧵
0
0
13