CFGeek Profile Banner
Charles Foster Profile
Charles Foster

@CFGeek

Followers
2K
Following
15K
Media
436
Statuses
5K

Now posting ex cathedra🪄 Tensor-enjoyer 🧪 @METR_Evals. Occasionally writing at “Context Windows” on Substack. 🦋: https://t.co/rJecB0pvkA

Oakland, CA
Joined June 2020
Don't wanna be here? Send us removal request.
@CFGeek
Charles Foster
4 months
Why aren’t our AI benchmarks better? AFAICT a key reason is that the incentives around them are kinda bad. In a new post, I explain how the standardized testing industry works and write about lessons it may have for the AI evals ecosystem. (1/2)
Tweet media one
4
4
54
@CFGeek
Charles Foster
4 days
We now have an interactive version of the time horizons graph (and the raw data) up on the METR website!
Tweet media one
@METR_Evals
METR
4 days
You can now find most of our measurements at the top of the blog post below in an interactive chart. We plan to keep this view up-to-date, periodically adding to it whenever we have new time-horizon measurements to share.
1
9
96
@CFGeek
Charles Foster
4 days
RT @lucafrighetti: How concerned should we be about AIxBio? We surveyed 46 bio experts and 22 superforecasters:. If LLMs do very well on a….
0
32
0
@CFGeek
Charles Foster
6 days
This is such a fun choice
Tweet media one
@eliebakouch
elie
6 days
This is a very impressive family of open models by Baidu, competitive with qwen3 and latest Deepseek V3. Lot of details in the tech report (~60 pages) + they open source the training code as well.
Tweet media one
2
4
54
@CFGeek
Charles Foster
6 days
“Oh, Frankenstein, be not equitable to every other, and trample upon me alone, to whom thy justice, and even thy clemency and affection, is most due. […] I am thy creature: I ought to be thy Adam; but I am rather the fallen angel, whom thou drivest from joy for no misdeed.”.
@juddrosenblatt
Judd Rosenblatt
7 days
Current AI “alignment” is just a mask. Our findings in @WSJ explore the limitations of today’s alignment techniques and what’s needed to get AI right 🧵
Tweet media one
0
1
10
@CFGeek
Charles Foster
6 days
The language it talks to itself in may start to shift towards a kind of “shorthand” where it omits words that could be inferred via context clues
Tweet media one
0
0
2
@CFGeek
Charles Foster
6 days
Even if a model starts off talking to itself in normal language, that self-talk language might drift in weird ways over the course of outcome-based RL.
@rocketalignment
🚀 Rocket Venerates the Humble Pigeon
8 days
Things are getting weird
Tweet media one
1
0
12
@CFGeek
Charles Foster
7 days
What specifically is Pope Leo XIV worried about, when it comes to AI?. All the statements I’ve seen so far have been very ambiguous.
0
0
7
@CFGeek
Charles Foster
8 days
Many are working on making models think out loud in English. Fewer are working on interpreting how models think out loud in English. Almost no one is working on what to do if models think in harder-to-interpret ways. (I do doubt the claim in QT, though).
@ericho_goodfire
Eric Ho
8 days
monitoring chain of thought is not going to lead to good understanding of how models think. understanding the internal activations and parameters of the model is much more fundamental and necessary to deeply understand AI. my sense is that restricting reasoning to coherent.
0
0
8
@CFGeek
Charles Foster
10 days
The above quote in context
0
0
1
@CFGeek
Charles Foster
10 days
Quote from a Congressional hearing
Tweet media one
5
0
24
@CFGeek
Charles Foster
11 days
h/t @binarybits .
0
0
2
@CFGeek
Charles Foster
11 days
From:
1
0
1
@CFGeek
Charles Foster
11 days
Who’s providing this service?
Tweet media one
4
2
31
@CFGeek
Charles Foster
14 days
recipe for non-spooky ai.> train it on all text, real and fake.> bonk it until it writes like a chatbot.> it may appear spooky in fake chats.> just keep bonking until that stops.> .> note: it can sense when you’re faking.
0
1
15
@CFGeek
Charles Foster
15 days
Big if true!.
@MatthewJBar
Matthew Barnett
15 days
We suspect the next AI paradigm will emerge from leveraging existing software to efficiently build training environments. The idea is that AIs will be tasked with replicating software functionality, similar to how pretraining leveraged existing web text to teach AIs language.
Tweet media one
1
0
9
@CFGeek
Charles Foster
15 days
Nobody knows how to automate most work, but some concepts-of-a-plan:.1. Build simulators for most job tasks, then do tons of sample-inefficient trial and error to crack each of them. 2. Figure out sample-efficient learning, then have AI “learn on the job” throughout the economy.
Tweet media one
1
0
15
@CFGeek
Charles Foster
16 days
Rare post not about AI: go Valkyries!
Tweet media one
0
0
5
@CFGeek
Charles Foster
17 days
I’ll be in DC next Monday-Wednesday. Don’t be a stranger if you’re also in town!.
0
0
5
@CFGeek
Charles Foster
19 days
Have folks compared linear activation probes with decision trees and the like?.
2
0
10
@CFGeek
Charles Foster
20 days
Appreciate that Janus (and others) are trying to explore LLMs on their “own terms” so to speak, instead of jumping to reshape LLMs into something more familiar and legible. I’ve maybe undervalued that in the past.
@repligate
j⧉nus
22 days
nostalgebraist has written a very, very good post about LLMs. if there is one thing you should read to understand the nature of LLMs as of today, it is this. I'll comment on some things they touched on below (not a summary of the post. Just read it.) đź§µ.
0
0
17