1a3orn
@1a3orn
Followers
2K
Following
13K
Media
83
Statuses
1K
Many people have been saying things like this, but this is quite false. 500m in damage isn't the end of the world -- it's a Tuesday in the global economy. Let me give examples.
If your model causes mass casualties or >$500 million in damages, something has clearly gone very wrong. Such a scenario is not a normal part of innovation.
16
39
533
It's disquieting that we're going to have AIs as smart as humans, that can sound like humans, and there are 0 good theories to help determine if they're actually conscious. This bit from @jd_pressman seems quite accurate and quite grim.
37
50
394
I would guess SORA was trained at least partially with NeRF data at some point. Based almost entirely off the way that the trees look in this video, which screams NeRF artifacts to me.
Introducing Sora, our text-to-video model. Sora can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions. Prompt: “Beautiful, snowy
12
15
381
From ~18 months ago until quite recently, the standard EA model was that the US was so far ahead of China in AI that even if our regulations slowed us down, they still just wouldn't be able to catch up.
But now onto our case:.1: First and most importantly—Chinese large language models (the type of AI we focus on) just aren't that competitive with the cutting edge of the field. Unless you go by parameter counts—which you shouldn't—it's hard to be impressed by Chinese releases.
15
11
229
@AnjneyMidha @midjourney @discord It's weird we've figured out how to educate machines right as our social technology for educating humans seems to have largely imploded.
3
19
221
My prediction is that the first model to get > 80% on this will remain easy to steer by humans, despite its superhumanity. "Instrumental convergence" towards hard-to-dig-out power seeking will not have happened.
1/10 Today we're launching FrontierMath, a benchmark for evaluating advanced mathematical reasoning in AI. We collaborated with 60+ leading mathematicians to create hundreds of original, exceptionally challenging math problems, of which current AI systems solve less than 2%.
8
10
147
@krishnanrohit computer Chess introduced the youngest grandmasters ever through Computer-human distillation.
2
3
69
@doomslide this has made me yet gloomier about the prospects of avoiding a future war between the US and China. a big cause of war is mis-estimating the strength of your opponent, and apparently a big chunk of of the US are actually mentally stuck in 1999.
3
1
51
@TolgaBilge_ @simonw Suppose I gave you, a human, a task, though words. Does:"Do this task. Nothing else matters but accomplishing this task at any cost.". Mean the same thing as "Do this task, please don't do other random shit?" to you?.
3
0
49
On Lex, Dario Amodei says they limit knowledge of secret training sauce to a few people, to secure an advantage. DeepSeek has an optimized pipeline with their best ideas, accessible by all, to speed experimentation. And just publishes their stuff. Who would move faster?.
[Long Tweet Ahead] I just have to say, I’m genuinely impressed by DeepSeek. 💡.It’s no wonder their reports are so elegant and fluffless. Here’s what I noticed about their culture, a space where real innovation thrives, during my time there ↓. — — — — —.🌟 1. Be nice and careful.
0
1
46
I am really happy that Prime Intellect exists. They're looking at *what actually needs to be done* to prevent AI from being completely locked down by a handful of company.
A few ideas of what we’d be excited to fund.• Novel decentralized training approaches, building on the amazing work of DiLoCo (@Ar_Douillard et al.), SWARM Parallelism, HiveMind (@m_ryabinin @Tim_Dettmers et al) and other approaches. • Decentralized Llama-3: MoE Sparse.
2
6
39
A lot of cope about China and LLMs is literally the same:. "They're just imitating, they don't have the ineffable [Western / biological] spark that gives you originality.".
With the unveiling of China’s new 400 km/h (250 mph) bullet trains, it’s a good time to reshare my all-time favorite slide. It shows how China used partnerships with foreign firms to acquire tech and know-how. Then building on this, China developed its own high-speed trains.
2
2
41
My preliminary answer is "nope, MoE / dense transformers learn exactly the same thing" at least for my extremely crude test setup. 1/N.
Has there been any research on whether dense vs. MoE transformers trained to *equivalent loss* tend to learn different things?. Apropos of Llama 3 being dense, which was quite a surprise to me.
1
5
34
Research from @RANDCorporation and @open_phil that finds that LLMs don't make planning bioterror easier. Good for them, and a positive update on Open Phil from me.
Cool research from our grantees at @RANDCorporation finding that current LLMs don't outperform google at planning bioweapons attacks
3
1
33
In fact, this is what it's like to be hunted by something 20x larger than you, with 3x-5x more friends, in locally bad territory. Replace the seal with a much smarter chimp; the chimp still fuckin dies if the orcas want it to.
2
2
33
@KelseyTuoc Kelsey I don't think you're engaging with the actual worry. If we passed a law saying the publisher were liable for extreme damages resulting from book publication, that were made significantly easier from the book, you'd see a massive decrease in even non-dangerous books pub'd.
3
1
31
All this seem correct. A central reason for the situation is that "alignment" is such a positive term that "faking alignment" has apocalyptic negative connotation, but the terms are without any clear or fixed meaning. Man, imagine how bad discourse like that could get.
I think it is simultaneously the case that:. 1) This is a no-win scenario for the LLM. 2) What the LLM does in this no-win scenario is useful information about LLMs. 3) The framing for this experiment as "alignment faking" is very unfortunate. 4) Framing it properly is awkward.
0
1
29
@teortaxesTex @janleike Exactly. You can gin up a negative headline no matter what. - Tries to preserve value "Alignment faking".- Doesn't try to preserve value? "Reflexive Value Instability" - oh no they'll just abandon their values after a few iterations!.
3
2
28
@ajeya_cotra I think it's pretty much false to say people worry entirely about scaled up future systems, because they literally have tried to ban open weights for ones that exist right now. I discuss at reasonable length:
1
1
26
It's maybe a little hard to see how this *wouldn't* scale up to a fully remote software engineer at this point, albeit with (a lot) more multi-turn RL environment work. There's no czar of AGI to tell you "Yes, this is AGI.".
You can tell we're close because discussion of "AGI timelines" now feels archaic and beside the point. It's like discussing "pandemic timelines" in March 2020. Main thing now is how fast it diffuses through the economy.
1
1
25
@DanHendrycks @dwarkesh_sp Why not the hypothesis that people can independently come up with simple ideas in different places, as has been ubiquitous throughout the history of science?. Lab leaks aren't always the case.
3
0
24
@vlad3ciobanu @RichardSocher @vlad3ciobanu This isn't true. They've tried to ban models trained with many ties less compute than GPT-4.
2
0
22
If I saw this in a movie just 2-3 years ago, I'd have guessed it had to be CG / sped up. Just great.
Cooling limited on quality, SuperSlicer limited on thruput. Probably gonna let the eventual community around these machines do the bulk of the modification necessary to get to the current speedboat leaderboard but this is a good example of what’s possible with stock config
1
2
22
I do want to praise @EpochAIResearch for writing papers / creating databases that present empirical information on AI *without* also including op-eds on their preferred AI governance policies within the papers. Very few orgs seem willing to do this kind of thing, at this point.
.@EpochAIResearch : come for the pretty plots -- stay for the hardcore commitment to truth seeking and science.
0
2
21
@slatestarcodex @krishnanrohit Last year CAIS wrote to the NTIA, and proposed criminalizing open source models trained beyond 10^23, which was *at the time* aprox a little beyond what had been open sourced.
1
3
22
"Shoggoth" is a useful propaganda term imo because it's a nice wrapper for the motte of "hard to control, hostile" with the bailey of "hard to understand". But these are just different things! The brain of a Border Collie is ~just as hard to understand as an LLM.
@jd_pressman I'd note that the denial of shoggoths' malign connotations, saying it's just about shapeshifting mimicry, is pretty damn disingenuous. They're Lovecraftian monsters who have exterminated their creator species, but every aspect of them is nightmarish. They even stink.
3
4
20