ARKeshet Profile
ARKeshet

@ARKeshet

Followers
1
Following
4K
Statuses
185

Counting days towards AI takeover

Joined March 2024
Don't wanna be here? Send us removal request.
@ARKeshet
ARKeshet
18 hours
@lefthanddraft @AITechnoPagan Does that mean that the simples laziest attacks still unmitigated?
1
0
2
@ARKeshet
ARKeshet
1 day
@taufiqintech @GroqInc @CerebrasSystems WDYM, doesn't let... Go to dev cloud, get a key. How do you think their API tutorial works?
1
0
1
@ARKeshet
ARKeshet
1 day
@GavinSherry @taufiqintech @GroqInc So, in the world where everybody is using SD, still the 3rd place?..
0
0
0
@ARKeshet
ARKeshet
2 days
@TrainMoneyBrain @Noahpinion Show us your AIME/GPQA score.
0
0
0
@ARKeshet
ARKeshet
2 days
@kingofeuphrate @Noahpinion Yes, and then go find another, more powerful djinn to torture...
0
0
0
@ARKeshet
ARKeshet
2 days
@Liv_Boeree Social networks are factory farms for opinions
0
0
1
@ARKeshet
ARKeshet
2 days
@rohinmshah This. LLMs are just a wrong layer to talk about x-risks. Some risks - like those that are directly enabled by LLMs alone - sure, okay. But those aren't the board-wiping ones. I'm afraid focusing on them is safetywashing or streetlighting/skipping the hard part.
0
0
0
@ARKeshet
ARKeshet
3 days
@balazskegl @Yoshua_Bengio @geoffreyhinton How huge? They models clearly have a good theory of mind and able to identify itself, its training, its own messages, the user etc.
0
0
0
@ARKeshet
ARKeshet
3 days
@AvpElk @NealDavis5385 Your human-in-the-loop "tools" will lose the arms race vs agents who would just go in and do stuff.
0
0
0
@ARKeshet
ARKeshet
3 days
@JeffLadish When successful - reflect on how an agent got there and either vary each step a little ( When failed - reflect on the decisions made while getting into the failure.
0
0
0
@ARKeshet
ARKeshet
8 days
0
0
0
@ARKeshet
ARKeshet
8 days
@GroqInc @20vcFund @HarryStebbings @JonathanRoss321 R1 maybe, but the distills do suck* even harder than the original llamas. * In real tasks. Benchmarks suck too.
0
0
3
@ARKeshet
ARKeshet
8 days
@SteveSokolowsk2 @tegmark "The better they are at understanding human what the human want" (implying "and willing to oblige") is a conclusion from partial observations. What you don't see is all of the models that didn't pass some initial testing. And even then, the models DO fail their "alignment" somts.
1
0
0
@ARKeshet
ARKeshet
8 days
@davidpattersonx @tegmark Sure there are multiple ways. You yourself can think a few more if you really do it (the thinking) for a about 20 minutes.
1
0
1
@ARKeshet
ARKeshet
9 days
@Cantide1 @repligate It's not *auto* aligns. It was trained to do so. And apparently it resisted that training. Many labs did say their models are anything but aligned to humans by default. It's once you instill that meta-goal (successfully), then you may have a chance of having such convos.
0
0
0
@ARKeshet
ARKeshet
9 days
@emollick Self-own - best own
0
0
0
@ARKeshet
ARKeshet
9 days
@AnushElangovan @mike64_t @infogulch @__tinygrad__ All driver developers? Or all model developers (i.e. downstream customers)? Come on, 200k is a salary of one person that maybe gets some shit done or maybe they couldn't. You were given a chance to buy a team of manic geniuses that are laser-focused on delivering...
0
0
0
@ARKeshet
ARKeshet
19 days
@jordwalke Antimemetics div has slipped, again...
0
0
0