RickLamers Profile Banner
Rick Lamers Profile
Rick Lamers

@RickLamers

Followers
5K
Following
3K
Statuses
2K

πŸ‘¨β€πŸ’» AI Research & Engineering @GroqInc. Occasional angel investor. I publish technical resources about LLMs every week. Opinions are my own.

Join 7,237+ readers β†’
Joined July 2009
Don't wanna be here? Send us removal request.
@RickLamers
Rick Lamers
3 months
"if you can compress data well, you understand its patterns" Low loss on training error + highly compressible hypothesis = strong generalization Large models tend to find compressible hypotheses in their weight space because of the optimization process (gradient descent) and because the training distribution can actually be compressed well. (Language is highly regular, i.e. a low length hypothesis/model exists.)
Tweet media one
@andrewgwils
Andrew Gordon Wilson
3 months
My ICLR talk β€œHow Do We Build a General Intelligence?” is now online!
1
0
21
@RickLamers
Rick Lamers
1 hour
RT @omarkilani: YOLO Monday I guess... 🀫
Tweet media one
0
1
0
@RickLamers
Rick Lamers
1 day
@ozenhati @fire @GroqInc Actively looking at various options 🀞
1
0
2
@RickLamers
Rick Lamers
2 days
RT @GavinSherry: Huge news here in Saudi Arabia β€” amazing partnership between @GroqInc and Saudi Arabia πŸ‡ΈπŸ‡¦
Tweet media one
0
28
0
@RickLamers
Rick Lamers
3 days
@TrelisResearch Yeah if RL on LLM base is learning how to reason for LLMs I hope their reasoning gets better over time too not just that it learns how to reason for longer (albeit coherently/productively – which to be fair also isn't trivial).
0
0
2
@RickLamers
Rick Lamers
3 days
RT @mbalunovic: We finally have an answer to the debate over whether LLMs generalize to new math problems or they merely memorized the answ…
0
162
0
@RickLamers
Rick Lamers
3 days
@noamgat Jevons paradox, more value per token increases the addressable market πŸ™Œ
0
0
1
@RickLamers
Rick Lamers
3 days
Read the entire thing, trust me, you won't regret it.
0
1
5
@RickLamers
Rick Lamers
3 days
Ah, did find: >Software Engineering Tasks: Due to the long evaluation times, which impact the efficiency of the RL process, large-scale RL has not been applied extensively in software engineering tasks. As a result, DeepSeek-R1 has not demonstrated a huge improvement over DeepSeek-V3 on software engineering benchmarks. Future versions will address this by implementing rejection sampling on software engineering data or incorporating asynchronous evaluations during the RL process to improve efficiency.
0
0
0
@RickLamers
Rick Lamers
4 days
And the crucial bit of what is "emerging on its own" in the long CoT RL paradigm: >backtracking, checking, error-correction and other branching-like behavior
0
0
1
@RickLamers
Rick Lamers
4 days
@0xs____ Will pass on this feedback re: rate limit handling. Ty
0
0
0
@RickLamers
Rick Lamers
4 days
@TrelisResearch @anaisbetts >things probably collapse into being one agent I feel like encapsulation and information/context overload contradict this. How to organize I don't know, but "putting everything into one thing" feels like it won't scale.
0
0
1
@RickLamers
Rick Lamers
4 days
Tweet media one
1
0
4
@RickLamers
Rick Lamers
4 days
I like to think this is inspired by the human brain that is pretty sample efficient also. Evolution = pre-training. Eduction = supervised learning. Trial and error self teaching = RL.
@awnihannun
Awni Hannun
4 days
Turns out you don't need that much data or compute to get a reasoning model from a high quality base model. @ylecun was right
Tweet media one
0
0
1
@RickLamers
Rick Lamers
4 days
77% woah higher than usual πŸ”₯
@paulgauthier
Paul Gauthier
4 days
Aider v0.74.0 is out: - Improved o3-mini, R1 and V3 support, especially via third-party API providers. - Better Ollama support, by dynamically adjusting the context window. - Bug fixes and QOL improvements. Aider wrote 77% of the code in this release.
0
0
1
@RickLamers
Rick Lamers
4 days
RT @hila_chefer: VideoJAM is our new framework for improved motion generation from @AIatMeta We show that video generators struggle with m…
0
192
0
@RickLamers
Rick Lamers
4 days
RT @ggerganov: @Olney1Ben @exolabs Or just use llama-rpc:
Tweet media one
0
2
0