juliusadml Profile Banner
Julius Adebayo Profile
Julius Adebayo

@juliusadml

Followers
2K
Following
2K
Statuses
3K

Building @guidelabsai - Engineering interpretable models and agents that are easy to audit. PhD in ML @MITEECS, Ex @Meta, Google Brain, & Prescient Design.

San Francisco, CA
Joined July 2011
Don't wanna be here? Send us removal request.
@juliusadml
Julius Adebayo
9 hours
RT @jon_barron: A meaty interview with Jascha about the history of diffusion, worth watching.
0
8
0
@juliusadml
Julius Adebayo
2 days
RT @nabeelqu: Trying to make any kind of "agent" work in a real enterprise is extremely blackpilling. Basically turns you into Gary Marcus.
0
19
0
@juliusadml
Julius Adebayo
3 days
RT @rdolmedo_: Self-reflection is not unique to “reasoning models” or to newer models. Here are some self-reflections produced by Llama 2…
0
3
0
@juliusadml
Julius Adebayo
7 days
RT @albertwenger: @elder_plinius Yet more evidence (as if that was needed) that if we want to have any hope of inner, non-resented alignmen…
0
2
0
@juliusadml
Julius Adebayo
8 days
RT @cloneofsimo: What students expect from ML job: analysis of sharpness effecting generalization bound analysis of NTK parameterization o…
0
24
0
@juliusadml
Julius Adebayo
12 days
RT @charliermarsh: We’re building a new static type checker for Python, from scratch, in Rust. From a technical perspective, it’s probably…
0
215
0
@juliusadml
Julius Adebayo
15 days
RT @jxmnop: most important thing we learned from R1? that there’s no secret revolutionary technique that’s only known by openAI. no magic…
0
109
0
@juliusadml
Julius Adebayo
15 days
RT @percyliang: While we celebrate @deepseek_ai 's release of open-weight models that we can all play with at home, just a friendly reminde…
0
467
0
@juliusadml
Julius Adebayo
28 days
RT @marimo_io: Sharing notebooks with data files is far harder than it should be. That's why we're announcing a new way to share Python no…
0
10
0
@juliusadml
Julius Adebayo
28 days
Excited to continue to @guidelabsai journey with @asalam_91. Interpretable models go brrr!
@asalam_91
Aya Abdelsalam Ismail
29 days
Exciting news! I've left Prescient to co-found @guidelabsai an interpretability startup with @juliusadml. We're building interpretable foundational models to address key AI challenges. While I loved my time at Prescient, I'm thrilled to build something I'm very passionate about! We're hiring! If you're an ML interpretability researcher, ML engineer, frontend engineer, or are generally curious about what we're doing, please reach out!
0
0
7
@juliusadml
Julius Adebayo
1 month
Writing for the AIs...I guess AI SEO is going to be a nightmare.
@AndrewCurran_
Andrew Curran
1 month
Tyler Cowen and Gwern have both seen it. 'But who reviews it? Is TLS going to pick it up? It doesn't matter anymore.'
Tweet media one
0
0
1
@juliusadml
Julius Adebayo
1 month
The empire strikes back. Ironic that the jax code uses a pytorch dataloader tho.
@cgarciae88
Cristian Garcia
1 month
Link:
0
0
1
@juliusadml
Julius Adebayo
1 month
RT @matthewjmandel: The Deep Tech Opportunity Far from permanently redefining venture capital, the software moment of the last thirty year…
0
16
0
@juliusadml
Julius Adebayo
1 month
RT @rdolmedo_: My favorite figure: Pythia performs at random chance on MMLU and GSM8K irrespective of scale. In contrast, Llama and Qwen s…
0
2
0
@juliusadml
Julius Adebayo
1 month
Really cool paper questioning all the 'incredible' progress we've seen recently: "after fine-tuning all models on the same amount of task data, performance per pre-training compute equalizes and newer models are no better than earlier models."
@rdolmedo_
Ricardo Dominguez-Olmedo
2 months
Models released after November 2023 strongly outperform earlier ones on MMLU and GSM8K. However, after fine-tuning all models on the same amount of task data, performance per pre-training compute equalizes and newer models are no better than earlier models.
Tweet media one
0
4
24
@juliusadml
Julius Adebayo
1 month
New nice report on LLM training for those that don’t have 10k h100s just chilling in some dungeon.
@soldni
Luca Soldaini 🎀
1 month
OLMo 2 tech report is out We get in the weeds with this one, with 50+ pages on 4 crucial components of LLM development pipeline:
Tweet media one
0
0
3
@juliusadml
Julius Adebayo
1 month
RT @jeremyphoward: SW eng manager: No real work gets done in Jupyter notebooks. Alex Radford: I invented GPT and CLIP in Jupyter notebooks.
0
202
0
@juliusadml
Julius Adebayo
2 months
People seem outraged about this. It is simple: LLMs have bulldozed the test vs train division we used to hold sacred in machine learning. A quote from the original test-time training paper (: "we hope this paper can encourage researchers to abandon the self-imposed constraint of a fixed decision boundary for testing, or even the artificial division between training and testing altogether." See this important talk for more discussion: GPT-3 ushered in a brave new world 😉.
@mikeknoop
Mike Knoop
2 months
Raising visibility on this note we added to address ARC "tuned" confusion: > OpenAI shared they trained the o3 we tested on 75% of the Public Training set. This is the explicit purpose of the training set. It is designed to expose a system to the core knowledge priors needed to beat the much harder eval set. The idea is each training task shows you an isolated single prior. And the eval set requires you to recombine and abstract from those priors on the fly. Broadly, the eval tasks require utilizing 3-5 priors. The eval sets are extremely resistant to just "memorizing" the training set. This is why o3 is impressive.
0
0
3
@juliusadml
Julius Adebayo
2 months
RT @liambolling: this is my feed every week
Tweet media one
0
16
0
@juliusadml
Julius Adebayo
2 months
We might all be going extinct soon 😁, but it is not too late to learn a little learning theory.
@BachFrancis
Francis Bach
2 months
My book is (at last) out, just in time for Christmas! A blog post to celebrate and present it:
Tweet media one
0
0
3