Bram Wallace Profile
Bram Wallace

@bram_wallace

Followers
378
Following
87
Statuses
63

Joined May 2022
Don't wanna be here? Send us removal request.
@bram_wallace
Bram Wallace
2 months
Stoked to finally be able to say what I've been working on this year! There's so much exciting post-training research ahead of us, we're just getting started over here
@billpeeb
Bill Peebles
2 months
Today’s deployment is only possible with our new model, Sora Turbo. Huge shoutout to @bram_wallace and @JureZbontar who spearheaded Sora model acceleration research!
1
5
63
@bram_wallace
Bram Wallace
2 months
RT @sama: we are ready for you! (and check out how good the featured feed is getting)
0
369
0
@bram_wallace
Bram Wallace
2 months
RT @rohanjamin: signups are fully open
0
40
0
@bram_wallace
Bram Wallace
2 months
Welcome to Sora y'all! #sora #openai
@OpenAI
OpenAI
2 months
Our holiday gift to you: Sora is here.
0
2
15
@bram_wallace
Bram Wallace
2 months
RT @sama: aditya (@model_mechanic) is a legend and visionary in the field, and runs a very special team.
0
236
0
@bram_wallace
Bram Wallace
2 months
RT @willdepue: sora is launching today to all chatgpt pro and plus users! it's been a big effort to make this possible + i think the pro…
0
79
0
@bram_wallace
Bram Wallace
2 months
RT @billpeeb: Today's the day! Included with ChatGPT Plus/Pro!
0
109
0
@bram_wallace
Bram Wallace
11 months
RT @guskamp: here's a clip of video set to my new song "sugarize" - every frame created from clips i made using Sora, by @OpenAI https://t…
0
14
0
@bram_wallace
Bram Wallace
11 months
RT @paultrillo: Made with Sora. The Golden Record - from raw earth material to a time capsule of human life on Earth. Using 11 different ge…
0
243
0
@bram_wallace
Bram Wallace
11 months
RT @shykids: 'air head' is one of the first short films made using #Sora by @OpenAI. the response so far has left us floating.🎈 https://…
0
601
0
@bram_wallace
Bram Wallace
1 year
Great work! Love the integration of dataset + self learning and the results look sweet. I think this style of method has tons of potential
@_akhaliq
AK
1 year
Self-Play Fine-Tuning of Diffusion Models for Text-to-Image Generation paper page: Fine-tuning Diffusion Models remains an underexplored frontier in generative artificial intelligence (GenAI), especially when compared with the remarkable progress made in fine-tuning Large Language Models (LLMs). While cutting-edge diffusion models such as Stable Diffusion (SD) and SDXL rely on supervised fine-tuning, their performance inevitably plateaus after seeing a certain volume of data. Recently, reinforcement learning (RL) has been employed to fine-tune diffusion models with human preference data, but it requires at least two images ("winner" and "loser" images) for each text prompt. In this paper, we introduce an innovative technique called self-play fine-tuning for diffusion models (SPIN-Diffusion), where the diffusion model engages in competition with its earlier versions, facilitating an iterative self-improvement process. Our approach offers an alternative to conventional supervised fine-tuning and RL strategies, significantly improving both model performance and alignment. Our experiments on the Pick-a-Pic dataset reveal that SPIN-Diffusion outperforms the existing supervised fine-tuning method in aspects of human preference alignment and visual appeal right from its first iteration. By the second iteration, it exceeds the performance of RLHF-based methods across all metrics, achieving these results with less data.
Tweet media one
0
0
6
@bram_wallace
Bram Wallace
1 year
1
0
1
@bram_wallace
Bram Wallace
1 year
@EMostaque Love it! Excited to see how DPO alignment can help!
0
0
0
@bram_wallace
Bram Wallace
1 year
@Suhail Finetune, nothing here is commercialized and the turbo experiment was just a proof-of-concept run to test how this works on distilled models
0
0
1
@bram_wallace
Bram Wallace
1 year
Our research codebase for Diffusion-DPO is now public! This is the exact (cleaned up) code we used for the primary results in our paper. It also has an SDXL-Turbo script which works well (see . Enjoy!
@bram_wallace
Bram Wallace
1 year
Turboing into 2024! Why let SDXL-Base have all the fun? DPO-tuning 4-step SDXL-Turbo turns out to work pretty well. Original SDXL-Turbo on left, Turb(DP)o on right. The video shown is sped up 2x to stop typing being the bottleneck
0
17
79
@bram_wallace
Bram Wallace
1 year
RT @SFResearch: Our blog for Diffusion-DPO is now live!🚀 In this project we brought the benefits of Reinforcement Learning from Human Feed…
0
8
0
@bram_wallace
Bram Wallace
1 year
Also, this training will soon be integrated into diffusers as well (
0
0
2