Aleksander Holynski Profile
Aleksander Holynski

@holynski_

Followers
1,683
Following
291
Media
35
Statuses
182
Explore trending content on Musk Viewer
Pinned Tweet
@holynski_
Aleksander Holynski
4 months
Videos are cool and all...but everything's more fun when it's interactive. Check out our new project, ✨CAT3D✨, that turns anything (text, image, & more) into interactive 3D scenes! Don't miss the demo!!
@RuiqiGao
Ruiqi Gao
4 months
🌟 Create anything in 3D! 🌟 Introducing CAT3D: a new method that generates high-fidelity 3D scenes from any number of real or generated images in one minute, powered by multi-view diffusion models. w/ lovely coauthors @holynski_ , @poolio and an amazing team!
24
64
407
8
35
219
@holynski_
Aleksander Holynski
1 year
Check out our new paper that turns a (single image) => (interactive dynamic scene)! I’ve had so much fun playing around with this demo. Try it out yourself on the website:
@zhengqi_li
Zhengqi Li
1 year
Excited to share our work on Generative Image Dynamics! We learn a generative image-space prior for scene dynamics, which can turn a still photo into a seamless looping video or let you interact with objects in the picture. Check out the interactive demo:
19
154
854
26
316
2K
@holynski_
Aleksander Holynski
4 years
Excited to show off our new project on single-image cinemagraphs. Our method automatically turns a _single image_ into a seamlessly looping video! Website: Video: w/ Brian Curless, Steve Seitz, Rick Szeliski More in thread! [1/5]
30
382
2K
@holynski_
Aleksander Holynski
9 months
Excited to share ReconFusion! 3D reconstruction of real-world scenes from only a few photos, powered by diffusion priors: w/ amazing team @ChrisWu6080 @BenMildenhall @philipphenzler @KeunhongP @RuiqiGao @watson_nn @_pratul_ @dorverbin @jon_barron @poolio
9
60
355
@holynski_
Aleksander Holynski
4 years
And also some pretty cool failure cases -- if a class of objects aren't seen during training, but share similar textural properties to fluids... [6/5]
15
40
294
@holynski_
Aleksander Holynski
11 months
We just posted a report on the state of the art in diffusion models for visual computing: If you're new to diffusion models, or maybe just want a recap of everything that's been going on lately---this is a great place to start.
Tweet media one
4
45
295
@holynski_
Aleksander Holynski
1 year
Excited to share self-guidance, a new method for controllable image generation that guides sampling using only the attention and activations of a pretrained diffusion model: Work led by Dave Epstein w/ @ajabri , @poolio , Alyosha Efros More in thread🧵
Tweet media one
9
58
252
@holynski_
Aleksander Holynski
1 year
Happy to finally be able to share our #CVPR2022 paper, InstructPix2Pix! We taught a diffusion model how to follow image editing instructions — just say how you want to edit an image, and it’ll do it! (w/ Tim Brooks & Alyosha Efros) More on Tim’s site: 🧵
Tweet media one
Tweet media two
Tweet media three
Tweet media four
4
28
194
@holynski_
Aleksander Holynski
11 months
. @QianqianWang5 's 🎉Best Student Paper🎉 is being presented at #ICCV2023 tomorrow (Friday)! ▶️"Tracking Everything Everywhere All At Once"◀️ w/ Yen-Yu Chang, @ruojin8 @zhengqi_li @BharathHarihar3 @Jimantha Friday Afternoon Oral & Poster! Come say hi!
1
21
170
@holynski_
Aleksander Holynski
9 months
We posted an updated version of Generative Image Dynamics to arXiv---the biggest change is to better contextualize our method with respect to prior work in image space motion analysis, especially the great work of @AbeDavis
@holynski_
Aleksander Holynski
1 year
Check out our new paper that turns a (single image) => (interactive dynamic scene)! I’ve had so much fun playing around with this demo. Try it out yourself on the website:
26
316
2K
4
21
149
@holynski_
Aleksander Holynski
3 months
I'll be presenting CAT3D tomorrow at CVPR. Come say hi! Monday 2:30pm at AI for 3D Generation (Summit Flex A)
@RuiqiGao
Ruiqi Gao
4 months
🌟 Create anything in 3D! 🌟 Introducing CAT3D: a new method that generates high-fidelity 3D scenes from any number of real or generated images in one minute, powered by multi-view diffusion models. w/ lovely coauthors @holynski_ , @poolio and an amazing team!
24
64
407
3
12
115
@holynski_
Aleksander Holynski
4 months
some more fun CAT3D results ✨ tons more in the gallery:
5
10
92
@holynski_
Aleksander Holynski
4 years
It turns out images contain lots of useful cues about how things should be flowing -- like ripples in water, turbulent streams, motion blur. An image-to-image GAN learns a lot of these subtle cues, and can synthesize pretty complex motion. [3/5] Here's another result:
5
5
80
@holynski_
Aleksander Holynski
9 months
Check out @xiaojuan_wang7 's new project! 🔎Generative Powers of Ten🔍 Use a pre-trained text-to-image model to generate deeeeep zoom videos! (Excuse Twitter's terrible compression, check the webpage instead: )
@xiaojuan_wang7
Xiaojuan Wang
9 months
Excited to share our work Generative Powers of Ten w/ @holynski_ @_pratul_ @BenMildenhall @dorverbin @kemelmi Given a set of prompts describing a scene at varying zoom levels, our method creates a seamless zooming video. Check it out here:
4
16
110
3
6
73
@holynski_
Aleksander Holynski
4 years
We focus on fluids (flowing water, billowing smoke, clouds), i.e., things well approximated by particle motion. So, instead of predicting a sequence of flow fields for a video, we can predict a single Eulerian motion field (a particle velocity field). [2/5]
2
6
70
@holynski_
Aleksander Holynski
3 years
Wow!! I've been a big fan of @twominutepapers for the longest time...it's such an incredible honor to have our paper featured.
3
5
57
@holynski_
Aleksander Holynski
4 years
To generate the video frames, we use a deep warping technique (encode-warp-decode). Since warping a single image usually leads to big holes, we use a novel symmetric splatting approach, which combines features from different points in time to produce more realistic images. [4/5]
2
4
53
@holynski_
Aleksander Holynski
4 months
Thanks for the tweet! Check out our project page:
@_akhaliq
AK
4 months
Google presents CAT3D Create Anything in 3D with Multi-View Diffusion Models Advances in 3D reconstruction have enabled high-quality 3D capture, but require a user to collect hundreds to thousands of images to create a 3D scene. We present CAT3D, a method for creating
9
125
557
2
9
52
@holynski_
Aleksander Holynski
4 years
We've tried our method on a large collection of images, and found it to be surprisingly robust on a pretty wide variety of scenes! [5/5]
1
1
49
@holynski_
Aleksander Holynski
3 months
Come hang out at our posters! 📅 Weds AM • Generative Powers of Ten ( #231 ) • Readout Guidance ( #332 ) • Video Interpolation with Diffusion Models ( #247 ) 📅 Fri • ReconFusion ( #193 ) • Generative Image Dynamics ( #117 ) • NerFiller ( #114 ) • ExtraNeRF ( #82 ) Links ⬇️
3
3
46
@holynski_
Aleksander Holynski
4 months
Super neat! An interactive diffusion-based Photoshop. A great example of how the right interfaces and controls can make a massive difference in the utility of these generative models.
@nneonneo
Robert Xiao
4 months
We are thrilled to announce "Layered Diffusion Brushes": a real-time training-free image editor powered by diffusion models. 🎨✨ This is new work from my PhD student Peyman Gholami @peymo0n . Explore the interactive demo and check out more videos at:
2
16
104
1
7
40
@holynski_
Aleksander Holynski
9 months
🔮Readout Guidance🔮 is a neat way of controlling diffusion models (in pretty complex ways!) See the site () for applications and interactive galleries. Here's one favorite: where we guide the identity in a generated image to match a reference image.
@graceluo_
Grace Luo
9 months
Guidance on top of diffusion models can now be used to drag and manipulate images, create pose-conditioned images, and so much more! Check out Readout Guidance: Work w/ @trevordarrell , @oliver_wang2 , @danbgoldman , @holynski_ . More in thread 🧵.
5
48
240
1
6
41
@holynski_
Aleksander Holynski
8 months
We're hosting a CVPR workshop on AI-assisted art---a big focus is to understand how AI models are currently being used in artistic workflows (to help inspire the next generation of better, more useful AI tools).
@raoanyi
Anyi Rao
8 months
We are excited to share that the AI for Creative Visual Content Generation Editing and Understanding Workshop @cveu_workshop has been accepted to #CVPR2024 @CVPRConf . See you in Seattle to meet art, tech, and creativity! It is the first time we come to the US🇺🇸
Tweet media one
Tweet media two
2
10
65
1
5
34
@holynski_
Aleksander Holynski
3 years
I'll be talking about our paper "Animating Pictures with Eulerian Motion Fields" this evening at Paper Session #5 (10pm-12a ET, 7pm-9pm PT). Come say hi!
@holynski_
Aleksander Holynski
4 years
Excited to show off our new project on single-image cinemagraphs. Our method automatically turns a _single image_ into a seamlessly looping video! Website: Video: w/ Brian Curless, Steve Seitz, Rick Szeliski More in thread! [1/5]
30
382
2K
0
2
33
@holynski_
Aleksander Holynski
3 months
Come hang out at this #CVPR2024 workshop we're organizing! Learn from researchers & artists about new creative applications, open technical challenges, & more. The event is in-person only---no recording, no streaming! Don't miss out! @CVPR
@jon_barron
Jon Barron
3 months
I'm co-organizing a CVPR workshop next Tuesday that is absolutely stacked with talent. If you're interested in anything related to art or generative video (eg Sora, Veo, Pika, Runway), be there.
Tweet media one
8
9
123
0
4
29
@holynski_
Aleksander Holynski
2 months
Congrats to @zhengqi_li , @Jimantha , & Richard!!!
@GoogleAI
Google AI
2 months
Congratulations to @zhengqi_li , Richard Tucker, @Jimantha , and @holynski_ . Their paper “Generative Image Dynamics” received the #CVPR2024 Best Paper Award. Read the paper:
Tweet media one
Tweet media two
8
28
178
0
0
26
@holynski_
Aleksander Holynski
4 years
Darn -- looks like twitter's encoding messed with the looping. Check the website for the full-quality results:
2
1
23
@holynski_
Aleksander Holynski
4 years
Very excited to mess around with this.
@JPKopf
Johannes Kopf
4 years
Our latest work on making Consistent Video Depth more ROBUST. Works great for casual phone videos that are really difficult for previous methods. Another great collaboration with @jastarex and @jbhuang0604 . arXiv: Project:
4
54
279
0
0
21
@holynski_
Aleksander Holynski
1 year
Come say hi tomorrow morning! 10:30-12:30 at poster #183 #CVPR2023
@holynski_
Aleksander Holynski
1 year
Happy to finally be able to share our #CVPR2022 paper, InstructPix2Pix! We taught a diffusion model how to follow image editing instructions — just say how you want to edit an image, and it’ll do it! (w/ Tim Brooks & Alyosha Efros) More on Tim’s site: 🧵
Tweet media one
Tweet media two
Tweet media three
Tweet media four
4
28
194
1
0
18
@holynski_
Aleksander Holynski
9 months
Seeing the world in a potato!
@dorverbin
Dor Verbin
9 months
Introducing Eclipse, a method for recovering lighting and materials even from diffuse objects! The key idea is that standard "NeRF-like" data has all we need: a photographer moving around a scene to capture it causes "accidental" lighting variations. (1/3)
5
34
328
1
0
12
@holynski_
Aleksander Holynski
6 months
check out dave's project! automatically decomposes complex 3D scenes into individual objects (without relying on per-object text descriptions or annotations!) a neat central insight: think of objects as "parts of a scene that can be moved around independently"
@daveepstein
dave
6 months
text-to-3d scenes that are automatically decomposed into the objects they contain, using only an image diffusion model & no other supervision: work w/ @poolio @BenMildenhall Alyosha Efros and @holynski_
3
21
162
1
0
12
@holynski_
Aleksander Holynski
4 years
Come check out our paper at 3DV today! (6a PST oral / 8:30a PST poster) We use vanishing points and planes to get rid of pose drift in SfM. "Reducing Drift in Structure from Motion Using Extended Features" Project page: Video:
Tweet media one
1
8
12
@holynski_
Aleksander Holynski
9 months
For those wondering, yes, we did try it on images from the original Powers of Ten 🙃
0
1
12
@holynski_
Aleksander Holynski
1 year
@jon_barron Or...you can wear it as a bolo tie
Tweet media one
0
0
10
@holynski_
Aleksander Holynski
3 years
@jbhuang0604 I can't get enough of these advice threads. This needs to be a class!! PHD101 "How to be a graphics+vision researcher", with Prof. Huang
0
1
10
@holynski_
Aleksander Holynski
4 years
Wonderfully trippy results!
@akanazawa
Angjoo Kanazawa
4 years
View synthesis is super cool! How can we push it further to generate the world *far* beyond the edges of an image? We present Infinite Nature, a method that combines image synthesis and 3D to generate long videos of natural scenes from a single image.
19
458
2K
0
1
9
@holynski_
Aleksander Holynski
1 year
We trained the model on a massive dataset of generated editing examples, with triplets containing: 1. input image 2. text editing instruction 3. output image How does one generate a dataset like this, you might ask?
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
0
8
@holynski_
Aleksander Holynski
9 months
If you’re interested in this stuff, I’d highly recommend reading Abe’s thesis, which includes a thorough and beautiful theory about the underlying frequency-space motion representation and how it connects to modeling object dynamics.
0
0
8
@holynski_
Aleksander Holynski
1 year
Diffusion models let you create amazing images given the right prompt. But some things are hard to express in text, like where objects should go or exactly how big they should be. How can we get this kind of control?
Tweet media one
1
0
8
@holynski_
Aleksander Holynski
1 year
Self-guidance offers a new way to control the generation process: Without any extra models or training, we can extract properties like object shape, size, and appearance from internal attention maps + activations. We can then guide these properties to edit generated images.
Tweet media one
1
1
8
@holynski_
Aleksander Holynski
1 year
Finally, fine-tune a text-to-image diffusion model to learn this transformation, conditioned on the input image and the instruction! Here's one of our favorites, but you can try it for yourself with the demo on the website:
Tweet media one
1
0
7
@holynski_
Aleksander Holynski
3 years
Robust Consistent Video Depth Estimation @JPKopf , @jastarex , @jbhuang0604 Jointly estimates camera pose & dense depth for challenging video captures of dynamic scenes
Tweet media one
0
1
7
@holynski_
Aleksander Holynski
7 months
Wow!
@_tim_brooks
Tim Brooks
7 months
Sora is our first video generation model - it can create HD videos up to 1 min long. AGI will be able to simulate the physical world, and Sora is a key step in that direction. thrilled to have worked on this with @billpeeb at @openai for the past year
152
160
1K
0
0
7
@holynski_
Aleksander Holynski
4 years
Everything's better in 3D!
@jbhuang0604
Jia-Bin Huang
4 years
I was blown away by the incredible results of animating fluid motion from a single image last week. I thought it would be fun to add a bit of 3D. Here are some results using 3D photo inpainting
3
31
221
0
1
7
@holynski_
Aleksander Holynski
2 months
I'm moderating the second session today---stop by!
@jon_barron
Jon Barron
3 months
I'm co-organizing a CVPR workshop next Tuesday that is absolutely stacked with talent. If you're interested in anything related to art or generative video (eg Sora, Veo, Pika, Runway), be there.
Tweet media one
8
9
123
0
0
7
@holynski_
Aleksander Holynski
4 years
Follow-up idea: a network that makes me look good in pictures
@rmbrualla
Ricardo Martin-Brualla
4 years
We present nerfies! We use selfie videos to create 3d free-viewpoint portrait visualizations of yourself using Deformable NeRFs! More details at and below (1/8)
9
126
751
0
0
7
@holynski_
Aleksander Holynski
1 year
Self-guidance also works on real images, which allows you to "borrow" real objects and stick them in new contexts, sort of like a zero-shot DreamBooth.
Tweet media one
1
0
7
@holynski_
Aleksander Holynski
4 years
No more chopped-off fingers!
@SenguptRoni
Roni Sengupta
4 years
Excited to release v2.0 of our Background Matting project, which is now REAL-TIME & BETTER quality: 60fps at FHD and 30fps at 4K! You can use this with Zoom, check out our demo! 👇 Webpage: Video: More in the thread! [1/6]
22
162
996
0
0
6
@holynski_
Aleksander Holynski
1 year
Come say hi next week at CVPR! We’ll be presenting the InstructPix2Pix poster on Thursday morning — and I’ll also be giving a talk on this at the Multimodal Learning (MULA) workshop on Sunday. @CVPR #CVPR2023
0
0
6
@holynski_
Aleksander Holynski
1 year
By combining a large language model (GPT-3) and a text-to-image model (Stable Diffusion)! First, fine-tune the LLM on a small collection of human-written examples, and use it to generate a dataset of text triplets: 1. image caption 2. edit instruction 3. caption after the edit
Tweet media one
1
0
6
@holynski_
Aleksander Holynski
1 year
For example, you can use self-guidance to move or resize an object (like this donut) — or even replace it with an item from a real image — all without changing the rest of the scene.
Tweet media one
1
0
6
@holynski_
Aleksander Holynski
1 year
We can use self-guidance to edit entire images at once, too. For example, we can copy the appearance or inherit the layout of another scene — basically re-styling or re-composing any image!
Tweet media one
2
0
6
@holynski_
Aleksander Holynski
4 years
@JPKopf But I'm curious what the best way is to handle transient objects, like the kids in most of the RCVD videos. Maybe something like action shots () Create a couple instances of the person, so there's always something interesting in every direction
Tweet media one
1
0
6
@holynski_
Aleksander Holynski
3 years
Learning To Recover 3D Scene Shape From a Single Image Wei Yin, @oliverwang81 @simon_niklaus & co Uses learned priors to estimate the unknown focal and shift required for unprojecting monocular depth estimates to structurally-correct point clouds
Tweet media one
1
0
6
@holynski_
Aleksander Holynski
3 months
@basilevanh Very cool, congrats!
0
0
5
@holynski_
Aleksander Holynski
2 years
Very cool! Check out the demos on the webpage.
@JulienPhilip2
Julien Philip
2 years
We are happy to finally share our (David Griffiths and Tobias Ritschel) latest Relighting paper, OutCast: Outdoor Single-image Relighting with Cast Shadows. To be presented next week at Eurographics 2022. Paper, results and more at: . (1/4)
9
80
365
0
0
4
@holynski_
Aleksander Holynski
1 year
Then, use the input and output caption to generate a pair of corresponding images, using Prompt-to-Prompt:
Tweet media one
1
0
4
@holynski_
Aleksander Holynski
3 months
@ericmchen1 Sure, here are a couple from a cool sculpture I saw at the MoMA. The samples aren't 100% consistent, but definitely within the ballpark of what a well-tuned NeRF pipeline can handle (just as it might handle inconsistencies in real captures, eg minor dynamics & lighting changes)
1
0
3
@holynski_
Aleksander Holynski
3 years
Real-time NeRF in your browser -- even works on a laptop, no hefty GPUs needed!
@jon_barron
Jon Barron
3 years
Rendering a NeRF is slow, so we found a way to "bake" NeRFs into something more GPU-friendly. The result: real time NeRF rendering in your browser. Try it now before the internet hugs our server to death. @PeterHedman3 @_pratul_ @BenMildenhall @debfx
5
36
194
1
1
4
@holynski_
Aleksander Holynski
9 months
Abe’s work on Image-Space Modal Bases was an important inspiration for our paper, and the new version of the paper better reflects this —connecting some of the terms we used in our original draft with those used in prior work
1
0
4
@holynski_
Aleksander Holynski
3 years
Super cool stuff from @yoni_kasten @dolevofri @oliverwang81 @talidekel Decomposes a video into two canonical templates/atlases (represented by MLPs) that can be edited --- allowing edits to propagate through a whole video. Sort of like MLP Unwrap Mosaics. Awesome results!
@yoni_kasten
Yoni Kasten
3 years
Excited to share our work "Layered Neural Atlases for Consistent Video Editing". Paper: Project page: (with @dolevofri , @oliverwang81 and @talidekel , to appear in SIGGRAPH Asia'21) (1/9)
3
4
42
0
0
4
@holynski_
Aleksander Holynski
9 months
Our work uses these insights about converting dense 2D motion trajectories to the frequency domain, and shows that this spectral volume representation is also an efficient and effective one for __generating__ long-term motion from a single, still image.
1
0
3
@holynski_
Aleksander Holynski
4 years
@jbhuang0604 Wow, this is awesome! This was exactly what was missing. When I was collecting the training data, I was specifically looking for stationary shots, but this made things difficult, since the vast majority of nature videos online have this kind of slow pan.
2
0
3
@holynski_
Aleksander Holynski
9 months
Among the insights in Abe’s work is that the motion trajectories observed in a 2D video of an object exhibiting oscillations/vibrations, when converted to the Fourier domain, are — under certain assumptions — a projection of the 3D vibration modes of that object.
1
0
3
@holynski_
Aleksander Holynski
3 years
Challenge accepted! Here are a couple of my first-day favorites:
@abursuc
Andrei Bursuc
3 years
I dare a challenge for #cvpr2021 twitter: for each paper of yours that you advertise on twitter, please share 1-3 interesting papers from other teams. Authors will appreciate it and the community will be stronger. Here are some champions: @ducha_aiki @CSProfKGD @artsiom_s
3
4
60
1
0
3
@holynski_
Aleksander Holynski
3 years
Pulsar: Efficient Sphere-Based Neural Rendering @chlassner , @MZollhoefer Super cool visualizations of the optimization process -- definitely check out the video.
Tweet media one
1
0
3
@holynski_
Aleksander Holynski
4 years
@jonathanfly @jbhuang0604 @Feldman1Michael Haven't tried it on pixel art yet, but I'll try that soon. I posted one example of non-nature in the original thread:
@holynski_
Aleksander Holynski
4 years
And also some pretty cool failure cases -- if a class of objects aren't seen during training, but share similar textural properties to fluids... [6/5]
15
40
294
1
0
3
@holynski_
Aleksander Holynski
4 years
Modern photos of the past!
@XuanLuo14
Xuan Luo
4 years
What did Abraham Lincoln really look like? Our new project simulates traveling back in time with a modern camera to rephotograph famous subjects. Web: Video: w/ @ceciliazhang77 , @RealPaulYoo , @rmbrualla , @jlawrence176 , Steve Seitz
5
63
248
0
0
3
@holynski_
Aleksander Holynski
4 years
@jonathanfly Yeah, default parameters. The main training Colab will periodically save a checkpoint model, and you have to use a second Colab () to render the video given one of those checkpoints. I think I used the first or second checkpoint.
0
0
2
@holynski_
Aleksander Holynski
2 years
What a cool idea! Excited to see more complex totems like 3D printed refractive structures with unknown internal geometry / material.
@phillip_isola
Phillip Isola
2 years
Tomorrow at ECCV, we are presenting “Totems: Physical Objects for Verifying Visual Integrity” Remember totems from Inception? We tried to make something *a bit* like that in reality. website: paper: 1/n
Tweet media one
3
18
110
1
0
2
@holynski_
Aleksander Holynski
3 months
• • • • • •
0
0
2
@holynski_
Aleksander Holynski
1 month
@bfl_ml Congrats!!!
0
0
2
@holynski_
Aleksander Holynski
3 months
0
0
2
@holynski_
Aleksander Holynski
9 months
@mrdoob Hey, not a bad idea...
0
0
1
@holynski_
Aleksander Holynski
4 years
@JPKopf Haha, yeah! It's improved a lot. You can catch some of our older results in the ablations ;-)
0
0
1
@holynski_
Aleksander Holynski
3 months
0
0
1
@holynski_
Aleksander Holynski
4 months
@CSProfKGD Thanks for checking it out! Send any notes our way!
0
0
1
@holynski_
Aleksander Holynski
1 year
@YifanJiang17 Whoops! 😆 It feels like we wrote it an eternity ago!
0
0
1
@holynski_
Aleksander Holynski
3 months
@albtaiuti Thanks for checking it out. Feel free to send any questions our way.
0
0
1
@holynski_
Aleksander Holynski
3 months
Update, it's now in room ARCH 4F!
0
0
1
@holynski_
Aleksander Holynski
3 years
0
0
1
@holynski_
Aleksander Holynski
4 years
@jbhuang0604 So I've been hoping to find a way to do something like this, to make the shots look more like the type of stuff you'd find online -- you beat me to it! Thanks, I'll have to try it out myself :-)
1
0
1
@holynski_
Aleksander Holynski
2 months
@c_valenzuelab 👀👀👀👀👀👀
0
0
1