Aliaksandr Siarohin

Welcome! I am a Staff Research Scientist leading video generation research at Snap. Previously, I was a Ph.D Student at the University of Trento under supervision of prof Nicu Sebe. During my Ph.D. years I created First Order Motion Model, a video animation technology that started several startups and for several years was one of the most popular models on Runway. First Order Motion Model was also one of the first AI technologies used for commercial media creation. My team and I developed SnapVideo, a family of foundational video generation models of unprecedented speed, low cost and the quality at the level of leading models, such as Veo and Sora. SnapVideo now powers all video generation applications at Snap. I have more than 30 works published in top computer vision and machine learning conferences.

Contact: aliaksandr [dot] siarohin [at] gmail [dot] com

[Google Scholar] [GitHub] [CV]

Publications:


DenseDPO: Fine-Grained Temporal Preference Optimization for Video Diffusion Models

Ziyi Wu, Anil Kag, Ivan Skorokhodov, Willi Menapace, Ashkan Mirzaei, Igor Gilitschenski, Sergey Tulyakov, Aliaksandr Siarohin NeurIPS 2025

[Paper] [Website]

4Real-Video-V2: Fused View-Time Attention and Feedforward Reconstruction for 4D Scene Generation

Chaoyang Wang, Ashkan Mirzaei, Vidit Goel, Willi Menapace, Aliaksandr Siarohin, Avalon Vinella, Michael Vasilkovsky, Ivan Skorokhodov, Vladislav Shakhrai, Sergey Korolev, Sergey Tulyakov, Peter Wonka NeurIPS 2025

[Paper] [Website]

Improving Progressive Generation with Decomposable Flow Matching

Moayed Haji-Ali, Willi Menapace, Ivan Skorokhodov, Arpit Sahni, Sergey Tulyakov, Vicente Ordonez, Aliaksandr Siarohin NeurIPS 2025

[Paper] [Website]

Improving the Diffusability of Autoencoders

Ivan Skorokhodov, Sharath Girish, Benran Hu, Willi Menapace, Yanyu Li, Rameen Abdal, Sergey Tulyakov, Aliaksandr Siarohin ICML 2025

[Paper] [Code]

AC3D: Analyzing and Improving 3D Camera Control in Video Diffusion Transformers

Sherwin Bahmani, Ivan Skorokhodov, Guocheng Qian, Aliaksandr Siarohin, Willi Menapace, Andrea Tagliasacchi, David B. Lindell, Sergey Tulyakov CVPR 2025

[Paper] [Website]

AV-Link: Temporally-Aligned Diffusion Features for Cross-Modal Audio-Video Generation

Moayed Haji-Ali, Willi Menapace, Aliaksandr Siarohin, Ivan Skorokhodov, Alper Canberk, Kwot Sin Lee, Vicente Ordonez, Sergey Tulyakov ICCV 2025

[Paper] [Website]

4Real-Video: Learning Generalizable Photo-Realistic 4D Video Diffusion

Chaoyang Wang, Peiye Zhuang, Tuan Duc Ngo, Willi Menapace, Aliaksandr Siarohin, Michael Vasilkovsky, Ivan Skorokhodov, Sergey Tulyakov, Peter Wonka, Hsin-Ying Lee CVPR 2025

[Paper] [Website]

VideoAlchemy: Open-set Personalization in Video Generation

Tsai-Shien Chen, Aliaksandr Siarohin, Willi Menapace, Yuwei Fang, Kwot Sin Lee, Ivan Skorokhodov, Kfir Aberman, Jun-Yan Zhu, Ming-Hsuan Yang, Sergey Tulyakov CVPR 2025

[Paper] [Website]

Mind the Time: Temporally-Controlled Multi-Event Video Generation

Ziyi Wu, Aliaksandr Siarohin, Willi Menapace, Ivan Skorokhodov, Yuwei Fang, Varnith Chordia, Igor Gilitschenski, Sergey Tulyakov CVPR 2025

[Paper] [Website]

SF-V: Single Forward Video Generation Model

Zhixing Zhang, Yanyu Li, Yushu Wu, Yanwu Xu, Anil Kag, Ivan Skorokhodov, Willi Menapace, Aliaksandr Siarohin, Junli Cao, Dimitris Metaxas, Sergey Tulyakov, Jian Ren NeurIPS 2024

[Paper] [Website]

4Real: Towards Photorealistic 4D Scene Generation via Video Diffusion Models

Heng Yu, Chaoyang Wang, Peiye Zhuang, Willi Menapace, Aliaksandr Siarohin, Junli Cao, Laszlo A Jeni, Sergey Tulyakov, Hsin-Ying Lee NeurIPS 2024

[Paper] [Website]

Panda-70M: Captioning 70M Videos with Multiple Cross-Modality Teachers

Tsai-Shien Chen, Aliaksandr Siarohin, Willi Menapace, Ekaterina Deyneka, Hsiang-wei Chao, Byung Eun Jeon, Yuwei Fang, Hsin-Ying Lee, Jian Ren, Ming-Hsuan Yang, Sergey Tulyakov CVPR 2024

[Paper] [Website]

Snap Video: Scaled Spatiotemporal Transformers for Text-to-video Synthesis

Willi Menapace, Aliaksandr Siarohin, Ivan Skorokhodov, Ekaterina Deyneka, Tsai-Shien Chen, Anil Kag, Yuwei Fang, Aleksei Stoliar, Elisa Ricci, Jian Ren, Sergey Tulyakov CVPR 2024

[Paper] [Website]

Hierarchical Patch Diffusion Models for High-Resolution Video Generation

Ivan Skorokhodov, Willi Menapace, Aliaksandr Siarohin, Sergey Tulyakov CVPR 2024

[Paper] [Website]

SPAD: Spatially Aware Multi-View Diffusers

Yash Kant, Aliaksandr Siarohin, Ziyi Wu, Michael Vasilkovsky, Guocheng Qian, Jian Ren, Riza Alp Guler, Bernard Ghanem, Sergey Tulyakov, Igor Gilitschenski CVPR 2024

[Paper] [Website]

Promptable Game Models: Text-guided Game Simulation via Masked Diffusion Models

Willi Menapace, Aliaksandr Siarohin, Stéphane Lathuilière, Panos Achlioptas, Vladislav Golyanik, Sergey Tulyakov, Elisa Ricci TOG 2024

[Paper] [Website]

Magic123: One Image to High-Quality 3D Object Generation Using Both 2D and 3D Diffusion Priors

Guocheng Qian, Jinjie Mai, Abdullah Hamdi, Jian Ren, Aliaksandr Siarohin, Bing Li, Hsin-Ying Lee, Ivan Skorokhodov, Peter Wonka, Sergey Tulyakov, Bernard Ghanem ICLR 2024

[Paper] [Website]

HyperHuman: Hyper-Realistic Human Generation with Latent Structural Diffusion

Xian Liu, Jian Ren, Aliaksandr Siarohin, Ivan Skorokhodov, Yanyu Li, Dahua Lin, Xihui Liu, Ziwei Liu, Sergey Tulyakov ICLR 2024

[Paper] [Website]

Text-Guided Synthesis of Eulerian Cinemagraphs

Aniruddha Mahapatra, Aliaksandr Siarohin, Hsin-Ying Lee, Sergey Tulyakov, Jun-Yan Zhu SIGGRAPH Asia 2023

[Paper] [Website]

Autodecoding latent 3d diffusion models

Evangelos Ntavelis, Aliaksandr Siarohin, Kyle Olszewski, Chaoyang Wang, Luc Van Gool, Sergey Tulyakov NeurIPS 2023

[Paper] [Website]

Unsupervised Volumetric Animation

Aliaksandr Siarohin, Willi Menapace, Ivan Skorokhodov, Kyle Olszewski, Jian Ren, Hsin-Ying Lee, Menglei Chai, Sergey Tulyakov CVPR 2023

[Paper] [Website]

Playable Environments: Video Manipulation in Space and Time

Willi Menapace, Stéphane Lathuilière, Aliaksandr Siarohin, Christian Theobalt, Sergey Tulyakov, Vladislav Golyanik, Elisa Ricci
CVPR 2022

[Paper] [Website] [Code]

Motion Representations for Articulated Animation

Aliaksandr Siarohin, Oliver Woodford, Jian Ren, Menglei Chai, Sergey Tulyakov
CVPR 2021

[Paper] [Website] [Code]

Playable Video Generation

Willi Menapace, Stéphane Lathuilière, Sergey Tulyakov, Aliaksandr Siarohin, Elisa Ricci
CVPR 2021

[Paper] [Website] [Code]

Motion-supervised Co-Part Segmentation

Aliaksandr Siarohin, Subhankar Roy, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, Nicu Sebe
ICPR 2021

[Paper] [Code]

TriGAN: image-to-image translation for multi-source domain adaptation

Subhankar Roy, Aliaksandr Siarohin, Enver Sangineto, Nicu Sebe, Elisa Ricci
Machine Vision and Applications 2021

[Paper]

First Order Motion Model for Image Animation

Aliaksandr Siarohin, Stephane Lathuillere, Sergey Tulyakov, Elisa Ricci, Nicu Sebe
NeurIPS 2019

[Paper] [Code]

Attention-based Fusion for Multi-source Human Image Generation

Stéphane Lathuilière, Enver Sangineto, Aliaksandr Siarohin, Nicu Sebe
WACV 2020

[Paper]

DwNet: Dense warp-based network for pose-guided human video generation

Polina Zablotskaia, Aliaksandr Siarohin, Bo Zhao, Leonid Sigal
BMVC 2019

[Paper] [Code]

Increasing image memorability with neural style transfer

Aliaksandr Siarohin, Gloria Zen, Cveta Majtanovic, Xavier Alameda-Pineda, Elisa Ricci, Nicu Sebe
TOMM 2019 (Best Paper Award)

[Paper] [Code]

Appearance and Pose-Conditioned Human Image Generation using Deformable GANs

Aliaksandr Siarohin, Stéphane Lathuilière, Enver Sangineto, Nicu Sebe
PAMI 2019

[Paper] [Code]

Unsupervised Domain Adaptation using Feature-Whitening and Consensus Loss

Subhankar Roy, Aliaksandr Siarohin, Enver Sangineto, Samuel Rota Bulo, Nicu Sebe, Elisa Ricci
CVPR 2019

[Paper] [Code]

Animating arbitrary objects via deep motion transfer

Aliaksandr Siarohin, Stéphane Lathuilière, Sergey Tulyakov, Elisa Ricci, Nicu Sebe
CVPR 2019

[Paper] [Code]

Whitening and Coloring Batch Transform for GANs

Aliaksandr Siarohin, Enver Sangineto, Nicu Sebe
ICLR 2019

[Paper] [Code]

Enhancing Perceptual Attributes with Bayesian Style Generation

Aliaksandr Siarohin, Gloria Zen, Nicu Sebe, Elisa Ricci
ACCV 2018

[Paper] [Code]

Deformable gans for pose-based human image generation

Aliaksandr Siarohin, Enver Sangineto, Stéphane Lathuilière, Nicu Sebe
CVPR 2018

[Paper] [Code]

How to make an image more memorable?: A deep style transfer approach

Aliaksandr Siarohin, Gloria Zen, Cveta Majtanovic, Xavier Alameda-Pineda, Elisa Ricci, Nicu Sebe
ICMR 2017

[Paper] [Code]