Timezone: »
We present a new implicit warping framework for image animation using sets of source images through the transfer of motion of a driving video. A single cross-modal attention layer is used to find correspondences between the source images and the driving image, choose the most appropriate features from different source images, and warp the selected features. This is in contrast to the existing methods that use explicit flow-based warping, which is designed for animation using a single source and does not extend well to multiple sources. The pick-and-choose capability of our framework helps it achieve state-of-the-art results on multiple datasets for image animation using both single and multiple source images.
Author Information
Arun Mallya (NVIDIA)
Ting-Chun Wang (NVIDIA)
Ming-Yu Liu (NVIDIA)
More from the Same Authors
-
2022 Poster: Implicit Neural Representations with Levels-of-Experts »
Zekun Hao · Arun Mallya · Serge Belongie · Ming-Yu Liu -
2022 Poster: Generating Long Videos of Dynamic Scenes »
Tim Brooks · Janne Hellsten · Miika Aittala · Ting-Chun Wang · Timo Aila · Jaakko Lehtinen · Ming-Yu Liu · Alexei Efros · Tero Karras -
2021 Poster: Deep Marching Tetrahedra: a Hybrid Representation for High-Resolution 3D Shape Synthesis »
Tianchang Shen · Jun Gao · Kangxue Yin · Ming-Yu Liu · Sanja Fidler -
2020 Poster: Learning compositional functions via multiplicative weight updates »
Jeremy Bernstein · Jiawei Zhao · Markus Meister · Ming-Yu Liu · Anima Anandkumar · Yisong Yue