Paper ID | ARS-10.8 | ||
Paper Title | MAT-NET: REPRESENTING APPEARANCE-IRRELEVANT WARP FIELD BY MULTIPLE AFFINE TRANSFORMATIONS | ||
Authors | Jingwei Liu, Longquan Dai, Nanjing University Of Science And Technology, China | ||
Session | ARS-10: Image and Video Analysis and Synthesis | ||
Location | Area H | ||
Session Time: | Monday, 20 September, 15:30 - 17:00 | ||
Presentation Time: | Monday, 20 September, 15:30 - 17:00 | ||
Presentation | Poster | ||
Topic | Image and Video Analysis, Synthesis, and Retrieval: Image & Video Synthesis, Rendering, and Visualization | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | Warp-based methods for image animation estimate a warp field what do a rearrangement on the pixels of the input image to roughly align with the target image. Current methods predict accurate warp field by using manually annotated data. In this paper, we propose a simple method (MAT-net) to predict more precise warp field in self-supervised way. MAT-net decomposes complex spatial object movement between two images into multiple simple local motions (i.e. affine transformation) occurring in different areas of images. Sequentially, our model calculates a warp field depicting complex object movement by combining all local motions. MAT-net encodes appearance-irrelevant object movement accurately. Compared to the state-of-the-art method, MAT-net generates more realistic images with faster inference speed. We published the source code of our project online. |