Self-Calibration Flow Guided Denoising Diffusion Model for Human Pose Transfer

Xue, Yu and Po, Lai-Man and Yu, Wing-Yin and Wu, Haoxuan and Xu, Xuyuan and Li, Kun and Liu, Yuyang (2024) Self-Calibration Flow Guided Denoising Diffusion Model for Human Pose Transfer. IEEE Transactions on Circuits and Systems for Video Technology, 34 (9). pp. 7896-7911. ISSN 1051-8215

Full text not available from this repository.

Abstract

The human pose transfer task aims to generate synthetic person images that preserve the style of reference images while accurately aligning them with the desired target pose. However, existing methods based on generative adversarial networks (GANs) struggle to produce realistic details and often face spatial misalignment issues. On the other hand, methods relying on denoising diffusion models require a large number of model parameters, resulting in slower convergence rates. To address these challenges, we propose a self-calibration flow-guided module (SCFM) to establish precise spatial correspondence between reference images and target poses. This module facilitates the denoising diffusion model in predicting the noise at each denoising step more effectively. Additionally, we introduce a multi-scale feature fusing module (MSFF) that enhances the denoising U-Net architecture through a cross-attention mechanism, achieving better performance with a reduced parameter count. Our proposed model outperforms state-of-the-art methods on the DeepFashion and Market-1501 datasets in terms of both the quantity and quality of the synthesized images. Our code is publicly available at https://github.com/zylwithxy/SCFM-guided-DDPM.

Item Type:
Journal Article
Journal or Publication Title:
IEEE Transactions on Circuits and Systems for Video Technology
Uncontrolled Keywords:
Research Output Funding/no_not_funded
Subjects:
?? no - not fundednomedia technologyelectrical and electronic engineering ??
ID Code:
229492
Deposited By:
Deposited On:
19 May 2025 13:45
Refereed?:
Yes
Published?:
Published
Last Modified:
20 May 2025 02:30