LongDiff : Training-Free Long Video Generation in One Go

Li, Zhuoling and Rahmani, Hossein and Ke, Qiuhong and Liu, Jun (2025) LongDiff : Training-Free Long Video Generation in One Go. Other. Arxiv.

Full text not available from this repository.

Abstract

Video diffusion models have recently achieved remarkable results in video generation. Despite their encouraging performance, most of these models are mainly designed and trained for short video generation, leading to challenges in maintaining temporal consistency and visual details in long video generation. In this paper, we propose LongDiff, a novel training-free method consisting of carefully designed components \ -- Position Mapping (PM) and Informative Frame Selection (IFS) \ -- to tackle two key challenges that hinder short-to-long video generation generalization: temporal position ambiguity and information dilution. Our LongDiff unlocks the potential of off-the-shelf video diffusion models to achieve high-quality long video generation in one go. Extensive experiments demonstrate the efficacy of our method.

Item Type:
Monograph (Other)
Additional Information:
DBLP's bibliographic metadata records provided through http://dblp.org/search/publ/api are distributed under a Creative Commons CC0 1.0 Universal Public Domain Dedication. Although the bibliographic metadata records are provided consistent with CC0 1.0 Dedication, the content described by the metadata records is not. Content may be subject to copyright, rights of privacy, rights of publicity and other restrictions.
Uncontrolled Keywords:
Research Output Funding/yes_externally_funded
Subjects:
?? yes - externally funded ??
ID Code:
230521
Deposited By:
Deposited On:
12 Nov 2025 14:25
Refereed?:
No
Published?:
Published
Last Modified:
12 Nov 2025 23:15