Unified Video Editing with Temporal Reasoner

1University of Technology Sydney 2Zhejiang University

A Short Video Introduction

Abstract

Existing video editing methods face a critical trade-off: expert models offer precision but rely on task-specific priors like masks, hindering unification; conversely, unified temporal in-context learning models are mask-free but lack explicit spatial cues, leading to weak instruction-to-region mapping and imprecise localization. To resolve this conflict, we propose VideoCoF, a novel Chain-of-Frames approach inspired by Chain-of-Thought reasoning.

VideoCoF enforces a “see → reason → edit” procedure by compelling the video diffusion model to first predict reasoning tokens (edit-region latents) before generating the target video tokens. This explicit reasoning step removes the need for user-provided masks while achieving precise instruction-to-region alignment and fine-grained video editing. Furthermore, we introduce a RoPE alignment strategy that leverages these reasoning tokens to ensure motion alignment and enable length extrapolation beyond the training duration. We demonstrate that with a minimal data cost of only 50k video pairs, VideoCoF achieves SOTA performance on VideoCoF-Bench, validating the efficiency and effectiveness of our approach.

Why We Need Reasoning Before Editing?

Current video editing methods typically follow two paths:
(1) expert models, which rely on external masks for precision but sacrifice unification;
(2) unified in-context learning models, which are mask-free but often struggle with spatial accuracy due to the lack of explicit cues.

This raises a critical question: Can we maintain the precision of expert models and the unification of in-context models without the mask dependency?

To resolve this conflict, we propose VideoCoF, a Chain-of-Frames approach that predicts reasoning tokens (edit-region latents) before generating the target video tokens, thereby removing the need for user-provided masks while achieving precise instruction-to-region alignment.

Length Extrapolation

Trained on only 50k data (33 frames), this following examples shows multi-shot editing and robust 4× length generalization.

Seeing, Reasoning, Editing

VideoCoF adopts a "seeing, reasoning, editing" approach, where it first reasons about the editing region before implementing the corresponding edit.

Citation

@article{yang2025videocof,
                title={Unified Video Editing with Temporal Reasoner},
                author={Yang, Xiangpeng and Xie, Ji and Yang, Yiyuan and Huang, Yan and Xu, Min and Wu, Qiang},
                journal={arXiv preprint arXiv:2512.07469},
                year={2025}
                }