CineTrans: Learning to Generate Videos with Cinematic Transitions via Masked Diffusion Models

 β€‚  β€‚  β€‚  β€‚ 

Xiaoxue Wu1,2*, Bingjie Gao2,3, Yu Qiao2†, Yaohui Wang2†, Xinyuan Chen2†

1Fudan University 2Shanghai Artificial Intelligence Laboratory 3Shanghai Jiao Tong University

*Work done during internship at Shanghai AI Laboratory †Corresponding author

πŸ“₯ Installation

  1. Clone the Repository
git clone https://github.com/UknowSth/CineTrans.git
cd CineTrans
  1. Set up Environment
conda create -n cinetrans python==3.11.9
conda activate cinetrans

pip install torch==2.5.1 torchvision==0.20.1 --index-url https://download.pytorch.org/whl/cu118
pip install -r requirements.txt

πŸ€— Checkpoint

CineTrans-DiT

Download the weights of Wan2.1-T2V-1.3B and lora weights. Place them as:

Wan2.1-T2V-1.3B/ # original weights
│── google/
β”‚   └── umt5-xxl/
│── config.json
│── diffusion_pytorch_model.safetensors
│── models_t5_umt5-xxl-enc-bf16.pth
│── Wan2.1_VAE.pth
ckpt/
└── weights.pt # lora weights

For more inference details, please refer to our GitHub repository.

πŸ“‘ BiTeX

If you find CineTrans useful for your research and applications, please cite using this BibTeX:

@misc{wu2025cinetranslearninggeneratevideos,
      title={CineTrans: Learning to Generate Videos with Cinematic Transitions via Masked Diffusion Models}, 
      author={Xiaoxue Wu and Bingjie Gao and Yu Qiao and Yaohui Wang and Xinyuan Chen},
      year={2025},
      eprint={2508.11484},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2508.11484}, 
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for NumlockUknowSth/CineTrans-DiT

Finetuned
(28)
this model

Dataset used to train NumlockUknowSth/CineTrans-DiT

Paper for NumlockUknowSth/CineTrans-DiT