Text-to-Video

CineScale: Free Lunch in High-Resolution Cinematic Visual Generation

This repository contains the CineScale models presented in the paper CineScale: Free Lunch in High-Resolution Cinematic Visual Generation.

CineScale proposes a novel inference paradigm to enable higher-resolution visual generation. It broadens the scope by enabling high-resolution I2V (Image-to-Video) and V2V (Video-to-Video) synthesis, built atop state-of-the-art open-source video generation frameworks, significantly improving upon existing methods which are prone to repetitive patterns in high-resolution outputs.

Project Page: https://eyeline-labs.github.io/CineScale/ Code & Detailed Usage: https://github.com/Eyeline-Labs/CineScale

Models

CineScale provides a family of models, including Text-to-Video (T2V) and Image-to-Video (I2V) variants, capable of generating videos up to 4K resolution.

Model Tuning Resolution Checkpoint Description
CineScale-1.3B-T2V 1088x1920 Hugging Face Supports 3K (1632x2880) inference on A100 x 1
CineScale-14B-T2V 1088x1920 Hugging Face Supports 4K (2176x3840) inference on A100 x 8
CineScale-14B-I2V 1088x1920 Hugging Face Supports 4K (2176x3840) inference on A100 x 8

Quick Start

To get started, you will need to set up the environment and download the model checkpoints as described in the GitHub repository.

Inference examples for various resolutions and tasks are provided in the GitHub repository's command-line scripts. For instance, to run 2K-resolution text-to-video inference:

# Example for 2K-Resolution Text-to-Video (Base Model Wan2.1-1.3B)
# Single GPU
CUDA_VISIBLE_DEVICES=0 python cinescale_t2v1.3b_single.py
# Multiple GPUs
torchrun --standalone --nproc_per_node=8 cinescale_t2v1.3b.py

Refer to the GitHub repository for more detailed instructions and examples for 3K and 4K video generation.

Citation

If you find our work useful, please consider citing our paper:

@article{qiu2025cinescale,
  title={CineScale: Free Lunch in High-Resolution Cinematic Visual Generation},
  author={Haonan Qiu and Ning Yu and Ziqi Huang and Paul Debevec and Ziwei Liu},
  journal={arXiv preprint arXiv:2508.15774},
  year={2025}
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support