Video-to-Video
liuhuadai commited on
Commit
ca09660
·
verified ·
1 Parent(s): e13db5f

Upload 3 files

Browse files
Files changed (4) hide show
  1. .gitattributes +2 -0
  2. README (1).md +35 -0
  3. model_structure.png +3 -0
  4. teaser.png +3 -0
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ model_structure.png filter=lfs diff=lfs merge=lfs -text
37
+ teaser.png filter=lfs diff=lfs merge=lfs -text
README (1).md ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: video-to-video
4
+ ---
5
+
6
+ This repository contains the weights of [ThinkSound: Chain-of-Thought Reasoning in Multimodal Large Language Models for Audio Generation and Editing](https://arxiv.org/abs/2506.21448).
7
+
8
+ Project Page: https://thinksound-project.github.io/.
9
+
10
+ Paper: https://huggingface.co/papers/2506.21448\
11
+
12
+ <img src="./teaser.png" alt="model_structure" style="zoom:20%;" />
13
+
14
+ ## Abstract
15
+ While end-to-end video-to-audio generation has greatly improved, producing high-fidelity audio that authentically captures the nuances of visual content remains challenging. Like professionals in the creative industries, such generation requires sophisticated reasoning about items such as visual dynamics, acoustic environments, and temporal relationships. We present ThinkSound, a novel framework that leverages Chain-of-Thought (CoT) reasoning to enable stepwise, interactive audio generation and editing for videos. Our approach decomposes the process into three complementary stages: foundational foley generation that creates semantically coherent soundscapes, interactive object-centric refinement through precise user interactions, and targeted editing guided by natural language instructions. At each stage, a multimodal large language model generates contextually aligned CoT reasoning that guides a unified audio foundation model. Furthermore, we introduce AudioCoT, a comprehensive dataset with structured reasoning annotations that establishes connections between visual content, textual descriptions, and sound synthesis. Experiments demonstrate that ThinkSound achieves state-of-the-art performance in video-to-audio generation across both audio metrics and CoT metrics and excels in out-of-distribution Movie Gen Audio benchmark. The demo page is available at https://ThinkSound-Project.github.io.
16
+
17
+ ## Model Overview
18
+
19
+ <img src="./model_structure.png" alt="model_structure" style="zoom:40%;" />
20
+
21
+ ## Citation
22
+
23
+ If you find our work useful, please cite our paper:
24
+
25
+ ```bibtex
26
+ @misc{liu2025thinksoundchainofthoughtreasoningmultimodal,
27
+ title={ThinkSound: Chain-of-Thought Reasoning in Multimodal Large Language Models for Audio Generation and Editing},
28
+ author={Huadai Liu and Jialei Wang and Kaicheng Luo and Wen Wang and Qian Chen and Zhou Zhao and Wei Xue},
29
+ year={2025},
30
+ eprint={2506.21448},
31
+ archivePrefix={arXiv},
32
+ primaryClass={eess.AS},
33
+ url={https://arxiv.org/abs/2506.21448},
34
+ }
35
+ ```
model_structure.png ADDED

Git LFS Details

  • SHA256: 373cdfd3c12d83d030a25ac7e2611a139ba6471b6b90972b69065f70ff1ad32e
  • Pointer size: 131 Bytes
  • Size of remote file: 478 kB
teaser.png ADDED

Git LFS Details

  • SHA256: 9161d9c92a067b33aa7f4080ebd11a7de4c8998c646b46430aeb3ee6bb1d593e
  • Pointer size: 132 Bytes
  • Size of remote file: 4.7 MB