nielsr HF Staff commited on
Commit
7273c83
·
verified ·
1 Parent(s): 804a6e5

Improve dataset card: Add metadata, paper link, code link, and sample usage

Browse files

This PR significantly improves the dataset card for `TiM-Toy-T2I-Dataset` by adding comprehensive information.

Key enhancements include:
- Adding `task_categories: text-to-image`, `license: apache-2.0`, `language: en`, and relevant tags (`diffusion-model`, `generative-ai`, `image-generation`, `toy-dataset`) to the metadata for better discoverability and clarity.
- Linking to the official Hugging Face paper page: [Transition Models: Rethinking the Generative Learning Objective](https://huggingface.co/papers/2509.04394).
- Including a link to the associated GitHub repository: [https://github.com/WZDTHU/TiM](https://github.com/WZDTHU/TiM).
- Providing an introductory description of the dataset and the Transition Models (TiM) paradigm.
- Adding a "Sample Usage" section with a code snippet from the GitHub README, demonstrating how to download this specific toy dataset for training.
- Including the BibTeX citation and a clear license statement.

These updates make the dataset card more informative, accessible, and compliant with Hugging Face Hub best practices.

Files changed (1) hide show
  1. README.md +49 -1
README.md CHANGED
@@ -1 +1,49 @@
1
- arxiv.org/abs/2509.04394
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - text-to-image
4
+ license: apache-2.0
5
+ language:
6
+ - en
7
+ tags:
8
+ - diffusion-model
9
+ - generative-ai
10
+ - image-generation
11
+ - toy-dataset
12
+ ---
13
+
14
+ # TiM-Toy-T2I-Dataset
15
+
16
+ This repository contains the toy dataset for text-to-image (T2I) generation used in the paper [Transition Models: Rethinking the Generative Learning Objective](https://huggingface.co/papers/2509.04394).
17
+
18
+ Transition Models (TiM) introduce an exact, continuous-time dynamics equation that analytically defines state transitions across any finite time interval. This leads to a novel generative paradigm, TiM, which adapts to arbitrary-step transitions, seamlessly traversing the generative trajectory from single leaps to fine-grained refinement with more steps. This specific dataset serves as a toy dataset for training and experimentation with TiM for text-to-image generation.
19
+
20
+ Code: [https://github.com/WZDTHU/TiM](https://github.com/WZDTHU/TiM)
21
+
22
+ ## Sample Usage
23
+
24
+ To download this toy text-to-image dataset for training TiM models, you can follow the instructions from the official GitHub repository under the "Dataset Setup" section:
25
+
26
+ ```bash
27
+ bash tools/download_toy_t2i_dataset.sh
28
+ ```
29
+
30
+ For more details on setting up the environment, downloading models, and other training/sampling scripts, please refer to the [official GitHub repository](https://github.com/WZDTHU/TiM).
31
+
32
+ ## Citation
33
+
34
+ If you find the project useful, please kindly cite:
35
+
36
+ ```bibtex
37
+ @article{wang2025transition,
38
+ title={Transition Models: Rethinking the Generative Learning Objective},
39
+ author={Wang, Zidong and Zhang, Yiyuan and Yue, Xiaoyu and Yue, Xiangyu and Li, Yangguang and Ouyang, Wanli and Bai, Lei},
40
+ year={2025},
41
+ eprint={2509.04394},
42
+ archivePrefix={arXiv},
43
+ primaryClass={cs.LG}
44
+ }
45
+ ```
46
+
47
+ ## License
48
+
49
+ This project is licensed under the Apache-2.0 license.