Improve model card with tags and paper link

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +39 -7
README.md CHANGED
@@ -1,14 +1,46 @@
1
  ---
2
  license: mit
 
 
3
  ---
4
 
5
- This is the checkpoint for Stable Diffusion fine-tuned with CXR-BERT on the MIMIC-CXR dataset.
6
- This checkpoint can be used to reproduce the results of the Generate to Ground paper.
7
- The corresponding code can be found here: https://github.com/Felix-012/generate_to_ground/.
8
 
9
- Training was performed for 30,000 steps using eight A100 GPUs.
10
 
 
 
 
 
11
 
12
- Stable Diffusion: https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5 \
13
- CXR-BERT: https://huggingface.co/microsoft/BiomedVLP-CXR-BERT-specialized \
14
- MIMIC-CXR: https://physionet.org/content/mimic-cxr/2.0.0/
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ pipeline_tag: zero-shot-object-detection
4
+ library_name: diffusers
5
  ---
6
 
7
+ # Generate to Ground: Multimodal Text Conditioning Boosts Phrase Grounding in Medical Vision-Language Models
 
 
8
 
9
+ This is a checkpoint for Stable Diffusion fine-tuned with CXR-BERT on the MIMIC-CXR dataset, as presented in the paper [Generate to Ground: Multimodal Text Conditioning Boosts Phrase Grounding in Medical Vision-Language Models](https://huggingface.co/papers/2507.12236).
10
 
11
+ This model introduces a novel approach to **phrase grounding** in medical imaging, demonstrating that generative text-to-image diffusion models, specifically fine-tuned Stable Diffusion, can achieve superior zero-shot performance compared to traditional discriminative methods. Key innovations include:
12
+ - Leveraging cross-attention maps from generative diffusion models for phrase grounding.
13
+ - Fine-tuning diffusion models with a frozen, domain-specific language model (CXR-BERT) to significantly improve performance in medical contexts.
14
+ - Introducing **Bimodal Bias Merging (BBM)**, a novel post-processing technique that aligns text and image biases to refine cross-attention maps and enhance localization accuracy.
15
 
16
+ The model aims to map natural language phrases from clinical reports to specific image regions, facilitating disease localization. Training was performed for 30,000 steps using eight A100 GPUs.
17
+
18
+ ## Base Models and Datasets
19
+ * **Stable Diffusion:** [runwayml/stable-diffusion-v1-5](https://huggingface.co/runwayml/stable-diffusion-v1-5)
20
+ * **CXR-BERT:** [microsoft/BiomedVLP-CXR-BERT-specialized](https://huggingface.co/microsoft/BiomedVLP-CXR-BERT-specialized)
21
+ * **MIMIC-CXR (dataset):** [physionet.org/content/mimic-cxr/2.0.0/](https://physionet.org/content/mimic-cxr/2.0.0/)
22
+
23
+ ## Usage and Reproduction
24
+
25
+ To reproduce the results of the "Generate to Ground" paper, including environment setup, data preparation, and execution of evaluation scripts (with and without Bimodal Bias Merging), please refer to the official GitHub repository. The repository provides comprehensive instructions and the corresponding code:
26
+
27
+ [https://github.com/Felix-012/generate_to_ground/](https://github.com/Felix-012/generate_to_ground/)
28
+
29
+ ## Citation
30
+
31
+ If you find this work helpful or inspiring, please consider citing the original paper:
32
+
33
+ ```bibtex
34
+ @inproceedings{
35
+ nutzel2025generate,
36
+ title={Generate to Ground: Multimodal Text Conditioning Boosts Phrase Grounding in Medical Vision-Language Models},
37
+ author={Felix N{\"u}tzel and Mischa Dombrowski and Bernhard Kainz},
38
+ booktitle={Medical Imaging with Deep Learning},
39
+ year={2025},
40
+ url={https://openreview.net/forum?id=yTjotBI30L}
41
+ }
42
+ ```
43
+
44
+ ## Acknowledgement
45
+
46
+ (Some) HPC resources were provided by the Erlangen National High Performance Computing Center (NHR@FAU) of the Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU) under the NHR projects b143dc and b180dc. NHR funding is provided by federal and Bavarian state authorities. NHR@FAU hardware is partially funded by the German Research Foundation (DFG) – 440719683.