Fix pipeline tag, add project page

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -1,10 +1,10 @@
1
  ---
2
  base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
3
- library_name: peft
4
- license: llama3.1
5
  language:
6
  - en
7
- pipeline_tag: text2text-generation
 
 
8
  ---
9
 
10
  # LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP Model Card
@@ -14,11 +14,12 @@ pipeline_tag: text2text-generation
14
  ## Model Description
15
  - **Repository:** [Code](https://github.com/ahnjaewoo/MAC)
16
  - **Paper:** [Can LLMs Deceive CLIP? Benchmarking Adversarial Compositionality of Pre-trained Multimodal Representation via Text Updates](https://arxiv.org/abs/2505.22943)
 
17
  - **Point of Contact:** [Jaewoo Ahn](mailto:[email protected]), [Heeseung Yun](mailto:[email protected])
18
 
19
  ## Model Details
20
- - **Model**: *LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP* is a deceptive caption generator built on **LLaMA-3.1-8B**, fine-tuned using LoRA (i.e., *self-training*, or more specifically, *rejection sampling fine-tuning (RFT)*) to deceive **CLIP** on the **COCO** dataset. It achieves an **attack success rate (ASR)** of **42.1%**.
21
  - **Architecture**: This model is based on [LLaMA-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) and utilizes [PEFT](https://github.com/huggingface/peft) v0.12.0 for efficient fine-tuning.
22
 
23
  ## How to Use
24
- See our GitHub [repository](https://github.com/ahnjaewoo/MAC) for full usage instructions and scripts.
 
1
  ---
2
  base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
 
 
3
  language:
4
  - en
5
+ library_name: peft
6
+ license: llama3.1
7
+ pipeline_tag: text-generation
8
  ---
9
 
10
  # LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP Model Card
 
14
  ## Model Description
15
  - **Repository:** [Code](https://github.com/ahnjaewoo/MAC)
16
  - **Paper:** [Can LLMs Deceive CLIP? Benchmarking Adversarial Compositionality of Pre-trained Multimodal Representation via Text Updates](https://arxiv.org/abs/2505.22943)
17
+ - **Project Page:** [Project Page](https://vision.snu.ac.kr/projects/mac)
18
  - **Point of Contact:** [Jaewoo Ahn](mailto:[email protected]), [Heeseung Yun](mailto:[email protected])
19
 
20
  ## Model Details
21
+ - **Model**: *LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP* is a deceptive caption generator built on **LLaMA-3.1-8B**, fine-tuned using LoRA (i.e., *self-training*, or more specifically, *rejection sampling fine-tuning (RFT)*) to deceive **CLIP** on the **COCO** dataset. It achieves an **attack success rate (ASR)** of **42.1%**.\
22
  - **Architecture**: This model is based on [LLaMA-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) and utilizes [PEFT](https://github.com/huggingface/peft) v0.12.0 for efficient fine-tuning.
23
 
24
  ## How to Use
25
+ See our GitHub [repository](https://github.com/ahnjaewoo/MAC) for full usage instructions and scripts.