Fix pipeline tag, add project page
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
---
|
2 |
base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
|
3 |
-
library_name: peft
|
4 |
-
license: llama3.1
|
5 |
language:
|
6 |
- en
|
7 |
-
|
|
|
|
|
8 |
---
|
9 |
|
10 |
# LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP Model Card
|
@@ -14,11 +14,12 @@ pipeline_tag: text2text-generation
|
|
14 |
## Model Description
|
15 |
- **Repository:** [Code](https://github.com/ahnjaewoo/MAC)
|
16 |
- **Paper:** [Can LLMs Deceive CLIP? Benchmarking Adversarial Compositionality of Pre-trained Multimodal Representation via Text Updates](https://arxiv.org/abs/2505.22943)
|
|
|
17 |
- **Point of Contact:** [Jaewoo Ahn](mailto:[email protected]), [Heeseung Yun](mailto:[email protected])
|
18 |
|
19 |
## Model Details
|
20 |
-
- **Model**: *LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP* is a deceptive caption generator built on **LLaMA-3.1-8B**, fine-tuned using LoRA (i.e., *self-training*, or more specifically, *rejection sampling fine-tuning (RFT)*) to deceive **CLIP** on the **COCO** dataset. It achieves an **attack success rate (ASR)** of **42.1
|
21 |
- **Architecture**: This model is based on [LLaMA-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) and utilizes [PEFT](https://github.com/huggingface/peft) v0.12.0 for efficient fine-tuning.
|
22 |
|
23 |
## How to Use
|
24 |
-
See our GitHub [repository](https://github.com/ahnjaewoo/MAC) for full usage instructions and scripts.
|
|
|
1 |
---
|
2 |
base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
|
|
|
|
|
3 |
language:
|
4 |
- en
|
5 |
+
library_name: peft
|
6 |
+
license: llama3.1
|
7 |
+
pipeline_tag: text-generation
|
8 |
---
|
9 |
|
10 |
# LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP Model Card
|
|
|
14 |
## Model Description
|
15 |
- **Repository:** [Code](https://github.com/ahnjaewoo/MAC)
|
16 |
- **Paper:** [Can LLMs Deceive CLIP? Benchmarking Adversarial Compositionality of Pre-trained Multimodal Representation via Text Updates](https://arxiv.org/abs/2505.22943)
|
17 |
+
- **Project Page:** [Project Page](https://vision.snu.ac.kr/projects/mac)
|
18 |
- **Point of Contact:** [Jaewoo Ahn](mailto:[email protected]), [Heeseung Yun](mailto:[email protected])
|
19 |
|
20 |
## Model Details
|
21 |
+
- **Model**: *LLaMA-3.1-8B-LoRA-COCO-Deceptive-CLIP* is a deceptive caption generator built on **LLaMA-3.1-8B**, fine-tuned using LoRA (i.e., *self-training*, or more specifically, *rejection sampling fine-tuning (RFT)*) to deceive **CLIP** on the **COCO** dataset. It achieves an **attack success rate (ASR)** of **42.1%**.\
|
22 |
- **Architecture**: This model is based on [LLaMA-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) and utilizes [PEFT](https://github.com/huggingface/peft) v0.12.0 for efficient fine-tuning.
|
23 |
|
24 |
## How to Use
|
25 |
+
See our GitHub [repository](https://github.com/ahnjaewoo/MAC) for full usage instructions and scripts.
|