Update README.md
Browse files
README.md
CHANGED
@@ -15,6 +15,9 @@ datasets:
|
|
15 |
- tomg-group-umd/DynaBench
|
16 |
base_model:
|
17 |
- Qwen/Qwen3-1.7B
|
|
|
|
|
|
|
18 |
---
|
19 |
|
20 |
# DynaGuard-1.7B 🛡️
|
@@ -22,9 +25,11 @@ base_model:
|
|
22 |
**The DynaGuard model series** is a family of guardian models designed to evaluate text against user-defined, natural language policies. They provide a flexible and powerful solution for moderating chatbot outputs beyond static, predefined harm categories. Developed by researchers at the University of Maryland and Capital One , the series includes three open-weight models of varying sizes:
|
23 |
1.7B, 4B, and 8B — allowing developers to choose the best balance of performance and efficiency for their needs.
|
24 |
Unlike traditional guardian models that screen for a fixed set of harms (e.g., violence or self-harm) , DynaGuard can enforce bespoke, application-specific rules. This includes scenarios like preventing a customer service bot from mistakenly issuing refunds or ensuring a medical bot avoids giving unauthorized advice.
|
25 |
-
The DynaGuard series achieves state-of-the-art performance across a wide range of safety and compliance benchmarks, with the flagship **
|
26 |
|
27 |
-
|
|
|
|
|
28 |
|
29 |
|
30 |
## Model Details
|
@@ -124,6 +129,10 @@ DynaGuard achieves state-of-the-art performance, outperforming other dedicated g
|
|
124 |
If you use DynaGuard or the DynaBench dataset in your research, please cite our work:
|
125 |
```
|
126 |
@article{hoover2025dynaguard,
|
127 |
-
|
|
|
|
|
|
|
|
|
128 |
}
|
129 |
```
|
|
|
15 |
- tomg-group-umd/DynaBench
|
16 |
base_model:
|
17 |
- Qwen/Qwen3-1.7B
|
18 |
+
repo_url: https://github.com/montehoover/DynaGuard
|
19 |
+
paper_url: https://arxiv.org/abs/2509.02563
|
20 |
+
project_page: https://github.com/taruschirag/DynaGuard
|
21 |
---
|
22 |
|
23 |
# DynaGuard-1.7B 🛡️
|
|
|
25 |
**The DynaGuard model series** is a family of guardian models designed to evaluate text against user-defined, natural language policies. They provide a flexible and powerful solution for moderating chatbot outputs beyond static, predefined harm categories. Developed by researchers at the University of Maryland and Capital One , the series includes three open-weight models of varying sizes:
|
26 |
1.7B, 4B, and 8B — allowing developers to choose the best balance of performance and efficiency for their needs.
|
27 |
Unlike traditional guardian models that screen for a fixed set of harms (e.g., violence or self-harm) , DynaGuard can enforce bespoke, application-specific rules. This includes scenarios like preventing a customer service bot from mistakenly issuing refunds or ensuring a medical bot avoids giving unauthorized advice.
|
28 |
+
The DynaGuard series achieves state-of-the-art performance across a wide range of safety and compliance benchmarks, with the flagship **DynaGuard-8B** model outperforming other guardian models and even strong generalist models like GPT-4o-mini.
|
29 |
|
30 |
+
| 🔖 | 💻 | 🌐 |
|
31 |
+
|----|----|---|
|
32 |
+
| [Paper (arXiv)](https://arxiv.org/abs/2509.02563) | [Code (GitHub)](https://github.com/montehoover/DynaGuard) | [Project page ](https://github.com/taruschirag/DynaGuard) |
|
33 |
|
34 |
|
35 |
## Model Details
|
|
|
129 |
If you use DynaGuard or the DynaBench dataset in your research, please cite our work:
|
130 |
```
|
131 |
@article{hoover2025dynaguard,
|
132 |
+
title={DynaGuard: A Dynamic Guardrail Model With User-Defined Policies},
|
133 |
+
author={Monte Hoover and Vatsal Baherwani and Neel Jain and Khalid Saifullah and Joseph Vincent and Chirag Jain and Melissa Kazemi Rad and C. Bayan Bruss and Ashwinee Panda and Tom Goldstein},
|
134 |
+
journal={arXiv preprint},
|
135 |
+
year={2025},
|
136 |
+
url={https://arxiv.org/abs/2509.02563},
|
137 |
}
|
138 |
```
|