montehoover commited on
Commit
026e85d
·
verified ·
1 Parent(s): 439e2e1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -3
README.md CHANGED
@@ -15,6 +15,9 @@ datasets:
15
  - tomg-group-umd/DynaBench
16
  base_model:
17
  - Qwen/Qwen3-1.7B
 
 
 
18
  ---
19
 
20
  # DynaGuard-1.7B 🛡️
@@ -22,9 +25,11 @@ base_model:
22
  **The DynaGuard model series** is a family of guardian models designed to evaluate text against user-defined, natural language policies. They provide a flexible and powerful solution for moderating chatbot outputs beyond static, predefined harm categories. Developed by researchers at the University of Maryland and Capital One , the series includes three open-weight models of varying sizes:
23
  1.7B, 4B, and 8B — allowing developers to choose the best balance of performance and efficiency for their needs.
24
  Unlike traditional guardian models that screen for a fixed set of harms (e.g., violence or self-harm) , DynaGuard can enforce bespoke, application-specific rules. This includes scenarios like preventing a customer service bot from mistakenly issuing refunds or ensuring a medical bot avoids giving unauthorized advice.
25
- The DynaGuard series achieves state-of-the-art performance across a wide range of safety and compliance benchmarks, with the flagship **[DynaGuard-8B](https://huggingface.co/tomg-group-umd/DynaGuard-8B)** model outperforming other guardian models and even strong generalist models like GPT-4o-mini.
26
 
27
- [https://arxiv.org/abs/2509.02563](https://arxiv.org/abs/2509.02563)
 
 
28
 
29
 
30
  ## Model Details
@@ -124,6 +129,10 @@ DynaGuard achieves state-of-the-art performance, outperforming other dedicated g
124
  If you use DynaGuard or the DynaBench dataset in your research, please cite our work:
125
  ```
126
  @article{hoover2025dynaguard,
127
- title={DynaGuard: A Dynamic Guardrail Model With User-Defined Policies},
 
 
 
 
128
  }
129
  ```
 
15
  - tomg-group-umd/DynaBench
16
  base_model:
17
  - Qwen/Qwen3-1.7B
18
+ repo_url: https://github.com/montehoover/DynaGuard
19
+ paper_url: https://arxiv.org/abs/2509.02563
20
+ project_page: https://github.com/taruschirag/DynaGuard
21
  ---
22
 
23
  # DynaGuard-1.7B 🛡️
 
25
  **The DynaGuard model series** is a family of guardian models designed to evaluate text against user-defined, natural language policies. They provide a flexible and powerful solution for moderating chatbot outputs beyond static, predefined harm categories. Developed by researchers at the University of Maryland and Capital One , the series includes three open-weight models of varying sizes:
26
  1.7B, 4B, and 8B — allowing developers to choose the best balance of performance and efficiency for their needs.
27
  Unlike traditional guardian models that screen for a fixed set of harms (e.g., violence or self-harm) , DynaGuard can enforce bespoke, application-specific rules. This includes scenarios like preventing a customer service bot from mistakenly issuing refunds or ensuring a medical bot avoids giving unauthorized advice.
28
+ The DynaGuard series achieves state-of-the-art performance across a wide range of safety and compliance benchmarks, with the flagship **DynaGuard-8B** model outperforming other guardian models and even strong generalist models like GPT-4o-mini.
29
 
30
+ | 🔖 | 💻 | 🌐 |
31
+ |----|----|---|
32
+ | [Paper (arXiv)](https://arxiv.org/abs/2509.02563) | [Code (GitHub)](https://github.com/montehoover/DynaGuard) | [Project page ](https://github.com/taruschirag/DynaGuard) |
33
 
34
 
35
  ## Model Details
 
129
  If you use DynaGuard or the DynaBench dataset in your research, please cite our work:
130
  ```
131
  @article{hoover2025dynaguard,
132
+ title={DynaGuard: A Dynamic Guardrail Model With User-Defined Policies},
133
+ author={Monte Hoover and Vatsal Baherwani and Neel Jain and Khalid Saifullah and Joseph Vincent and Chirag Jain and Melissa Kazemi Rad and C. Bayan Bruss and Ashwinee Panda and Tom Goldstein},
134
+ journal={arXiv preprint},
135
+ year={2025},
136
+ url={https://arxiv.org/abs/2509.02563},
137
  }
138
  ```