You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Primus: A Pioneering Collection of Open-Source Datasets for Cybersecurity LLM Training

Primus Overview

TL;DR: Llama-Primus-Base is a foundation model based on Llama-3.1-8B-Instruct, continually pre-trained on Primus-Seed (0.2B) and Primus-FineWeb (2.57B). Primus-Seed is a high-quality, manually curated cybersecurity text dataset, while Primus-FineWeb consists of cybersecurity texts filtered from FineWeb, a refined version of Common Crawl. By pretraining on such a large-scale cybersecurity corpus, it achieves a 🚀15.88% improvement in aggregated scores across multiple cybersecurity benchmarks, demonstrating the effectiveness of cybersecurity-specific pretraining.

🔥 For more details, please refer to the paper: [📄Paper].

Introduction

Large Language Models (LLMs) have demonstrated remarkable versatility in recent years, with promising applications in specialized domains such as finance, law, and biomedicine. However, in the domain of cybersecurity, we noticed a lack of open-source datasets specifically designed for LLM pre-training—even though much research has shown that LLMs acquire their knowledge during pre-training. To fill this gap, we present a collection of datasets covering multiple stages of cybersecurity LLM training, including pre-training (Primus-Seed and Primus-FineWeb), instruction fine-tuning (Primus-Instruct), and reasoning data for distillation (Primus-Reasoning). Based on these datasets and Llama-3.1-8B-Instruct, we developed Llama-Primus-Base, Llama-Primus-Merged, and Llama-Primus-Reasoning. This model card is Llama-Primus-Base.

Note: No TrendMicro customer information is included.

Cybersecurity Benchmark Results

Metric (5-shot, w/o CoT) Llama-3.1-8B-Instruct Llama-Primus-Base
CISSP (Exams in book) 0.7073 0.7230
CTI-Bench (MCQ) 0.6420 0.6676
CTI-Bench (CVE → CWE) 0.5910 0.6780
CTI-Bench (CVSS, lower is better) 1.2712 1.0912
CTI-Bench (ATE) 0.2721 0.3140
CyberMetric (500) 0.8560 0.8660
SecEval 0.4966 0.5007
Agg. 2.29 2.66 ↑15.88% 🔥

CTI-Bench (CVSS) is scored using Mean Absolute Deviation (lower is better), CTI-ATE uses F1 score, and the others use accuracy. The aggregate score (Agg.) is the sum of all benchmarks, with CTI-Bench (CVSS) negated.

References:

License

This model is based on the MIT license, but you must also comply with the Llama 3.1 Community License Agreement.

Downloads last month
8
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for trendmicro-ailab/Llama-Primus-Base

Finetuned
(879)
this model
Finetunes
1 model

Dataset used to train trendmicro-ailab/Llama-Primus-Base

Collection including trendmicro-ailab/Llama-Primus-Base