Update README.md
Browse files
README.md
CHANGED
@@ -1,26 +1,67 @@
|
|
1 |
-
---
|
2 |
-
license: mit
|
3 |
-
language:
|
4 |
-
- en
|
5 |
-
---
|
6 |
|
7 |
-
#
|
8 |
|
9 |
-
|
|
|
10 |
|
11 |
-
|
12 |
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
-
This model is
|
16 |
|
17 |
- **Developed by:** Canstralian
|
18 |
-
- **
|
19 |
-
- **
|
20 |
-
- **
|
21 |
-
|
22 |
-
|
23 |
-
- **
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
|
25 |
### Model Sources [optional]
|
26 |
|
@@ -32,35 +73,46 @@ This model is designed for detecting cyber attacks, focusing on identifying patt
|
|
32 |
|
33 |
### Direct Use
|
34 |
|
35 |
-
This model can be used directly for detecting cyber attacks by analyzing network traffic or system logs. It
|
36 |
|
37 |
### Downstream Use [optional]
|
38 |
|
39 |
-
The model can be fine-tuned further for specific types of cyber attacks or to suit different environments (e.g., enterprise networks, small businesses).
|
40 |
|
41 |
### Out-of-Scope Use
|
42 |
|
43 |
-
The model is not intended for detecting non-cyber attacks or for use outside cybersecurity applications. It may not perform well with highly specialized or obscure types of attacks.
|
44 |
|
45 |
## Bias, Risks, and Limitations
|
46 |
|
47 |
-
The model
|
48 |
|
49 |
### Recommendations
|
50 |
|
51 |
-
Users should
|
52 |
|
53 |
## How to Get Started with the Model
|
54 |
|
55 |
-
|
56 |
|
57 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
58 |
|
59 |
## Training Details
|
60 |
|
61 |
### Training Data
|
62 |
|
63 |
-
The model was trained
|
|
|
|
|
|
|
|
|
64 |
|
65 |
### Training Procedure
|
66 |
|
@@ -82,11 +134,13 @@ The model was trained on publicly available datasets for cybersecurity attacks,
|
|
82 |
|
83 |
#### Testing Data
|
84 |
|
85 |
-
|
|
|
|
|
86 |
|
87 |
#### Factors
|
88 |
|
89 |
-
The evaluation
|
90 |
|
91 |
#### Metrics
|
92 |
|
@@ -101,7 +155,7 @@ The evaluation was done based on the detection of known attack types, network tr
|
|
101 |
|
102 |
#### Summary
|
103 |
|
104 |
-
The model
|
105 |
|
106 |
## Model Examination [optional]
|
107 |
|
@@ -119,7 +173,7 @@ The model has demonstrated reasonable success in detecting a variety of cyber at
|
|
119 |
|
120 |
### Model Architecture and Objective
|
121 |
|
122 |
-
The model
|
123 |
|
124 |
### Compute Infrastructure
|
125 |
|
@@ -155,4 +209,4 @@ The model utilizes deep learning algorithms, such as neural networks, to analyze
|
|
155 |
|
156 |
## Model Card Contact
|
157 |
|
158 |
-
[More Information Needed]
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
---
|
6 |
|
7 |
+
# **Canstralian/CyberAttackDetection - AI Model Overview**
|
8 |
|
9 |
+
## **Model Description**
|
10 |
+
**CyberAttackDetection** is a cutting-edge machine learning model designed to detect and classify a wide range of cyberattacks in real-time. Built using advanced algorithms and a comprehensive dataset of known attack signatures, the model can effectively identify abnormal behaviors, intrusion attempts, and potential threats in network traffic and system logs.
|
11 |
|
12 |
+
The model is optimized for high accuracy and low latency, making it ideal for use in real-time network monitoring, incident response, and security operations centers. By leveraging **WhiteRabbitNeo** (based on Llama-3.1), it offers high adaptability to new attack vectors and ensures robust protection against both common and sophisticated threats.
|
13 |
|
14 |
+
**Key Features:**
|
15 |
+
- Real-time detection and classification of cyberattacks
|
16 |
+
- Identification of vulnerabilities and exploits, including zero-day attacks
|
17 |
+
- Adaptive learning capabilities to recognize new threats
|
18 |
+
- High accuracy and low false-positive rates
|
19 |
+
- Scalable for deployment in diverse environments, from small businesses to large enterprises
|
20 |
|
21 |
+
This model is tailored for penetration testers, cybersecurity professionals, and organizations looking to enhance their security posture with AI-powered attack detection.
|
22 |
|
23 |
- **Developed by:** Canstralian
|
24 |
+
- **Model type:** Cyberattack Detection
|
25 |
+
- **License:** MIT
|
26 |
+
- **Finetuned from model:** [WhiteRabbitNeo/Llama-3.1-WhiteRabbitNeo-2-70B](https://huggingface.co/WhiteRabbitNeo/Llama-3.1-WhiteRabbitNeo-2-70B)
|
27 |
+
|
28 |
+
## **WhiteRabbitNeo License + Usage Restrictions**
|
29 |
+
The **CyberAttackDetection** model is built using **WhiteRabbitNeo**, and it adheres to the Llama-3.1 License, with an extended version specific to **WhiteRabbitNeo**. By using this model, you agree to the following usage restrictions:
|
30 |
+
|
31 |
+
You may not use the model or its derivatives in any way that:
|
32 |
+
- Violates any applicable national or international law or infringes upon third-party rights.
|
33 |
+
- Is intended for military use or harm to minors.
|
34 |
+
- Generates false information or disseminates inappropriate content.
|
35 |
+
- Exploits or harms individuals based on protected characteristics.
|
36 |
+
- Discriminates against individuals or groups based on personal characteristics or legal protections.
|
37 |
+
|
38 |
+
For further details on the licensing and restrictions, refer to the [WhiteRabbitNeo License Agreement](https://www.whiterabbitneo.com/license).
|
39 |
+
|
40 |
+
## **Topics Covered in Cyberattack Detection**
|
41 |
+
The **CyberAttackDetection** model helps identify vulnerabilities that attackers commonly exploit, including but not limited to:
|
42 |
+
|
43 |
+
- **Open Ports:** Identifying entry points like HTTP (80, 443), FTP (21), SSH (22), and SMB (445).
|
44 |
+
- **Outdated Software:** Vulnerabilities arising from outdated systems and third-party services.
|
45 |
+
- **Default Credentials:** Risks posed by common factory-installed usernames and passwords.
|
46 |
+
- **Misconfigurations:** Insecure service configurations that can open up attack vectors.
|
47 |
+
- **Injection Flaws:** Common web vulnerabilities like SQL injection, XSS, and command injections.
|
48 |
+
- **Unencrypted Services:** Identifying services without encryption (e.g., HTTP vs HTTPS).
|
49 |
+
- **Known Software Vulnerabilities:** Checking for outdated software vulnerabilities using resources like the NVD or tools like Nessus and OpenVAS.
|
50 |
+
- **Cross-Site Request Forgery (CSRF):** Unauthorized command transmission in web apps.
|
51 |
+
- **API Vulnerabilities:** Detecting insecure API endpoints and data leakage.
|
52 |
+
- **Denial of Service (DoS):** Identifying DoS vulnerabilities that impact system availability.
|
53 |
+
- **Sensitive Data Exposure:** Identifying vulnerabilities that expose personal or financial data.
|
54 |
+
|
55 |
+
## **Terms of Use**
|
56 |
+
By accessing and using this AI model, you acknowledge that you are solely responsible for its usage and the outcomes that result. You agree to indemnify, defend, and hold harmless the creators and any affiliated entities from any liabilities, damages, or losses incurred as a result of using the model.
|
57 |
+
|
58 |
+
This AI model is provided "as is" and "as available" without any warranties, express or implied. The creators make no guarantee that the model will meet your requirements or be available without interruption, security breaches, or errors.
|
59 |
+
|
60 |
+
**Disclaimer:** Use this model at your own risk. The creators will not be liable for any damages, including loss of data or system failures, resulting from the use of this model.
|
61 |
+
|
62 |
+
---
|
63 |
+
|
64 |
+
Let me know if you need any more modifications!
|
65 |
|
66 |
### Model Sources [optional]
|
67 |
|
|
|
73 |
|
74 |
### Direct Use
|
75 |
|
76 |
+
This model can be used directly for detecting cyber attacks by analyzing network traffic or system logs. It is especially useful for network administrators and cybersecurity experts who need real-time or historical analysis of potentially malicious activities.
|
77 |
|
78 |
### Downstream Use [optional]
|
79 |
|
80 |
+
The model can be fine-tuned further for specific types of cyber attacks or to suit different environments (e.g., enterprise networks, small businesses). It can also be integrated into larger security ecosystems that perform continuous monitoring and threat analysis.
|
81 |
|
82 |
### Out-of-Scope Use
|
83 |
|
84 |
+
The model is not intended for detecting non-cyber attacks or for use outside cybersecurity applications. It may not perform well with highly specialized or obscure types of attacks that are not well-represented in the training data.
|
85 |
|
86 |
## Bias, Risks, and Limitations
|
87 |
|
88 |
+
The model’s performance is influenced by the quality and diversity of the training data. Misclassifications may occur, particularly when encountering novel attack patterns or environments not well-represented in the dataset. Furthermore, the model may generate false positives or miss complex attack vectors.
|
89 |
|
90 |
### Recommendations
|
91 |
|
92 |
+
Users should regularly update the model with new data and threat intelligence to keep it relevant. The model should be used in conjunction with human oversight and other detection mechanisms to minimize the risk of undetected threats.
|
93 |
|
94 |
## How to Get Started with the Model
|
95 |
|
96 |
+
To get started with the model, use the following code:
|
97 |
|
98 |
+
```python
|
99 |
+
from transformers import pipeline
|
100 |
+
|
101 |
+
model = pipeline("cyber_attack_detection", model="Canstralian/CyberAttackDetection")
|
102 |
+
# Example usage: Pass network traffic or system log data to the model
|
103 |
+
result = model("Example log data or network traffic")
|
104 |
+
print(result)
|
105 |
+
```
|
106 |
|
107 |
## Training Details
|
108 |
|
109 |
### Training Data
|
110 |
|
111 |
+
The model was trained using a combination of datasets related to penetration testing, shell commands, and wordlists, which are essential for recognizing attack vectors and behaviors in real-world environments.
|
112 |
+
|
113 |
+
- **Pentesting Dataset**: [Canstralian/pentesting_dataset](https://huggingface.co/datasets/Canstralian/pentesting_dataset)
|
114 |
+
- **Shell Commands Dataset**: [Canstralian/ShellCommands](https://huggingface.co/datasets/Canstralian/ShellCommands)
|
115 |
+
- **Wordlists Dataset**: [Canstralian/Wordlists](https://huggingface.co/datasets/Canstralian/Wordlists)
|
116 |
|
117 |
### Training Procedure
|
118 |
|
|
|
134 |
|
135 |
#### Testing Data
|
136 |
|
137 |
+
- **Pentesting Dataset**: Used for testing the model’s ability to detect attack behaviors.
|
138 |
+
- **Shell Commands Dataset**: Assessed the model’s effectiveness in recognizing shell-related attack commands.
|
139 |
+
- **Wordlists Dataset**: Evaluated the model’s proficiency in detecting dictionary-based attacks.
|
140 |
|
141 |
#### Factors
|
142 |
|
143 |
+
The evaluation tests for the model’s ability to detect common attack vectors, unusual patterns, and malicious behaviors across different datasets.
|
144 |
|
145 |
#### Metrics
|
146 |
|
|
|
155 |
|
156 |
#### Summary
|
157 |
|
158 |
+
The model performs well at detecting common types of cyber attacks but is subject to limitations in environments where the attack types differ significantly from those seen in the training datasets.
|
159 |
|
160 |
## Model Examination [optional]
|
161 |
|
|
|
173 |
|
174 |
### Model Architecture and Objective
|
175 |
|
176 |
+
The model uses deep learning techniques to classify and identify malicious patterns in system logs and network traffic.
|
177 |
|
178 |
### Compute Infrastructure
|
179 |
|
|
|
209 |
|
210 |
## Model Card Contact
|
211 |
|
212 |
+
[More Information Needed]
|