Update paper link to Hugging Face Papers
Browse filesThis PR improves the model card by updating the technical report link from arXiv to the corresponding Hugging Face Papers page ([Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct Technical Report](https://huggingface.co/papers/2508.01059)). This enhances the discoverability and consistency of the model documentation directly on the Hugging Face Hub.
README.md
CHANGED
@@ -23,7 +23,7 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
|
|
23 |
|
24 |
- **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
|
25 |
- **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
|
26 |
-
- **Technical Report:** [
|
27 |
- **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
|
28 |
- **Model Release Date:** August 1st, 2025
|
29 |
- **Supported Language(s):** English
|
@@ -32,7 +32,6 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
|
|
32 |
- **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
|
33 |
- **License:** See NOTICE.md
|
34 |
|
35 |
-
|
36 |
|
37 |
## Intended Use
|
38 |
|
@@ -75,22 +74,22 @@ For questions or assistance with fine-tuning Foundation-Sec-8B-Instruct, please
|
|
75 |
|
76 |
The following uses are out-of-scope and are neither recommended nor intended use cases:
|
77 |
|
78 |
-
1.
|
79 |
-
-
|
80 |
-
-
|
81 |
-
-
|
82 |
-
-
|
83 |
-
2.
|
84 |
-
-
|
85 |
-
-
|
86 |
-
-
|
87 |
-
-
|
88 |
-
3.
|
89 |
-
-
|
90 |
-
-
|
91 |
-
-
|
92 |
-
4.
|
93 |
-
5.
|
94 |
|
95 |
## How to Get Started with the Model
|
96 |
|
@@ -187,49 +186,49 @@ It is recommended to deploy this model with additional safeguards (such as Llama
|
|
187 |
|
188 |
Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
|
189 |
|
190 |
-
1.
|
191 |
-
-
|
192 |
-
-
|
193 |
-
2.
|
194 |
-
-
|
195 |
-
-
|
196 |
-
-
|
197 |
-
-
|
198 |
-
3.
|
199 |
-
-
|
200 |
-
-
|
201 |
-
-
|
202 |
-
4.
|
203 |
-
-
|
204 |
-
5.
|
205 |
-
-
|
206 |
-
-
|
207 |
-
-
|
208 |
-
-
|
209 |
-
6.
|
210 |
-
-
|
211 |
|
212 |
|
213 |
### Recommendations
|
214 |
|
215 |
To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
|
216 |
|
217 |
-
1.
|
218 |
-
-
|
219 |
-
-
|
220 |
-
-
|
221 |
-
2.
|
222 |
-
-
|
223 |
-
-
|
224 |
-
-
|
225 |
-
3.
|
226 |
-
-
|
227 |
-
-
|
228 |
-
-
|
229 |
-
4.
|
230 |
-
-
|
231 |
-
-
|
232 |
-
5.
|
233 |
-
-
|
234 |
-
-
|
235 |
-
-
|
|
|
23 |
|
24 |
- **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
|
25 |
- **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
|
26 |
+
- **Technical Report:** [Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct Technical Report](https://huggingface.co/papers/2508.01059)
|
27 |
- **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
|
28 |
- **Model Release Date:** August 1st, 2025
|
29 |
- **Supported Language(s):** English
|
|
|
32 |
- **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
|
33 |
- **License:** See NOTICE.md
|
34 |
|
|
|
35 |
|
36 |
## Intended Use
|
37 |
|
|
|
74 |
|
75 |
The following uses are out-of-scope and are neither recommended nor intended use cases:
|
76 |
|
77 |
+
1. **Generating harmful content** - The model should not be used to:
|
78 |
+
- Generate malware or other malicious code
|
79 |
+
- Create phishing content or social engineering scripts
|
80 |
+
- Develop attack plans targeting specific organizations
|
81 |
+
- Design exploitation techniques for vulnerabilities without legitimate security research purposes
|
82 |
+
2. **Critical security decisions without human oversight** - The model should not be used for:
|
83 |
+
- Autonomous security decision-making without human review
|
84 |
+
- Critical infrastructure protection without expert supervision
|
85 |
+
- Final determination of security compliance without human verification
|
86 |
+
- Autonomous vulnerability remediation without testing
|
87 |
+
3. **Legal or medical advice** - The model is not qualified to provide:
|
88 |
+
- Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
|
89 |
+
- Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
|
90 |
+
- Medical advice regarding health impacts of security incidents
|
91 |
+
4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
|
92 |
+
5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
|
93 |
|
94 |
## How to Get Started with the Model
|
95 |
|
|
|
186 |
|
187 |
Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
|
188 |
|
189 |
+
1. **Domain-specific knowledge limitations**:
|
190 |
+
- Foundation-Sec-8B-Instruct may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
|
191 |
+
- Knowledge of specialized or proprietary security systems or tools may be limited
|
192 |
+
2. **Potential biases**:
|
193 |
+
- The model may reflect biases present in security literature and documentation
|
194 |
+
- The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
|
195 |
+
- Security practices and recommendations may be biased toward certain technological ecosystems
|
196 |
+
- Geographic and cultural biases in security approaches may be present
|
197 |
+
3. **Security risks**:
|
198 |
+
- The model cannot verify the identity or intentions of users
|
199 |
+
- Adversarial prompting techniques might potentially bypass safety mechanisms
|
200 |
+
- The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
|
201 |
+
4. **Contextual blindness:**
|
202 |
+
- The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
|
203 |
+
5. **Technical limitations**:
|
204 |
+
- Performance varies based on how security concepts are described in prompts
|
205 |
+
- May not fully understand complex, multi-step security scenarios without clear explanation
|
206 |
+
- Cannot access external systems or actively scan environments
|
207 |
+
- Cannot independently verify factual accuracy of its outputs
|
208 |
+
6. **Ethical considerations**:
|
209 |
+
- Dual-use nature of security knowledge requires careful consideration of appropriate use cases
|
210 |
|
211 |
|
212 |
### Recommendations
|
213 |
|
214 |
To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
|
215 |
|
216 |
+
1. **Human oversight**:
|
217 |
+
- Always have qualified security professionals review model outputs before implementation
|
218 |
+
- Use the model as an assistive tool rather than a replacement for expert human judgment
|
219 |
+
- Implement a human-in-the-loop approach for security-critical applications
|
220 |
+
2. **System design safeguards**:
|
221 |
+
- Implement additional validation layers for applications built with this model
|
222 |
+
- Consider architectural constraints that limit the model's ability to perform potentially harmful actions (excessive agency)
|
223 |
+
- Deploy the model in environments with appropriate access controls
|
224 |
+
3. **Prompt engineering**:
|
225 |
+
- Use carefully designed prompts that encourage ethical security practices
|
226 |
+
- Include explicit instructions regarding responsible disclosure and ethical hacking principles
|
227 |
+
- Structure interactions to minimize the risk of inadvertently harmful outputs
|
228 |
+
4. **Knowledge supplementation**:
|
229 |
+
- Supplement the model with up-to-date security feeds and databases
|
230 |
+
- Implement retrieval-augmented generation for current threat intelligence sources
|
231 |
+
5. **Usage policies**:
|
232 |
+
- Develop and enforce clear acceptable use policies for applications using this model
|
233 |
+
- Implement monitoring and auditing for high-risk applications
|
234 |
+
- Create documentation for end users about the model's limitations
|