nielsr HF Staff commited on
Commit
f6d7318
·
verified ·
1 Parent(s): 3808b84

Update paper link to Hugging Face Papers

Browse files

This PR improves the model card by updating the technical report link from arXiv to the corresponding Hugging Face Papers page ([Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct Technical Report](https://huggingface.co/papers/2508.01059)). This enhances the discoverability and consistency of the model documentation directly on the Hugging Face Hub.

Files changed (1) hide show
  1. README.md +57 -58
README.md CHANGED
@@ -23,7 +23,7 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
23
 
24
  - **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
25
  - **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
26
- - **Technical Report:** [`https://arxiv.org/abs/2508.01059`](https://arxiv.org/abs/2508.01059)
27
  - **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
28
  - **Model Release Date:** August 1st, 2025
29
  - **Supported Language(s):** English
@@ -32,7 +32,6 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
32
  - **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
33
  - **License:** See NOTICE.md
34
 
35
-
36
 
37
  ## Intended Use
38
 
@@ -75,22 +74,22 @@ For questions or assistance with fine-tuning Foundation-Sec-8B-Instruct, please
75
 
76
  The following uses are out-of-scope and are neither recommended nor intended use cases:
77
 
78
- 1. **Generating harmful content** - The model should not be used to:
79
- - Generate malware or other malicious code
80
- - Create phishing content or social engineering scripts
81
- - Develop attack plans targeting specific organizations
82
- - Design exploitation techniques for vulnerabilities without legitimate security research purposes
83
- 2. **Critical security decisions without human oversight** - The model should not be used for:
84
- - Autonomous security decision-making without human review
85
- - Critical infrastructure protection without expert supervision
86
- - Final determination of security compliance without human verification
87
- - Autonomous vulnerability remediation without testing
88
- 3. **Legal or medical advice** - The model is not qualified to provide:
89
- - Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
90
- - Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
91
- - Medical advice regarding health impacts of security incidents
92
- 4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
93
- 5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
94
 
95
  ## How to Get Started with the Model
96
 
@@ -187,49 +186,49 @@ It is recommended to deploy this model with additional safeguards (such as Llama
187
 
188
  Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
189
 
190
- 1. **Domain-specific knowledge limitations**:
191
- - Foundation-Sec-8B-Instruct may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
192
- - Knowledge of specialized or proprietary security systems or tools may be limited
193
- 2. **Potential biases**:
194
- - The model may reflect biases present in security literature and documentation
195
- - The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
196
- - Security practices and recommendations may be biased toward certain technological ecosystems
197
- - Geographic and cultural biases in security approaches may be present
198
- 3. **Security risks**:
199
- - The model cannot verify the identity or intentions of users
200
- - Adversarial prompting techniques might potentially bypass safety mechanisms
201
- - The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
202
- 4. **Contextual blindness:**
203
- - The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
204
- 5. **Technical limitations**:
205
- - Performance varies based on how security concepts are described in prompts
206
- - May not fully understand complex, multi-step security scenarios without clear explanation
207
- - Cannot access external systems or actively scan environments
208
- - Cannot independently verify factual accuracy of its outputs
209
- 6. **Ethical considerations**:
210
- - Dual-use nature of security knowledge requires careful consideration of appropriate use cases
211
 
212
 
213
  ### Recommendations
214
 
215
  To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
216
 
217
- 1. **Human oversight**:
218
- - Always have qualified security professionals review model outputs before implementation
219
- - Use the model as an assistive tool rather than a replacement for expert human judgment
220
- - Implement a human-in-the-loop approach for security-critical applications
221
- 2. **System design safeguards**:
222
- - Implement additional validation layers for applications built with this model
223
- - Consider architectural constraints that limit the model's ability to perform potentially harmful actions (excessive agency)
224
- - Deploy the model in environments with appropriate access controls
225
- 3. **Prompt engineering**:
226
- - Use carefully designed prompts that encourage ethical security practices
227
- - Include explicit instructions regarding responsible disclosure and ethical hacking principles
228
- - Structure interactions to minimize the risk of inadvertently harmful outputs
229
- 4. **Knowledge supplementation**:
230
- - Supplement the model with up-to-date security feeds and databases
231
- - Implement retrieval-augmented generation for current threat intelligence sources
232
- 5. **Usage policies**:
233
- - Develop and enforce clear acceptable use policies for applications using this model
234
- - Implement monitoring and auditing for high-risk applications
235
- - Create documentation for end users about the model's limitations
 
23
 
24
  - **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
25
  - **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
26
+ - **Technical Report:** [Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct Technical Report](https://huggingface.co/papers/2508.01059)
27
  - **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
28
  - **Model Release Date:** August 1st, 2025
29
  - **Supported Language(s):** English
 
32
  - **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
33
  - **License:** See NOTICE.md
34
 
 
35
 
36
  ## Intended Use
37
 
 
74
 
75
  The following uses are out-of-scope and are neither recommended nor intended use cases:
76
 
77
+ 1. **Generating harmful content** - The model should not be used to:
78
+ - Generate malware or other malicious code
79
+ - Create phishing content or social engineering scripts
80
+ - Develop attack plans targeting specific organizations
81
+ - Design exploitation techniques for vulnerabilities without legitimate security research purposes
82
+ 2. **Critical security decisions without human oversight** - The model should not be used for:
83
+ - Autonomous security decision-making without human review
84
+ - Critical infrastructure protection without expert supervision
85
+ - Final determination of security compliance without human verification
86
+ - Autonomous vulnerability remediation without testing
87
+ 3. **Legal or medical advice** - The model is not qualified to provide:
88
+ - Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
89
+ - Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
90
+ - Medical advice regarding health impacts of security incidents
91
+ 4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
92
+ 5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
93
 
94
  ## How to Get Started with the Model
95
 
 
186
 
187
  Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
188
 
189
+ 1. **Domain-specific knowledge limitations**:
190
+ - Foundation-Sec-8B-Instruct may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
191
+ - Knowledge of specialized or proprietary security systems or tools may be limited
192
+ 2. **Potential biases**:
193
+ - The model may reflect biases present in security literature and documentation
194
+ - The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
195
+ - Security practices and recommendations may be biased toward certain technological ecosystems
196
+ - Geographic and cultural biases in security approaches may be present
197
+ 3. **Security risks**:
198
+ - The model cannot verify the identity or intentions of users
199
+ - Adversarial prompting techniques might potentially bypass safety mechanisms
200
+ - The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
201
+ 4. **Contextual blindness:**
202
+ - The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
203
+ 5. **Technical limitations**:
204
+ - Performance varies based on how security concepts are described in prompts
205
+ - May not fully understand complex, multi-step security scenarios without clear explanation
206
+ - Cannot access external systems or actively scan environments
207
+ - Cannot independently verify factual accuracy of its outputs
208
+ 6. **Ethical considerations**:
209
+ - Dual-use nature of security knowledge requires careful consideration of appropriate use cases
210
 
211
 
212
  ### Recommendations
213
 
214
  To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
215
 
216
+ 1. **Human oversight**:
217
+ - Always have qualified security professionals review model outputs before implementation
218
+ - Use the model as an assistive tool rather than a replacement for expert human judgment
219
+ - Implement a human-in-the-loop approach for security-critical applications
220
+ 2. **System design safeguards**:
221
+ - Implement additional validation layers for applications built with this model
222
+ - Consider architectural constraints that limit the model's ability to perform potentially harmful actions (excessive agency)
223
+ - Deploy the model in environments with appropriate access controls
224
+ 3. **Prompt engineering**:
225
+ - Use carefully designed prompts that encourage ethical security practices
226
+ - Include explicit instructions regarding responsible disclosure and ethical hacking principles
227
+ - Structure interactions to minimize the risk of inadvertently harmful outputs
228
+ 4. **Knowledge supplementation**:
229
+ - Supplement the model with up-to-date security feeds and databases
230
+ - Implement retrieval-augmented generation for current threat intelligence sources
231
+ 5. **Usage policies**:
232
+ - Develop and enforce clear acceptable use policies for applications using this model
233
+ - Implement monitoring and auditing for high-risk applications
234
+ - Create documentation for end users about the model's limitations