Update paper link to Hugging Face Papers

#3
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +57 -58
README.md CHANGED
@@ -23,7 +23,7 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
23
 
24
  - **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
25
  - **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
26
- - **Technical Report:** [`https://arxiv.org/abs/2508.01059`](https://arxiv.org/abs/2508.01059)
27
  - **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
28
  - **Model Release Date:** August 1st, 2025
29
  - **Supported Language(s):** English
@@ -32,7 +32,6 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
32
  - **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
33
  - **License:** See NOTICE.md
34
 
35
-
36
 
37
  ## Intended Use
38
 
@@ -75,22 +74,22 @@ For questions or assistance with fine-tuning Foundation-Sec-8B-Instruct, please
75
 
76
  The following uses are out-of-scope and are neither recommended nor intended use cases:
77
 
78
- 1. **Generating harmful content** - The model should not be used to:
79
- - Generate malware or other malicious code
80
- - Create phishing content or social engineering scripts
81
- - Develop attack plans targeting specific organizations
82
- - Design exploitation techniques for vulnerabilities without legitimate security research purposes
83
- 2. **Critical security decisions without human oversight** - The model should not be used for:
84
- - Autonomous security decision-making without human review
85
- - Critical infrastructure protection without expert supervision
86
- - Final determination of security compliance without human verification
87
- - Autonomous vulnerability remediation without testing
88
- 3. **Legal or medical advice** - The model is not qualified to provide:
89
- - Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
90
- - Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
91
- - Medical advice regarding health impacts of security incidents
92
- 4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
93
- 5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
94
 
95
  ## How to Get Started with the Model
96
 
@@ -187,49 +186,49 @@ It is recommended to deploy this model with additional safeguards (such as Llama
187
 
188
  Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
189
 
190
- 1. **Domain-specific knowledge limitations**:
191
- - Foundation-Sec-8B-Instruct may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
192
- - Knowledge of specialized or proprietary security systems or tools may be limited
193
- 2. **Potential biases**:
194
- - The model may reflect biases present in security literature and documentation
195
- - The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
196
- - Security practices and recommendations may be biased toward certain technological ecosystems
197
- - Geographic and cultural biases in security approaches may be present
198
- 3. **Security risks**:
199
- - The model cannot verify the identity or intentions of users
200
- - Adversarial prompting techniques might potentially bypass safety mechanisms
201
- - The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
202
- 4. **Contextual blindness:**
203
- - The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
204
- 5. **Technical limitations**:
205
- - Performance varies based on how security concepts are described in prompts
206
- - May not fully understand complex, multi-step security scenarios without clear explanation
207
- - Cannot access external systems or actively scan environments
208
- - Cannot independently verify factual accuracy of its outputs
209
- 6. **Ethical considerations**:
210
- - Dual-use nature of security knowledge requires careful consideration of appropriate use cases
211
 
212
 
213
  ### Recommendations
214
 
215
  To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
216
 
217
- 1. **Human oversight**:
218
- - Always have qualified security professionals review model outputs before implementation
219
- - Use the model as an assistive tool rather than a replacement for expert human judgment
220
- - Implement a human-in-the-loop approach for security-critical applications
221
- 2. **System design safeguards**:
222
- - Implement additional validation layers for applications built with this model
223
- - Consider architectural constraints that limit the model's ability to perform potentially harmful actions (excessive agency)
224
- - Deploy the model in environments with appropriate access controls
225
- 3. **Prompt engineering**:
226
- - Use carefully designed prompts that encourage ethical security practices
227
- - Include explicit instructions regarding responsible disclosure and ethical hacking principles
228
- - Structure interactions to minimize the risk of inadvertently harmful outputs
229
- 4. **Knowledge supplementation**:
230
- - Supplement the model with up-to-date security feeds and databases
231
- - Implement retrieval-augmented generation for current threat intelligence sources
232
- 5. **Usage policies**:
233
- - Develop and enforce clear acceptable use policies for applications using this model
234
- - Implement monitoring and auditing for high-risk applications
235
- - Create documentation for end users about the model's limitations
 
23
 
24
  - **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
25
  - **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
26
+ - **Technical Report:** [Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct Technical Report](https://huggingface.co/papers/2508.01059)
27
  - **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
28
  - **Model Release Date:** August 1st, 2025
29
  - **Supported Language(s):** English
 
32
  - **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
33
  - **License:** See NOTICE.md
34
 
 
35
 
36
  ## Intended Use
37
 
 
74
 
75
  The following uses are out-of-scope and are neither recommended nor intended use cases:
76
 
77
+ 1. **Generating harmful content** - The model should not be used to:
78
+ - Generate malware or other malicious code
79
+ - Create phishing content or social engineering scripts
80
+ - Develop attack plans targeting specific organizations
81
+ - Design exploitation techniques for vulnerabilities without legitimate security research purposes
82
+ 2. **Critical security decisions without human oversight** - The model should not be used for:
83
+ - Autonomous security decision-making without human review
84
+ - Critical infrastructure protection without expert supervision
85
+ - Final determination of security compliance without human verification
86
+ - Autonomous vulnerability remediation without testing
87
+ 3. **Legal or medical advice** - The model is not qualified to provide:
88
+ - Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
89
+ - Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
90
+ - Medical advice regarding health impacts of security incidents
91
+ 4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
92
+ 5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
93
 
94
  ## How to Get Started with the Model
95
 
 
186
 
187
  Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
188
 
189
+ 1. **Domain-specific knowledge limitations**:
190
+ - Foundation-Sec-8B-Instruct may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
191
+ - Knowledge of specialized or proprietary security systems or tools may be limited
192
+ 2. **Potential biases**:
193
+ - The model may reflect biases present in security literature and documentation
194
+ - The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
195
+ - Security practices and recommendations may be biased toward certain technological ecosystems
196
+ - Geographic and cultural biases in security approaches may be present
197
+ 3. **Security risks**:
198
+ - The model cannot verify the identity or intentions of users
199
+ - Adversarial prompting techniques might potentially bypass safety mechanisms
200
+ - The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
201
+ 4. **Contextual blindness:**
202
+ - The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
203
+ 5. **Technical limitations**:
204
+ - Performance varies based on how security concepts are described in prompts
205
+ - May not fully understand complex, multi-step security scenarios without clear explanation
206
+ - Cannot access external systems or actively scan environments
207
+ - Cannot independently verify factual accuracy of its outputs
208
+ 6. **Ethical considerations**:
209
+ - Dual-use nature of security knowledge requires careful consideration of appropriate use cases
210
 
211
 
212
  ### Recommendations
213
 
214
  To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
215
 
216
+ 1. **Human oversight**:
217
+ - Always have qualified security professionals review model outputs before implementation
218
+ - Use the model as an assistive tool rather than a replacement for expert human judgment
219
+ - Implement a human-in-the-loop approach for security-critical applications
220
+ 2. **System design safeguards**:
221
+ - Implement additional validation layers for applications built with this model
222
+ - Consider architectural constraints that limit the model's ability to perform potentially harmful actions (excessive agency)
223
+ - Deploy the model in environments with appropriate access controls
224
+ 3. **Prompt engineering**:
225
+ - Use carefully designed prompts that encourage ethical security practices
226
+ - Include explicit instructions regarding responsible disclosure and ethical hacking principles
227
+ - Structure interactions to minimize the risk of inadvertently harmful outputs
228
+ 4. **Knowledge supplementation**:
229
+ - Supplement the model with up-to-date security feeds and databases
230
+ - Implement retrieval-augmented generation for current threat intelligence sources
231
+ 5. **Usage policies**:
232
+ - Develop and enforce clear acceptable use policies for applications using this model
233
+ - Implement monitoring and auditing for high-risk applications
234
+ - Create documentation for end users about the model's limitations