{ "bomFormat": "CycloneDX", "specVersion": "1.6", "serialNumber": "urn:uuid:68e31c27-69fa-4eba-98f4-af6c900705dd", "version": 1, "metadata": { "timestamp": "2025-06-05T09:40:41.667208+00:00", "component": { "type": "machine-learning-model", "bom-ref": "Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2-326049c7-7efd-54be-8c7a-90827ace56c1", "name": "Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2", "externalReferences": [ { "url": "https://huggingface.co/Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2", "type": "documentation" } ], "modelCard": { "modelParameters": { "task": "text-generation", "architectureFamily": "llama", "modelArchitecture": "LlamaForCausalLM" }, "properties": [ { "name": "library_name", "value": "transformers" } ], "quantitativeAnalysis": { "performanceMetrics": [ { "slice": "dataset: HuggingFaceH4/ifeval", "type": "inst_level_strict_acc and prompt_level_strict_acc", "value": 77.92 }, { "slice": "dataset: BBH", "type": "acc_norm", "value": 29.69 }, { "slice": "dataset: hendrycks/competition_math", "type": "exact_match", "value": 16.92 }, { "slice": "dataset: Idavidrein/gpqa", "type": "acc_norm", "value": 4.36 }, { "slice": "dataset: TAUR-Lab/MuSR", "type": "acc_norm", "value": 7.77 }, { "slice": "dataset: TIGER-Lab/MMLU-Pro, split: test, config: main", "type": "acc", "value": 30.9 } ] } }, "authors": [ { "name": "Orenguteng" } ], "licenses": [ { "license": { "name": "llama3.1" } } ], "tags": [ "transformers", "safetensors", "llama", "text-generation", "conversational", "license:llama3.1", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ] } } }