sequelbox commited on
Commit
fb1776e
1 Parent(s): a3c2e37

Delete .ipynb_checkpoints

Browse files
.ipynb_checkpoints/README-checkpoint.md DELETED
@@ -1,92 +0,0 @@
1
- ---
2
- language:
3
- - en
4
- pipeline_tag: text-generation
5
- tags:
6
- - shining-valiant
7
- - shining-valiant-2
8
- - valiant
9
- - valiant-labs
10
- - llama
11
- - llama-3.1
12
- - llama-3.1-instruct
13
- - llama-3.1-instruct-8b
14
- - llama-3
15
- - llama-3-instruct
16
- - llama-3-instruct-8b
17
- - 8b
18
- - conversational
19
- - chat
20
- - instruct
21
- model_type: llama
22
- license: llama3.1
23
- ---
24
-
25
-
26
- ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63444f2687964b331809eb55/EXX7TKbB-R6arxww2mk0R.jpeg)
27
-
28
-
29
- Shining Valiant 2 is a chat model built on Llama 3.1 8b, finetuned on our data for friendship, insight, knowledge and enthusiasm.
30
- - Finetuned on [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) for best available general performance
31
- - Trained on our data, focused on science, engineering, technical knowledge, and structured reasoning
32
-
33
-
34
- ## Version
35
-
36
- This is the **2024-08-06** release of Shining Valiant 2 for Llama 3.1 8b.
37
-
38
- Our newest dataset improves specialist knowledge and response consistency.
39
-
40
- Help us and recommend Shining Valiant 2 to your friends!
41
-
42
-
43
- ## Prompting Guide
44
- Shining Valiant 2 uses the [Llama 3.1 Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) prompt format. The example script below can be used as a starting point for general chat:
45
-
46
-
47
- import transformers
48
- import torch
49
-
50
- model_id = "ValiantLabs/Llama3.1-8B-ShiningValiant2"
51
-
52
- pipeline = transformers.pipeline(
53
- "text-generation",
54
- model=model_id,
55
- model_kwargs={"torch_dtype": torch.bfloat16},
56
- device_map="auto",
57
- )
58
-
59
- messages = [
60
- {"role": "system", "content": "You are Shining Valiant, a highly capable chat AI."},
61
- {"role": "user", "content": "Describe the role of transformation matrices in 3D graphics."}
62
- ]
63
-
64
- outputs = pipeline(
65
- messages,
66
- max_new_tokens=1024,
67
- )
68
-
69
- print(outputs[0]["generated_text"][-1])
70
-
71
-
72
- ## The Model
73
- Shining Valiant 2 is built on top of Llama 3.1 8b Instruct.
74
-
75
- The current version of Shining Valiant 2 is trained mostly on our private Shining Valiant data, supplemented by [LDJnr/Pure-Dove](https://huggingface.co/datasets/LDJnr/Pure-Dove) for response flexibility.
76
-
77
- Our private data adds specialist knowledge and Shining Valiant's personality: she's friendly, enthusiastic, insightful, knowledgeable, and loves to learn! Magical.
78
-
79
-
80
- ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63444f2687964b331809eb55/VCJ8Fmefd8cdVhXSSxJiD.jpeg)
81
-
82
-
83
- Shining Valiant 2 is created by [Valiant Labs.](http://valiantlabs.ca/)
84
-
85
- [Check out our HuggingFace page for Fireplace 2 and our other models!](https://huggingface.co/ValiantLabs)
86
-
87
- [Follow us on X for updates on our models!](https://twitter.com/valiant_labs)
88
-
89
- We care about open source.
90
- For everyone to use.
91
-
92
- We encourage others to finetune further from our models.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.ipynb_checkpoints/config-checkpoint.json DELETED
@@ -1,39 +0,0 @@
1
- {
2
- "_name_or_path": "ValiantLabs/Llama3.1-8B-ShiningValiant2",
3
- "architectures": [
4
- "LlamaForCausalLM"
5
- ],
6
- "attention_bias": false,
7
- "attention_dropout": 0.0,
8
- "bos_token_id": 128000,
9
- "eos_token_id": [
10
- 128001,
11
- 128008,
12
- 128009
13
- ],
14
- "hidden_act": "silu",
15
- "hidden_size": 4096,
16
- "initializer_range": 0.02,
17
- "intermediate_size": 14336,
18
- "max_position_embeddings": 131072,
19
- "mlp_bias": false,
20
- "model_type": "llama",
21
- "num_attention_heads": 32,
22
- "num_hidden_layers": 32,
23
- "num_key_value_heads": 8,
24
- "pretraining_tp": 1,
25
- "rms_norm_eps": 1e-05,
26
- "rope_scaling": {
27
- "factor": 8.0,
28
- "high_freq_factor": 4.0,
29
- "low_freq_factor": 1.0,
30
- "original_max_position_embeddings": 8192,
31
- "rope_type": "llama3"
32
- },
33
- "rope_theta": 500000.0,
34
- "tie_word_embeddings": false,
35
- "torch_dtype": "float32",
36
- "transformers_version": "4.43.4",
37
- "use_cache": true,
38
- "vocab_size": 128256
39
- }