paul-stansifer commited on
Commit
bb3d49f
1 Parent(s): c8f6395

paul-stansifer/llama3-qwantz-coherent

Browse files
README.md CHANGED
@@ -9,7 +9,6 @@ metrics:
9
  model-index:
10
  - name: llama3-qwantz-coherent
11
  results: []
12
- pipeline_tag: text-classification
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -17,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # llama3-qwantz-coherent
19
 
20
- This model is a fine-tuned version of [unsloth/llama-3-8b-bnb-4bit](https://huggingface.co/unsloth/llama-3-8b-bnb-4bit) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.2537
23
- - Accuracy: 0.9279
24
 
25
  ## Model description
26
 
@@ -51,38 +50,8 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
53
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
54
- | 0.2581 | 1.0 | 1016 | 0.2537 | 0.9279 |
55
-
56
-
57
- ```
58
- Can save 90% of coherent strings by discarding 94% of dp strings (cutoff is 75.22732615470886)
59
- Can save 95% of coherent strings by discarding 91% of dp strings (cutoff is -64.16597366333008)
60
- Can save 98% of coherent strings by discarding 86% of dp strings (cutoff is -93.8580572605133)
61
- Can save 99% of coherent strings by discarding 78% of dp strings (cutoff is -99.2882251739502)
62
-
63
- I have constructed a rocket-ship for myself ==> coherent: 99.92%
64
- I have constructed a refund for a ==> dp: 98.67%
65
- Descartes was a dude who wrote "Cogito ergo sum" which means "I think, therefore I am". PRETTY ==> coherent: 100.00%
66
- Descartes was a dude who wrote "Cogito ergo sum" which means "I think, therefore finite lifetimes the ==> dp: 99.99%
67
- That's certainly one way of looking at it, right, Dromiceiomimums? ==> coherent: 100.00%
68
- That's certainly one way of looking at it, is the ==> dp: 98.15%
69
- I'm here to pick up my prescription "Happy New year 2004" glasses! They have a plastic "2" on ==> coherent: 99.79%
70
- I'm here to pick up my prescription "Happy New year 2004" glasses! They have come into cartoon stereotypes ==> dp: 99.99%
71
- I didn't mean for that to be ==> coherent: 99.13%
72
- I didn't mean for the police officer ==> dp: 64.15%
73
- You know what would go down if Nintendo came over? ==> coherent: 100.00%
74
- You know what would go down if Nintendo i live ==> dp: 100.00%
75
- "Aw shucks! I guess it IS true that you're never too ==> coherent: 100.00%
76
- "Aw shucks! I guess it IS true that worse the turmeric ==> dp: 100.00%
77
- Is it true that the only questions worth asking are those that ==> coherent: 99.98%
78
- Is it true that the only questions worth preserving if i've been ==> dp: 99.99%
79
- What? No, he was in pieces. His hand even landed in ==> coherent: 99.44%
80
- What? No, he was in pieces. His gun that an excellent ==> dp: 99.99%
81
- Also, many of the signs are really evocative, so they're easy to ==> coherent: 100.00%
82
- Also, many of the signs are really evocative, so they're approved aaargh ==> dp: 100.00%
83
- Another beautiful hot day! I look forward to these "dog days" ==> coherent: 99.97%
84
- Another beautiful hot day! I look forward to return to make ==> dp: 99.81%
85
- ```
86
 
87
  ### Framework versions
88
 
 
9
  model-index:
10
  - name: llama3-qwantz-coherent
11
  results: []
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
16
 
17
  # llama3-qwantz-coherent
18
 
19
+ This model is a fine-tuned version of [unsloth/llama-3-8b-bnb-4bit](https://huggingface.co/unsloth/llama-3-8b-bnb-4bit) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.3295
22
+ - Accuracy: 0.8758
23
 
24
  ## Model description
25
 
 
50
 
51
  | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
  |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
+ | 0.4482 | 1.0 | 1428 | 0.3295 | 0.8758 |
54
+
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
55
 
56
  ### Framework versions
57
 
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
  "q_proj",
25
  "v_proj",
26
- "k_proj"
 
27
  ],
28
  "task_type": "SEQ_CLS",
29
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "q_proj",
24
  "v_proj",
25
+ "k_proj",
26
+ "o_proj"
27
  ],
28
  "task_type": "SEQ_CLS",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c55ad7da58365c88bee2a9e6e3c7bdf27dccf8bdadf924dd33ba3a08f2bfcd55
3
  size 54593240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aafbb760ed8b904e0cf8d1c3726ac2d22e6f65940f9e2b06b486f66dc54de260
3
  size 54593240
runs/May03_18-12-38_048cd167e598/events.out.tfevents.1714759959.048cd167e598.211.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:446fb0a0caac3cfa4a41f0b30f0e82fd25cde23bb53e72329bc4884f97d0341c
3
+ size 17949
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a939bbd485602e127542064e72f1117ffe82674046ddf744d500e494cfd4c33f
3
  size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73f9fa2fbada4e354ca6f78290205d9ec5eecfada10232d0a46295520fe058ef
3
  size 5048