boda commited on
Commit
8bf69a4
·
verified ·
1 Parent(s): e92df20

Model save

Browse files
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: allenai/scitulu-7b
3
+ library_name: transformers
4
+ model_name: review_evaluation_all
5
+ tags:
6
+ - generated_from_trainer
7
+ - trl
8
+ - sft
9
+ licence: license
10
+ ---
11
+
12
+ # Model Card for review_evaluation_all
13
+
14
+ This model is a fine-tuned version of [allenai/scitulu-7b](https://huggingface.co/allenai/scitulu-7b).
15
+ It has been trained using [TRL](https://github.com/huggingface/trl).
16
+
17
+ ## Quick start
18
+
19
+ ```python
20
+ from transformers import pipeline
21
+
22
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="boda/review_evaluation_all", device="cuda")
24
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
+ print(output["generated_text"])
26
+ ```
27
+
28
+ ## Training procedure
29
+
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/bodasadallah2/review_evaluation/runs/qh4fwvyk)
31
+
32
+
33
+ This model was trained with SFT.
34
+
35
+ ### Framework versions
36
+
37
+ - TRL: 0.15.1
38
+ - Transformers: 4.49.0
39
+ - Pytorch: 2.4.0
40
+ - Datasets: 3.3.2
41
+ - Tokenizers: 0.21.0
42
+
43
+ ## Citations
44
+
45
+
46
+
47
+ Cite TRL as:
48
+
49
+ ```bibtex
50
+ @misc{vonwerra2022trl,
51
+ title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
53
+ year = 2020,
54
+ journal = {GitHub repository},
55
+ publisher = {GitHub},
56
+ howpublished = {\url{https://github.com/huggingface/trl}}
57
+ }
58
+ ```
adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "allenai/scitulu-7b",
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": true,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 16,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 16,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "up_proj",
27
+ "down_proj",
28
+ "q_proj",
29
+ "o_proj",
30
+ "gate_proj",
31
+ "k_proj",
32
+ "v_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a47a4671015ecaf760615c3f90857a849eadf11cf80466d56259a25f006b15b3
3
+ size 80014016
all_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss": 2.0348782539367676,
3
+ "eval_runtime": 1.2317,
4
+ "eval_samples": 100,
5
+ "eval_samples_per_second": 4.059,
6
+ "eval_steps_per_second": 0.812,
7
+ "total_flos": 7336072979677184.0,
8
+ "train_loss": 0.0,
9
+ "train_runtime": 0.0077,
10
+ "train_samples": 898,
11
+ "train_samples_per_second": 25878.664,
12
+ "train_steps_per_second": 3136.808
13
+ }
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "allenai/scitulu-7b",
4
+ "architectures": [
5
+ "LlamaForCausalLM"
6
+ ],
7
+ "attention_bias": false,
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "head_dim": 128,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 11008,
16
+ "max_position_embeddings": 4096,
17
+ "mlp_bias": false,
18
+ "model_type": "llama",
19
+ "num_attention_heads": 32,
20
+ "num_hidden_layers": 32,
21
+ "num_key_value_heads": 32,
22
+ "pad_token_id": 0,
23
+ "pretraining_tp": 1,
24
+ "rms_norm_eps": 1e-06,
25
+ "rope_scaling": null,
26
+ "rope_theta": 10000.0,
27
+ "tie_word_embeddings": false,
28
+ "torch_dtype": "bfloat16",
29
+ "transformers_version": "4.49.0",
30
+ "use_cache": true,
31
+ "vocab_size": 32000
32
+ }
eval_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_loss": 2.0348782539367676,
3
+ "eval_runtime": 1.2317,
4
+ "eval_samples": 100,
5
+ "eval_samples_per_second": 4.059,
6
+ "eval_steps_per_second": 0.812
7
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": true,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": true,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "extra_special_tokens": {},
36
+ "legacy": true,
37
+ "model_max_length": 2048,
38
+ "pad_token": "</s>",
39
+ "sp_model_kwargs": {},
40
+ "spaces_between_special_tokens": false,
41
+ "tokenizer_class": "LlamaTokenizer",
42
+ "unk_token": "<unk>",
43
+ "use_default_system_prompt": false
44
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "total_flos": 7336072979677184.0,
3
+ "train_loss": 0.0,
4
+ "train_runtime": 0.0077,
5
+ "train_samples": 898,
6
+ "train_samples_per_second": 25878.664,
7
+ "train_steps_per_second": 3136.808
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.7058823529411766,
5
+ "eval_steps": 500,
6
+ "global_step": 24,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.11764705882352941,
13
+ "grad_norm": 0.5514968283268096,
14
+ "learning_rate": 6.666666666666667e-05,
15
+ "loss": 2.103,
16
+ "mean_token_accuracy": 0.6410358250141144,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.5882352941176471,
21
+ "grad_norm": 0.5564948741796076,
22
+ "learning_rate": 0.0001955572805786141,
23
+ "loss": 1.95,
24
+ "mean_token_accuracy": 0.6562966406345367,
25
+ "step": 5
26
+ },
27
+ {
28
+ "epoch": 1.0,
29
+ "eval_loss": 2.2298309803009033,
30
+ "eval_mean_token_accuracy": 0.6953375414013863,
31
+ "eval_runtime": 1.1412,
32
+ "eval_samples_per_second": 4.381,
33
+ "eval_steps_per_second": 0.876,
34
+ "step": 9
35
+ },
36
+ {
37
+ "epoch": 1.1176470588235294,
38
+ "grad_norm": 0.3550428109424067,
39
+ "learning_rate": 0.00015000000000000001,
40
+ "loss": 1.5318,
41
+ "mean_token_accuracy": 0.6940737068653107,
42
+ "step": 10
43
+ },
44
+ {
45
+ "epoch": 1.7058823529411766,
46
+ "grad_norm": 0.15172839418805506,
47
+ "learning_rate": 7.774790660436858e-05,
48
+ "loss": 1.4025,
49
+ "mean_token_accuracy": 0.7238408386707306,
50
+ "step": 15
51
+ },
52
+ {
53
+ "epoch": 2.0,
54
+ "eval_loss": 2.040649652481079,
55
+ "eval_mean_token_accuracy": 0.6979705393314362,
56
+ "eval_runtime": 1.278,
57
+ "eval_samples_per_second": 3.912,
58
+ "eval_steps_per_second": 0.782,
59
+ "step": 18
60
+ },
61
+ {
62
+ "epoch": 2.235294117647059,
63
+ "grad_norm": 0.16155482865148485,
64
+ "learning_rate": 1.7376122568400532e-05,
65
+ "loss": 1.4263,
66
+ "mean_token_accuracy": 0.7188252657651901,
67
+ "step": 20
68
+ },
69
+ {
70
+ "epoch": 2.7058823529411766,
71
+ "step": 24,
72
+ "total_flos": 7336072979677184.0,
73
+ "train_loss": 0.0,
74
+ "train_runtime": 0.0077,
75
+ "train_samples_per_second": 25878.664,
76
+ "train_steps_per_second": 3136.808
77
+ }
78
+ ],
79
+ "logging_steps": 5,
80
+ "max_steps": 24,
81
+ "num_input_tokens_seen": 0,
82
+ "num_train_epochs": 3,
83
+ "save_steps": 100,
84
+ "stateful_callbacks": {
85
+ "TrainerControl": {
86
+ "args": {
87
+ "should_epoch_stop": false,
88
+ "should_evaluate": false,
89
+ "should_log": false,
90
+ "should_save": true,
91
+ "should_training_stop": true
92
+ },
93
+ "attributes": {}
94
+ }
95
+ },
96
+ "total_flos": 7336072979677184.0,
97
+ "train_batch_size": 2,
98
+ "trial_name": null,
99
+ "trial_params": null
100
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d15422a8cc312f27243bbe037d59674ce9812dd2538bb912d4172e959462e158
3
+ size 7288