shoubing35 commited on
Commit
199eaff
·
verified ·
1 Parent(s): 0533f83

Training in progress, step 500

Browse files
README.md CHANGED
@@ -1,18 +1,17 @@
1
  ---
2
  base_model: openai-community/gpt2
3
- datasets: rajpurkar/squad
4
  library_name: transformers
5
  model_name: gpt2-qat
6
  tags:
7
  - generated_from_trainer
8
- - sft
9
  - trl
 
10
  licence: license
11
  ---
12
 
13
  # Model Card for gpt2-qat
14
 
15
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [rajpurkar/squad](https://huggingface.co/datasets/rajpurkar/squad) dataset.
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
@@ -28,7 +27,7 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shoubing-apple/huggingface/runs/3bg819bc)
32
 
33
 
34
  This model was trained with SFT.
@@ -39,7 +38,7 @@ This model was trained with SFT.
39
  - Transformers: 4.52.4
40
  - Pytorch: 2.6.0+cu124
41
  - Datasets: 3.6.0
42
- - Tokenizers: 0.21.1
43
 
44
  ## Citations
45
 
 
1
  ---
2
  base_model: openai-community/gpt2
 
3
  library_name: transformers
4
  model_name: gpt2-qat
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
+ - sft
9
  licence: license
10
  ---
11
 
12
  # Model Card for gpt2-qat
13
 
14
+ This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shoubing-apple/huggingface/runs/afjzjglt)
31
 
32
 
33
  This model was trained with SFT.
 
38
  - Transformers: 4.52.4
39
  - Pytorch: 2.6.0+cu124
40
  - Datasets: 3.6.0
41
+ - Tokenizers: 0.21.2
42
 
43
  ## Citations
44
 
adapter_config.json CHANGED
@@ -15,7 +15,7 @@
15
  "loftq_config": {},
16
  "lora_alpha": 32,
17
  "lora_bias": false,
18
- "lora_dropout": 0.0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": [
@@ -26,9 +26,9 @@
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
29
  "c_fc",
30
- "c_attn",
31
- "c_proj"
32
  ],
33
  "task_type": "CAUSAL_LM",
34
  "trainable_token_indices": null,
 
15
  "loftq_config": {},
16
  "lora_alpha": 32,
17
  "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": [
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "c_proj",
30
  "c_fc",
31
+ "c_attn"
 
32
  ],
33
  "task_type": "CAUSAL_LM",
34
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:91538ea8849e8c41ef014ee2023e93bcf3b83e5bc72fc44e49721073c2ece758
3
  size 9449344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:876eb5b688ece899825f4020d4c01c332367893bc951b01e32e98b80606861c9
3
  size 9449344
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5ca29b0a7fd4bc87f72ad65d9c892946b134c6918b192864fab23f0362017db
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87c1755905df6180d4ab8a0eb718e64129434adb9d08a083a83430d2c722ba15
3
  size 5752