shoubing35 commited on
Commit
1f8fb13
·
verified ·
1 Parent(s): f60ee6e

Training in progress, step 40

Browse files
README.md CHANGED
@@ -1,18 +1,17 @@
1
  ---
2
  base_model: openai-community/gpt2
3
- datasets: rajpurkar/squad
4
  library_name: transformers
5
  model_name: gpt2-qat
6
  tags:
7
  - generated_from_trainer
8
- - trl
9
  - sft
 
10
  licence: license
11
  ---
12
 
13
  # Model Card for gpt2-qat
14
 
15
- This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2) on the [rajpurkar/squad](https://huggingface.co/datasets/rajpurkar/squad) dataset.
16
  It has been trained using [TRL](https://github.com/huggingface/trl).
17
 
18
  ## Quick start
@@ -28,7 +27,7 @@ print(output["generated_text"])
28
 
29
  ## Training procedure
30
 
31
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shoubing-apple/huggingface/runs/8j5wiaig)
32
 
33
 
34
  This model was trained with SFT.
 
1
  ---
2
  base_model: openai-community/gpt2
 
3
  library_name: transformers
4
  model_name: gpt2-qat
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
+ - trl
9
  licence: license
10
  ---
11
 
12
  # Model Card for gpt2-qat
13
 
14
+ This model is a fine-tuned version of [openai-community/gpt2](https://huggingface.co/openai-community/gpt2).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/shoubing-apple/huggingface/runs/7fvvdarg)
31
 
32
 
33
  This model was trained with SFT.
adapter_config.json CHANGED
@@ -15,7 +15,7 @@
15
  "loftq_config": {},
16
  "lora_alpha": 32,
17
  "lora_bias": false,
18
- "lora_dropout": 0.05,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": [
@@ -26,9 +26,9 @@
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
29
  "c_attn",
30
- "c_proj",
31
- "c_fc"
32
  ],
33
  "task_type": "CAUSAL_LM",
34
  "trainable_token_indices": null,
 
15
  "loftq_config": {},
16
  "lora_alpha": 32,
17
  "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": [
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "c_fc",
30
  "c_attn",
31
+ "c_proj"
 
32
  ],
33
  "task_type": "CAUSAL_LM",
34
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cd7aecb971ad3e3952bd7ee70bb0686045bdcf0aa0dfd282151d92cae72148c
3
  size 9449344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af18ac57a03f9302f156bdb886cbd59bf4dee2798128f81aa7a7cdce59130a33
3
  size 9449344
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b1ba0f042464a14cbe6d0413d6be19e4fdf55507b5502d2009d53d7c79213f4
3
  size 497774208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78d0583001c3a841c3dfe9e14c938b81b9b8b47bcf0822bc152218da98197450
3
  size 497774208
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4c4dc2ccc1968527222ae4b48d920c514561a5ec43d5513429641c2f5fbdce7
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a92452418d5296c17d6031c229cce7bd0be37688510d53bf398a2e36e5f9120
3
  size 5752