MisterAI commited on
Commit
53b0574
·
verified ·
1 Parent(s): 05c9131

Upload folder using huggingface_hub

Browse files
checkpoint-316/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "MisterAI/bigscience_bloom-560m",
3
+ "apply_residual_connection_post_layernorm": false,
4
+ "architectures": [
5
+ "BloomForCausalLM"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "attention_softmax_in_fp32": true,
9
+ "bias_dropout_fusion": true,
10
+ "bos_token_id": 1,
11
+ "eos_token_id": 2,
12
+ "hidden_dropout": 0.0,
13
+ "hidden_size": 1024,
14
+ "initializer_range": 0.02,
15
+ "layer_norm_epsilon": 1e-05,
16
+ "masked_softmax_fusion": true,
17
+ "model_type": "bloom",
18
+ "n_head": 16,
19
+ "n_inner": null,
20
+ "n_layer": 24,
21
+ "offset_alibi": 100,
22
+ "pad_token_id": 3,
23
+ "pretraining_tp": 1,
24
+ "skip_bias_add": true,
25
+ "skip_bias_add_qkv": false,
26
+ "slow_but_exact": false,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.49.0",
29
+ "unk_token_id": 0,
30
+ "use_cache": true,
31
+ "vocab_size": 250880
32
+ }
checkpoint-316/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 3,
6
+ "transformers_version": "4.49.0"
7
+ }
checkpoint-316/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adfbb72d053017486fb307a2cd9cf83dcd8d915f84797c1106fdb50766696788
3
+ size 2236892304
checkpoint-316/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3ba5f9c9c1c4264f97668c459b9d743ec3e79436157a64a3d598f460ed063e6
3
+ size 4473954638
checkpoint-316/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4864c89835cbf79e96fc0a2d2de7044506fea9c48cfcbba91fdab915a1eb9808
3
+ size 13990
checkpoint-316/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9031796aa97525ded3c7342519833c7e15ce42382c4b2d3b8c3226274a7d42d
3
+ size 1064
checkpoint-316/trainer_state.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 316,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [],
11
+ "logging_steps": 500,
12
+ "max_steps": 316,
13
+ "num_input_tokens_seen": 0,
14
+ "num_train_epochs": 1,
15
+ "save_steps": 10000,
16
+ "stateful_callbacks": {
17
+ "TrainerControl": {
18
+ "args": {
19
+ "should_epoch_stop": false,
20
+ "should_evaluate": false,
21
+ "should_log": false,
22
+ "should_save": true,
23
+ "should_training_stop": true
24
+ },
25
+ "attributes": {}
26
+ }
27
+ },
28
+ "total_flos": 586014108155904.0,
29
+ "train_batch_size": 2,
30
+ "trial_name": null,
31
+ "trial_params": null
32
+ }
checkpoint-316/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27131aab987486cc0abf734c35f1790dbb9af39e0d4051e29646035b8e709763
3
+ size 5432
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9370f4a70b0a63c182e280ba6cc11d8d2de19073f4b79feed1677dfabbf8f6e8
3
  size 2236892304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adfbb72d053017486fb307a2cd9cf83dcd8d915f84797c1106fdb50766696788
3
  size 2236892304
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:943aef7a9045f9ab551b6790c00162e434f16ed88f19f903212dbc517c4cc0bf
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27131aab987486cc0abf734c35f1790dbb9af39e0d4051e29646035b8e709763
3
  size 5432