Upload LlamaForCausalLM

#6
by WesleySantos - opened
config.json CHANGED
@@ -19,7 +19,7 @@
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
  "tie_word_embeddings": false,
22
- "torch_dtype": "bfloat16",
23
  "transformers_version": "4.31.0",
24
  "use_cache": true,
25
  "vocab_size": 32000
 
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
21
  "tie_word_embeddings": false,
22
+ "torch_dtype": "float16",
23
  "transformers_version": "4.31.0",
24
  "use_cache": true,
25
  "vocab_size": 32000
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e5b9108916566f9bfbbc5832c04ff140fe0873cbc596b49c001fb7a624aff383
3
- size 9976634558
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8acdaa6474b353649d433d94a292d661b364e200e1abb66c1013dade5fe85454
3
+ size 9976637886
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:626b1ae30cb7806b5b9f6c087adaa7c5addb4774db539ed610ec7e55aece486b
3
- size 3500315539
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7075a5d76f093d71b01b77fc82ed023564c8cecf22dc6cc7c90881771bf53cf3
3
+ size 3500316627