Training in progress, step 10
Browse files- adapter_config.json +0 -2
- adapter_model.safetensors +2 -2
- runs/Jun13_17-03-02_lambda-hyperplane/events.out.tfevents.1749823396.lambda-hyperplane +3 -0
- runs/Jun14_01-06-51_lambda-hyperplane/events.out.tfevents.1749852413.lambda-hyperplane +3 -0
- runs/Jun14_01-16-06_lambda-hyperplane/events.out.tfevents.1749852970.lambda-hyperplane +3 -0
- runs/Jun14_01-17-42_lambda-hyperplane/events.out.tfevents.1749853065.lambda-hyperplane +3 -0
- runs/Jun14_01-19-55_lambda-hyperplane/events.out.tfevents.1749853199.lambda-hyperplane +3 -0
- runs/Jun14_01-36-59_lambda-hyperplane/events.out.tfevents.1749854335.lambda-hyperplane +3 -0
- runs/Jun14_01-40-48_lambda-hyperplane/events.out.tfevents.1749854532.lambda-hyperplane +3 -0
- training_args.bin +2 -2
adapter_config.json
CHANGED
@@ -3,7 +3,6 @@
|
|
3 |
"auto_mapping": null,
|
4 |
"base_model_name_or_path": "QCRI/Fanar-1-9B-Instruct",
|
5 |
"bias": "none",
|
6 |
-
"corda_config": null,
|
7 |
"eva_config": null,
|
8 |
"exclude_modules": null,
|
9 |
"fan_in_fan_out": false,
|
@@ -28,7 +27,6 @@
|
|
28 |
"v_proj"
|
29 |
],
|
30 |
"task_type": "CAUSAL_LM",
|
31 |
-
"trainable_token_indices": null,
|
32 |
"use_dora": false,
|
33 |
"use_rslora": false
|
34 |
}
|
|
|
3 |
"auto_mapping": null,
|
4 |
"base_model_name_or_path": "QCRI/Fanar-1-9B-Instruct",
|
5 |
"bias": "none",
|
|
|
6 |
"eva_config": null,
|
7 |
"exclude_modules": null,
|
8 |
"fan_in_fan_out": false,
|
|
|
27 |
"v_proj"
|
28 |
],
|
29 |
"task_type": "CAUSAL_LM",
|
|
|
30 |
"use_dora": false,
|
31 |
"use_rslora": false
|
32 |
}
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e212a7603724537b4229550b2b2cf1aae733482503975db113807b957ba78aa
|
3 |
+
size 8967968
|
runs/Jun13_17-03-02_lambda-hyperplane/events.out.tfevents.1749823396.lambda-hyperplane
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b12526d04e132bfbfff82230f29179cadc6d05776c76c7f0f4189d90abd8d3bd
|
3 |
+
size 5494
|
runs/Jun14_01-06-51_lambda-hyperplane/events.out.tfevents.1749852413.lambda-hyperplane
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c2c226647119b65a676059aa87bc10beb4f62686e842a163e770cadeb47918c
|
3 |
+
size 5494
|
runs/Jun14_01-16-06_lambda-hyperplane/events.out.tfevents.1749852970.lambda-hyperplane
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:109e366e3c25c9c1afeb3349bb353b69b1bbd23485c5a0ff0ff0c2c85806586b
|
3 |
+
size 5495
|
runs/Jun14_01-17-42_lambda-hyperplane/events.out.tfevents.1749853065.lambda-hyperplane
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8a25ac021883fdf74d67f2ee64ade7a8b9f161d98ca0be327e350280b07c9cb4
|
3 |
+
size 5494
|
runs/Jun14_01-19-55_lambda-hyperplane/events.out.tfevents.1749853199.lambda-hyperplane
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f6aa4155885ad246429e8038023227493c0b80be314f7c1df5f6735ddb6d71a6
|
3 |
+
size 5494
|
runs/Jun14_01-36-59_lambda-hyperplane/events.out.tfevents.1749854335.lambda-hyperplane
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e7d1c70b4741789522fa483639317668b9a62a8f8e5b78f1be62c3319b03486d
|
3 |
+
size 5494
|
runs/Jun14_01-40-48_lambda-hyperplane/events.out.tfevents.1749854532.lambda-hyperplane
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7c3e98c8d592772ef527c2011e6f3b5057ac3525996532738b39909df4d634fe
|
3 |
+
size 5494
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79fcbedab2c3bfd2cd26cb6ed7545e77d8b85250e33bced4a1bd77819e730720
|
3 |
+
size 6776
|