Naela00 commited on
Commit
1643c23
·
1 Parent(s): 481f266

adding odal,oeal,rdal,real,rebl

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/README.md +202 -0
  2. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/adapter_config.json +3 -0
  3. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/adapter_model.safetensors +3 -0
  4. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/added_tokens.json +3 -0
  5. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/chat_template.jinja +3 -0
  6. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/global_step96/mp_rank_00_model_states.pt +3 -0
  7. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/global_step96/zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  8. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/global_step96/zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  9. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/latest +1 -0
  10. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/merges.txt +3 -0
  11. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/rng_state_0.pth +3 -0
  12. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/rng_state_1.pth +3 -0
  13. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/special_tokens_map.json +3 -0
  14. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/tokenizer.json +3 -0
  15. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/tokenizer_config.json +3 -0
  16. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/trainer_state.json +3 -0
  17. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/training_args.bin +3 -0
  18. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/vocab.json +3 -0
  19. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/README.md +202 -0
  20. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/adapter_config.json +3 -0
  21. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/adapter_model.safetensors +3 -0
  22. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/added_tokens.json +3 -0
  23. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/chat_template.jinja +3 -0
  24. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/global_step134/mp_rank_00_model_states.pt +3 -0
  25. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/global_step134/zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  26. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/global_step134/zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  27. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/latest +1 -0
  28. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/merges.txt +3 -0
  29. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/rng_state_0.pth +3 -0
  30. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/rng_state_1.pth +3 -0
  31. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/special_tokens_map.json +3 -0
  32. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/tokenizer.json +3 -0
  33. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/tokenizer_config.json +3 -0
  34. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/trainer_state.json +3 -0
  35. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/training_args.bin +3 -0
  36. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/vocab.json +3 -0
  37. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/README.md +202 -0
  38. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/adapter_config.json +3 -0
  39. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/adapter_model.safetensors +3 -0
  40. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/added_tokens.json +3 -0
  41. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/chat_template.jinja +3 -0
  42. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/global_step48/mp_rank_00_model_states.pt +3 -0
  43. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/global_step48/zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  44. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/global_step48/zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  45. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/latest +1 -0
  46. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/merges.txt +3 -0
  47. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/rng_state_0.pth +3 -0
  48. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/rng_state_1.pth +3 -0
  49. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/special_tokens_map.json +3 -0
  50. output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/tokenizer.json +3 -0
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d24c948342ca579064152f31c52b4669fab0b418f1f45d441e54177b8feb8a94
3
+ size 828
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fd5dce1e22abf8efabaf6830723a55ab608fa3a814109a0d7b42b9ab0bfcd51
3
+ size 1579102000
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0284b582e14987fbd3d5a2cb2bd139084371ed9acbae488829a1c900833c680
3
+ size 707
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55ee1b1660128b7098723e0abcd92caa0788061051c62d51cbe87d9cf1974d8
3
+ size 4168
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/global_step96/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c624fdbebc56ffc5db66d490af16ee1801c1bd8c0e2b27c093a1e3934b57b54
3
+ size 861284860
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/global_step96/zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:292c5081b17f7bbb1b4a6769850b7644c3ae57ce0fe3d2f09258a2ba38b6d45c
3
+ size 77883757
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/global_step96/zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df484e4cb7c85ab56f01df01e5869b5b5bcfa1b08417479d8615af79e4630846
3
+ size 77883885
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step96
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/merges.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5
3
+ size 1671853
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ef236769919c6937bbb9c1f58b007c30e89350e3f1c178f734e603513816799
3
+ size 14512
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03f6e54ffc615d317ae648550964e00d4c66ee3004542a0f2f166d92b7f90a9f
3
+ size 14512
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76862e765266b85aa9459767e33cbaf13970f327a0e88d1c65846c2ddd3a1ecd
3
+ size 613
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:443bfa629eb16387a12edbf92a76f6a6f10b2af3b53d87ba1550adfcf45f7fa0
3
+ size 5404
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3d84d939b68fc7e50c8c32d359b8a583f4d28965cd7b3624b1b26be45b89dd9
3
+ size 2217
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffbe4648d61459b6bf1672387ea81ae4255c9993a6f3c26550f43d5c98b3dbc9
3
+ size 7544
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-100/vocab.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca10d7e9fb3ed18575dd1e277a2579c16d108e32f27439684afa0e10b1440910
3
+ size 2776833
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d24c948342ca579064152f31c52b4669fab0b418f1f45d441e54177b8feb8a94
3
+ size 828
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2a938c79caabffb0faf44ac38aa77f435e75f083431b1149c4ca9f177fb48aa
3
+ size 1579102000
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0284b582e14987fbd3d5a2cb2bd139084371ed9acbae488829a1c900833c680
3
+ size 707
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55ee1b1660128b7098723e0abcd92caa0788061051c62d51cbe87d9cf1974d8
3
+ size 4168
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/global_step134/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74f51c5147bba5721a86fd9ff0a3b725435042cb88cb0726c2b80173d928ec22
3
+ size 861284860
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/global_step134/zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:702b15d5d9fae9f57bd08e74866d230eef1ed93b39e2b687db2e0a3899fbec19
3
+ size 77883757
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/global_step134/zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9c68e33d6ad6b4520a0f4f489c958de69a1054e762a125b9f3760529d53a9f8
3
+ size 77883885
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step134
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/merges.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5
3
+ size 1671853
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9587c53f6ce81b18f8f1a262d403903aa458ae4eda15d4bf707a77f52d32c7e
3
+ size 14512
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5089f39a8fc3c18d44c5b96748259bcaabae188d5a795d860afd41814dd0bf79
3
+ size 14512
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76862e765266b85aa9459767e33cbaf13970f327a0e88d1c65846c2ddd3a1ecd
3
+ size 613
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:443bfa629eb16387a12edbf92a76f6a6f10b2af3b53d87ba1550adfcf45f7fa0
3
+ size 5404
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39ad065f5b9e629b1ef33ba6f05fa9ff471de5a498a79a444f347fb8ffcde2e3
3
+ size 2203
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffbe4648d61459b6bf1672387ea81ae4255c9993a6f3c26550f43d5c98b3dbc9
3
+ size 7544
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-140/vocab.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca10d7e9fb3ed18575dd1e277a2579c16d108e32f27439684afa0e10b1440910
3
+ size 2776833
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d24c948342ca579064152f31c52b4669fab0b418f1f45d441e54177b8feb8a94
3
+ size 828
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c6ef29a3a75bcda2d3e5d0a13f7c74d85b30e226e357d7d843319fcb3df8785
3
+ size 1579102000
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0284b582e14987fbd3d5a2cb2bd139084371ed9acbae488829a1c900833c680
3
+ size 707
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55ee1b1660128b7098723e0abcd92caa0788061051c62d51cbe87d9cf1974d8
3
+ size 4168
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/global_step48/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09fe5eed8f11859594b1c1a0359867e4f7cc9f7510f8d8842a0c3b1528f20301
3
+ size 861284860
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/global_step48/zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:343118874248e9d4bd40e2048ec2c6aab53394252b41e70123ad340ed39d1da2
3
+ size 77883757
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/global_step48/zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc42140fcf64959d9c0b47e8acd42f6f071020f44b611ed302da5cc5a1398dde
3
+ size 77883885
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step48
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/merges.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5
3
+ size 1671853
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a60e03c0c99e198e68453659512103d39892f5141c5d625a20b87a8bb3bae15
3
+ size 14512
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30373d33a42c7b4e5ad53e2185900863617a890513a8634a8f9a57a767f18f7f
3
+ size 14512
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76862e765266b85aa9459767e33cbaf13970f327a0e88d1c65846c2ddd3a1ecd
3
+ size 613
output_oebl_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-50/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654