Naela00 commited on
Commit
481f266
·
1 Parent(s): d98087e

adding odal,oeal,rdal,real,rebl

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +4 -0
  2. .gitignore +5 -0
  3. README.md +52 -0
  4. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/README.md +202 -0
  5. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/adapter_config.json +3 -0
  6. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/adapter_model.safetensors +3 -0
  7. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/added_tokens.json +3 -0
  8. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/chat_template.jinja +3 -0
  9. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/global_step417/mp_rank_00_model_states.pt +3 -0
  10. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/global_step417/zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  11. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/global_step417/zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  12. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/latest +1 -0
  13. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/merges.txt +3 -0
  14. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/rng_state_0.pth +3 -0
  15. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/rng_state_1.pth +3 -0
  16. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/special_tokens_map.json +3 -0
  17. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/tokenizer.json +3 -0
  18. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/tokenizer_config.json +3 -0
  19. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/trainer_state.json +3 -0
  20. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/training_args.bin +3 -0
  21. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/vocab.json +3 -0
  22. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/README.md +202 -0
  23. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/adapter_config.json +3 -0
  24. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/adapter_model.safetensors +3 -0
  25. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/added_tokens.json +3 -0
  26. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/chat_template.jinja +3 -0
  27. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/global_step537/mp_rank_00_model_states.pt +3 -0
  28. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/global_step537/zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  29. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/global_step537/zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  30. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/latest +1 -0
  31. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/merges.txt +3 -0
  32. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/rng_state_0.pth +3 -0
  33. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/rng_state_1.pth +3 -0
  34. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/special_tokens_map.json +3 -0
  35. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/tokenizer.json +3 -0
  36. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/tokenizer_config.json +3 -0
  37. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/trainer_state.json +3 -0
  38. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/training_args.bin +3 -0
  39. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/vocab.json +3 -0
  40. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/README.md +202 -0
  41. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/adapter_config.json +3 -0
  42. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/adapter_model.safetensors +3 -0
  43. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/added_tokens.json +3 -0
  44. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/chat_template.jinja +3 -0
  45. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/global_step544/mp_rank_00_model_states.pt +3 -0
  46. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/global_step544/zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
  47. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/global_step544/zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
  48. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/latest +1 -0
  49. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/merges.txt +3 -0
  50. output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/rng_state_0.pth +3 -0
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.json filter=lfs diff=lfs merge=lfs -text
37
+ *..jinja filter=lfs diff=lfs merge=lfs -text
38
+ *.jinja filter=lfs diff=lfs merge=lfs -text
39
+ *.txt filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ runs
2
+ *.ipynb
3
+ utils
4
+ __pycache__
5
+ *.py
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ task_categories:
4
+ - text-classification
5
+ language:
6
+ - fr
7
+ tags:
8
+ - text-classification
9
+ - toxicity
10
+ - hate-speech
11
+ - content-moderation
12
+ - chain-of-thought
13
+ - curriculum-learning
14
+ - nlp
15
+ - french-dataset
16
+ - classification
17
+ pretty_name: ToxiFrench
18
+ datasets:
19
+ - Naela00/ToxiFrenchFinetuning
20
+ base_model:
21
+ - Qwen/Qwen3-4B
22
+ ---
23
+
24
+ # ToxiFrench Model
25
+
26
+ This repository contains the **ToxiFrench** model, a **French language model** fine-tuned for **toxic comment classification**. It is based on the [**Qwen/Qwen3-4B**](https://huggingface.co/Qwen/Qwen3-4B) architecture and is designed to detect and classify toxic comments in French text.
27
+
28
+ We performed a series of experiments to evaluate the model's performance under different fine-tuning configurations, focusing on the impact of **data selection strategies** and **Chain-of-Thought (CoT)** annotations.
29
+
30
+ ## Finetuning notations
31
+
32
+ Each experiment follows a naming scheme like: **(r/o)(e/d)(a/b)(s/m/l)**
33
+ Where:
34
+
35
+ - `r` = random order, `o` = ordered (curriculum)
36
+ - `e` = equal toxic/non-toxic, `d` = real-world imbalance
37
+ - `a` = with CoT finetuning, `b` = without CoT
38
+ - `s` = small (100), `m` = medium (1000), `l` = large (all)
39
+
40
+ > e.g. `rdal` is the model trained on the natural distribution of toxicity (`d`), on an arbitrary order (`r`), with CoT annotations (`a`), and on the whole dataset (`l`).
41
+
42
+ If a label like `<cot-step>` is present in the checkpoint name, it indicates that the CoT that was used during training did not include this specific reasoning step.
43
+
44
+ ## Citation
45
+
46
+ ```
47
+ @misc{toxifrench2025,
48
+ title={ToxiFrench},
49
+ author={Delaval Axel},
50
+ year={2025},
51
+ }
52
+ ```
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7230bbb1d85421684376a57f17cb053f867a001c5ca3ff30f574021e3e3c167
3
+ size 828
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a220a58078aa7dbeedaece01045895702dc1f4341e6eb68eb372147c69363b1
3
+ size 1579102000
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0284b582e14987fbd3d5a2cb2bd139084371ed9acbae488829a1c900833c680
3
+ size 707
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55ee1b1660128b7098723e0abcd92caa0788061051c62d51cbe87d9cf1974d8
3
+ size 4168
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/global_step417/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3fb11f9b7ee3aed9731967a002a28af3f236f98a4518f832085713aa935c238
3
+ size 861285289
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/global_step417/zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08ecc800d2de12d403727fce406819aa8759a2d327e157b34d6c8d2ede2f9e75
3
+ size 77884216
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/global_step417/zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e243d774fb8843c0a02feeee03e3cc25352a2de0ba0f907b01121a83571dbe4
3
+ size 77884344
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step417
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/merges.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5
3
+ size 1671853
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64120fecc819570fdb1d5359518c94d15c1bc479ba296968352e83038a4c9806
3
+ size 14917
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ba0dc98ec5329a2b43a51e909701a70a2c4641578ad65d679147f68b4736f49
3
+ size 14917
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76862e765266b85aa9459767e33cbaf13970f327a0e88d1c65846c2ddd3a1ecd
3
+ size 613
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:443bfa629eb16387a12edbf92a76f6a6f10b2af3b53d87ba1550adfcf45f7fa0
3
+ size 5404
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7edf6e7b712340e695675a44c607e6b28efccb857b15c2f5e66703c80219d157
3
+ size 12181
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a27bfb6651c77744bd50f9d980f357ec5e3e1a52718e7cd7cbf37b8716ffd1a0
3
+ size 7889
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-420/vocab.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca10d7e9fb3ed18575dd1e277a2579c16d108e32f27439684afa0e10b1440910
3
+ size 2776833
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7230bbb1d85421684376a57f17cb053f867a001c5ca3ff30f574021e3e3c167
3
+ size 828
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c471f33c6fc5d1b0487803f39021414584e6964957f7dcb91fdb3d4da087e4a
3
+ size 1579102000
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0284b582e14987fbd3d5a2cb2bd139084371ed9acbae488829a1c900833c680
3
+ size 707
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55ee1b1660128b7098723e0abcd92caa0788061051c62d51cbe87d9cf1974d8
3
+ size 4168
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/global_step537/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13b33c193c2b7d902bb1ae931585374f558951cc8f5b1f09751d7e4f11efee8b
3
+ size 861285289
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/global_step537/zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:204b03392e17cccf47207ab9b7349f85e22381962ecc6d97cfccfe2d0d4d6aef
3
+ size 77884216
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/global_step537/zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10c42677818f4adf001cbf6c00c7b76b8d896d4a9cb17ef3016b2d207d60befe
3
+ size 77884344
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step537
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/merges.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5
3
+ size 1671853
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efc3b787ee0ff48b74e4adb7f937bf74a2a5d3e1b0b23c24d989611139cd39a8
3
+ size 14917
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab9aa2a52ba9cac6bf490c27ae75cc2de3c5f6ea1426cba772af0e6552d96aaa
3
+ size 14917
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76862e765266b85aa9459767e33cbaf13970f327a0e88d1c65846c2ddd3a1ecd
3
+ size 613
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:443bfa629eb16387a12edbf92a76f6a6f10b2af3b53d87ba1550adfcf45f7fa0
3
+ size 5404
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/trainer_state.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01345551b3a74614e68a28d05230f65c1885f628069ad92d27efbc75568fafd3
3
+ size 15288
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a27bfb6651c77744bd50f9d980f357ec5e3e1a52718e7cd7cbf37b8716ffd1a0
3
+ size 7889
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-540/vocab.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca10d7e9fb3ed18575dd1e277a2579c16d108e32f27439684afa0e10b1440910
3
+ size 2776833
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B
3
+ library_name: peft
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
200
+ ### Framework versions
201
+
202
+ - PEFT 0.15.2
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7230bbb1d85421684376a57f17cb053f867a001c5ca3ff30f574021e3e3c167
3
+ size 828
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26893ada07476813e552f5b735e06edccb0cd089b10d53d94551e2d2500e1e97
3
+ size 1579102000
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0284b582e14987fbd3d5a2cb2bd139084371ed9acbae488829a1c900833c680
3
+ size 707
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/chat_template.jinja ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a55ee1b1660128b7098723e0abcd92caa0788061051c62d51cbe87d9cf1974d8
3
+ size 4168
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/global_step544/mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a1c3516b4a07c885d519f0cc8a1565eb7b7ba12a68e505ef91ca0c1c7de3051
3
+ size 861285289
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/global_step544/zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:867cffbbad20a1e9bff6de7622f9decd1c4c67297eb52c36c79d0979836295b6
3
+ size 77884216
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/global_step544/zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c1d9114cd659bdf29f4fd0586888c9f537af1932aa28489a994545fb0747f9
3
+ size 77884344
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/latest ADDED
@@ -0,0 +1 @@
 
 
1
+ global_step544
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/merges.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8831e4f1a044471340f7c0a83d7bd71306a5b867e95fd870f74d0c5308a904d5
3
+ size 1671853
output_odal_Qwen3-4B_<cot_intention>-<cot_categorie_list>/checkpoint-548/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b7e02beb04282fa93748a1dd67ed95925f835cc32b9c3072cb8deb9f80998f9
3
+ size 14917