nightmedia commited on
Commit
aa98465
Β·
verified Β·
1 Parent(s): 7664bae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -17,7 +17,7 @@ Here's a precise analysis of YOYO-V2-dwq5's performance compared to the other qu
17
 
18
  Comparison Table (YOYO-V2 Quantized Variants)
19
  ```bash
20
- Task YOYO-V2-dwq5 YOYO-V2-dwq4 YOYO-V2-dwq3 YOYO-V2-q6
21
  arc_challenge 0.523 0.511 0.497 0.532
22
  arc_easy 0.682 0.655 0.657 0.685
23
  boolq 0.883 0.879 0.876 0.886
@@ -31,8 +31,8 @@ YOYO-V2-q6 scores are highest across all tasks in this dataset.
31
 
32
  πŸ“Š Critical Insights from YOYO-V2's Internal Quantization Comparison
33
 
34
- ```bash
35
  YOYO-V2-dwq5 Consistently Improves Over Lower-DWQ Variants
 
36
  DWQ5 surpasses dwq4 in all tasks (e.g., +0.002 on arc_easy, +0.007 on boolq).
37
  DWQ5 surpasses dwq3 in all tasks (e.g., +0.026 on arc_easy, +0.014 on boolq).
38
  ```
@@ -62,16 +62,16 @@ piqa: q6’s lead (0.782 vs dwq5’s 0.778) is 1.3% – critical for logic reas
62
 
63
  🎯 Practical Takeaways for Model Selection
64
  ```bash
65
- Quantization Type Best For Why
66
- YOYO-V2-dwq5 Hardware with moderate resources Best balance between speed and accuracy (e.g., 5-bit DWQ)
67
- YOYO-V2-q6 High-precision tasks (e.g., reasoning) Slightly better than dwq5 in 4+ tasks; optimal for stability
68
  ```
69
 
70
- For most use cases, YOYO-V2-q6 is still the top performer (1.3–2.0% edge over dwq5 in tasks like boolq and piqa).
71
 
72
- YOYO-V2-dwq5 is ideal if you need to reduce memory footprint while still achieving near-q6 performance (e.g., in edge devices).
73
 
74
- YOYO-V2-dwq5 outperforms the lower-DWQ quantizations (dwq3, dwq4) across all tasks, showing a clear progression in precision as the DWQ bitwidth increases from 3 β†’ 5 bits. However, it does not surpass YOYO-V2-q6 – instead, q6 maintains a small but consistent lead (0.005–0.013) in high-precision tasks like boolq and piqa.
75
 
76
  This confirms that YOYO-V2’s performance steadily improves with higher quantization fidelity within its own variants, but the fixed Q6 quantization still delivers edge gains for critical tasks where minor precision losses are unacceptable.
77
 
 
17
 
18
  Comparison Table (YOYO-V2 Quantized Variants)
19
  ```bash
20
+ Task dwq5 dwq4 dwq3 q6
21
  arc_challenge 0.523 0.511 0.497 0.532
22
  arc_easy 0.682 0.655 0.657 0.685
23
  boolq 0.883 0.879 0.876 0.886
 
31
 
32
  πŸ“Š Critical Insights from YOYO-V2's Internal Quantization Comparison
33
 
 
34
  YOYO-V2-dwq5 Consistently Improves Over Lower-DWQ Variants
35
+ ```bash
36
  DWQ5 surpasses dwq4 in all tasks (e.g., +0.002 on arc_easy, +0.007 on boolq).
37
  DWQ5 surpasses dwq3 in all tasks (e.g., +0.026 on arc_easy, +0.014 on boolq).
38
  ```
 
62
 
63
  🎯 Practical Takeaways for Model Selection
64
  ```bash
65
+ Quant Best For Why
66
+ dwq5 Hardware with moderate resources Best balance between speed and accuracy (e.g., 5-bit DWQ)
67
+ q6 High-precision tasks (e.g., reasoning) Slightly better than dwq5 in 4+ tasks; optimal for stability
68
  ```
69
 
70
+ For most use cases, q6 is still the top performer (1.3–2.0% edge over dwq5 in tasks like boolq and piqa).
71
 
72
+ dwq5 is ideal if you need to reduce memory footprint while still achieving near-q6 performance (e.g., in edge devices).
73
 
74
+ dwq5 outperforms the lower-DWQ quantizations (dwq3, dwq4) across all tasks, showing a clear progression in precision as the DWQ bitwidth increases from 3 β†’ 5 bits. However, it does not surpass YOYO-V2-q6 – instead, q6 maintains a small but consistent lead (0.005–0.013) in high-precision tasks like boolq and piqa.
75
 
76
  This confirms that YOYO-V2’s performance steadily improves with higher quantization fidelity within its own variants, but the fixed Q6 quantization still delivers edge gains for critical tasks where minor precision losses are unacceptable.
77