kesenZhaoNTU commited on
Commit
5ee53d9
·
verified ·
1 Parent(s): 47bf470

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -22,14 +22,14 @@ Chain-of-thought (CoT) reasoning greatly improves the interpretability and probl
22
 
23
  UV-CoT achieves this by performing preference comparisons between model-generated bounding boxes. It generates preference data automatically, then uses an evaluator MLLM (e.g., OmniLLM-12B) to rank responses, which serves as supervision to train the target MLLM (e.g., LLaVA-1.5-7B). This approach emulates human perception—identifying key regions and reasoning based on them—thereby improving visual comprehension, particularly in spatial reasoning tasks.
24
 
25
- ![Figure 1: UV-CoT Overview](https://raw.githubusercontent.com/UV-CoT/UV-CoT/main/images/fig1.png)
26
 
27
  ## Visualizations
28
 
29
  Qualitative examples demonstrating UV-CoT's visual reasoning:
30
 
31
- ![Figure 5: UV-CoT Visualization 1](https://raw.githubusercontent.com/UV-CoT/UV-CoT/main/images/fig5.png)
32
- ![Figure 6: UV-CoT Visualization 2](https://raw.githubusercontent.com/UV-CoT/UV-CoT/main/images/fig6.png)
33
 
34
  ## Installation
35
 
 
22
 
23
  UV-CoT achieves this by performing preference comparisons between model-generated bounding boxes. It generates preference data automatically, then uses an evaluator MLLM (e.g., OmniLLM-12B) to rank responses, which serves as supervision to train the target MLLM (e.g., LLaVA-1.5-7B). This approach emulates human perception—identifying key regions and reasoning based on them—thereby improving visual comprehension, particularly in spatial reasoning tasks.
24
 
25
+ ![Figure 1: UV-CoT Overview](./images/fig1.svg)
26
 
27
  ## Visualizations
28
 
29
  Qualitative examples demonstrating UV-CoT's visual reasoning:
30
 
31
+ ![Figure 5: UV-CoT Visualization 1](./images/fig5_v1.2.svg)
32
+ ![Figure 6: UV-CoT Visualization 2](./images/fig6_v1.2.svg)
33
 
34
  ## Installation
35