czczup commited on
Commit
80a1f7b
1 Parent(s): 4b5ced9

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ examples/red-panda.mp4 filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,659 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ pipeline_tag: image-text-to-text
4
+ library_name: transformers
5
+ base_model:
6
+ - OpenGVLab/InternViT-300M-448px-V2_5
7
+ - Qwen/Qwen2.5-0.5B-Instruct
8
+ base_model_relation: merge
9
+ language:
10
+ - multilingual
11
+ tags:
12
+ - internvl
13
+ - custom_code
14
+ ---
15
+
16
+ # InternVL2_5-1B
17
+
18
+ [\[📂 GitHub\]](https://github.com/OpenGVLab/InternVL) [\[📜 InternVL 1.0\]](https://huggingface.co/papers/2312.14238) [\[📜 InternVL 1.5\]](https://huggingface.co/papers/2404.16821) [\[📜 Mini-InternVL\]](https://arxiv.org/abs/2410.16261) [\[📜 InternVL 2.5\]](https://huggingface.co/papers/2412.05271) [\[📜 InternVL2.5-MPO\]](https://huggingface.co/papers/2411.10442)
19
+
20
+ [\[🆕 Blog\]](https://internvl.github.io/blog/) [\[🗨️ Chat Demo\]](https://internvl.opengvlab.com/) [\[🤗 HF Demo\]](https://huggingface.co/spaces/OpenGVLab/InternVL) [\[🚀 Quick Start\]](#quick-start) [\[📖 Documents\]](https://internvl.readthedocs.io/en/latest/)
21
+
22
+ <div align="center">
23
+ <img width="500" alt="image" src="https://cdn-uploads.huggingface.co/production/uploads/64006c09330a45b03605bba3/zJsd2hqd3EevgXo6fNgC-.png">
24
+ </div>
25
+
26
+ ## Introduction
27
+
28
+ We introduce InternVL2.5-MPO, an advanced multimodal large language model (MLLM) series that demonstrates superior overall performance. This series builds upon InternVL2.5 and Mixed Preference Optimization.
29
+
30
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/619507e7b74b6c591f794340/WRe2VxD7bi-YMh1cLz9E2.jpeg)
31
+
32
+ ## InternVL 2.5 Family
33
+
34
+ In the following table, we provide an overview of the InternVL2.5-MPO series.
35
+
36
+ | Model Name | Vision Part | Language Part | HF Link |
37
+ | :-----------------: | :-------------------------------------------------------------------------------------: | :----------------------------------------------------------------------------: | :------------------------------------------------------------: |
38
+ | InternVL2_5-1B-MPO | [InternViT-300M-448px-V2_5](https://huggingface.co/OpenGVLab/InternViT-300M-448px-V2_5) | [Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) | [🤗 link](https://huggingface.co/OpenGVLab/InternVL2_5-1B-MPO) |
39
+ | InternVL2_5-2B-MPO | [InternViT-300M-448px-V2_5](https://huggingface.co/OpenGVLab/InternViT-300M-448px-V2_5) | [internlm2_5-1_8b-chat](https://huggingface.co/internlm/internlm2_5-1_8b-chat) | [🤗 link](https://huggingface.co/OpenGVLab/InternVL2_5-2B-MPO) |
40
+ | InternVL2_5-4B-MPO | [InternViT-300M-448px-V2_5](https://huggingface.co/OpenGVLab/InternViT-300M-448px-V2_5) | [Qwen2.5-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-3B-Instruct) | [🤗 link](https://huggingface.co/OpenGVLab/InternVL2_5-4B-MPO) |
41
+ | InternVL2_5-8B-MPO | [InternViT-300M-448px-V2_5](https://huggingface.co/OpenGVLab/InternViT-300M-448px-V2_5) | [internlm2_5-7b-chat](https://huggingface.co/internlm/internlm2_5-7b-chat) | [🤗 link](https://huggingface.co/OpenGVLab/InternVL2_5-8B-MPO) |
42
+ | InternVL2_5-26B-MPO | [InternViT-6B-448px-V2_5](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V2_5) | [internlm2_5-20b-chat](https://huggingface.co/internlm/internlm2_5-20b-chat) | [🤗 link](https://huggingface.co/OpenGVLab/InternVL2_5-26B-MPO) |
43
+ | InternVL2_5-38B-MPO | [InternViT-6B-448px-V2_5](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V2_5) | [Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) | [🤗 link](https://huggingface.co/OpenGVLab/InternVL2_5-38B-MPO) |
44
+ | InternVL2_5-78B-MPO | [InternViT-6B-448px-V2_5](https://huggingface.co/OpenGVLab/InternViT-6B-448px-V2_5) | [Qwen2.5-72B-Instruct](https://huggingface.co/Qwen/Qwen2.5-72B-Instruct) | [🤗 link](https://huggingface.co/OpenGVLab/InternVL2_5-78B-MPO) |
45
+
46
+ ## Model Architecture
47
+
48
+ As shown in the following figure, [InternVL2.5-MPO](https://internvl.github.io/blog/2024-12-20-InternVL-2.5-MPO/) retains the same model architecture as [InternVL 2.5](https://internvl.github.io/blog/2024-12-05-InternVL-2.5/) and its predecessors, InternVL 1.5 and 2.0, following the "ViT-MLP-LLM" paradigm. In this new version, we integrate a newly incrementally pre-trained InternViT with various pre-trained LLMs, including InternLM 2.5 and Qwen 2.5, using a randomly initialized MLP projector.
49
+
50
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64119264f0f81eb569e0d569/BiiyXN6NOk0p-3rl3ueyL.png)
51
+
52
+ As in the previous version, we applied a pixel unshuffle operation, reducing the number of visual tokens to one-quarter of the original. Besides, we adopted a similar dynamic resolution strategy as InternVL 1.5, dividing images into tiles of 448×448 pixels. The key difference, starting from InternVL 2.0, is that we additionally introduced support for multi-image and video data.
53
+
54
+ ## Key Designs
55
+
56
+ ### Multi-Modal Preference Dataset
57
+
58
+ MMPR is a large-scale and high-quality multimodal reasoning preference dataset. This dataset includes about 3 million samples.
59
+
60
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/619507e7b74b6c591f794340/mmXL47UPDFwYOWdn9Z6j5.jpeg)
61
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/619507e7b74b6c591f794340/6fnvI_wCd9JXAs6vYthaG.jpeg)
62
+
63
+ To construct this dataset, we propose an efficient data construction pipeline. Specifically, we categorize the multimodal data into **samples with clear ground truths** and **samples without clear ground truths**.
64
+
65
+ - **For samples with clear ground truths:**
66
+ the model is prompted to first provide the reasoning process and then give the final answer in the format like `Final Answer: ***`.
67
+ Responses matching the ground truth answer constitute the positive set $\mathcal{Y}_p$, while those that do not match make up the negative set $\mathcal{Y}_n$. Additionally, responses that fail to provide a clear final answer are also merged into $\mathcal{Y}_n$.
68
+ Given these responses labeled as positive or negative, we build the preference pairs by selecting a chosen response $y_c$ from $\mathcal{Y}_p$ and a negative response $y_r$ from $\mathcal{Y}_n$.
69
+
70
+ - **For samples without clear ground truths:**
71
+ we propose a simple yet effective method: Dropout Next-Token Prediction (Dropout NTP).
72
+ Specifically, we use the responses generated by InternVL2-8B as chosen answers.
73
+ Given the chosen answer, we truncate it by half and then prompt InternVL2-8B to complete the remaining
74
+ portion of the truncated answer without access to the image input.
75
+ This generated completion serves as the rejected answer for the paired sample.
76
+ It is worth noting that while the responses generated by InternVL2-8B may not be perfect,
77
+ the completions generated without the image input will introduce more hallucinations than those
78
+ generated with the image input.
79
+ Therefore, the partial order relationship between the chosen and rejected responses holds true.
80
+
81
+ The data construction pipeline is open-sourced, see more details in our [document](https://internvl.readthedocs.io/en/latest/internvl2.0/preference_optimization.html#generate-additional-preference-data).
82
+
83
+
84
+ ### Mixed Preference Optimization
85
+
86
+ The key insight behind MPO is that *an effective PO process should enable the model to learn the relative preference between pairs of responses, the absolute quality of individual responses, and the process for generating preferred responses.* We define the training objective as a combination of
87
+ preference loss $\mathcal{L}_{\text{p}}$,
88
+ quality loss $\mathcal{L}_{\text{q}}$,
89
+ and generation loss $\mathcal{L}_{\text{g}}$,
90
+ referred to as Mixed Preference Optimization:
91
+
92
+ $$
93
+ \mathcal{L}=w_{p}\cdot\mathcal{L}_{\text{p}} + w_{q}\cdot\mathcal{L}_{\text{q}} + w_{g}\cdot\mathcal{L}_{\text{g}},
94
+ $$
95
+
96
+ where $w_{*}$ represents the weight assigned to each loss component.
97
+ In this work, we empirically compare different variants of preference loss.
98
+ Based on the experimental results, we use DPO as our preference loss and BCO as our quality loss.
99
+
100
+ Specifically, the DPO serves as the preference loss to enable the model to learn the
101
+ relative preference between chosen and rejected responses.
102
+ This algorithm optimizes the following loss function:
103
+
104
+ $$
105
+ \mathcal{L}_{\text{p}}=-\log \sigma\left(\beta \log \frac{\pi_\theta\left(y_c \mid x\right)}{\pi_0\left(y_c \mid x\right)}-\beta \log \frac{\pi_\theta\left(y_r \mid x\right)}{\pi_0\left(y_r \mid x\right)}\right),
106
+ $$
107
+
108
+ where $\beta$ is the KL penalty coefficient, and $x$, $y_c$, and $y_r$ are user query, chosen response, and rejected response, respectively.
109
+ The policy model $\pi_\theta$ is initialized from model $\pi_0$.
110
+
111
+ Additionally, the BCO loss is employed as the quality loss, which helps the model to understand the absolute quality of individual responses.
112
+ The loss function is defined as:
113
+
114
+ $$
115
+ \mathcal{L}_{\text{q}}=\mathcal{L}_{\text{q}}^+ + \mathcal{L}_{\text{q}}^-,
116
+ $$
117
+
118
+ where $\mathcal{L}_{\text{q}}^{+}$ and $\mathcal{L}_{\text{q}}^{+}$ represent the loss for chosen and rejected responses, respectively.
119
+ Each response type's loss is calculated independently, requiring the model to differentiate the absolute quality of individual responses. The loss terms are given by:
120
+
121
+ $$
122
+ \mathcal{L}_{\text{q}}^+=-\log \sigma\left(\beta \log \frac{\pi_\theta\left(y_c \mid x\right)}{\pi_0\left(y_c \mid x\right)} - \delta\right),
123
+ $$
124
+
125
+ $$
126
+ \mathcal{L}_{\text{q}}^-=-\log \sigma\left(-\left(\beta \log \frac{\pi_\theta\left(y_r \mid x\right)}{\pi_0\left(y_r \mid x\right)} - \delta\right) \right),
127
+ $$
128
+
129
+ where $\delta$ represents the reward shift, calculated as the moving average of previous rewards to stabilize training.
130
+
131
+ Finally, the SFT loss is used as the generation loss to help the model learn the generation process of preferred responses.
132
+ The loss function is defined as:
133
+
134
+ $$
135
+ \mathcal{L}_{\text{gen}}=-\frac{\log\pi_\theta\left(y_c \mid x\right)}{\left| y_c \right|}.
136
+ $$
137
+
138
+ ## Evaluation on Multimodal Capability
139
+
140
+ To comprehensively compare InternVL's performance before and after MPO, we employ the benchmarks from OpenCompass Learderboard, including both well-established classic datasets and newly introduced ones. These benchmarks span a wide range of categories, aiming to provide a thorough and balanced assessment of InternVL’s capabilities across various multimodal tasks. We provide the evaluation results in the tables behind.
141
+
142
+ | Model | Avg. | MMBench v1.1 | MMStar | MMMU | MathVista | HallusionBench | AI2D | OCRBench | MMVet |
143
+ | ------------------- | ---- | ------------ | ------ | ---- | --------- | -------------- | ---- | -------- | ----- |
144
+ | InternVL2-5-1B | 54.9 | 66.5 | 51.3 | 41.2 | 47.1 | 39.4 | 69.0 | 77.4 | 47.2 |
145
+ | InternVL2-5-1B-MPO | 56.4 | 67.2 | 49.7 | 40.8 | 53.0 | 40.0 | 69.4 | 83.6 | 47.2 |
146
+ | InternVL2-5-2B | 59.9 | 70.9 | 54.3 | 43.2 | 51.1 | 42.3 | 74.9 | 80.2 | 62.6 |
147
+ | InternVL2-5-2B-MPO | 62.0 | 71.6 | 55.0 | 45.0 | 56.4 | 43.0 | 75.3 | 84.2 | 65.4 |
148
+ | InternVL2-5-4B | 65.1 | 78.2 | 58.7 | 51.8 | 60.8 | 46.6 | 81.4 | 82.0 | 61.5 |
149
+ | InternVL2-5-4B-MPO | 67.6 | 78.6 | 60.2 | 51.6 | 65.3 | 47.8 | 82.0 | 88.0 | 67.1 |
150
+ | InternVL2-5-8B | 68.9 | 82.5 | 63.2 | 56.2 | 64.5 | 49.0 | 84.6 | 82.1 | 62.8 |
151
+ | InternVL2-5-8B-MPO | 70.4 | 82.4 | 65.7 | 54.9 | 68.9 | 51.4 | 84.5 | 88.3 | 66.9 |
152
+ | InternVL2-5-26B | 71.6 | 84.6 | 66.5 | 60.7 | 68.0 | 55.8 | 86.2 | 85.4 | 65.4 |
153
+ | InternVL2-5-26B-MPO | 72.7 | 84.2 | 67.2 | 57.7 | 72.8 | 55.3 | 86.2 | 91.2 | 67.1 |
154
+ | InternVL2-5-38B | 73.5 | 85.4 | 68.5 | 64.6 | 72.4 | 57.9 | 87.6 | 84.1 | 67.2 |
155
+ | InternVL2-5-38B-MPO | 75.5 | 85.6 | 69.8 | 64.1 | 73.8 | 61.5 | 88.1 | 88.5 | 72.5 |
156
+ | InternVL2-5-78B | 75.2 | 87.5 | 69.5 | 70.0 | 70.6 | 57.4 | 89.1 | 85.3 | 71.8 |
157
+ | InternVL2-5-78B-MPO | 76.6 | 87.3 | 73.1 | 68.3 | 73.8 | 58.7 | 89.3 | 91.2 | 71.4 |
158
+
159
+
160
+ ## Quick Start
161
+
162
+ We provide an example code to run `InternVL2_5-1B` using `transformers`.
163
+
164
+ > Please use transformers>=4.37.2 to ensure the model works normally.
165
+
166
+ ### Model Loading
167
+
168
+ #### 16-bit (bf16 / fp16)
169
+
170
+ ```python
171
+ import torch
172
+ from transformers import AutoTokenizer, AutoModel
173
+ path = "OpenGVLab/InternVL2_5-1B"
174
+ model = AutoModel.from_pretrained(
175
+ path,
176
+ torch_dtype=torch.bfloat16,
177
+ low_cpu_mem_usage=True,
178
+ use_flash_attn=True,
179
+ trust_remote_code=True).eval().cuda()
180
+ ```
181
+
182
+ #### BNB 8-bit Quantization
183
+
184
+ ```python
185
+ import torch
186
+ from transformers import AutoTokenizer, AutoModel
187
+ path = "OpenGVLab/InternVL2_5-1B"
188
+ model = AutoModel.from_pretrained(
189
+ path,
190
+ torch_dtype=torch.bfloat16,
191
+ load_in_8bit=True,
192
+ low_cpu_mem_usage=True,
193
+ use_flash_attn=True,
194
+ trust_remote_code=True).eval()
195
+ ```
196
+
197
+ #### Multiple GPUs
198
+
199
+ The reason for writing the code this way is to avoid errors that occur during multi-GPU inference due to tensors not being on the same device. By ensuring that the first and last layers of the large language model (LLM) are on the same device, we prevent such errors.
200
+
201
+ ```python
202
+ import math
203
+ import torch
204
+ from transformers import AutoTokenizer, AutoModel
205
+
206
+ def split_model(model_name):
207
+ device_map = {}
208
+ world_size = torch.cuda.device_count()
209
+ num_layers = {
210
+ 'InternVL2_5-1B': 24, 'InternVL2_5-2B': 24, 'InternVL2_5-4B': 36, 'InternVL2_5-8B': 32,
211
+ 'InternVL2_5-26B': 48, 'InternVL2_5-38B': 64, 'InternVL2_5-78B': 80}[model_name]
212
+ # Since the first GPU will be used for ViT, treat it as half a GPU.
213
+ num_layers_per_gpu = math.ceil(num_layers / (world_size - 0.5))
214
+ num_layers_per_gpu = [num_layers_per_gpu] * world_size
215
+ num_layers_per_gpu[0] = math.ceil(num_layers_per_gpu[0] * 0.5)
216
+ layer_cnt = 0
217
+ for i, num_layer in enumerate(num_layers_per_gpu):
218
+ for j in range(num_layer):
219
+ device_map[f'language_model.model.layers.{layer_cnt}'] = i
220
+ layer_cnt += 1
221
+ device_map['vision_model'] = 0
222
+ device_map['mlp1'] = 0
223
+ device_map['language_model.model.tok_embeddings'] = 0
224
+ device_map['language_model.model.embed_tokens'] = 0
225
+ device_map['language_model.output'] = 0
226
+ device_map['language_model.model.norm'] = 0
227
+ device_map['language_model.lm_head'] = 0
228
+ device_map[f'language_model.model.layers.{num_layers - 1}'] = 0
229
+
230
+ return device_map
231
+
232
+ path = "OpenGVLab/InternVL2_5-1B"
233
+ device_map = split_model('InternVL2_5-1B')
234
+ model = AutoModel.from_pretrained(
235
+ path,
236
+ torch_dtype=torch.bfloat16,
237
+ low_cpu_mem_usage=True,
238
+ use_flash_attn=True,
239
+ trust_remote_code=True,
240
+ device_map=device_map).eval()
241
+ ```
242
+
243
+ ### Inference with Transformers
244
+
245
+ ```python
246
+ import numpy as np
247
+ import torch
248
+ import torchvision.transforms as T
249
+ from decord import VideoReader, cpu
250
+ from PIL import Image
251
+ from torchvision.transforms.functional import InterpolationMode
252
+ from transformers import AutoModel, AutoTokenizer
253
+
254
+ IMAGENET_MEAN = (0.485, 0.456, 0.406)
255
+ IMAGENET_STD = (0.229, 0.224, 0.225)
256
+
257
+ def build_transform(input_size):
258
+ MEAN, STD = IMAGENET_MEAN, IMAGENET_STD
259
+ transform = T.Compose([
260
+ T.Lambda(lambda img: img.convert('RGB') if img.mode != 'RGB' else img),
261
+ T.Resize((input_size, input_size), interpolation=InterpolationMode.BICUBIC),
262
+ T.ToTensor(),
263
+ T.Normalize(mean=MEAN, std=STD)
264
+ ])
265
+ return transform
266
+
267
+ def find_closest_aspect_ratio(aspect_ratio, target_ratios, width, height, image_size):
268
+ best_ratio_diff = float('inf')
269
+ best_ratio = (1, 1)
270
+ area = width * height
271
+ for ratio in target_ratios:
272
+ target_aspect_ratio = ratio[0] / ratio[1]
273
+ ratio_diff = abs(aspect_ratio - target_aspect_ratio)
274
+ if ratio_diff < best_ratio_diff:
275
+ best_ratio_diff = ratio_diff
276
+ best_ratio = ratio
277
+ elif ratio_diff == best_ratio_diff:
278
+ if area > 0.5 * image_size * image_size * ratio[0] * ratio[1]:
279
+ best_ratio = ratio
280
+ return best_ratio
281
+
282
+ def dynamic_preprocess(image, min_num=1, max_num=12, image_size=448, use_thumbnail=False):
283
+ orig_width, orig_height = image.size
284
+ aspect_ratio = orig_width / orig_height
285
+
286
+ # calculate the existing image aspect ratio
287
+ target_ratios = set(
288
+ (i, j) for n in range(min_num, max_num + 1) for i in range(1, n + 1) for j in range(1, n + 1) if
289
+ i * j <= max_num and i * j >= min_num)
290
+ target_ratios = sorted(target_ratios, key=lambda x: x[0] * x[1])
291
+
292
+ # find the closest aspect ratio to the target
293
+ target_aspect_ratio = find_closest_aspect_ratio(
294
+ aspect_ratio, target_ratios, orig_width, orig_height, image_size)
295
+
296
+ # calculate the target width and height
297
+ target_width = image_size * target_aspect_ratio[0]
298
+ target_height = image_size * target_aspect_ratio[1]
299
+ blocks = target_aspect_ratio[0] * target_aspect_ratio[1]
300
+
301
+ # resize the image
302
+ resized_img = image.resize((target_width, target_height))
303
+ processed_images = []
304
+ for i in range(blocks):
305
+ box = (
306
+ (i % (target_width // image_size)) * image_size,
307
+ (i // (target_width // image_size)) * image_size,
308
+ ((i % (target_width // image_size)) + 1) * image_size,
309
+ ((i // (target_width // image_size)) + 1) * image_size
310
+ )
311
+ # split the image
312
+ split_img = resized_img.crop(box)
313
+ processed_images.append(split_img)
314
+ assert len(processed_images) == blocks
315
+ if use_thumbnail and len(processed_images) != 1:
316
+ thumbnail_img = image.resize((image_size, image_size))
317
+ processed_images.append(thumbnail_img)
318
+ return processed_images
319
+
320
+ def load_image(image_file, input_size=448, max_num=12):
321
+ image = Image.open(image_file).convert('RGB')
322
+ transform = build_transform(input_size=input_size)
323
+ images = dynamic_preprocess(image, image_size=input_size, use_thumbnail=True, max_num=max_num)
324
+ pixel_values = [transform(image) for image in images]
325
+ pixel_values = torch.stack(pixel_values)
326
+ return pixel_values
327
+
328
+ # If you want to load a model using multiple GPUs, please refer to the `Multiple GPUs` section.
329
+ path = 'OpenGVLab/InternVL2_5-1B'
330
+ model = AutoModel.from_pretrained(
331
+ path,
332
+ torch_dtype=torch.bfloat16,
333
+ low_cpu_mem_usage=True,
334
+ use_flash_attn=True,
335
+ trust_remote_code=True).eval().cuda()
336
+ tokenizer = AutoTokenizer.from_pretrained(path, trust_remote_code=True, use_fast=False)
337
+
338
+ # set the max number of tiles in `max_num`
339
+ pixel_values = load_image('./examples/image1.jpg', max_num=12).to(torch.bfloat16).cuda()
340
+ generation_config = dict(max_new_tokens=1024, do_sample=True)
341
+
342
+ # pure-text conversation (纯文本对话)
343
+ question = 'Hello, who are you?'
344
+ response, history = model.chat(tokenizer, None, question, generation_config, history=None, return_history=True)
345
+ print(f'User: {question}\nAssistant: {response}')
346
+
347
+ question = 'Can you tell me a story?'
348
+ response, history = model.chat(tokenizer, None, question, generation_config, history=history, return_history=True)
349
+ print(f'User: {question}\nAssistant: {response}')
350
+
351
+ # single-image single-round conversation (单图单轮对话)
352
+ question = '<image>\nPlease describe the image shortly.'
353
+ response = model.chat(tokenizer, pixel_values, question, generation_config)
354
+ print(f'User: {question}\nAssistant: {response}')
355
+
356
+ # single-image multi-round conversation (单图多轮对话)
357
+ question = '<image>\nPlease describe the image in detail.'
358
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config, history=None, return_history=True)
359
+ print(f'User: {question}\nAssistant: {response}')
360
+
361
+ question = 'Please write a poem according to the image.'
362
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config, history=history, return_history=True)
363
+ print(f'User: {question}\nAssistant: {response}')
364
+
365
+ # multi-image multi-round conversation, combined images (多图多轮对话,拼接图像)
366
+ pixel_values1 = load_image('./examples/image1.jpg', max_num=12).to(torch.bfloat16).cuda()
367
+ pixel_values2 = load_image('./examples/image2.jpg', max_num=12).to(torch.bfloat16).cuda()
368
+ pixel_values = torch.cat((pixel_values1, pixel_values2), dim=0)
369
+
370
+ question = '<image>\nDescribe the two images in detail.'
371
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config,
372
+ history=None, return_history=True)
373
+ print(f'User: {question}\nAssistant: {response}')
374
+
375
+ question = 'What are the similarities and differences between these two images.'
376
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config,
377
+ history=history, return_history=True)
378
+ print(f'User: {question}\nAssistant: {response}')
379
+
380
+ # multi-image multi-round conversation, separate images (多图多轮对话,独立图像)
381
+ pixel_values1 = load_image('./examples/image1.jpg', max_num=12).to(torch.bfloat16).cuda()
382
+ pixel_values2 = load_image('./examples/image2.jpg', max_num=12).to(torch.bfloat16).cuda()
383
+ pixel_values = torch.cat((pixel_values1, pixel_values2), dim=0)
384
+ num_patches_list = [pixel_values1.size(0), pixel_values2.size(0)]
385
+
386
+ question = 'Image-1: <image>\nImage-2: <image>\nDescribe the two images in detail.'
387
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config,
388
+ num_patches_list=num_patches_list,
389
+ history=None, return_history=True)
390
+ print(f'User: {question}\nAssistant: {response}')
391
+
392
+ question = 'What are the similarities and differences between these two images.'
393
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config,
394
+ num_patches_list=num_patches_list,
395
+ history=history, return_history=True)
396
+ print(f'User: {question}\nAssistant: {response}')
397
+
398
+ # batch inference, single image per sample (单图批处理)
399
+ pixel_values1 = load_image('./examples/image1.jpg', max_num=12).to(torch.bfloat16).cuda()
400
+ pixel_values2 = load_image('./examples/image2.jpg', max_num=12).to(torch.bfloat16).cuda()
401
+ num_patches_list = [pixel_values1.size(0), pixel_values2.size(0)]
402
+ pixel_values = torch.cat((pixel_values1, pixel_values2), dim=0)
403
+
404
+ questions = ['<image>\nDescribe the image in detail.'] * len(num_patches_list)
405
+ responses = model.batch_chat(tokenizer, pixel_values,
406
+ num_patches_list=num_patches_list,
407
+ questions=questions,
408
+ generation_config=generation_config)
409
+ for question, response in zip(questions, responses):
410
+ print(f'User: {question}\nAssistant: {response}')
411
+
412
+ # video multi-round conversation (视频多轮对话)
413
+ def get_index(bound, fps, max_frame, first_idx=0, num_segments=32):
414
+ if bound:
415
+ start, end = bound[0], bound[1]
416
+ else:
417
+ start, end = -100000, 100000
418
+ start_idx = max(first_idx, round(start * fps))
419
+ end_idx = min(round(end * fps), max_frame)
420
+ seg_size = float(end_idx - start_idx) / num_segments
421
+ frame_indices = np.array([
422
+ int(start_idx + (seg_size / 2) + np.round(seg_size * idx))
423
+ for idx in range(num_segments)
424
+ ])
425
+ return frame_indices
426
+
427
+ def load_video(video_path, bound=None, input_size=448, max_num=1, num_segments=32):
428
+ vr = VideoReader(video_path, ctx=cpu(0), num_threads=1)
429
+ max_frame = len(vr) - 1
430
+ fps = float(vr.get_avg_fps())
431
+
432
+ pixel_values_list, num_patches_list = [], []
433
+ transform = build_transform(input_size=input_size)
434
+ frame_indices = get_index(bound, fps, max_frame, first_idx=0, num_segments=num_segments)
435
+ for frame_index in frame_indices:
436
+ img = Image.fromarray(vr[frame_index].asnumpy()).convert('RGB')
437
+ img = dynamic_preprocess(img, image_size=input_size, use_thumbnail=True, max_num=max_num)
438
+ pixel_values = [transform(tile) for tile in img]
439
+ pixel_values = torch.stack(pixel_values)
440
+ num_patches_list.append(pixel_values.shape[0])
441
+ pixel_values_list.append(pixel_values)
442
+ pixel_values = torch.cat(pixel_values_list)
443
+ return pixel_values, num_patches_list
444
+
445
+ video_path = './examples/red-panda.mp4'
446
+ pixel_values, num_patches_list = load_video(video_path, num_segments=8, max_num=1)
447
+ pixel_values = pixel_values.to(torch.bfloat16).cuda()
448
+ video_prefix = ''.join([f'Frame{i+1}: <image>\n' for i in range(len(num_patches_list))])
449
+ question = video_prefix + 'What is the red panda doing?'
450
+ # Frame1: <image>\nFrame2: <image>\n...\nFrame8: <image>\n{question}
451
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config,
452
+ num_patches_list=num_patches_list, history=None, return_history=True)
453
+ print(f'User: {question}\nAssistant: {response}')
454
+
455
+ question = 'Describe this video in detail.'
456
+ response, history = model.chat(tokenizer, pixel_values, question, generation_config,
457
+ num_patches_list=num_patches_list, history=history, return_history=True)
458
+ print(f'User: {question}\nAssistant: {response}')
459
+ ```
460
+
461
+ #### Streaming Output
462
+
463
+ Besides this method, you can also use the following code to get streamed output.
464
+
465
+ ```python
466
+ from transformers import TextIteratorStreamer
467
+ from threading import Thread
468
+
469
+ # Initialize the streamer
470
+ streamer = TextIteratorStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True, timeout=10)
471
+ # Define the generation configuration
472
+ generation_config = dict(max_new_tokens=1024, do_sample=False, streamer=streamer)
473
+ # Start the model chat in a separate thread
474
+ thread = Thread(target=model.chat, kwargs=dict(
475
+ tokenizer=tokenizer, pixel_values=pixel_values, question=question,
476
+ history=None, return_history=False, generation_config=generation_config,
477
+ ))
478
+ thread.start()
479
+
480
+ # Initialize an empty string to store the generated text
481
+ generated_text = ''
482
+ # Loop through the streamer to get the new text as it is generated
483
+ for new_text in streamer:
484
+ if new_text == model.conv_template.sep:
485
+ break
486
+ generated_text += new_text
487
+ print(new_text, end='', flush=True) # Print each new chunk of generated text on the same line
488
+ ```
489
+
490
+ ## Finetune
491
+
492
+ Many repositories now support fine-tuning of the InternVL series models, including [InternVL](https://github.com/OpenGVLab/InternVL), [SWIFT](https://github.com/modelscope/ms-swift), [XTurner](https://github.com/InternLM/xtuner), and others. Please refer to their documentation for more details on fine-tuning.
493
+
494
+ ## Deployment
495
+
496
+ ### LMDeploy
497
+
498
+ LMDeploy is a toolkit for compressing, deploying, and serving LLMs & VLMs.
499
+
500
+ ```sh
501
+ pip install lmdeploy>=0.6.4
502
+ ```
503
+
504
+ LMDeploy abstracts the complex inference process of multi-modal Vision-Language Models (VLM) into an easy-to-use pipeline, similar to the Large Language Model (LLM) inference pipeline.
505
+
506
+ #### A 'Hello, world' Example
507
+
508
+ ```python
509
+ from lmdeploy import pipeline, TurbomindEngineConfig
510
+ from lmdeploy.vl import load_image
511
+
512
+ model = 'OpenGVLab/InternVL2_5-1B'
513
+ image = load_image('https://raw.githubusercontent.com/open-mmlab/mmdeploy/main/tests/data/tiger.jpeg')
514
+ pipe = pipeline(model, backend_config=TurbomindEngineConfig(session_len=8192))
515
+ response = pipe(('describe this image', image))
516
+ print(response.text)
517
+ ```
518
+
519
+ If `ImportError` occurs while executing this case, please install the required dependency packages as prompted.
520
+
521
+ #### Multi-images Inference
522
+
523
+ When dealing with multiple images, you can put them all in one list. Keep in mind that multiple images will lead to a higher number of input tokens, and as a result, the size of the context window typically needs to be increased.
524
+
525
+ ```python
526
+ from lmdeploy import pipeline, TurbomindEngineConfig
527
+ from lmdeploy.vl import load_image
528
+ from lmdeploy.vl.constants import IMAGE_TOKEN
529
+
530
+ model = 'OpenGVLab/InternVL2_5-1B'
531
+ pipe = pipeline(model, backend_config=TurbomindEngineConfig(session_len=8192))
532
+
533
+ image_urls=[
534
+ 'https://raw.githubusercontent.com/open-mmlab/mmdeploy/main/demo/resources/human-pose.jpg',
535
+ 'https://raw.githubusercontent.com/open-mmlab/mmdeploy/main/demo/resources/det.jpg'
536
+ ]
537
+
538
+ images = [load_image(img_url) for img_url in image_urls]
539
+ # Numbering images improves multi-image conversations
540
+ response = pipe((f'Image-1: {IMAGE_TOKEN}\nImage-2: {IMAGE_TOKEN}\ndescribe these two images', images))
541
+ print(response.text)
542
+ ```
543
+
544
+ #### Batch Prompts Inference
545
+
546
+ Conducting inference with batch prompts is quite straightforward; just place them within a list structure:
547
+
548
+ ```python
549
+ from lmdeploy import pipeline, TurbomindEngineConfig
550
+ from lmdeploy.vl import load_image
551
+
552
+ model = 'OpenGVLab/InternVL2_5-1B'
553
+ pipe = pipeline(model, backend_config=TurbomindEngineConfig(session_len=8192))
554
+
555
+ image_urls=[
556
+ "https://raw.githubusercontent.com/open-mmlab/mmdeploy/main/demo/resources/human-pose.jpg",
557
+ "https://raw.githubusercontent.com/open-mmlab/mmdeploy/main/demo/resources/det.jpg"
558
+ ]
559
+ prompts = [('describe this image', load_image(img_url)) for img_url in image_urls]
560
+ response = pipe(prompts)
561
+ print(response)
562
+ ```
563
+
564
+ #### Multi-turn Conversation
565
+
566
+ There are two ways to do the multi-turn conversations with the pipeline. One is to construct messages according to the format of OpenAI and use above introduced method, the other is to use the `pipeline.chat` interface.
567
+
568
+ ```python
569
+ from lmdeploy import pipeline, TurbomindEngineConfig, GenerationConfig
570
+ from lmdeploy.vl import load_image
571
+
572
+ model = 'OpenGVLab/InternVL2_5-1B'
573
+ pipe = pipeline(model, backend_config=TurbomindEngineConfig(session_len=8192))
574
+
575
+ image = load_image('https://raw.githubusercontent.com/open-mmlab/mmdeploy/main/demo/resources/human-pose.jpg')
576
+ gen_config = GenerationConfig(top_k=40, top_p=0.8, temperature=0.8)
577
+ sess = pipe.chat(('describe this image', image), gen_config=gen_config)
578
+ print(sess.response.text)
579
+ sess = pipe.chat('What is the woman doing?', session=sess, gen_config=gen_config)
580
+ print(sess.response.text)
581
+ ```
582
+
583
+ #### Service
584
+
585
+ LMDeploy's `api_server` enables models to be easily packed into services with a single command. The provided RESTful APIs are compatible with OpenAI's interfaces. Below are an example of service startup:
586
+
587
+ ```shell
588
+ lmdeploy serve api_server OpenGVLab/InternVL2_5-1B --server-port 23333
589
+ ```
590
+
591
+ To use the OpenAI-style interface, you need to install OpenAI:
592
+
593
+ ```shell
594
+ pip install openai
595
+ ```
596
+
597
+ Then, use the code below to make the API call:
598
+
599
+ ```python
600
+ from openai import OpenAI
601
+
602
+ client = OpenAI(api_key='YOUR_API_KEY', base_url='http://0.0.0.0:23333/v1')
603
+ model_name = client.models.list().data[0].id
604
+ response = client.chat.completions.create(
605
+ model=model_name,
606
+ messages=[{
607
+ 'role':
608
+ 'user',
609
+ 'content': [{
610
+ 'type': 'text',
611
+ 'text': 'describe this image',
612
+ }, {
613
+ 'type': 'image_url',
614
+ 'image_url': {
615
+ 'url':
616
+ 'https://modelscope.oss-cn-beijing.aliyuncs.com/resource/tiger.jpeg',
617
+ },
618
+ }],
619
+ }],
620
+ temperature=0.8,
621
+ top_p=0.8)
622
+ print(response)
623
+ ```
624
+
625
+ ## License
626
+
627
+ This project is released under the MIT License. This project uses the pre-trained Qwen2.5-0.5B-Instruct as a component, which is licensed under the Apache License 2.0.
628
+
629
+ ## Citation
630
+
631
+ If you find this project useful in your research, please consider citing:
632
+
633
+ ```BibTeX
634
+ @article{wang2024mpo,
635
+ title={Enhancing the Reasoning Ability of Multimodal Large Language Models via Mixed Preference Optimization},
636
+ author={Wang, Weiyun and Chen, Zhe and Wang, Wenhai and Cao, Yue and Liu, Yangzhou and Gao, Zhangwei and Zhu, Jinguo and Zhu, Xizhou and Lu, Lewei and Qiao, Yu and Dai, Jifeng},
637
+ journal={arXiv preprint arXiv:2411.10442},
638
+ year={2024}
639
+ }
640
+ @article{chen2024expanding,
641
+ title={Expanding Performance Boundaries of Open-Source Multimodal Models with Model, Data, and Test-Time Scaling},
642
+ author={Chen, Zhe and Wang, Weiyun and Cao, Yue and Liu, Yangzhou and Gao, Zhangwei and Cui, Erfei and Zhu, Jinguo and Ye, Shenglong and Tian, Hao and Liu, Zhaoyang and others},
643
+ journal={arXiv preprint arXiv:2412.05271},
644
+ year={2024}
645
+ }
646
+ @article{chen2024far,
647
+ title={How Far Are We to GPT-4V? Closing the Gap to Commercial Multimodal Models with Open-Source Suites},
648
+ author={Chen, Zhe and Wang, Weiyun and Tian, Hao and Ye, Shenglong and Gao, Zhangwei and Cui, Erfei and Tong, Wenwen and Hu, Kongzhi and Luo, Jiapeng and Ma, Zheng and others},
649
+ journal={arXiv preprint arXiv:2404.16821},
650
+ year={2024}
651
+ }
652
+ @inproceedings{chen2024internvl,
653
+ title={Internvl: Scaling up vision foundation models and aligning for generic visual-linguistic tasks},
654
+ author={Chen, Zhe and Wu, Jiannan and Wang, Wenhai and Su, Weijie and Chen, Guo and Xing, Sen and Zhong, Muyan and Zhang, Qinglong and Zhu, Xizhou and Lu, Lewei and others},
655
+ booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
656
+ pages={24185--24198},
657
+ year={2024}
658
+ }
659
+ ```
added_tokens.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</box>": 151673,
3
+ "</img>": 151666,
4
+ "</quad>": 151669,
5
+ "</ref>": 151671,
6
+ "</tool_call>": 151658,
7
+ "<IMG_CONTEXT>": 151667,
8
+ "<box>": 151672,
9
+ "<img>": 151665,
10
+ "<quad>": 151668,
11
+ "<ref>": 151670,
12
+ "<tool_call>": 151657,
13
+ "<|box_end|>": 151649,
14
+ "<|box_start|>": 151648,
15
+ "<|endoftext|>": 151643,
16
+ "<|file_sep|>": 151664,
17
+ "<|fim_middle|>": 151660,
18
+ "<|fim_pad|>": 151662,
19
+ "<|fim_prefix|>": 151659,
20
+ "<|fim_suffix|>": 151661,
21
+ "<|im_end|>": 151645,
22
+ "<|im_start|>": 151644,
23
+ "<|image_pad|>": 151655,
24
+ "<|object_ref_end|>": 151647,
25
+ "<|object_ref_start|>": 151646,
26
+ "<|quad_end|>": 151651,
27
+ "<|quad_start|>": 151650,
28
+ "<|repo_name|>": 151663,
29
+ "<|video_pad|>": 151656,
30
+ "<|vision_end|>": 151653,
31
+ "<|vision_pad|>": 151654,
32
+ "<|vision_start|>": 151652
33
+ }
config.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": null,
3
+ "architectures": [
4
+ "InternVLChatModel"
5
+ ],
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_internvl_chat.InternVLChatConfig",
8
+ "AutoModel": "modeling_internvl_chat.InternVLChatModel",
9
+ "AutoModelForCausalLM": "modeling_internvl_chat.InternVLChatModel"
10
+ },
11
+ "downsample_ratio": 0.5,
12
+ "dynamic_image_size": true,
13
+ "force_image_size": 448,
14
+ "llm_config": {
15
+ "_name_or_path": "Qwen/Qwen2.5-3B-Instruct",
16
+ "add_cross_attention": false,
17
+ "architectures": [
18
+ "Qwen2ForCausalLM"
19
+ ],
20
+ "_attn_implementation": "flash_attention_2",
21
+ "attention_dropout": 0.0,
22
+ "bad_words_ids": null,
23
+ "begin_suppress_tokens": null,
24
+ "bos_token_id": 151643,
25
+ "chunk_size_feed_forward": 0,
26
+ "cross_attention_hidden_size": null,
27
+ "decoder_start_token_id": null,
28
+ "diversity_penalty": 0.0,
29
+ "do_sample": false,
30
+ "early_stopping": false,
31
+ "encoder_no_repeat_ngram_size": 0,
32
+ "eos_token_id": 151645,
33
+ "exponential_decay_length_penalty": null,
34
+ "finetuning_task": null,
35
+ "forced_bos_token_id": null,
36
+ "forced_eos_token_id": null,
37
+ "hidden_act": "silu",
38
+ "hidden_size": 2048,
39
+ "id2label": {
40
+ "0": "LABEL_0",
41
+ "1": "LABEL_1"
42
+ },
43
+ "initializer_range": 0.02,
44
+ "intermediate_size": 11008,
45
+ "is_decoder": false,
46
+ "is_encoder_decoder": false,
47
+ "label2id": {
48
+ "LABEL_0": 0,
49
+ "LABEL_1": 1
50
+ },
51
+ "length_penalty": 1.0,
52
+ "max_length": 20,
53
+ "max_position_embeddings": 32768,
54
+ "max_window_layers": 70,
55
+ "min_length": 0,
56
+ "model_type": "qwen2",
57
+ "no_repeat_ngram_size": 0,
58
+ "num_attention_heads": 16,
59
+ "num_beam_groups": 1,
60
+ "num_beams": 1,
61
+ "num_hidden_layers": 36,
62
+ "num_key_value_heads": 2,
63
+ "num_return_sequences": 1,
64
+ "output_attentions": false,
65
+ "output_hidden_states": false,
66
+ "output_scores": false,
67
+ "pad_token_id": null,
68
+ "prefix": null,
69
+ "problem_type": null,
70
+ "pruned_heads": {},
71
+ "remove_invalid_values": false,
72
+ "repetition_penalty": 1.0,
73
+ "return_dict": true,
74
+ "return_dict_in_generate": false,
75
+ "rms_norm_eps": 1e-06,
76
+ "rope_theta": 1000000.0,
77
+ "sep_token_id": null,
78
+ "sliding_window": 32768,
79
+ "suppress_tokens": null,
80
+ "task_specific_params": null,
81
+ "temperature": 1.0,
82
+ "tf_legacy_loss": false,
83
+ "tie_encoder_decoder": false,
84
+ "tie_word_embeddings": false,
85
+ "tokenizer_class": null,
86
+ "top_k": 50,
87
+ "top_p": 1.0,
88
+ "torch_dtype": "bfloat16",
89
+ "torchscript": false,
90
+ "transformers_version": "4.37.2",
91
+ "typical_p": 1.0,
92
+ "use_bfloat16": true,
93
+ "use_cache": true,
94
+ "use_sliding_window": false,
95
+ "vocab_size": 151674
96
+ },
97
+ "max_dynamic_patch": 12,
98
+ "min_dynamic_patch": 1,
99
+ "model_type": "internvl_chat",
100
+ "ps_version": "v2",
101
+ "select_layer": -1,
102
+ "template": "internvl2_5",
103
+ "torch_dtype": "bfloat16",
104
+ "use_backbone_lora": 0,
105
+ "use_llm_lora": 0,
106
+ "use_thumbnail": true,
107
+ "vision_config": {
108
+ "architectures": [
109
+ "InternVisionModel"
110
+ ],
111
+ "attention_dropout": 0.0,
112
+ "drop_path_rate": 0.0,
113
+ "dropout": 0.0,
114
+ "hidden_act": "gelu",
115
+ "hidden_size": 1024,
116
+ "image_size": 448,
117
+ "initializer_factor": 1.0,
118
+ "initializer_range": 0.02,
119
+ "intermediate_size": 4096,
120
+ "layer_norm_eps": 1e-06,
121
+ "model_type": "intern_vit_6b",
122
+ "norm_type": "layer_norm",
123
+ "num_attention_heads": 16,
124
+ "num_channels": 3,
125
+ "num_hidden_layers": 24,
126
+ "output_attentions": false,
127
+ "output_hidden_states": false,
128
+ "patch_size": 14,
129
+ "qk_normalization": false,
130
+ "qkv_bias": true,
131
+ "return_dict": true,
132
+ "torch_dtype": "bfloat16",
133
+ "transformers_version": "4.37.2",
134
+ "use_bfloat16": true,
135
+ "use_flash_attn": true
136
+ }
137
+ }
configuration_intern_vit.py ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # --------------------------------------------------------
2
+ # InternVL
3
+ # Copyright (c) 2024 OpenGVLab
4
+ # Licensed under The MIT License [see LICENSE for details]
5
+ # --------------------------------------------------------
6
+
7
+ import os
8
+ from typing import Union
9
+
10
+ from transformers.configuration_utils import PretrainedConfig
11
+ from transformers.utils import logging
12
+
13
+ logger = logging.get_logger(__name__)
14
+
15
+
16
+ class InternVisionConfig(PretrainedConfig):
17
+ r"""
18
+ This is the configuration class to store the configuration of a [`InternVisionModel`]. It is used to
19
+ instantiate a vision encoder according to the specified arguments, defining the model architecture.
20
+
21
+ Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the
22
+ documentation from [`PretrainedConfig`] for more information.
23
+
24
+ Args:
25
+ num_channels (`int`, *optional*, defaults to 3):
26
+ Number of color channels in the input images (e.g., 3 for RGB).
27
+ patch_size (`int`, *optional*, defaults to 14):
28
+ The size (resolution) of each patch.
29
+ image_size (`int`, *optional*, defaults to 224):
30
+ The size (resolution) of each image.
31
+ qkv_bias (`bool`, *optional*, defaults to `False`):
32
+ Whether to add a bias to the queries and values in the self-attention layers.
33
+ hidden_size (`int`, *optional*, defaults to 3200):
34
+ Dimensionality of the encoder layers and the pooler layer.
35
+ num_attention_heads (`int`, *optional*, defaults to 25):
36
+ Number of attention heads for each attention layer in the Transformer encoder.
37
+ intermediate_size (`int`, *optional*, defaults to 12800):
38
+ Dimensionality of the "intermediate" (i.e., feed-forward) layer in the Transformer encoder.
39
+ qk_normalization (`bool`, *optional*, defaults to `True`):
40
+ Whether to normalize the queries and keys in the self-attention layers.
41
+ num_hidden_layers (`int`, *optional*, defaults to 48):
42
+ Number of hidden layers in the Transformer encoder.
43
+ use_flash_attn (`bool`, *optional*, defaults to `True`):
44
+ Whether to use flash attention mechanism.
45
+ hidden_act (`str` or `function`, *optional*, defaults to `"gelu"`):
46
+ The non-linear activation function (function or string) in the encoder and pooler. If string, `"gelu"`,
47
+ `"relu"`, `"selu"` and `"gelu_new"` ``"gelu"` are supported.
48
+ layer_norm_eps (`float`, *optional*, defaults to 1e-6):
49
+ The epsilon used by the layer normalization layers.
50
+ dropout (`float`, *optional*, defaults to 0.0):
51
+ The dropout probability for all fully connected layers in the embeddings, encoder, and pooler.
52
+ drop_path_rate (`float`, *optional*, defaults to 0.0):
53
+ Dropout rate for stochastic depth.
54
+ attention_dropout (`float`, *optional*, defaults to 0.0):
55
+ The dropout ratio for the attention probabilities.
56
+ initializer_range (`float`, *optional*, defaults to 0.02):
57
+ The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
58
+ initializer_factor (`float`, *optional*, defaults to 0.1):
59
+ A factor for layer scale.
60
+ """
61
+
62
+ model_type = 'intern_vit_6b'
63
+
64
+ def __init__(
65
+ self,
66
+ num_channels=3,
67
+ patch_size=14,
68
+ image_size=224,
69
+ qkv_bias=False,
70
+ hidden_size=3200,
71
+ num_attention_heads=25,
72
+ intermediate_size=12800,
73
+ qk_normalization=True,
74
+ num_hidden_layers=48,
75
+ use_flash_attn=True,
76
+ hidden_act='gelu',
77
+ norm_type='rms_norm',
78
+ layer_norm_eps=1e-6,
79
+ dropout=0.0,
80
+ drop_path_rate=0.0,
81
+ attention_dropout=0.0,
82
+ initializer_range=0.02,
83
+ initializer_factor=0.1,
84
+ **kwargs,
85
+ ):
86
+ super().__init__(**kwargs)
87
+
88
+ self.hidden_size = hidden_size
89
+ self.intermediate_size = intermediate_size
90
+ self.dropout = dropout
91
+ self.drop_path_rate = drop_path_rate
92
+ self.num_hidden_layers = num_hidden_layers
93
+ self.num_attention_heads = num_attention_heads
94
+ self.num_channels = num_channels
95
+ self.patch_size = patch_size
96
+ self.image_size = image_size
97
+ self.initializer_range = initializer_range
98
+ self.initializer_factor = initializer_factor
99
+ self.attention_dropout = attention_dropout
100
+ self.layer_norm_eps = layer_norm_eps
101
+ self.hidden_act = hidden_act
102
+ self.norm_type = norm_type
103
+ self.qkv_bias = qkv_bias
104
+ self.qk_normalization = qk_normalization
105
+ self.use_flash_attn = use_flash_attn
106
+
107
+ @classmethod
108
+ def from_pretrained(cls, pretrained_model_name_or_path: Union[str, os.PathLike], **kwargs) -> 'PretrainedConfig':
109
+ config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
110
+
111
+ if 'vision_config' in config_dict:
112
+ config_dict = config_dict['vision_config']
113
+
114
+ if 'model_type' in config_dict and hasattr(cls, 'model_type') and config_dict['model_type'] != cls.model_type:
115
+ logger.warning(
116
+ f"You are using a model of type {config_dict['model_type']} to instantiate a model of type "
117
+ f'{cls.model_type}. This is not supported for all configurations of models and can yield errors.'
118
+ )
119
+
120
+ return cls.from_dict(config_dict, **kwargs)
configuration_internvl_chat.py ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # --------------------------------------------------------
2
+ # InternVL
3
+ # Copyright (c) 2024 OpenGVLab
4
+ # Licensed under The MIT License [see LICENSE for details]
5
+ # --------------------------------------------------------
6
+
7
+ import copy
8
+
9
+ from transformers import AutoConfig, LlamaConfig, Qwen2Config
10
+ from transformers.configuration_utils import PretrainedConfig
11
+ from transformers.utils import logging
12
+
13
+ from .configuration_intern_vit import InternVisionConfig
14
+
15
+ logger = logging.get_logger(__name__)
16
+
17
+
18
+ class InternVLChatConfig(PretrainedConfig):
19
+ model_type = 'internvl_chat'
20
+ is_composition = True
21
+
22
+ def __init__(
23
+ self,
24
+ vision_config=None,
25
+ llm_config=None,
26
+ use_backbone_lora=0,
27
+ use_llm_lora=0,
28
+ select_layer=-1,
29
+ force_image_size=None,
30
+ downsample_ratio=0.5,
31
+ template=None,
32
+ dynamic_image_size=False,
33
+ use_thumbnail=False,
34
+ ps_version='v1',
35
+ min_dynamic_patch=1,
36
+ max_dynamic_patch=6,
37
+ **kwargs):
38
+ super().__init__(**kwargs)
39
+
40
+ if vision_config is None:
41
+ vision_config = {'architectures': ['InternVisionModel']}
42
+ logger.info('vision_config is None. Initializing the InternVisionConfig with default values.')
43
+
44
+ if llm_config is None:
45
+ llm_config = {'architectures': ['Qwen2ForCausalLM']}
46
+ logger.info('llm_config is None. Initializing the LlamaConfig config with default values (`LlamaConfig`).')
47
+
48
+ self.vision_config = InternVisionConfig(**vision_config)
49
+ if llm_config.get('architectures')[0] == 'LlamaForCausalLM':
50
+ self.llm_config = LlamaConfig(**llm_config)
51
+ elif llm_config.get('architectures')[0] == 'Qwen2ForCausalLM':
52
+ self.llm_config = Qwen2Config(**llm_config)
53
+ else:
54
+ raise ValueError('Unsupported architecture: {}'.format(llm_config.get('architectures')[0]))
55
+ self.use_backbone_lora = use_backbone_lora
56
+ self.use_llm_lora = use_llm_lora
57
+ self.select_layer = select_layer
58
+ self.force_image_size = force_image_size
59
+ self.downsample_ratio = downsample_ratio
60
+ self.template = template
61
+ self.dynamic_image_size = dynamic_image_size
62
+ self.use_thumbnail = use_thumbnail
63
+ self.ps_version = ps_version # pixel shuffle version
64
+ self.min_dynamic_patch = min_dynamic_patch
65
+ self.max_dynamic_patch = max_dynamic_patch
66
+
67
+ logger.info(f'vision_select_layer: {self.select_layer}')
68
+ logger.info(f'ps_version: {self.ps_version}')
69
+ logger.info(f'min_dynamic_patch: {self.min_dynamic_patch}')
70
+ logger.info(f'max_dynamic_patch: {self.max_dynamic_patch}')
71
+
72
+ def to_dict(self):
73
+ """
74
+ Serializes this instance to a Python dictionary. Override the default [`~PretrainedConfig.to_dict`].
75
+
76
+ Returns:
77
+ `Dict[str, any]`: Dictionary of all the attributes that make up this configuration instance,
78
+ """
79
+ output = copy.deepcopy(self.__dict__)
80
+ output['vision_config'] = self.vision_config.to_dict()
81
+ output['llm_config'] = self.llm_config.to_dict()
82
+ output['model_type'] = self.__class__.model_type
83
+ output['use_backbone_lora'] = self.use_backbone_lora
84
+ output['use_llm_lora'] = self.use_llm_lora
85
+ output['select_layer'] = self.select_layer
86
+ output['force_image_size'] = self.force_image_size
87
+ output['downsample_ratio'] = self.downsample_ratio
88
+ output['template'] = self.template
89
+ output['dynamic_image_size'] = self.dynamic_image_size
90
+ output['use_thumbnail'] = self.use_thumbnail
91
+ output['ps_version'] = self.ps_version
92
+ output['min_dynamic_patch'] = self.min_dynamic_patch
93
+ output['max_dynamic_patch'] = self.max_dynamic_patch
94
+
95
+ return output
conversation.py ADDED
@@ -0,0 +1,391 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """
2
+ Conversation prompt templates.
3
+
4
+ We kindly request that you import fastchat instead of copying this file if you wish to use it.
5
+ If you have changes in mind, please contribute back so the community can benefit collectively and continue to maintain these valuable templates.
6
+
7
+ Modified from https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py
8
+ """
9
+
10
+ import dataclasses
11
+ from enum import IntEnum, auto
12
+ from typing import Dict, List, Tuple, Union
13
+
14
+
15
+ class SeparatorStyle(IntEnum):
16
+ """Separator styles."""
17
+
18
+ ADD_COLON_SINGLE = auto()
19
+ ADD_COLON_TWO = auto()
20
+ ADD_COLON_SPACE_SINGLE = auto()
21
+ NO_COLON_SINGLE = auto()
22
+ NO_COLON_TWO = auto()
23
+ ADD_NEW_LINE_SINGLE = auto()
24
+ LLAMA2 = auto()
25
+ CHATGLM = auto()
26
+ CHATML = auto()
27
+ CHATINTERN = auto()
28
+ DOLLY = auto()
29
+ RWKV = auto()
30
+ PHOENIX = auto()
31
+ ROBIN = auto()
32
+ FALCON_CHAT = auto()
33
+ CHATGLM3 = auto()
34
+ INTERNVL_ZH = auto()
35
+ MPT = auto()
36
+
37
+
38
+ @dataclasses.dataclass
39
+ class Conversation:
40
+ """A class that manages prompt templates and keeps all conversation history."""
41
+
42
+ # The name of this template
43
+ name: str
44
+ # The template of the system prompt
45
+ system_template: str = '{system_message}'
46
+ # The system message
47
+ system_message: str = ''
48
+ # The names of two roles
49
+ roles: Tuple[str] = ('USER', 'ASSISTANT')
50
+ # All messages. Each item is (role, message).
51
+ messages: List[List[str]] = ()
52
+ # The number of few shot examples
53
+ offset: int = 0
54
+ # The separator style and configurations
55
+ sep_style: SeparatorStyle = SeparatorStyle.ADD_COLON_SINGLE
56
+ sep: str = '\n'
57
+ sep2: str = None
58
+ # Stop criteria (the default one is EOS token)
59
+ stop_str: Union[str, List[str]] = None
60
+ # Stops generation if meeting any token in this list
61
+ stop_token_ids: List[int] = None
62
+
63
+ def get_prompt(self) -> str:
64
+ """Get the prompt for generation."""
65
+ system_prompt = self.system_template.format(system_message=self.system_message)
66
+ if self.sep_style == SeparatorStyle.ADD_COLON_SINGLE:
67
+ ret = system_prompt + self.sep
68
+ for role, message in self.messages:
69
+ if message:
70
+ ret += role + ': ' + message + self.sep
71
+ else:
72
+ ret += role + ':'
73
+ return ret
74
+ elif self.sep_style == SeparatorStyle.ADD_COLON_TWO:
75
+ seps = [self.sep, self.sep2]
76
+ ret = system_prompt + seps[0]
77
+ for i, (role, message) in enumerate(self.messages):
78
+ if message:
79
+ ret += role + ': ' + message + seps[i % 2]
80
+ else:
81
+ ret += role + ':'
82
+ return ret
83
+ elif self.sep_style == SeparatorStyle.ADD_COLON_SPACE_SINGLE:
84
+ ret = system_prompt + self.sep
85
+ for role, message in self.messages:
86
+ if message:
87
+ ret += role + ': ' + message + self.sep
88
+ else:
89
+ ret += role + ': ' # must be end with a space
90
+ return ret
91
+ elif self.sep_style == SeparatorStyle.ADD_NEW_LINE_SINGLE:
92
+ ret = '' if system_prompt == '' else system_prompt + self.sep
93
+ for role, message in self.messages:
94
+ if message:
95
+ ret += role + '\n' + message + self.sep
96
+ else:
97
+ ret += role + '\n'
98
+ return ret
99
+ elif self.sep_style == SeparatorStyle.NO_COLON_SINGLE:
100
+ ret = system_prompt
101
+ for role, message in self.messages:
102
+ if message:
103
+ ret += role + message + self.sep
104
+ else:
105
+ ret += role
106
+ return ret
107
+ elif self.sep_style == SeparatorStyle.NO_COLON_TWO:
108
+ seps = [self.sep, self.sep2]
109
+ ret = system_prompt
110
+ for i, (role, message) in enumerate(self.messages):
111
+ if message:
112
+ ret += role + message + seps[i % 2]
113
+ else:
114
+ ret += role
115
+ return ret
116
+ elif self.sep_style == SeparatorStyle.RWKV:
117
+ ret = system_prompt
118
+ for i, (role, message) in enumerate(self.messages):
119
+ if message:
120
+ ret += (
121
+ role
122
+ + ': '
123
+ + message.replace('\r\n', '\n').replace('\n\n', '\n')
124
+ )
125
+ ret += '\n\n'
126
+ else:
127
+ ret += role + ':'
128
+ return ret
129
+ elif self.sep_style == SeparatorStyle.LLAMA2:
130
+ seps = [self.sep, self.sep2]
131
+ if self.system_message:
132
+ ret = system_prompt
133
+ else:
134
+ ret = '[INST] '
135
+ for i, (role, message) in enumerate(self.messages):
136
+ tag = self.roles[i % 2]
137
+ if message:
138
+ if i == 0:
139
+ ret += message + ' '
140
+ else:
141
+ ret += tag + ' ' + message + seps[i % 2]
142
+ else:
143
+ ret += tag
144
+ return ret
145
+ elif self.sep_style == SeparatorStyle.CHATGLM:
146
+ # source: https://huggingface.co/THUDM/chatglm-6b/blob/1d240ba371910e9282298d4592532d7f0f3e9f3e/modeling_chatglm.py#L1302-L1308
147
+ # source2: https://huggingface.co/THUDM/chatglm2-6b/blob/e186c891cf64310ac66ef10a87e6635fa6c2a579/modeling_chatglm.py#L926
148
+ round_add_n = 1 if self.name == 'chatglm2' else 0
149
+ if system_prompt:
150
+ ret = system_prompt + self.sep
151
+ else:
152
+ ret = ''
153
+
154
+ for i, (role, message) in enumerate(self.messages):
155
+ if i % 2 == 0:
156
+ ret += f'[Round {i//2 + round_add_n}]{self.sep}'
157
+
158
+ if message:
159
+ ret += f'{role}:{message}{self.sep}'
160
+ else:
161
+ ret += f'{role}:'
162
+ return ret
163
+ elif self.sep_style == SeparatorStyle.CHATML:
164
+ ret = '' if system_prompt == '' else system_prompt + self.sep + '\n'
165
+ for role, message in self.messages:
166
+ if message:
167
+ ret += role + '\n' + message + self.sep + '\n'
168
+ else:
169
+ ret += role + '\n'
170
+ return ret
171
+ elif self.sep_style == SeparatorStyle.CHATGLM3:
172
+ ret = ''
173
+ if self.system_message:
174
+ ret += system_prompt
175
+ for role, message in self.messages:
176
+ if message:
177
+ ret += role + '\n' + ' ' + message
178
+ else:
179
+ ret += role
180
+ return ret
181
+ elif self.sep_style == SeparatorStyle.CHATINTERN:
182
+ # source: https://huggingface.co/internlm/internlm-chat-7b-8k/blob/bd546fa984b4b0b86958f56bf37f94aa75ab8831/modeling_internlm.py#L771
183
+ seps = [self.sep, self.sep2]
184
+ ret = system_prompt
185
+ for i, (role, message) in enumerate(self.messages):
186
+ # if i % 2 == 0:
187
+ # ret += "<s>"
188
+ if message:
189
+ ret += role + ':' + message + seps[i % 2] + '\n'
190
+ else:
191
+ ret += role + ':'
192
+ return ret
193
+ elif self.sep_style == SeparatorStyle.DOLLY:
194
+ seps = [self.sep, self.sep2]
195
+ ret = system_prompt
196
+ for i, (role, message) in enumerate(self.messages):
197
+ if message:
198
+ ret += role + ':\n' + message + seps[i % 2]
199
+ if i % 2 == 1:
200
+ ret += '\n\n'
201
+ else:
202
+ ret += role + ':\n'
203
+ return ret
204
+ elif self.sep_style == SeparatorStyle.PHOENIX:
205
+ ret = system_prompt
206
+ for role, message in self.messages:
207
+ if message:
208
+ ret += role + ': ' + '<s>' + message + '</s>'
209
+ else:
210
+ ret += role + ': ' + '<s>'
211
+ return ret
212
+ elif self.sep_style == SeparatorStyle.ROBIN:
213
+ ret = system_prompt + self.sep
214
+ for role, message in self.messages:
215
+ if message:
216
+ ret += role + ':\n' + message + self.sep
217
+ else:
218
+ ret += role + ':\n'
219
+ return ret
220
+ elif self.sep_style == SeparatorStyle.FALCON_CHAT:
221
+ ret = ''
222
+ if self.system_message:
223
+ ret += system_prompt + self.sep
224
+ for role, message in self.messages:
225
+ if message:
226
+ ret += role + ': ' + message + self.sep
227
+ else:
228
+ ret += role + ':'
229
+
230
+ return ret
231
+ elif self.sep_style == SeparatorStyle.INTERNVL_ZH:
232
+ seps = [self.sep, self.sep2]
233
+ ret = self.system_message + seps[0]
234
+ for i, (role, message) in enumerate(self.messages):
235
+ if message:
236
+ ret += role + ': ' + message + seps[i % 2]
237
+ else:
238
+ ret += role + ':'
239
+ return ret
240
+ elif self.sep_style == SeparatorStyle.MPT:
241
+ ret = system_prompt + self.sep
242
+ for role, message in self.messages:
243
+ if message:
244
+ if type(message) is tuple:
245
+ message, _, _ = message
246
+ ret += role + message + self.sep
247
+ else:
248
+ ret += role
249
+ return ret
250
+ else:
251
+ raise ValueError(f'Invalid style: {self.sep_style}')
252
+
253
+ def set_system_message(self, system_message: str):
254
+ """Set the system message."""
255
+ self.system_message = system_message
256
+
257
+ def append_message(self, role: str, message: str):
258
+ """Append a new message."""
259
+ self.messages.append([role, message])
260
+
261
+ def update_last_message(self, message: str):
262
+ """Update the last output.
263
+
264
+ The last message is typically set to be None when constructing the prompt,
265
+ so we need to update it in-place after getting the response from a model.
266
+ """
267
+ self.messages[-1][1] = message
268
+
269
+ def to_gradio_chatbot(self):
270
+ """Convert the conversation to gradio chatbot format."""
271
+ ret = []
272
+ for i, (role, msg) in enumerate(self.messages[self.offset :]):
273
+ if i % 2 == 0:
274
+ ret.append([msg, None])
275
+ else:
276
+ ret[-1][-1] = msg
277
+ return ret
278
+
279
+ def to_openai_api_messages(self):
280
+ """Convert the conversation to OpenAI chat completion format."""
281
+ ret = [{'role': 'system', 'content': self.system_message}]
282
+
283
+ for i, (_, msg) in enumerate(self.messages[self.offset :]):
284
+ if i % 2 == 0:
285
+ ret.append({'role': 'user', 'content': msg})
286
+ else:
287
+ if msg is not None:
288
+ ret.append({'role': 'assistant', 'content': msg})
289
+ return ret
290
+
291
+ def copy(self):
292
+ return Conversation(
293
+ name=self.name,
294
+ system_template=self.system_template,
295
+ system_message=self.system_message,
296
+ roles=self.roles,
297
+ messages=[[x, y] for x, y in self.messages],
298
+ offset=self.offset,
299
+ sep_style=self.sep_style,
300
+ sep=self.sep,
301
+ sep2=self.sep2,
302
+ stop_str=self.stop_str,
303
+ stop_token_ids=self.stop_token_ids,
304
+ )
305
+
306
+ def dict(self):
307
+ return {
308
+ 'template_name': self.name,
309
+ 'system_message': self.system_message,
310
+ 'roles': self.roles,
311
+ 'messages': self.messages,
312
+ 'offset': self.offset,
313
+ }
314
+
315
+
316
+ # A global registry for all conversation templates
317
+ conv_templates: Dict[str, Conversation] = {}
318
+
319
+
320
+ def register_conv_template(template: Conversation, override: bool = False):
321
+ """Register a new conversation template."""
322
+ if not override:
323
+ assert (
324
+ template.name not in conv_templates
325
+ ), f'{template.name} has been registered.'
326
+
327
+ conv_templates[template.name] = template
328
+
329
+
330
+ def get_conv_template(name: str) -> Conversation:
331
+ """Get a conversation template."""
332
+ return conv_templates[name].copy()
333
+
334
+
335
+ # Both Hermes-2 and internlm2-chat are chatml-format conversation templates. The difference
336
+ # is that during training, the preprocessing function for the Hermes-2 template doesn't add
337
+ # <s> at the beginning of the tokenized sequence, while the internlm2-chat template does.
338
+ # Therefore, they are completely equivalent during inference.
339
+ register_conv_template(
340
+ Conversation(
341
+ name='Hermes-2',
342
+ system_template='<|im_start|>system\n{system_message}',
343
+ # note: The new system prompt was not used here to avoid changes in benchmark performance.
344
+ # system_message='我是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。',
345
+ system_message='你是由上海人工智能实验室联合商汤科技开发的书生多模态大模型,英文名叫InternVL, 是一个有用无害的人工智能助手。',
346
+ roles=('<|im_start|>user\n', '<|im_start|>assistant\n'),
347
+ sep_style=SeparatorStyle.MPT,
348
+ sep='<|im_end|>',
349
+ stop_str='<|endoftext|>',
350
+ )
351
+ )
352
+
353
+
354
+ register_conv_template(
355
+ Conversation(
356
+ name='internlm2-chat',
357
+ system_template='<|im_start|>system\n{system_message}',
358
+ # note: The new system prompt was not used here to avoid changes in benchmark performance.
359
+ # system_message='我是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。',
360
+ system_message='你是由上海人工智能实验室联合商汤科技开发的书生多模态大模型,英文名叫InternVL, 是一个有用无害的人工智能助手。',
361
+ roles=('<|im_start|>user\n', '<|im_start|>assistant\n'),
362
+ sep_style=SeparatorStyle.MPT,
363
+ sep='<|im_end|>',
364
+ )
365
+ )
366
+
367
+
368
+ register_conv_template(
369
+ Conversation(
370
+ name='phi3-chat',
371
+ system_template='<|system|>\n{system_message}',
372
+ # note: The new system prompt was not used here to avoid changes in benchmark performance.
373
+ # system_message='我是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。',
374
+ system_message='你是由上海人工智能实验室联合商汤科技开发的书生多模态大模型,英文名叫InternVL, 是一个有用无害的人工智能助手。',
375
+ roles=('<|user|>\n', '<|assistant|>\n'),
376
+ sep_style=SeparatorStyle.MPT,
377
+ sep='<|end|>',
378
+ )
379
+ )
380
+
381
+
382
+ register_conv_template(
383
+ Conversation(
384
+ name='internvl2_5',
385
+ system_template='<|im_start|>system\n{system_message}',
386
+ system_message='你是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。',
387
+ roles=('<|im_start|>user\n', '<|im_start|>assistant\n'),
388
+ sep_style=SeparatorStyle.MPT,
389
+ sep='<|im_end|>\n',
390
+ )
391
+ )
examples/image1.jpg ADDED
examples/image2.jpg ADDED
examples/red-panda.mp4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d921c07bb97224d65a37801541d246067f0d506f08723ffa1ad85c217907ccb8
3
+ size 1867237
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.45.1",
4
+ "eos_token_id": [
5
+ 151644,
6
+ 151645,
7
+ 151643
8
+ ]
9
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11f8962c7f75ad4335acc0d88c56044004786806c1ea9d2d907f1ad3f221c96b
3
+ size 4993023560
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:677123c8ad102208df677464eeac5d663ecf967308ee67eca83824ac207f8c64
3
+ size 2432349168
model.safetensors.index.json ADDED
@@ -0,0 +1,788 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 7425275904
4
+ },
5
+ "weight_map": {
6
+ "language_model.lm_head.weight": "model-00002-of-00002.safetensors",
7
+ "language_model.model.embed_tokens.weight": "model-00001-of-00002.safetensors",
8
+ "language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
9
+ "language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
10
+ "language_model.model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
11
+ "language_model.model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
12
+ "language_model.model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
13
+ "language_model.model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
14
+ "language_model.model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
15
+ "language_model.model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
16
+ "language_model.model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
17
+ "language_model.model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
18
+ "language_model.model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
19
+ "language_model.model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
20
+ "language_model.model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors",
21
+ "language_model.model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
22
+ "language_model.model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
23
+ "language_model.model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
24
+ "language_model.model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
25
+ "language_model.model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
26
+ "language_model.model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
27
+ "language_model.model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
28
+ "language_model.model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
29
+ "language_model.model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
30
+ "language_model.model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
31
+ "language_model.model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
32
+ "language_model.model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors",
33
+ "language_model.model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
34
+ "language_model.model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
35
+ "language_model.model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
36
+ "language_model.model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
37
+ "language_model.model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
38
+ "language_model.model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
39
+ "language_model.model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
40
+ "language_model.model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
41
+ "language_model.model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
42
+ "language_model.model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
43
+ "language_model.model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
44
+ "language_model.model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors",
45
+ "language_model.model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
46
+ "language_model.model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
47
+ "language_model.model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
48
+ "language_model.model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
49
+ "language_model.model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
50
+ "language_model.model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
51
+ "language_model.model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
52
+ "language_model.model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
53
+ "language_model.model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
54
+ "language_model.model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
55
+ "language_model.model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
56
+ "language_model.model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors",
57
+ "language_model.model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
58
+ "language_model.model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
59
+ "language_model.model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
60
+ "language_model.model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
61
+ "language_model.model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
62
+ "language_model.model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
63
+ "language_model.model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
64
+ "language_model.model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
65
+ "language_model.model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
66
+ "language_model.model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
67
+ "language_model.model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
68
+ "language_model.model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors",
69
+ "language_model.model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
70
+ "language_model.model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
71
+ "language_model.model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
72
+ "language_model.model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
73
+ "language_model.model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
74
+ "language_model.model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
75
+ "language_model.model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
76
+ "language_model.model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
77
+ "language_model.model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
78
+ "language_model.model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
79
+ "language_model.model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
80
+ "language_model.model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors",
81
+ "language_model.model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
82
+ "language_model.model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
83
+ "language_model.model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
84
+ "language_model.model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
85
+ "language_model.model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
86
+ "language_model.model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
87
+ "language_model.model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
88
+ "language_model.model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
89
+ "language_model.model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
90
+ "language_model.model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
91
+ "language_model.model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
92
+ "language_model.model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors",
93
+ "language_model.model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
94
+ "language_model.model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
95
+ "language_model.model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
96
+ "language_model.model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
97
+ "language_model.model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
98
+ "language_model.model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
99
+ "language_model.model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
100
+ "language_model.model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
101
+ "language_model.model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
102
+ "language_model.model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
103
+ "language_model.model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
104
+ "language_model.model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors",
105
+ "language_model.model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
106
+ "language_model.model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
107
+ "language_model.model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
108
+ "language_model.model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
109
+ "language_model.model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
110
+ "language_model.model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
111
+ "language_model.model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
112
+ "language_model.model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
113
+ "language_model.model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
114
+ "language_model.model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
115
+ "language_model.model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
116
+ "language_model.model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors",
117
+ "language_model.model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
118
+ "language_model.model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
119
+ "language_model.model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
120
+ "language_model.model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
121
+ "language_model.model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
122
+ "language_model.model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
123
+ "language_model.model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
124
+ "language_model.model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
125
+ "language_model.model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
126
+ "language_model.model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
127
+ "language_model.model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
128
+ "language_model.model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
129
+ "language_model.model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
130
+ "language_model.model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
131
+ "language_model.model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
132
+ "language_model.model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
133
+ "language_model.model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
134
+ "language_model.model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
135
+ "language_model.model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
136
+ "language_model.model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
137
+ "language_model.model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
138
+ "language_model.model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
139
+ "language_model.model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
140
+ "language_model.model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
141
+ "language_model.model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
142
+ "language_model.model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
143
+ "language_model.model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
144
+ "language_model.model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
145
+ "language_model.model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
146
+ "language_model.model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
147
+ "language_model.model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
148
+ "language_model.model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
149
+ "language_model.model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
150
+ "language_model.model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
151
+ "language_model.model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
152
+ "language_model.model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
153
+ "language_model.model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
154
+ "language_model.model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
155
+ "language_model.model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
156
+ "language_model.model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
157
+ "language_model.model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
158
+ "language_model.model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
159
+ "language_model.model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
160
+ "language_model.model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
161
+ "language_model.model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
162
+ "language_model.model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
163
+ "language_model.model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
164
+ "language_model.model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors",
165
+ "language_model.model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
166
+ "language_model.model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
167
+ "language_model.model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
168
+ "language_model.model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
169
+ "language_model.model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
170
+ "language_model.model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
171
+ "language_model.model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
172
+ "language_model.model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
173
+ "language_model.model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
174
+ "language_model.model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
175
+ "language_model.model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
176
+ "language_model.model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors",
177
+ "language_model.model.layers.21.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
178
+ "language_model.model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
179
+ "language_model.model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
180
+ "language_model.model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
181
+ "language_model.model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
182
+ "language_model.model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
183
+ "language_model.model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
184
+ "language_model.model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
185
+ "language_model.model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
186
+ "language_model.model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
187
+ "language_model.model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
188
+ "language_model.model.layers.22.input_layernorm.weight": "model-00001-of-00002.safetensors",
189
+ "language_model.model.layers.22.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
190
+ "language_model.model.layers.22.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
191
+ "language_model.model.layers.22.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
192
+ "language_model.model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
193
+ "language_model.model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
194
+ "language_model.model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
195
+ "language_model.model.layers.22.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
196
+ "language_model.model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
197
+ "language_model.model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
198
+ "language_model.model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
199
+ "language_model.model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
200
+ "language_model.model.layers.23.input_layernorm.weight": "model-00001-of-00002.safetensors",
201
+ "language_model.model.layers.23.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
202
+ "language_model.model.layers.23.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
203
+ "language_model.model.layers.23.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
204
+ "language_model.model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
205
+ "language_model.model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
206
+ "language_model.model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
207
+ "language_model.model.layers.23.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
208
+ "language_model.model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
209
+ "language_model.model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
210
+ "language_model.model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
211
+ "language_model.model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
212
+ "language_model.model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors",
213
+ "language_model.model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
214
+ "language_model.model.layers.24.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
215
+ "language_model.model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
216
+ "language_model.model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
217
+ "language_model.model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
218
+ "language_model.model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
219
+ "language_model.model.layers.24.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
220
+ "language_model.model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
221
+ "language_model.model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
222
+ "language_model.model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
223
+ "language_model.model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
224
+ "language_model.model.layers.25.input_layernorm.weight": "model-00002-of-00002.safetensors",
225
+ "language_model.model.layers.25.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
226
+ "language_model.model.layers.25.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
227
+ "language_model.model.layers.25.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
228
+ "language_model.model.layers.25.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
229
+ "language_model.model.layers.25.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
230
+ "language_model.model.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
231
+ "language_model.model.layers.25.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
232
+ "language_model.model.layers.25.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
233
+ "language_model.model.layers.25.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
234
+ "language_model.model.layers.25.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
235
+ "language_model.model.layers.25.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
236
+ "language_model.model.layers.26.input_layernorm.weight": "model-00002-of-00002.safetensors",
237
+ "language_model.model.layers.26.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
238
+ "language_model.model.layers.26.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
239
+ "language_model.model.layers.26.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
240
+ "language_model.model.layers.26.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
241
+ "language_model.model.layers.26.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
242
+ "language_model.model.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
243
+ "language_model.model.layers.26.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
244
+ "language_model.model.layers.26.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
245
+ "language_model.model.layers.26.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
246
+ "language_model.model.layers.26.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
247
+ "language_model.model.layers.26.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
248
+ "language_model.model.layers.27.input_layernorm.weight": "model-00002-of-00002.safetensors",
249
+ "language_model.model.layers.27.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
250
+ "language_model.model.layers.27.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
251
+ "language_model.model.layers.27.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
252
+ "language_model.model.layers.27.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
253
+ "language_model.model.layers.27.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
254
+ "language_model.model.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
255
+ "language_model.model.layers.27.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
256
+ "language_model.model.layers.27.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
257
+ "language_model.model.layers.27.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
258
+ "language_model.model.layers.27.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
259
+ "language_model.model.layers.27.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
260
+ "language_model.model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors",
261
+ "language_model.model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
262
+ "language_model.model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
263
+ "language_model.model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
264
+ "language_model.model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
265
+ "language_model.model.layers.28.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
266
+ "language_model.model.layers.28.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
267
+ "language_model.model.layers.28.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
268
+ "language_model.model.layers.28.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
269
+ "language_model.model.layers.28.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
270
+ "language_model.model.layers.28.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
271
+ "language_model.model.layers.28.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
272
+ "language_model.model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors",
273
+ "language_model.model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
274
+ "language_model.model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
275
+ "language_model.model.layers.29.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
276
+ "language_model.model.layers.29.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
277
+ "language_model.model.layers.29.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
278
+ "language_model.model.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
279
+ "language_model.model.layers.29.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
280
+ "language_model.model.layers.29.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
281
+ "language_model.model.layers.29.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
282
+ "language_model.model.layers.29.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
283
+ "language_model.model.layers.29.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
284
+ "language_model.model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
285
+ "language_model.model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
286
+ "language_model.model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
287
+ "language_model.model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
288
+ "language_model.model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
289
+ "language_model.model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
290
+ "language_model.model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
291
+ "language_model.model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
292
+ "language_model.model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
293
+ "language_model.model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
294
+ "language_model.model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
295
+ "language_model.model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
296
+ "language_model.model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors",
297
+ "language_model.model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
298
+ "language_model.model.layers.30.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
299
+ "language_model.model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
300
+ "language_model.model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
301
+ "language_model.model.layers.30.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
302
+ "language_model.model.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
303
+ "language_model.model.layers.30.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
304
+ "language_model.model.layers.30.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
305
+ "language_model.model.layers.30.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
306
+ "language_model.model.layers.30.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
307
+ "language_model.model.layers.30.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
308
+ "language_model.model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors",
309
+ "language_model.model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
310
+ "language_model.model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
311
+ "language_model.model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
312
+ "language_model.model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
313
+ "language_model.model.layers.31.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
314
+ "language_model.model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
315
+ "language_model.model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
316
+ "language_model.model.layers.31.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
317
+ "language_model.model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
318
+ "language_model.model.layers.31.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
319
+ "language_model.model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
320
+ "language_model.model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors",
321
+ "language_model.model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
322
+ "language_model.model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
323
+ "language_model.model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
324
+ "language_model.model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
325
+ "language_model.model.layers.32.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
326
+ "language_model.model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
327
+ "language_model.model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
328
+ "language_model.model.layers.32.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
329
+ "language_model.model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
330
+ "language_model.model.layers.32.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
331
+ "language_model.model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
332
+ "language_model.model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors",
333
+ "language_model.model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
334
+ "language_model.model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
335
+ "language_model.model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
336
+ "language_model.model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
337
+ "language_model.model.layers.33.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
338
+ "language_model.model.layers.33.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
339
+ "language_model.model.layers.33.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
340
+ "language_model.model.layers.33.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
341
+ "language_model.model.layers.33.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
342
+ "language_model.model.layers.33.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
343
+ "language_model.model.layers.33.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
344
+ "language_model.model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors",
345
+ "language_model.model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
346
+ "language_model.model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
347
+ "language_model.model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
348
+ "language_model.model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
349
+ "language_model.model.layers.34.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
350
+ "language_model.model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
351
+ "language_model.model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
352
+ "language_model.model.layers.34.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
353
+ "language_model.model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
354
+ "language_model.model.layers.34.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
355
+ "language_model.model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
356
+ "language_model.model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors",
357
+ "language_model.model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
358
+ "language_model.model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
359
+ "language_model.model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
360
+ "language_model.model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
361
+ "language_model.model.layers.35.self_attn.k_proj.bias": "model-00002-of-00002.safetensors",
362
+ "language_model.model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
363
+ "language_model.model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
364
+ "language_model.model.layers.35.self_attn.q_proj.bias": "model-00002-of-00002.safetensors",
365
+ "language_model.model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
366
+ "language_model.model.layers.35.self_attn.v_proj.bias": "model-00002-of-00002.safetensors",
367
+ "language_model.model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
368
+ "language_model.model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
369
+ "language_model.model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
370
+ "language_model.model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
371
+ "language_model.model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
372
+ "language_model.model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
373
+ "language_model.model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
374
+ "language_model.model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
375
+ "language_model.model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
376
+ "language_model.model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
377
+ "language_model.model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
378
+ "language_model.model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
379
+ "language_model.model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
380
+ "language_model.model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors",
381
+ "language_model.model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
382
+ "language_model.model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
383
+ "language_model.model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
384
+ "language_model.model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
385
+ "language_model.model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
386
+ "language_model.model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
387
+ "language_model.model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
388
+ "language_model.model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
389
+ "language_model.model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
390
+ "language_model.model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
391
+ "language_model.model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
392
+ "language_model.model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors",
393
+ "language_model.model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
394
+ "language_model.model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
395
+ "language_model.model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
396
+ "language_model.model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
397
+ "language_model.model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
398
+ "language_model.model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
399
+ "language_model.model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
400
+ "language_model.model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
401
+ "language_model.model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
402
+ "language_model.model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
403
+ "language_model.model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
404
+ "language_model.model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors",
405
+ "language_model.model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
406
+ "language_model.model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
407
+ "language_model.model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
408
+ "language_model.model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
409
+ "language_model.model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
410
+ "language_model.model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
411
+ "language_model.model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
412
+ "language_model.model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
413
+ "language_model.model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
414
+ "language_model.model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
415
+ "language_model.model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
416
+ "language_model.model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors",
417
+ "language_model.model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
418
+ "language_model.model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
419
+ "language_model.model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
420
+ "language_model.model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
421
+ "language_model.model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
422
+ "language_model.model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
423
+ "language_model.model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
424
+ "language_model.model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
425
+ "language_model.model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
426
+ "language_model.model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
427
+ "language_model.model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
428
+ "language_model.model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors",
429
+ "language_model.model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
430
+ "language_model.model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
431
+ "language_model.model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
432
+ "language_model.model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
433
+ "language_model.model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
434
+ "language_model.model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
435
+ "language_model.model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
436
+ "language_model.model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
437
+ "language_model.model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
438
+ "language_model.model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
439
+ "language_model.model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
440
+ "language_model.model.norm.weight": "model-00002-of-00002.safetensors",
441
+ "mlp1.0.bias": "model-00002-of-00002.safetensors",
442
+ "mlp1.0.weight": "model-00002-of-00002.safetensors",
443
+ "mlp1.1.bias": "model-00002-of-00002.safetensors",
444
+ "mlp1.1.weight": "model-00002-of-00002.safetensors",
445
+ "mlp1.3.bias": "model-00002-of-00002.safetensors",
446
+ "mlp1.3.weight": "model-00002-of-00002.safetensors",
447
+ "vision_model.embeddings.class_embedding": "model-00001-of-00002.safetensors",
448
+ "vision_model.embeddings.patch_embedding.bias": "model-00001-of-00002.safetensors",
449
+ "vision_model.embeddings.patch_embedding.weight": "model-00001-of-00002.safetensors",
450
+ "vision_model.embeddings.position_embedding": "model-00001-of-00002.safetensors",
451
+ "vision_model.encoder.layers.0.attn.proj.bias": "model-00001-of-00002.safetensors",
452
+ "vision_model.encoder.layers.0.attn.proj.weight": "model-00001-of-00002.safetensors",
453
+ "vision_model.encoder.layers.0.attn.qkv.bias": "model-00001-of-00002.safetensors",
454
+ "vision_model.encoder.layers.0.attn.qkv.weight": "model-00001-of-00002.safetensors",
455
+ "vision_model.encoder.layers.0.ls1": "model-00001-of-00002.safetensors",
456
+ "vision_model.encoder.layers.0.ls2": "model-00001-of-00002.safetensors",
457
+ "vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00002.safetensors",
458
+ "vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00002.safetensors",
459
+ "vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00002.safetensors",
460
+ "vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00002.safetensors",
461
+ "vision_model.encoder.layers.0.norm1.bias": "model-00001-of-00002.safetensors",
462
+ "vision_model.encoder.layers.0.norm1.weight": "model-00001-of-00002.safetensors",
463
+ "vision_model.encoder.layers.0.norm2.bias": "model-00001-of-00002.safetensors",
464
+ "vision_model.encoder.layers.0.norm2.weight": "model-00001-of-00002.safetensors",
465
+ "vision_model.encoder.layers.1.attn.proj.bias": "model-00001-of-00002.safetensors",
466
+ "vision_model.encoder.layers.1.attn.proj.weight": "model-00001-of-00002.safetensors",
467
+ "vision_model.encoder.layers.1.attn.qkv.bias": "model-00001-of-00002.safetensors",
468
+ "vision_model.encoder.layers.1.attn.qkv.weight": "model-00001-of-00002.safetensors",
469
+ "vision_model.encoder.layers.1.ls1": "model-00001-of-00002.safetensors",
470
+ "vision_model.encoder.layers.1.ls2": "model-00001-of-00002.safetensors",
471
+ "vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00002.safetensors",
472
+ "vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00002.safetensors",
473
+ "vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00002.safetensors",
474
+ "vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00002.safetensors",
475
+ "vision_model.encoder.layers.1.norm1.bias": "model-00001-of-00002.safetensors",
476
+ "vision_model.encoder.layers.1.norm1.weight": "model-00001-of-00002.safetensors",
477
+ "vision_model.encoder.layers.1.norm2.bias": "model-00001-of-00002.safetensors",
478
+ "vision_model.encoder.layers.1.norm2.weight": "model-00001-of-00002.safetensors",
479
+ "vision_model.encoder.layers.10.attn.proj.bias": "model-00001-of-00002.safetensors",
480
+ "vision_model.encoder.layers.10.attn.proj.weight": "model-00001-of-00002.safetensors",
481
+ "vision_model.encoder.layers.10.attn.qkv.bias": "model-00001-of-00002.safetensors",
482
+ "vision_model.encoder.layers.10.attn.qkv.weight": "model-00001-of-00002.safetensors",
483
+ "vision_model.encoder.layers.10.ls1": "model-00001-of-00002.safetensors",
484
+ "vision_model.encoder.layers.10.ls2": "model-00001-of-00002.safetensors",
485
+ "vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00002.safetensors",
486
+ "vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00002.safetensors",
487
+ "vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00002.safetensors",
488
+ "vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00002.safetensors",
489
+ "vision_model.encoder.layers.10.norm1.bias": "model-00001-of-00002.safetensors",
490
+ "vision_model.encoder.layers.10.norm1.weight": "model-00001-of-00002.safetensors",
491
+ "vision_model.encoder.layers.10.norm2.bias": "model-00001-of-00002.safetensors",
492
+ "vision_model.encoder.layers.10.norm2.weight": "model-00001-of-00002.safetensors",
493
+ "vision_model.encoder.layers.11.attn.proj.bias": "model-00001-of-00002.safetensors",
494
+ "vision_model.encoder.layers.11.attn.proj.weight": "model-00001-of-00002.safetensors",
495
+ "vision_model.encoder.layers.11.attn.qkv.bias": "model-00001-of-00002.safetensors",
496
+ "vision_model.encoder.layers.11.attn.qkv.weight": "model-00001-of-00002.safetensors",
497
+ "vision_model.encoder.layers.11.ls1": "model-00001-of-00002.safetensors",
498
+ "vision_model.encoder.layers.11.ls2": "model-00001-of-00002.safetensors",
499
+ "vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00002.safetensors",
500
+ "vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00002.safetensors",
501
+ "vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00002.safetensors",
502
+ "vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00002.safetensors",
503
+ "vision_model.encoder.layers.11.norm1.bias": "model-00001-of-00002.safetensors",
504
+ "vision_model.encoder.layers.11.norm1.weight": "model-00001-of-00002.safetensors",
505
+ "vision_model.encoder.layers.11.norm2.bias": "model-00001-of-00002.safetensors",
506
+ "vision_model.encoder.layers.11.norm2.weight": "model-00001-of-00002.safetensors",
507
+ "vision_model.encoder.layers.12.attn.proj.bias": "model-00001-of-00002.safetensors",
508
+ "vision_model.encoder.layers.12.attn.proj.weight": "model-00001-of-00002.safetensors",
509
+ "vision_model.encoder.layers.12.attn.qkv.bias": "model-00001-of-00002.safetensors",
510
+ "vision_model.encoder.layers.12.attn.qkv.weight": "model-00001-of-00002.safetensors",
511
+ "vision_model.encoder.layers.12.ls1": "model-00001-of-00002.safetensors",
512
+ "vision_model.encoder.layers.12.ls2": "model-00001-of-00002.safetensors",
513
+ "vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00002.safetensors",
514
+ "vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00002.safetensors",
515
+ "vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00002.safetensors",
516
+ "vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00002.safetensors",
517
+ "vision_model.encoder.layers.12.norm1.bias": "model-00001-of-00002.safetensors",
518
+ "vision_model.encoder.layers.12.norm1.weight": "model-00001-of-00002.safetensors",
519
+ "vision_model.encoder.layers.12.norm2.bias": "model-00001-of-00002.safetensors",
520
+ "vision_model.encoder.layers.12.norm2.weight": "model-00001-of-00002.safetensors",
521
+ "vision_model.encoder.layers.13.attn.proj.bias": "model-00001-of-00002.safetensors",
522
+ "vision_model.encoder.layers.13.attn.proj.weight": "model-00001-of-00002.safetensors",
523
+ "vision_model.encoder.layers.13.attn.qkv.bias": "model-00001-of-00002.safetensors",
524
+ "vision_model.encoder.layers.13.attn.qkv.weight": "model-00001-of-00002.safetensors",
525
+ "vision_model.encoder.layers.13.ls1": "model-00001-of-00002.safetensors",
526
+ "vision_model.encoder.layers.13.ls2": "model-00001-of-00002.safetensors",
527
+ "vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00002.safetensors",
528
+ "vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00002.safetensors",
529
+ "vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00002.safetensors",
530
+ "vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00002.safetensors",
531
+ "vision_model.encoder.layers.13.norm1.bias": "model-00001-of-00002.safetensors",
532
+ "vision_model.encoder.layers.13.norm1.weight": "model-00001-of-00002.safetensors",
533
+ "vision_model.encoder.layers.13.norm2.bias": "model-00001-of-00002.safetensors",
534
+ "vision_model.encoder.layers.13.norm2.weight": "model-00001-of-00002.safetensors",
535
+ "vision_model.encoder.layers.14.attn.proj.bias": "model-00001-of-00002.safetensors",
536
+ "vision_model.encoder.layers.14.attn.proj.weight": "model-00001-of-00002.safetensors",
537
+ "vision_model.encoder.layers.14.attn.qkv.bias": "model-00001-of-00002.safetensors",
538
+ "vision_model.encoder.layers.14.attn.qkv.weight": "model-00001-of-00002.safetensors",
539
+ "vision_model.encoder.layers.14.ls1": "model-00001-of-00002.safetensors",
540
+ "vision_model.encoder.layers.14.ls2": "model-00001-of-00002.safetensors",
541
+ "vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00002.safetensors",
542
+ "vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00002.safetensors",
543
+ "vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00002.safetensors",
544
+ "vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00002.safetensors",
545
+ "vision_model.encoder.layers.14.norm1.bias": "model-00001-of-00002.safetensors",
546
+ "vision_model.encoder.layers.14.norm1.weight": "model-00001-of-00002.safetensors",
547
+ "vision_model.encoder.layers.14.norm2.bias": "model-00001-of-00002.safetensors",
548
+ "vision_model.encoder.layers.14.norm2.weight": "model-00001-of-00002.safetensors",
549
+ "vision_model.encoder.layers.15.attn.proj.bias": "model-00001-of-00002.safetensors",
550
+ "vision_model.encoder.layers.15.attn.proj.weight": "model-00001-of-00002.safetensors",
551
+ "vision_model.encoder.layers.15.attn.qkv.bias": "model-00001-of-00002.safetensors",
552
+ "vision_model.encoder.layers.15.attn.qkv.weight": "model-00001-of-00002.safetensors",
553
+ "vision_model.encoder.layers.15.ls1": "model-00001-of-00002.safetensors",
554
+ "vision_model.encoder.layers.15.ls2": "model-00001-of-00002.safetensors",
555
+ "vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00002.safetensors",
556
+ "vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00002.safetensors",
557
+ "vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00002.safetensors",
558
+ "vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00002.safetensors",
559
+ "vision_model.encoder.layers.15.norm1.bias": "model-00001-of-00002.safetensors",
560
+ "vision_model.encoder.layers.15.norm1.weight": "model-00001-of-00002.safetensors",
561
+ "vision_model.encoder.layers.15.norm2.bias": "model-00001-of-00002.safetensors",
562
+ "vision_model.encoder.layers.15.norm2.weight": "model-00001-of-00002.safetensors",
563
+ "vision_model.encoder.layers.16.attn.proj.bias": "model-00001-of-00002.safetensors",
564
+ "vision_model.encoder.layers.16.attn.proj.weight": "model-00001-of-00002.safetensors",
565
+ "vision_model.encoder.layers.16.attn.qkv.bias": "model-00001-of-00002.safetensors",
566
+ "vision_model.encoder.layers.16.attn.qkv.weight": "model-00001-of-00002.safetensors",
567
+ "vision_model.encoder.layers.16.ls1": "model-00001-of-00002.safetensors",
568
+ "vision_model.encoder.layers.16.ls2": "model-00001-of-00002.safetensors",
569
+ "vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00002.safetensors",
570
+ "vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00002.safetensors",
571
+ "vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00002.safetensors",
572
+ "vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00002.safetensors",
573
+ "vision_model.encoder.layers.16.norm1.bias": "model-00001-of-00002.safetensors",
574
+ "vision_model.encoder.layers.16.norm1.weight": "model-00001-of-00002.safetensors",
575
+ "vision_model.encoder.layers.16.norm2.bias": "model-00001-of-00002.safetensors",
576
+ "vision_model.encoder.layers.16.norm2.weight": "model-00001-of-00002.safetensors",
577
+ "vision_model.encoder.layers.17.attn.proj.bias": "model-00001-of-00002.safetensors",
578
+ "vision_model.encoder.layers.17.attn.proj.weight": "model-00001-of-00002.safetensors",
579
+ "vision_model.encoder.layers.17.attn.qkv.bias": "model-00001-of-00002.safetensors",
580
+ "vision_model.encoder.layers.17.attn.qkv.weight": "model-00001-of-00002.safetensors",
581
+ "vision_model.encoder.layers.17.ls1": "model-00001-of-00002.safetensors",
582
+ "vision_model.encoder.layers.17.ls2": "model-00001-of-00002.safetensors",
583
+ "vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00002.safetensors",
584
+ "vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00002.safetensors",
585
+ "vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00002.safetensors",
586
+ "vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00002.safetensors",
587
+ "vision_model.encoder.layers.17.norm1.bias": "model-00001-of-00002.safetensors",
588
+ "vision_model.encoder.layers.17.norm1.weight": "model-00001-of-00002.safetensors",
589
+ "vision_model.encoder.layers.17.norm2.bias": "model-00001-of-00002.safetensors",
590
+ "vision_model.encoder.layers.17.norm2.weight": "model-00001-of-00002.safetensors",
591
+ "vision_model.encoder.layers.18.attn.proj.bias": "model-00001-of-00002.safetensors",
592
+ "vision_model.encoder.layers.18.attn.proj.weight": "model-00001-of-00002.safetensors",
593
+ "vision_model.encoder.layers.18.attn.qkv.bias": "model-00001-of-00002.safetensors",
594
+ "vision_model.encoder.layers.18.attn.qkv.weight": "model-00001-of-00002.safetensors",
595
+ "vision_model.encoder.layers.18.ls1": "model-00001-of-00002.safetensors",
596
+ "vision_model.encoder.layers.18.ls2": "model-00001-of-00002.safetensors",
597
+ "vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00002.safetensors",
598
+ "vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00002.safetensors",
599
+ "vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00002.safetensors",
600
+ "vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00002.safetensors",
601
+ "vision_model.encoder.layers.18.norm1.bias": "model-00001-of-00002.safetensors",
602
+ "vision_model.encoder.layers.18.norm1.weight": "model-00001-of-00002.safetensors",
603
+ "vision_model.encoder.layers.18.norm2.bias": "model-00001-of-00002.safetensors",
604
+ "vision_model.encoder.layers.18.norm2.weight": "model-00001-of-00002.safetensors",
605
+ "vision_model.encoder.layers.19.attn.proj.bias": "model-00001-of-00002.safetensors",
606
+ "vision_model.encoder.layers.19.attn.proj.weight": "model-00001-of-00002.safetensors",
607
+ "vision_model.encoder.layers.19.attn.qkv.bias": "model-00001-of-00002.safetensors",
608
+ "vision_model.encoder.layers.19.attn.qkv.weight": "model-00001-of-00002.safetensors",
609
+ "vision_model.encoder.layers.19.ls1": "model-00001-of-00002.safetensors",
610
+ "vision_model.encoder.layers.19.ls2": "model-00001-of-00002.safetensors",
611
+ "vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00002.safetensors",
612
+ "vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00002.safetensors",
613
+ "vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00002.safetensors",
614
+ "vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00002.safetensors",
615
+ "vision_model.encoder.layers.19.norm1.bias": "model-00001-of-00002.safetensors",
616
+ "vision_model.encoder.layers.19.norm1.weight": "model-00001-of-00002.safetensors",
617
+ "vision_model.encoder.layers.19.norm2.bias": "model-00001-of-00002.safetensors",
618
+ "vision_model.encoder.layers.19.norm2.weight": "model-00001-of-00002.safetensors",
619
+ "vision_model.encoder.layers.2.attn.proj.bias": "model-00001-of-00002.safetensors",
620
+ "vision_model.encoder.layers.2.attn.proj.weight": "model-00001-of-00002.safetensors",
621
+ "vision_model.encoder.layers.2.attn.qkv.bias": "model-00001-of-00002.safetensors",
622
+ "vision_model.encoder.layers.2.attn.qkv.weight": "model-00001-of-00002.safetensors",
623
+ "vision_model.encoder.layers.2.ls1": "model-00001-of-00002.safetensors",
624
+ "vision_model.encoder.layers.2.ls2": "model-00001-of-00002.safetensors",
625
+ "vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00002.safetensors",
626
+ "vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00002.safetensors",
627
+ "vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00002.safetensors",
628
+ "vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00002.safetensors",
629
+ "vision_model.encoder.layers.2.norm1.bias": "model-00001-of-00002.safetensors",
630
+ "vision_model.encoder.layers.2.norm1.weight": "model-00001-of-00002.safetensors",
631
+ "vision_model.encoder.layers.2.norm2.bias": "model-00001-of-00002.safetensors",
632
+ "vision_model.encoder.layers.2.norm2.weight": "model-00001-of-00002.safetensors",
633
+ "vision_model.encoder.layers.20.attn.proj.bias": "model-00001-of-00002.safetensors",
634
+ "vision_model.encoder.layers.20.attn.proj.weight": "model-00001-of-00002.safetensors",
635
+ "vision_model.encoder.layers.20.attn.qkv.bias": "model-00001-of-00002.safetensors",
636
+ "vision_model.encoder.layers.20.attn.qkv.weight": "model-00001-of-00002.safetensors",
637
+ "vision_model.encoder.layers.20.ls1": "model-00001-of-00002.safetensors",
638
+ "vision_model.encoder.layers.20.ls2": "model-00001-of-00002.safetensors",
639
+ "vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00002.safetensors",
640
+ "vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00002.safetensors",
641
+ "vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00002.safetensors",
642
+ "vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00002.safetensors",
643
+ "vision_model.encoder.layers.20.norm1.bias": "model-00001-of-00002.safetensors",
644
+ "vision_model.encoder.layers.20.norm1.weight": "model-00001-of-00002.safetensors",
645
+ "vision_model.encoder.layers.20.norm2.bias": "model-00001-of-00002.safetensors",
646
+ "vision_model.encoder.layers.20.norm2.weight": "model-00001-of-00002.safetensors",
647
+ "vision_model.encoder.layers.21.attn.proj.bias": "model-00001-of-00002.safetensors",
648
+ "vision_model.encoder.layers.21.attn.proj.weight": "model-00001-of-00002.safetensors",
649
+ "vision_model.encoder.layers.21.attn.qkv.bias": "model-00001-of-00002.safetensors",
650
+ "vision_model.encoder.layers.21.attn.qkv.weight": "model-00001-of-00002.safetensors",
651
+ "vision_model.encoder.layers.21.ls1": "model-00001-of-00002.safetensors",
652
+ "vision_model.encoder.layers.21.ls2": "model-00001-of-00002.safetensors",
653
+ "vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00002.safetensors",
654
+ "vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00002.safetensors",
655
+ "vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00002.safetensors",
656
+ "vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00002.safetensors",
657
+ "vision_model.encoder.layers.21.norm1.bias": "model-00001-of-00002.safetensors",
658
+ "vision_model.encoder.layers.21.norm1.weight": "model-00001-of-00002.safetensors",
659
+ "vision_model.encoder.layers.21.norm2.bias": "model-00001-of-00002.safetensors",
660
+ "vision_model.encoder.layers.21.norm2.weight": "model-00001-of-00002.safetensors",
661
+ "vision_model.encoder.layers.22.attn.proj.bias": "model-00001-of-00002.safetensors",
662
+ "vision_model.encoder.layers.22.attn.proj.weight": "model-00001-of-00002.safetensors",
663
+ "vision_model.encoder.layers.22.attn.qkv.bias": "model-00001-of-00002.safetensors",
664
+ "vision_model.encoder.layers.22.attn.qkv.weight": "model-00001-of-00002.safetensors",
665
+ "vision_model.encoder.layers.22.ls1": "model-00001-of-00002.safetensors",
666
+ "vision_model.encoder.layers.22.ls2": "model-00001-of-00002.safetensors",
667
+ "vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00002.safetensors",
668
+ "vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00002.safetensors",
669
+ "vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00002.safetensors",
670
+ "vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00002.safetensors",
671
+ "vision_model.encoder.layers.22.norm1.bias": "model-00001-of-00002.safetensors",
672
+ "vision_model.encoder.layers.22.norm1.weight": "model-00001-of-00002.safetensors",
673
+ "vision_model.encoder.layers.22.norm2.bias": "model-00001-of-00002.safetensors",
674
+ "vision_model.encoder.layers.22.norm2.weight": "model-00001-of-00002.safetensors",
675
+ "vision_model.encoder.layers.23.attn.proj.bias": "model-00001-of-00002.safetensors",
676
+ "vision_model.encoder.layers.23.attn.proj.weight": "model-00001-of-00002.safetensors",
677
+ "vision_model.encoder.layers.23.attn.qkv.bias": "model-00001-of-00002.safetensors",
678
+ "vision_model.encoder.layers.23.attn.qkv.weight": "model-00001-of-00002.safetensors",
679
+ "vision_model.encoder.layers.23.ls1": "model-00001-of-00002.safetensors",
680
+ "vision_model.encoder.layers.23.ls2": "model-00001-of-00002.safetensors",
681
+ "vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00002.safetensors",
682
+ "vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00002.safetensors",
683
+ "vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00002.safetensors",
684
+ "vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00002.safetensors",
685
+ "vision_model.encoder.layers.23.norm1.bias": "model-00001-of-00002.safetensors",
686
+ "vision_model.encoder.layers.23.norm1.weight": "model-00001-of-00002.safetensors",
687
+ "vision_model.encoder.layers.23.norm2.bias": "model-00001-of-00002.safetensors",
688
+ "vision_model.encoder.layers.23.norm2.weight": "model-00001-of-00002.safetensors",
689
+ "vision_model.encoder.layers.3.attn.proj.bias": "model-00001-of-00002.safetensors",
690
+ "vision_model.encoder.layers.3.attn.proj.weight": "model-00001-of-00002.safetensors",
691
+ "vision_model.encoder.layers.3.attn.qkv.bias": "model-00001-of-00002.safetensors",
692
+ "vision_model.encoder.layers.3.attn.qkv.weight": "model-00001-of-00002.safetensors",
693
+ "vision_model.encoder.layers.3.ls1": "model-00001-of-00002.safetensors",
694
+ "vision_model.encoder.layers.3.ls2": "model-00001-of-00002.safetensors",
695
+ "vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00002.safetensors",
696
+ "vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00002.safetensors",
697
+ "vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00002.safetensors",
698
+ "vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00002.safetensors",
699
+ "vision_model.encoder.layers.3.norm1.bias": "model-00001-of-00002.safetensors",
700
+ "vision_model.encoder.layers.3.norm1.weight": "model-00001-of-00002.safetensors",
701
+ "vision_model.encoder.layers.3.norm2.bias": "model-00001-of-00002.safetensors",
702
+ "vision_model.encoder.layers.3.norm2.weight": "model-00001-of-00002.safetensors",
703
+ "vision_model.encoder.layers.4.attn.proj.bias": "model-00001-of-00002.safetensors",
704
+ "vision_model.encoder.layers.4.attn.proj.weight": "model-00001-of-00002.safetensors",
705
+ "vision_model.encoder.layers.4.attn.qkv.bias": "model-00001-of-00002.safetensors",
706
+ "vision_model.encoder.layers.4.attn.qkv.weight": "model-00001-of-00002.safetensors",
707
+ "vision_model.encoder.layers.4.ls1": "model-00001-of-00002.safetensors",
708
+ "vision_model.encoder.layers.4.ls2": "model-00001-of-00002.safetensors",
709
+ "vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00002.safetensors",
710
+ "vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00002.safetensors",
711
+ "vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00002.safetensors",
712
+ "vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00002.safetensors",
713
+ "vision_model.encoder.layers.4.norm1.bias": "model-00001-of-00002.safetensors",
714
+ "vision_model.encoder.layers.4.norm1.weight": "model-00001-of-00002.safetensors",
715
+ "vision_model.encoder.layers.4.norm2.bias": "model-00001-of-00002.safetensors",
716
+ "vision_model.encoder.layers.4.norm2.weight": "model-00001-of-00002.safetensors",
717
+ "vision_model.encoder.layers.5.attn.proj.bias": "model-00001-of-00002.safetensors",
718
+ "vision_model.encoder.layers.5.attn.proj.weight": "model-00001-of-00002.safetensors",
719
+ "vision_model.encoder.layers.5.attn.qkv.bias": "model-00001-of-00002.safetensors",
720
+ "vision_model.encoder.layers.5.attn.qkv.weight": "model-00001-of-00002.safetensors",
721
+ "vision_model.encoder.layers.5.ls1": "model-00001-of-00002.safetensors",
722
+ "vision_model.encoder.layers.5.ls2": "model-00001-of-00002.safetensors",
723
+ "vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00002.safetensors",
724
+ "vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00002.safetensors",
725
+ "vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00002.safetensors",
726
+ "vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00002.safetensors",
727
+ "vision_model.encoder.layers.5.norm1.bias": "model-00001-of-00002.safetensors",
728
+ "vision_model.encoder.layers.5.norm1.weight": "model-00001-of-00002.safetensors",
729
+ "vision_model.encoder.layers.5.norm2.bias": "model-00001-of-00002.safetensors",
730
+ "vision_model.encoder.layers.5.norm2.weight": "model-00001-of-00002.safetensors",
731
+ "vision_model.encoder.layers.6.attn.proj.bias": "model-00001-of-00002.safetensors",
732
+ "vision_model.encoder.layers.6.attn.proj.weight": "model-00001-of-00002.safetensors",
733
+ "vision_model.encoder.layers.6.attn.qkv.bias": "model-00001-of-00002.safetensors",
734
+ "vision_model.encoder.layers.6.attn.qkv.weight": "model-00001-of-00002.safetensors",
735
+ "vision_model.encoder.layers.6.ls1": "model-00001-of-00002.safetensors",
736
+ "vision_model.encoder.layers.6.ls2": "model-00001-of-00002.safetensors",
737
+ "vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00002.safetensors",
738
+ "vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00002.safetensors",
739
+ "vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00002.safetensors",
740
+ "vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00002.safetensors",
741
+ "vision_model.encoder.layers.6.norm1.bias": "model-00001-of-00002.safetensors",
742
+ "vision_model.encoder.layers.6.norm1.weight": "model-00001-of-00002.safetensors",
743
+ "vision_model.encoder.layers.6.norm2.bias": "model-00001-of-00002.safetensors",
744
+ "vision_model.encoder.layers.6.norm2.weight": "model-00001-of-00002.safetensors",
745
+ "vision_model.encoder.layers.7.attn.proj.bias": "model-00001-of-00002.safetensors",
746
+ "vision_model.encoder.layers.7.attn.proj.weight": "model-00001-of-00002.safetensors",
747
+ "vision_model.encoder.layers.7.attn.qkv.bias": "model-00001-of-00002.safetensors",
748
+ "vision_model.encoder.layers.7.attn.qkv.weight": "model-00001-of-00002.safetensors",
749
+ "vision_model.encoder.layers.7.ls1": "model-00001-of-00002.safetensors",
750
+ "vision_model.encoder.layers.7.ls2": "model-00001-of-00002.safetensors",
751
+ "vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00002.safetensors",
752
+ "vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00002.safetensors",
753
+ "vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00002.safetensors",
754
+ "vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00002.safetensors",
755
+ "vision_model.encoder.layers.7.norm1.bias": "model-00001-of-00002.safetensors",
756
+ "vision_model.encoder.layers.7.norm1.weight": "model-00001-of-00002.safetensors",
757
+ "vision_model.encoder.layers.7.norm2.bias": "model-00001-of-00002.safetensors",
758
+ "vision_model.encoder.layers.7.norm2.weight": "model-00001-of-00002.safetensors",
759
+ "vision_model.encoder.layers.8.attn.proj.bias": "model-00001-of-00002.safetensors",
760
+ "vision_model.encoder.layers.8.attn.proj.weight": "model-00001-of-00002.safetensors",
761
+ "vision_model.encoder.layers.8.attn.qkv.bias": "model-00001-of-00002.safetensors",
762
+ "vision_model.encoder.layers.8.attn.qkv.weight": "model-00001-of-00002.safetensors",
763
+ "vision_model.encoder.layers.8.ls1": "model-00001-of-00002.safetensors",
764
+ "vision_model.encoder.layers.8.ls2": "model-00001-of-00002.safetensors",
765
+ "vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00002.safetensors",
766
+ "vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00002.safetensors",
767
+ "vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00002.safetensors",
768
+ "vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00002.safetensors",
769
+ "vision_model.encoder.layers.8.norm1.bias": "model-00001-of-00002.safetensors",
770
+ "vision_model.encoder.layers.8.norm1.weight": "model-00001-of-00002.safetensors",
771
+ "vision_model.encoder.layers.8.norm2.bias": "model-00001-of-00002.safetensors",
772
+ "vision_model.encoder.layers.8.norm2.weight": "model-00001-of-00002.safetensors",
773
+ "vision_model.encoder.layers.9.attn.proj.bias": "model-00001-of-00002.safetensors",
774
+ "vision_model.encoder.layers.9.attn.proj.weight": "model-00001-of-00002.safetensors",
775
+ "vision_model.encoder.layers.9.attn.qkv.bias": "model-00001-of-00002.safetensors",
776
+ "vision_model.encoder.layers.9.attn.qkv.weight": "model-00001-of-00002.safetensors",
777
+ "vision_model.encoder.layers.9.ls1": "model-00001-of-00002.safetensors",
778
+ "vision_model.encoder.layers.9.ls2": "model-00001-of-00002.safetensors",
779
+ "vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00002.safetensors",
780
+ "vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00002.safetensors",
781
+ "vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00002.safetensors",
782
+ "vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00002.safetensors",
783
+ "vision_model.encoder.layers.9.norm1.bias": "model-00001-of-00002.safetensors",
784
+ "vision_model.encoder.layers.9.norm1.weight": "model-00001-of-00002.safetensors",
785
+ "vision_model.encoder.layers.9.norm2.bias": "model-00001-of-00002.safetensors",
786
+ "vision_model.encoder.layers.9.norm2.weight": "model-00001-of-00002.safetensors"
787
+ }
788
+ }
modeling_intern_vit.py ADDED
@@ -0,0 +1,430 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # --------------------------------------------------------
2
+ # InternVL
3
+ # Copyright (c) 2024 OpenGVLab
4
+ # Licensed under The MIT License [see LICENSE for details]
5
+ # --------------------------------------------------------
6
+
7
+ from typing import Optional, Tuple, Union
8
+
9
+ import torch
10
+ import torch.nn.functional as F
11
+ import torch.utils.checkpoint
12
+ from einops import rearrange
13
+ from timm.models.layers import DropPath
14
+ from torch import nn
15
+ from transformers.activations import ACT2FN
16
+ from transformers.modeling_outputs import (BaseModelOutput,
17
+ BaseModelOutputWithPooling)
18
+ from transformers.modeling_utils import PreTrainedModel
19
+ from transformers.utils import logging
20
+
21
+ from .configuration_intern_vit import InternVisionConfig
22
+
23
+ try:
24
+ from flash_attn.bert_padding import pad_input, unpad_input
25
+ from flash_attn.flash_attn_interface import \
26
+ flash_attn_varlen_qkvpacked_func
27
+ has_flash_attn = True
28
+ except:
29
+ print('FlashAttention2 is not installed.')
30
+ has_flash_attn = False
31
+
32
+ logger = logging.get_logger(__name__)
33
+
34
+
35
+ class FlashAttention(nn.Module):
36
+ """Implement the scaled dot product attention with softmax.
37
+ Arguments
38
+ ---------
39
+ softmax_scale: The temperature to use for the softmax attention.
40
+ (default: 1/sqrt(d_keys) where d_keys is computed at
41
+ runtime)
42
+ attention_dropout: The dropout rate to apply to the attention
43
+ (default: 0.0)
44
+ """
45
+
46
+ def __init__(self, softmax_scale=None, attention_dropout=0.0, device=None, dtype=None):
47
+ super().__init__()
48
+ self.softmax_scale = softmax_scale
49
+ self.dropout_p = attention_dropout
50
+
51
+ def forward(self, qkv, key_padding_mask=None, causal=False, cu_seqlens=None,
52
+ max_s=None, need_weights=False):
53
+ """Implements the multihead softmax attention.
54
+ Arguments
55
+ ---------
56
+ qkv: The tensor containing the query, key, and value. (B, S, 3, H, D) if key_padding_mask is None
57
+ if unpadded: (nnz, 3, h, d)
58
+ key_padding_mask: a bool tensor of shape (B, S)
59
+ """
60
+ assert not need_weights
61
+ assert qkv.dtype in [torch.float16, torch.bfloat16]
62
+ assert qkv.is_cuda
63
+
64
+ if cu_seqlens is None:
65
+ batch_size = qkv.shape[0]
66
+ seqlen = qkv.shape[1]
67
+ if key_padding_mask is None:
68
+ qkv = rearrange(qkv, 'b s ... -> (b s) ...')
69
+ max_s = seqlen
70
+ cu_seqlens = torch.arange(0, (batch_size + 1) * seqlen, step=seqlen, dtype=torch.int32,
71
+ device=qkv.device)
72
+ output = flash_attn_varlen_qkvpacked_func(
73
+ qkv, cu_seqlens, max_s, self.dropout_p if self.training else 0.0,
74
+ softmax_scale=self.softmax_scale, causal=causal
75
+ )
76
+ output = rearrange(output, '(b s) ... -> b s ...', b=batch_size)
77
+ else:
78
+ nheads = qkv.shape[-2]
79
+ x = rearrange(qkv, 'b s three h d -> b s (three h d)')
80
+ x_unpad, indices, cu_seqlens, max_s = unpad_input(x, key_padding_mask)
81
+ x_unpad = rearrange(x_unpad, 'nnz (three h d) -> nnz three h d', three=3, h=nheads)
82
+ output_unpad = flash_attn_varlen_qkvpacked_func(
83
+ x_unpad, cu_seqlens, max_s, self.dropout_p if self.training else 0.0,
84
+ softmax_scale=self.softmax_scale, causal=causal
85
+ )
86
+ output = rearrange(pad_input(rearrange(output_unpad, 'nnz h d -> nnz (h d)'),
87
+ indices, batch_size, seqlen),
88
+ 'b s (h d) -> b s h d', h=nheads)
89
+ else:
90
+ assert max_s is not None
91
+ output = flash_attn_varlen_qkvpacked_func(
92
+ qkv, cu_seqlens, max_s, self.dropout_p if self.training else 0.0,
93
+ softmax_scale=self.softmax_scale, causal=causal
94
+ )
95
+
96
+ return output, None
97
+
98
+
99
+ class InternRMSNorm(nn.Module):
100
+ def __init__(self, hidden_size, eps=1e-6):
101
+ super().__init__()
102
+ self.weight = nn.Parameter(torch.ones(hidden_size))
103
+ self.variance_epsilon = eps
104
+
105
+ def forward(self, hidden_states):
106
+ input_dtype = hidden_states.dtype
107
+ hidden_states = hidden_states.to(torch.float32)
108
+ variance = hidden_states.pow(2).mean(-1, keepdim=True)
109
+ hidden_states = hidden_states * torch.rsqrt(variance + self.variance_epsilon)
110
+ return self.weight * hidden_states.to(input_dtype)
111
+
112
+
113
+ try:
114
+ from apex.normalization import FusedRMSNorm
115
+
116
+ InternRMSNorm = FusedRMSNorm # noqa
117
+
118
+ logger.info('Discovered apex.normalization.FusedRMSNorm - will use it instead of InternRMSNorm')
119
+ except ImportError:
120
+ # using the normal InternRMSNorm
121
+ pass
122
+ except Exception:
123
+ logger.warning('discovered apex but it failed to load, falling back to InternRMSNorm')
124
+ pass
125
+
126
+
127
+ NORM2FN = {
128
+ 'rms_norm': InternRMSNorm,
129
+ 'layer_norm': nn.LayerNorm,
130
+ }
131
+
132
+
133
+ class InternVisionEmbeddings(nn.Module):
134
+ def __init__(self, config: InternVisionConfig):
135
+ super().__init__()
136
+ self.config = config
137
+ self.embed_dim = config.hidden_size
138
+ self.image_size = config.image_size
139
+ self.patch_size = config.patch_size
140
+
141
+ self.class_embedding = nn.Parameter(
142
+ torch.randn(1, 1, self.embed_dim),
143
+ )
144
+
145
+ self.patch_embedding = nn.Conv2d(
146
+ in_channels=3, out_channels=self.embed_dim, kernel_size=self.patch_size, stride=self.patch_size
147
+ )
148
+
149
+ self.num_patches = (self.image_size // self.patch_size) ** 2
150
+ self.num_positions = self.num_patches + 1
151
+
152
+ self.position_embedding = nn.Parameter(torch.randn(1, self.num_positions, self.embed_dim))
153
+
154
+ def _get_pos_embed(self, pos_embed, H, W):
155
+ target_dtype = pos_embed.dtype
156
+ pos_embed = pos_embed.float().reshape(
157
+ 1, self.image_size // self.patch_size, self.image_size // self.patch_size, -1).permute(0, 3, 1, 2)
158
+ pos_embed = F.interpolate(pos_embed, size=(H, W), mode='bicubic', align_corners=False). \
159
+ reshape(1, -1, H * W).permute(0, 2, 1).to(target_dtype)
160
+ return pos_embed
161
+
162
+ def forward(self, pixel_values: torch.FloatTensor) -> torch.Tensor:
163
+ target_dtype = self.patch_embedding.weight.dtype
164
+ patch_embeds = self.patch_embedding(pixel_values) # shape = [*, channel, width, height]
165
+ batch_size, _, height, width = patch_embeds.shape
166
+ patch_embeds = patch_embeds.flatten(2).transpose(1, 2)
167
+ class_embeds = self.class_embedding.expand(batch_size, 1, -1).to(target_dtype)
168
+ embeddings = torch.cat([class_embeds, patch_embeds], dim=1)
169
+ position_embedding = torch.cat([
170
+ self.position_embedding[:, :1, :],
171
+ self._get_pos_embed(self.position_embedding[:, 1:, :], height, width)
172
+ ], dim=1)
173
+ embeddings = embeddings + position_embedding.to(target_dtype)
174
+ return embeddings
175
+
176
+
177
+ class InternAttention(nn.Module):
178
+ """Multi-headed attention from 'Attention Is All You Need' paper"""
179
+
180
+ def __init__(self, config: InternVisionConfig):
181
+ super().__init__()
182
+ self.config = config
183
+ self.embed_dim = config.hidden_size
184
+ self.num_heads = config.num_attention_heads
185
+ self.use_flash_attn = config.use_flash_attn and has_flash_attn
186
+ if config.use_flash_attn and not has_flash_attn:
187
+ print('Warning: Flash Attention is not available, use_flash_attn is set to False.')
188
+ self.head_dim = self.embed_dim // self.num_heads
189
+ if self.head_dim * self.num_heads != self.embed_dim:
190
+ raise ValueError(
191
+ f'embed_dim must be divisible by num_heads (got `embed_dim`: {self.embed_dim} and `num_heads`:'
192
+ f' {self.num_heads}).'
193
+ )
194
+
195
+ self.scale = self.head_dim ** -0.5
196
+ self.qkv = nn.Linear(self.embed_dim, 3 * self.embed_dim, bias=config.qkv_bias)
197
+ self.attn_drop = nn.Dropout(config.attention_dropout)
198
+ self.proj_drop = nn.Dropout(config.dropout)
199
+
200
+ self.qk_normalization = config.qk_normalization
201
+
202
+ if self.qk_normalization:
203
+ self.q_norm = InternRMSNorm(self.embed_dim, eps=config.layer_norm_eps)
204
+ self.k_norm = InternRMSNorm(self.embed_dim, eps=config.layer_norm_eps)
205
+
206
+ if self.use_flash_attn:
207
+ self.inner_attn = FlashAttention(attention_dropout=config.attention_dropout)
208
+ self.proj = nn.Linear(self.embed_dim, self.embed_dim)
209
+
210
+ def _naive_attn(self, x):
211
+ B, N, C = x.shape
212
+ qkv = self.qkv(x).reshape(B, N, 3, self.num_heads, C // self.num_heads).permute(2, 0, 3, 1, 4)
213
+ q, k, v = qkv.unbind(0) # make torchscript happy (cannot use tensor as tuple)
214
+
215
+ if self.qk_normalization:
216
+ B_, H_, N_, D_ = q.shape
217
+ q = self.q_norm(q.transpose(1, 2).flatten(-2, -1)).view(B_, N_, H_, D_).transpose(1, 2)
218
+ k = self.k_norm(k.transpose(1, 2).flatten(-2, -1)).view(B_, N_, H_, D_).transpose(1, 2)
219
+
220
+ attn = ((q * self.scale) @ k.transpose(-2, -1))
221
+ attn = attn.softmax(dim=-1)
222
+ attn = self.attn_drop(attn)
223
+
224
+ x = (attn @ v).transpose(1, 2).reshape(B, N, C)
225
+ x = self.proj(x)
226
+ x = self.proj_drop(x)
227
+ return x
228
+
229
+ def _flash_attn(self, x, key_padding_mask=None, need_weights=False):
230
+ qkv = self.qkv(x)
231
+ qkv = rearrange(qkv, 'b s (three h d) -> b s three h d', three=3, h=self.num_heads)
232
+
233
+ if self.qk_normalization:
234
+ q, k, v = qkv.unbind(2)
235
+ q = self.q_norm(q.flatten(-2, -1)).view(q.shape)
236
+ k = self.k_norm(k.flatten(-2, -1)).view(k.shape)
237
+ qkv = torch.stack([q, k, v], dim=2)
238
+
239
+ context, _ = self.inner_attn(
240
+ qkv, key_padding_mask=key_padding_mask, need_weights=need_weights, causal=False
241
+ )
242
+ outs = self.proj(rearrange(context, 'b s h d -> b s (h d)'))
243
+ outs = self.proj_drop(outs)
244
+ return outs
245
+
246
+ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
247
+ x = self._naive_attn(hidden_states) if not self.use_flash_attn else self._flash_attn(hidden_states)
248
+ return x
249
+
250
+
251
+ class InternMLP(nn.Module):
252
+ def __init__(self, config: InternVisionConfig):
253
+ super().__init__()
254
+ self.config = config
255
+ self.act = ACT2FN[config.hidden_act]
256
+ self.fc1 = nn.Linear(config.hidden_size, config.intermediate_size)
257
+ self.fc2 = nn.Linear(config.intermediate_size, config.hidden_size)
258
+
259
+ def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
260
+ hidden_states = self.fc1(hidden_states)
261
+ hidden_states = self.act(hidden_states)
262
+ hidden_states = self.fc2(hidden_states)
263
+ return hidden_states
264
+
265
+
266
+ class InternVisionEncoderLayer(nn.Module):
267
+ def __init__(self, config: InternVisionConfig, drop_path_rate: float):
268
+ super().__init__()
269
+ self.embed_dim = config.hidden_size
270
+ self.intermediate_size = config.intermediate_size
271
+ self.norm_type = config.norm_type
272
+
273
+ self.attn = InternAttention(config)
274
+ self.mlp = InternMLP(config)
275
+ self.norm1 = NORM2FN[self.norm_type](self.embed_dim, eps=config.layer_norm_eps)
276
+ self.norm2 = NORM2FN[self.norm_type](self.embed_dim, eps=config.layer_norm_eps)
277
+
278
+ self.ls1 = nn.Parameter(config.initializer_factor * torch.ones(self.embed_dim))
279
+ self.ls2 = nn.Parameter(config.initializer_factor * torch.ones(self.embed_dim))
280
+ self.drop_path1 = DropPath(drop_path_rate) if drop_path_rate > 0. else nn.Identity()
281
+ self.drop_path2 = DropPath(drop_path_rate) if drop_path_rate > 0. else nn.Identity()
282
+
283
+ def forward(
284
+ self,
285
+ hidden_states: torch.Tensor,
286
+ ) -> Tuple[torch.FloatTensor, Optional[torch.FloatTensor], Optional[Tuple[torch.FloatTensor]]]:
287
+ """
288
+ Args:
289
+ hidden_states (`Tuple[torch.FloatTensor, Optional[torch.FloatTensor]]`): input to the layer of shape `(batch, seq_len, embed_dim)`
290
+ """
291
+ hidden_states = hidden_states + self.drop_path1(self.attn(self.norm1(hidden_states).to(hidden_states.dtype)) * self.ls1)
292
+
293
+ hidden_states = hidden_states + self.drop_path2(self.mlp(self.norm2(hidden_states).to(hidden_states.dtype)) * self.ls2)
294
+
295
+ return hidden_states
296
+
297
+
298
+ class InternVisionEncoder(nn.Module):
299
+ """
300
+ Transformer encoder consisting of `config.num_hidden_layers` self attention layers. Each layer is a
301
+ [`InternEncoderLayer`].
302
+
303
+ Args:
304
+ config (`InternConfig`):
305
+ The corresponding vision configuration for the `InternEncoder`.
306
+ """
307
+
308
+ def __init__(self, config: InternVisionConfig):
309
+ super().__init__()
310
+ self.config = config
311
+ # stochastic depth decay rule
312
+ dpr = [x.item() for x in torch.linspace(0, config.drop_path_rate, config.num_hidden_layers)]
313
+ self.layers = nn.ModuleList([
314
+ InternVisionEncoderLayer(config, dpr[idx]) for idx in range(config.num_hidden_layers)])
315
+ self.gradient_checkpointing = True
316
+
317
+ def forward(
318
+ self,
319
+ inputs_embeds,
320
+ output_hidden_states: Optional[bool] = None,
321
+ return_dict: Optional[bool] = None,
322
+ ) -> Union[Tuple, BaseModelOutput]:
323
+ r"""
324
+ Args:
325
+ inputs_embeds (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`):
326
+ Embedded representation of the inputs. Should be float, not int tokens.
327
+ output_hidden_states (`bool`, *optional*):
328
+ Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors
329
+ for more detail.
330
+ return_dict (`bool`, *optional*):
331
+ Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.
332
+ """
333
+ output_hidden_states = (
334
+ output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
335
+ )
336
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
337
+
338
+ encoder_states = () if output_hidden_states else None
339
+ hidden_states = inputs_embeds
340
+
341
+ for idx, encoder_layer in enumerate(self.layers):
342
+ if output_hidden_states:
343
+ encoder_states = encoder_states + (hidden_states,)
344
+ if self.gradient_checkpointing and self.training:
345
+ layer_outputs = torch.utils.checkpoint.checkpoint(
346
+ encoder_layer,
347
+ hidden_states)
348
+ else:
349
+ layer_outputs = encoder_layer(
350
+ hidden_states,
351
+ )
352
+ hidden_states = layer_outputs
353
+
354
+ if output_hidden_states:
355
+ encoder_states = encoder_states + (hidden_states,)
356
+
357
+ if not return_dict:
358
+ return tuple(v for v in [hidden_states, encoder_states] if v is not None)
359
+ return BaseModelOutput(
360
+ last_hidden_state=hidden_states, hidden_states=encoder_states
361
+ )
362
+
363
+
364
+ class InternVisionModel(PreTrainedModel):
365
+ main_input_name = 'pixel_values'
366
+ _supports_flash_attn_2 = True
367
+ config_class = InternVisionConfig
368
+ _no_split_modules = ['InternVisionEncoderLayer']
369
+
370
+ def __init__(self, config: InternVisionConfig):
371
+ super().__init__(config)
372
+ self.config = config
373
+
374
+ self.embeddings = InternVisionEmbeddings(config)
375
+ self.encoder = InternVisionEncoder(config)
376
+
377
+ def resize_pos_embeddings(self, old_size, new_size, patch_size):
378
+ pos_emb = self.embeddings.position_embedding
379
+ _, num_positions, embed_dim = pos_emb.shape
380
+ cls_emb = pos_emb[:, :1, :]
381
+ pos_emb = pos_emb[:, 1:, :].reshape(1, old_size // patch_size, old_size // patch_size, -1).permute(0, 3, 1, 2)
382
+ pos_emb = F.interpolate(pos_emb.float(), size=new_size // patch_size, mode='bicubic', align_corners=False)
383
+ pos_emb = pos_emb.to(cls_emb.dtype).reshape(1, embed_dim, -1).permute(0, 2, 1)
384
+ pos_emb = torch.cat([cls_emb, pos_emb], dim=1)
385
+ self.embeddings.position_embedding = nn.Parameter(pos_emb)
386
+ self.embeddings.image_size = new_size
387
+ logger.info('Resized position embeddings from {} to {}'.format(old_size, new_size))
388
+
389
+ def get_input_embeddings(self):
390
+ return self.embeddings
391
+
392
+ def forward(
393
+ self,
394
+ pixel_values: Optional[torch.FloatTensor] = None,
395
+ output_hidden_states: Optional[bool] = None,
396
+ return_dict: Optional[bool] = None,
397
+ pixel_embeds: Optional[torch.FloatTensor] = None,
398
+ ) -> Union[Tuple, BaseModelOutputWithPooling]:
399
+ output_hidden_states = (
400
+ output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states
401
+ )
402
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
403
+
404
+ if pixel_values is None and pixel_embeds is None:
405
+ raise ValueError('You have to specify pixel_values or pixel_embeds')
406
+
407
+ if pixel_embeds is not None:
408
+ hidden_states = pixel_embeds
409
+ else:
410
+ if len(pixel_values.shape) == 4:
411
+ hidden_states = self.embeddings(pixel_values)
412
+ else:
413
+ raise ValueError(f'wrong pixel_values size: {pixel_values.shape}')
414
+ encoder_outputs = self.encoder(
415
+ inputs_embeds=hidden_states,
416
+ output_hidden_states=output_hidden_states,
417
+ return_dict=return_dict,
418
+ )
419
+ last_hidden_state = encoder_outputs.last_hidden_state
420
+ pooled_output = last_hidden_state[:, 0, :]
421
+
422
+ if not return_dict:
423
+ return (last_hidden_state, pooled_output) + encoder_outputs[1:]
424
+
425
+ return BaseModelOutputWithPooling(
426
+ last_hidden_state=last_hidden_state,
427
+ pooler_output=pooled_output,
428
+ hidden_states=encoder_outputs.hidden_states,
429
+ attentions=encoder_outputs.attentions,
430
+ )
modeling_internvl_chat.py ADDED
@@ -0,0 +1,348 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # --------------------------------------------------------
2
+ # InternVL
3
+ # Copyright (c) 2024 OpenGVLab
4
+ # Licensed under The MIT License [see LICENSE for details]
5
+ # --------------------------------------------------------
6
+
7
+ import warnings
8
+ from typing import List, Optional, Tuple, Union
9
+
10
+ import torch.utils.checkpoint
11
+ import transformers
12
+ from torch import nn
13
+ from torch.nn import CrossEntropyLoss
14
+ from transformers import (AutoModel, GenerationConfig, LlamaForCausalLM,
15
+ Qwen2ForCausalLM)
16
+ from transformers.modeling_outputs import CausalLMOutputWithPast
17
+ from transformers.modeling_utils import PreTrainedModel
18
+ from transformers.utils import ModelOutput, logging
19
+
20
+ from .configuration_internvl_chat import InternVLChatConfig
21
+ from .conversation import get_conv_template
22
+ from .modeling_intern_vit import InternVisionModel, has_flash_attn
23
+
24
+ logger = logging.get_logger(__name__)
25
+
26
+
27
+ def version_cmp(v1, v2, op='eq'):
28
+ import operator
29
+
30
+ from packaging import version
31
+ op_func = getattr(operator, op)
32
+ return op_func(version.parse(v1), version.parse(v2))
33
+
34
+
35
+ class InternVLChatModel(PreTrainedModel):
36
+ config_class = InternVLChatConfig
37
+ main_input_name = 'pixel_values'
38
+ base_model_prefix = 'language_model'
39
+ _supports_flash_attn_2 = True
40
+ _no_split_modules = ['InternVisionModel', 'LlamaDecoderLayer', 'Qwen2DecoderLayer']
41
+
42
+ def __init__(self, config: InternVLChatConfig, vision_model=None, language_model=None, use_flash_attn=True):
43
+ super().__init__(config)
44
+
45
+ assert version_cmp(transformers.__version__, '4.37.0', 'ge')
46
+ image_size = config.force_image_size or config.vision_config.image_size
47
+ patch_size = config.vision_config.patch_size
48
+ self.patch_size = patch_size
49
+ self.select_layer = config.select_layer
50
+ self.template = config.template
51
+ self.num_image_token = int((image_size // patch_size) ** 2 * (config.downsample_ratio ** 2))
52
+ self.downsample_ratio = config.downsample_ratio
53
+ self.ps_version = config.ps_version
54
+ use_flash_attn = use_flash_attn if has_flash_attn else False
55
+ config.vision_config.use_flash_attn = True if use_flash_attn else False
56
+ config.llm_config._attn_implementation = 'flash_attention_2' if use_flash_attn else 'eager'
57
+
58
+ logger.info(f'num_image_token: {self.num_image_token}')
59
+ logger.info(f'ps_version: {self.ps_version}')
60
+ if vision_model is not None:
61
+ self.vision_model = vision_model
62
+ else:
63
+ self.vision_model = InternVisionModel(config.vision_config)
64
+ if language_model is not None:
65
+ self.language_model = language_model
66
+ else:
67
+ if config.llm_config.architectures[0] == 'LlamaForCausalLM':
68
+ self.language_model = LlamaForCausalLM(config.llm_config)
69
+ elif config.llm_config.architectures[0] == 'Qwen2ForCausalLM':
70
+ self.language_model = Qwen2ForCausalLM(config.llm_config)
71
+ else:
72
+ raise NotImplementedError(f'{config.llm_config.architectures[0]} is not implemented.')
73
+
74
+ vit_hidden_size = config.vision_config.hidden_size
75
+ llm_hidden_size = config.llm_config.hidden_size
76
+
77
+ self.mlp1 = nn.Sequential(
78
+ nn.LayerNorm(vit_hidden_size * int(1 / self.downsample_ratio) ** 2),
79
+ nn.Linear(vit_hidden_size * int(1 / self.downsample_ratio) ** 2, llm_hidden_size),
80
+ nn.GELU(),
81
+ nn.Linear(llm_hidden_size, llm_hidden_size)
82
+ )
83
+
84
+ self.img_context_token_id = None
85
+ self.conv_template = get_conv_template(self.template)
86
+ self.system_message = self.conv_template.system_message
87
+
88
+ def forward(
89
+ self,
90
+ pixel_values: torch.FloatTensor,
91
+ input_ids: torch.LongTensor = None,
92
+ attention_mask: Optional[torch.Tensor] = None,
93
+ position_ids: Optional[torch.LongTensor] = None,
94
+ image_flags: Optional[torch.LongTensor] = None,
95
+ past_key_values: Optional[List[torch.FloatTensor]] = None,
96
+ labels: Optional[torch.LongTensor] = None,
97
+ use_cache: Optional[bool] = None,
98
+ output_attentions: Optional[bool] = None,
99
+ output_hidden_states: Optional[bool] = None,
100
+ return_dict: Optional[bool] = None,
101
+ ) -> Union[Tuple, CausalLMOutputWithPast]:
102
+ return_dict = return_dict if return_dict is not None else self.config.use_return_dict
103
+
104
+ image_flags = image_flags.squeeze(-1)
105
+ input_embeds = self.language_model.get_input_embeddings()(input_ids).clone()
106
+
107
+ vit_embeds = self.extract_feature(pixel_values)
108
+ vit_embeds = vit_embeds[image_flags == 1]
109
+ vit_batch_size = pixel_values.shape[0]
110
+
111
+ B, N, C = input_embeds.shape
112
+ input_embeds = input_embeds.reshape(B * N, C)
113
+
114
+ if torch.distributed.is_initialized() and torch.distributed.get_rank() == 0:
115
+ print(f'dynamic ViT batch size: {vit_batch_size}, images per sample: {vit_batch_size / B}, dynamic token length: {N}')
116
+
117
+ input_ids = input_ids.reshape(B * N)
118
+ selected = (input_ids == self.img_context_token_id)
119
+ try:
120
+ input_embeds[selected] = input_embeds[selected] * 0.0 + vit_embeds.reshape(-1, C)
121
+ except Exception as e:
122
+ vit_embeds = vit_embeds.reshape(-1, C)
123
+ print(f'warning: {e}, input_embeds[selected].shape={input_embeds[selected].shape}, '
124
+ f'vit_embeds.shape={vit_embeds.shape}')
125
+ n_token = selected.sum()
126
+ input_embeds[selected] = input_embeds[selected] * 0.0 + vit_embeds[:n_token]
127
+
128
+ input_embeds = input_embeds.reshape(B, N, C)
129
+
130
+ outputs = self.language_model(
131
+ inputs_embeds=input_embeds,
132
+ attention_mask=attention_mask,
133
+ position_ids=position_ids,
134
+ past_key_values=past_key_values,
135
+ use_cache=use_cache,
136
+ output_attentions=output_attentions,
137
+ output_hidden_states=output_hidden_states,
138
+ return_dict=return_dict,
139
+ )
140
+ logits = outputs.logits
141
+
142
+ loss = None
143
+ if labels is not None:
144
+ # Shift so that tokens < n predict n
145
+ shift_logits = logits[..., :-1, :].contiguous()
146
+ shift_labels = labels[..., 1:].contiguous()
147
+ # Flatten the tokens
148
+ loss_fct = CrossEntropyLoss()
149
+ shift_logits = shift_logits.view(-1, self.language_model.config.vocab_size)
150
+ shift_labels = shift_labels.view(-1)
151
+ # Enable model parallelism
152
+ shift_labels = shift_labels.to(shift_logits.device)
153
+ loss = loss_fct(shift_logits, shift_labels)
154
+
155
+ if not return_dict:
156
+ output = (logits,) + outputs[1:]
157
+ return (loss,) + output if loss is not None else output
158
+
159
+ return CausalLMOutputWithPast(
160
+ loss=loss,
161
+ logits=logits,
162
+ past_key_values=outputs.past_key_values,
163
+ hidden_states=outputs.hidden_states,
164
+ attentions=outputs.attentions,
165
+ )
166
+
167
+ def pixel_shuffle(self, x, scale_factor=0.5):
168
+ n, w, h, c = x.size()
169
+ # N, W, H, C --> N, W, H * scale, C // scale
170
+ x = x.view(n, w, int(h * scale_factor), int(c / scale_factor))
171
+ # N, W, H * scale, C // scale --> N, H * scale, W, C // scale
172
+ x = x.permute(0, 2, 1, 3).contiguous()
173
+ # N, H * scale, W, C // scale --> N, H * scale, W * scale, C // (scale ** 2)
174
+ x = x.view(n, int(h * scale_factor), int(w * scale_factor),
175
+ int(c / (scale_factor * scale_factor)))
176
+ if self.ps_version == 'v1':
177
+ warnings.warn("In ps_version 'v1', the height and width have not been swapped back, "
178
+ 'which results in a transposed image.')
179
+ else:
180
+ x = x.permute(0, 2, 1, 3).contiguous()
181
+ return x
182
+
183
+ def extract_feature(self, pixel_values):
184
+ if self.select_layer == -1:
185
+ vit_embeds = self.vision_model(
186
+ pixel_values=pixel_values,
187
+ output_hidden_states=False,
188
+ return_dict=True).last_hidden_state
189
+ else:
190
+ vit_embeds = self.vision_model(
191
+ pixel_values=pixel_values,
192
+ output_hidden_states=True,
193
+ return_dict=True).hidden_states[self.select_layer]
194
+ vit_embeds = vit_embeds[:, 1:, :]
195
+
196
+ h = w = int(vit_embeds.shape[1] ** 0.5)
197
+ vit_embeds = vit_embeds.reshape(vit_embeds.shape[0], h, w, -1)
198
+ vit_embeds = self.pixel_shuffle(vit_embeds, scale_factor=self.downsample_ratio)
199
+ vit_embeds = vit_embeds.reshape(vit_embeds.shape[0], -1, vit_embeds.shape[-1])
200
+ vit_embeds = self.mlp1(vit_embeds)
201
+ return vit_embeds
202
+
203
+ def batch_chat(self, tokenizer, pixel_values, questions, generation_config, num_patches_list=None,
204
+ history=None, return_history=False, IMG_START_TOKEN='<img>', IMG_END_TOKEN='</img>',
205
+ IMG_CONTEXT_TOKEN='<IMG_CONTEXT>', verbose=False, image_counts=None):
206
+ if history is not None or return_history:
207
+ print('Now multi-turn chat is not supported in batch_chat.')
208
+ raise NotImplementedError
209
+
210
+ if image_counts is not None:
211
+ num_patches_list = image_counts
212
+ print('Warning: `image_counts` is deprecated. Please use `num_patches_list` instead.')
213
+
214
+ img_context_token_id = tokenizer.convert_tokens_to_ids(IMG_CONTEXT_TOKEN)
215
+ self.img_context_token_id = img_context_token_id
216
+
217
+ if verbose and pixel_values is not None:
218
+ image_bs = pixel_values.shape[0]
219
+ print(f'dynamic ViT batch size: {image_bs}')
220
+
221
+ queries = []
222
+ for idx, num_patches in enumerate(num_patches_list):
223
+ question = questions[idx]
224
+ if pixel_values is not None and '<image>' not in question:
225
+ question = '<image>\n' + question
226
+ template = get_conv_template(self.template)
227
+ template.system_message = self.system_message
228
+ template.append_message(template.roles[0], question)
229
+ template.append_message(template.roles[1], None)
230
+ query = template.get_prompt()
231
+
232
+ image_tokens = IMG_START_TOKEN + IMG_CONTEXT_TOKEN * self.num_image_token * num_patches + IMG_END_TOKEN
233
+ query = query.replace('<image>', image_tokens, 1)
234
+ queries.append(query)
235
+
236
+ tokenizer.padding_side = 'left'
237
+ model_inputs = tokenizer(queries, return_tensors='pt', padding=True)
238
+ input_ids = model_inputs['input_ids'].to(self.device)
239
+ attention_mask = model_inputs['attention_mask'].to(self.device)
240
+ eos_token_id = tokenizer.convert_tokens_to_ids(template.sep.strip())
241
+ generation_config['eos_token_id'] = eos_token_id
242
+ generation_output = self.generate(
243
+ pixel_values=pixel_values,
244
+ input_ids=input_ids,
245
+ attention_mask=attention_mask,
246
+ **generation_config
247
+ )
248
+ responses = tokenizer.batch_decode(generation_output, skip_special_tokens=True)
249
+ responses = [response.split(template.sep.strip())[0].strip() for response in responses]
250
+ return responses
251
+
252
+ def chat(self, tokenizer, pixel_values, question, generation_config, history=None, return_history=False,
253
+ num_patches_list=None, IMG_START_TOKEN='<img>', IMG_END_TOKEN='</img>', IMG_CONTEXT_TOKEN='<IMG_CONTEXT>',
254
+ verbose=False):
255
+
256
+ if history is None and pixel_values is not None and '<image>' not in question:
257
+ question = '<image>\n' + question
258
+
259
+ if num_patches_list is None:
260
+ num_patches_list = [pixel_values.shape[0]] if pixel_values is not None else []
261
+ assert pixel_values is None or len(pixel_values) == sum(num_patches_list)
262
+
263
+ img_context_token_id = tokenizer.convert_tokens_to_ids(IMG_CONTEXT_TOKEN)
264
+ self.img_context_token_id = img_context_token_id
265
+
266
+ template = get_conv_template(self.template)
267
+ template.system_message = self.system_message
268
+ eos_token_id = tokenizer.convert_tokens_to_ids(template.sep.strip())
269
+
270
+ history = [] if history is None else history
271
+ for (old_question, old_answer) in history:
272
+ template.append_message(template.roles[0], old_question)
273
+ template.append_message(template.roles[1], old_answer)
274
+ template.append_message(template.roles[0], question)
275
+ template.append_message(template.roles[1], None)
276
+ query = template.get_prompt()
277
+
278
+ if verbose and pixel_values is not None:
279
+ image_bs = pixel_values.shape[0]
280
+ print(f'dynamic ViT batch size: {image_bs}')
281
+
282
+ for num_patches in num_patches_list:
283
+ image_tokens = IMG_START_TOKEN + IMG_CONTEXT_TOKEN * self.num_image_token * num_patches + IMG_END_TOKEN
284
+ query = query.replace('<image>', image_tokens, 1)
285
+
286
+ model_inputs = tokenizer(query, return_tensors='pt')
287
+ input_ids = model_inputs['input_ids'].to(self.device)
288
+ attention_mask = model_inputs['attention_mask'].to(self.device)
289
+ generation_config['eos_token_id'] = eos_token_id
290
+ generation_output = self.generate(
291
+ pixel_values=pixel_values,
292
+ input_ids=input_ids,
293
+ attention_mask=attention_mask,
294
+ **generation_config
295
+ )
296
+ response = tokenizer.batch_decode(generation_output, skip_special_tokens=True)[0]
297
+ response = response.split(template.sep.strip())[0].strip()
298
+ history.append((question, response))
299
+ if return_history:
300
+ return response, history
301
+ else:
302
+ query_to_print = query.replace(IMG_CONTEXT_TOKEN, '')
303
+ query_to_print = query_to_print.replace(f'{IMG_START_TOKEN}{IMG_END_TOKEN}', '<image>')
304
+ if verbose:
305
+ print(query_to_print, response)
306
+ return response
307
+
308
+ @torch.no_grad()
309
+ def generate(
310
+ self,
311
+ pixel_values: Optional[torch.FloatTensor] = None,
312
+ input_ids: Optional[torch.FloatTensor] = None,
313
+ attention_mask: Optional[torch.LongTensor] = None,
314
+ visual_features: Optional[torch.FloatTensor] = None,
315
+ generation_config: Optional[GenerationConfig] = None,
316
+ output_hidden_states: Optional[bool] = None,
317
+ **generate_kwargs,
318
+ ) -> torch.LongTensor:
319
+
320
+ assert self.img_context_token_id is not None
321
+ if pixel_values is not None:
322
+ if visual_features is not None:
323
+ vit_embeds = visual_features
324
+ else:
325
+ vit_embeds = self.extract_feature(pixel_values)
326
+ input_embeds = self.language_model.get_input_embeddings()(input_ids)
327
+ B, N, C = input_embeds.shape
328
+ input_embeds = input_embeds.reshape(B * N, C)
329
+
330
+ input_ids = input_ids.reshape(B * N)
331
+ selected = (input_ids == self.img_context_token_id)
332
+ assert selected.sum() != 0
333
+ input_embeds[selected] = vit_embeds.reshape(-1, C).to(input_embeds.device)
334
+
335
+ input_embeds = input_embeds.reshape(B, N, C)
336
+ else:
337
+ input_embeds = self.language_model.get_input_embeddings()(input_ids)
338
+
339
+ outputs = self.language_model.generate(
340
+ inputs_embeds=input_embeds,
341
+ attention_mask=attention_mask,
342
+ generation_config=generation_config,
343
+ output_hidden_states=output_hidden_states,
344
+ use_cache=True,
345
+ **generate_kwargs,
346
+ )
347
+
348
+ return outputs
preprocessor_config.json ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "crop_size": 448,
3
+ "do_center_crop": true,
4
+ "do_normalize": true,
5
+ "do_resize": true,
6
+ "feature_extractor_type": "CLIPFeatureExtractor",
7
+ "image_mean": [
8
+ 0.485,
9
+ 0.456,
10
+ 0.406
11
+ ],
12
+ "image_std": [
13
+ 0.229,
14
+ 0.224,
15
+ 0.225
16
+ ],
17
+ "resample": 3,
18
+ "size": 448
19
+ }
runs/Dec20_06-31-42_HOST-10-140-60-160/events.out.tfevents.1734647947.HOST-10-140-60-160.117020.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dde20b186f89d8928910b499d0232dd5c1489cd7db199f9417aa620389d63dff
3
+ size 1393451
special_tokens_map.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>",
16
+ "<img>",
17
+ "</img>",
18
+ "<IMG_CONTEXT>",
19
+ "<quad>",
20
+ "</quad>",
21
+ "<ref>",
22
+ "</ref>",
23
+ "<box>",
24
+ "</box>"
25
+ ],
26
+ "eos_token": {
27
+ "content": "<|im_end|>",
28
+ "lstrip": false,
29
+ "normalized": false,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ },
33
+ "pad_token": {
34
+ "content": "<|endoftext|>",
35
+ "lstrip": false,
36
+ "normalized": false,
37
+ "rstrip": false,
38
+ "single_word": false
39
+ }
40
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,289 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": false,
5
+ "added_tokens_decoder": {
6
+ "151643": {
7
+ "content": "<|endoftext|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "151644": {
15
+ "content": "<|im_start|>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "151645": {
23
+ "content": "<|im_end|>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "151646": {
31
+ "content": "<|object_ref_start|>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "151647": {
39
+ "content": "<|object_ref_end|>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ },
46
+ "151648": {
47
+ "content": "<|box_start|>",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false,
52
+ "special": true
53
+ },
54
+ "151649": {
55
+ "content": "<|box_end|>",
56
+ "lstrip": false,
57
+ "normalized": false,
58
+ "rstrip": false,
59
+ "single_word": false,
60
+ "special": true
61
+ },
62
+ "151650": {
63
+ "content": "<|quad_start|>",
64
+ "lstrip": false,
65
+ "normalized": false,
66
+ "rstrip": false,
67
+ "single_word": false,
68
+ "special": true
69
+ },
70
+ "151651": {
71
+ "content": "<|quad_end|>",
72
+ "lstrip": false,
73
+ "normalized": false,
74
+ "rstrip": false,
75
+ "single_word": false,
76
+ "special": true
77
+ },
78
+ "151652": {
79
+ "content": "<|vision_start|>",
80
+ "lstrip": false,
81
+ "normalized": false,
82
+ "rstrip": false,
83
+ "single_word": false,
84
+ "special": true
85
+ },
86
+ "151653": {
87
+ "content": "<|vision_end|>",
88
+ "lstrip": false,
89
+ "normalized": false,
90
+ "rstrip": false,
91
+ "single_word": false,
92
+ "special": true
93
+ },
94
+ "151654": {
95
+ "content": "<|vision_pad|>",
96
+ "lstrip": false,
97
+ "normalized": false,
98
+ "rstrip": false,
99
+ "single_word": false,
100
+ "special": true
101
+ },
102
+ "151655": {
103
+ "content": "<|image_pad|>",
104
+ "lstrip": false,
105
+ "normalized": false,
106
+ "rstrip": false,
107
+ "single_word": false,
108
+ "special": true
109
+ },
110
+ "151656": {
111
+ "content": "<|video_pad|>",
112
+ "lstrip": false,
113
+ "normalized": false,
114
+ "rstrip": false,
115
+ "single_word": false,
116
+ "special": true
117
+ },
118
+ "151657": {
119
+ "content": "<tool_call>",
120
+ "lstrip": false,
121
+ "normalized": false,
122
+ "rstrip": false,
123
+ "single_word": false,
124
+ "special": false
125
+ },
126
+ "151658": {
127
+ "content": "</tool_call>",
128
+ "lstrip": false,
129
+ "normalized": false,
130
+ "rstrip": false,
131
+ "single_word": false,
132
+ "special": false
133
+ },
134
+ "151659": {
135
+ "content": "<|fim_prefix|>",
136
+ "lstrip": false,
137
+ "normalized": false,
138
+ "rstrip": false,
139
+ "single_word": false,
140
+ "special": false
141
+ },
142
+ "151660": {
143
+ "content": "<|fim_middle|>",
144
+ "lstrip": false,
145
+ "normalized": false,
146
+ "rstrip": false,
147
+ "single_word": false,
148
+ "special": false
149
+ },
150
+ "151661": {
151
+ "content": "<|fim_suffix|>",
152
+ "lstrip": false,
153
+ "normalized": false,
154
+ "rstrip": false,
155
+ "single_word": false,
156
+ "special": false
157
+ },
158
+ "151662": {
159
+ "content": "<|fim_pad|>",
160
+ "lstrip": false,
161
+ "normalized": false,
162
+ "rstrip": false,
163
+ "single_word": false,
164
+ "special": false
165
+ },
166
+ "151663": {
167
+ "content": "<|repo_name|>",
168
+ "lstrip": false,
169
+ "normalized": false,
170
+ "rstrip": false,
171
+ "single_word": false,
172
+ "special": false
173
+ },
174
+ "151664": {
175
+ "content": "<|file_sep|>",
176
+ "lstrip": false,
177
+ "normalized": false,
178
+ "rstrip": false,
179
+ "single_word": false,
180
+ "special": false
181
+ },
182
+ "151665": {
183
+ "content": "<img>",
184
+ "lstrip": false,
185
+ "normalized": false,
186
+ "rstrip": false,
187
+ "single_word": false,
188
+ "special": true
189
+ },
190
+ "151666": {
191
+ "content": "</img>",
192
+ "lstrip": false,
193
+ "normalized": false,
194
+ "rstrip": false,
195
+ "single_word": false,
196
+ "special": true
197
+ },
198
+ "151667": {
199
+ "content": "<IMG_CONTEXT>",
200
+ "lstrip": false,
201
+ "normalized": false,
202
+ "rstrip": false,
203
+ "single_word": false,
204
+ "special": true
205
+ },
206
+ "151668": {
207
+ "content": "<quad>",
208
+ "lstrip": false,
209
+ "normalized": false,
210
+ "rstrip": false,
211
+ "single_word": false,
212
+ "special": true
213
+ },
214
+ "151669": {
215
+ "content": "</quad>",
216
+ "lstrip": false,
217
+ "normalized": false,
218
+ "rstrip": false,
219
+ "single_word": false,
220
+ "special": true
221
+ },
222
+ "151670": {
223
+ "content": "<ref>",
224
+ "lstrip": false,
225
+ "normalized": false,
226
+ "rstrip": false,
227
+ "single_word": false,
228
+ "special": true
229
+ },
230
+ "151671": {
231
+ "content": "</ref>",
232
+ "lstrip": false,
233
+ "normalized": false,
234
+ "rstrip": false,
235
+ "single_word": false,
236
+ "special": true
237
+ },
238
+ "151672": {
239
+ "content": "<box>",
240
+ "lstrip": false,
241
+ "normalized": false,
242
+ "rstrip": false,
243
+ "single_word": false,
244
+ "special": true
245
+ },
246
+ "151673": {
247
+ "content": "</box>",
248
+ "lstrip": false,
249
+ "normalized": false,
250
+ "rstrip": false,
251
+ "single_word": false,
252
+ "special": true
253
+ }
254
+ },
255
+ "additional_special_tokens": [
256
+ "<|im_start|>",
257
+ "<|im_end|>",
258
+ "<|object_ref_start|>",
259
+ "<|object_ref_end|>",
260
+ "<|box_start|>",
261
+ "<|box_end|>",
262
+ "<|quad_start|>",
263
+ "<|quad_end|>",
264
+ "<|vision_start|>",
265
+ "<|vision_end|>",
266
+ "<|vision_pad|>",
267
+ "<|image_pad|>",
268
+ "<|video_pad|>",
269
+ "<img>",
270
+ "</img>",
271
+ "<IMG_CONTEXT>",
272
+ "<quad>",
273
+ "</quad>",
274
+ "<ref>",
275
+ "</ref>",
276
+ "<box>",
277
+ "</box>"
278
+ ],
279
+ "bos_token": null,
280
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
281
+ "clean_up_tokenization_spaces": false,
282
+ "eos_token": "<|im_end|>",
283
+ "errors": "replace",
284
+ "model_max_length": 8192,
285
+ "pad_token": "<|endoftext|>",
286
+ "split_special_tokens": false,
287
+ "tokenizer_class": "Qwen2Tokenizer",
288
+ "unk_token": null
289
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff