fbaldassarri commited on
Commit
c2843e7
·
verified ·
1 Parent(s): 925ac1b

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -56,38 +56,38 @@ Finally, an INT8 quantized version of the Imange Enconder only can be find as op
56
 
57
  I suggest to install requirements into a dedicated python-virtualenv or a conda enviroment.
58
 
59
- '''
60
  pip install -q "torch>=2.1" "torchvision" "Pillow" "tqdm" "datasets>=2.14.6" "gradio>=4.36" "nncf>=2.13.0" --extra-index-url https://download.pytorch.org/whl/cpu
61
 
62
  pip install -q "transformers>=4.45" --extra-index-url https://download.pytorch.org/whl/cpu
63
 
64
  pip install -Uq --pre "openvino>2024.4.0" --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
65
- '''
66
 
67
  ### Step 2 Convert the model in OpenVINO Intermediate Representation (IR)
68
 
69
- '''
70
  from pathlib import Path
71
  from ov_mllama_helper import convert_mllama
72
  model_id = "meta-llama/Llama-3.2-11B-Vision-Instruct"
73
  model_dir = Path(model_id.split("/")[-1]) / "OpenVino"
74
  convert_mllama(model_id, model_dir)
75
- '''
76
 
77
  ### Step 3 INT4 Compression
78
 
79
- '''
80
  from ov_mllama_compression import compress
81
  from ov_mllama_compression import compression_widgets_helper
82
  compression_scenario, compress_args = compression_widgets_helper()
83
  compression_scenario
84
  compression_kwargs = {key: value.value for key, value in compress_args.items()}
85
  language_model_path = compress(model_dir, **compression_kwargs)
86
- '''
87
 
88
  ### Step 4 INT8 Image Enconder Optimization
89
 
90
- '''
91
  from ov_mllama_compression import vision_encoder_selection_widget
92
  vision_encoder_options = vision_encoder_selection_widget(device.value)
93
  vision_encoder_options
@@ -117,7 +117,7 @@ del calibration_dataset
117
  del calibration_data
118
  gc.collect()
119
  vision_encoder_path = int8_vision_encoder_path
120
- '''
121
 
122
  ## License
123
 
 
56
 
57
  I suggest to install requirements into a dedicated python-virtualenv or a conda enviroment.
58
 
59
+ ```
60
  pip install -q "torch>=2.1" "torchvision" "Pillow" "tqdm" "datasets>=2.14.6" "gradio>=4.36" "nncf>=2.13.0" --extra-index-url https://download.pytorch.org/whl/cpu
61
 
62
  pip install -q "transformers>=4.45" --extra-index-url https://download.pytorch.org/whl/cpu
63
 
64
  pip install -Uq --pre "openvino>2024.4.0" --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
65
+ ```
66
 
67
  ### Step 2 Convert the model in OpenVINO Intermediate Representation (IR)
68
 
69
+ ```
70
  from pathlib import Path
71
  from ov_mllama_helper import convert_mllama
72
  model_id = "meta-llama/Llama-3.2-11B-Vision-Instruct"
73
  model_dir = Path(model_id.split("/")[-1]) / "OpenVino"
74
  convert_mllama(model_id, model_dir)
75
+ ```
76
 
77
  ### Step 3 INT4 Compression
78
 
79
+ ```
80
  from ov_mllama_compression import compress
81
  from ov_mllama_compression import compression_widgets_helper
82
  compression_scenario, compress_args = compression_widgets_helper()
83
  compression_scenario
84
  compression_kwargs = {key: value.value for key, value in compress_args.items()}
85
  language_model_path = compress(model_dir, **compression_kwargs)
86
+ ```
87
 
88
  ### Step 4 INT8 Image Enconder Optimization
89
 
90
+ ```
91
  from ov_mllama_compression import vision_encoder_selection_widget
92
  vision_encoder_options = vision_encoder_selection_widget(device.value)
93
  vision_encoder_options
 
117
  del calibration_data
118
  gc.collect()
119
  vision_encoder_path = int8_vision_encoder_path
120
+ ```
121
 
122
  ## License
123