prince-canuma commited on
Commit
0f9964a
·
verified ·
1 Parent(s): 9c215d8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -4
README.md CHANGED
@@ -7,18 +7,22 @@ tags:
7
  - multimodal
8
  - qwen
9
  - mlx
 
10
  ---
11
 
12
  # mlx-community/nanoLLaVA
13
  This model was converted to MLX format from [`qnguyen3/nanoLLaVA`]() using mlx-vllm version **0.0.0**.
 
 
 
14
  Refer to the [original model card](https://huggingface.co/qnguyen3/nanoLLaVA) for more details on the model.
15
  ## Use with mlx
16
 
17
  ```bash
18
- git clone https://github.com/ml-explore/mlx-examples.git
19
  ```
20
 
21
  ```bash
22
- cd mlx-lm/vllms
23
- python -m generate --model mlx_model --max-tokens 10 --temp 0.0
24
- ```
 
7
  - multimodal
8
  - qwen
9
  - mlx
10
+ pipeline_tag: image-text-to-text
11
  ---
12
 
13
  # mlx-community/nanoLLaVA
14
  This model was converted to MLX format from [`qnguyen3/nanoLLaVA`]() using mlx-vllm version **0.0.0**.
15
+
16
+ Model added by [Prince Canuma](https://twitter.com/Prince_Canuma).
17
+
18
  Refer to the [original model card](https://huggingface.co/qnguyen3/nanoLLaVA) for more details on the model.
19
  ## Use with mlx
20
 
21
  ```bash
22
+ pip install mlx-vlm
23
  ```
24
 
25
  ```bash
26
+
27
+ python -m mlx-vlm.generate --model mlx_model --max-tokens 10 --temp 0.0
28
+ ```