ysn-rfd commited on
Commit
69332ff
·
verified ·
1 Parent(s): 85ee300

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +20 -31
README.md CHANGED
@@ -1,67 +1,56 @@
1
  ---
2
- base_model: universitytehran/PersianMind-v1.0
3
  language:
4
  - multilingual
5
  - fa
6
  - en
7
- library_name: transformers
8
  license: cc-by-nc-sa-4.0
 
 
 
 
 
9
  metrics:
10
  - bleu
11
  - comet
12
  - accuracy
13
  - perplexity
14
  - spearmanr
15
- pipeline_tag: text-generation
16
- tags:
17
- - text-generation-inference
18
- - llama-cpp
19
- - gguf-my-repo
20
  inference: false
 
21
  co2_eq_emissions:
22
  emissions: 232380
 
 
 
 
23
  ---
24
 
25
  # ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF
26
  This model was converted to GGUF format from [`universitytehran/PersianMind-v1.0`](https://huggingface.co/universitytehran/PersianMind-v1.0) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
27
  Refer to the [original model card](https://huggingface.co/universitytehran/PersianMind-v1.0) for more details on the model.
28
-
29
  ## Use with llama.cpp
30
- Install llama.cpp through brew (works on Mac and Linux)
31
 
32
- ```bash
33
- brew install llama.cpp
34
 
 
 
35
  ```
36
  Invoke the llama.cpp server or the CLI.
37
 
38
- ### CLI:
39
- ```bash
40
- llama-cli --hf-repo ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF --hf-file persianmind-v1.0-q4_k_m.gguf -p "The meaning to life and the universe is"
41
- ```
42
 
43
- ### Server:
44
  ```bash
45
- llama-server --hf-repo ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF --hf-file persianmind-v1.0-q4_k_m.gguf -c 2048
46
  ```
47
 
48
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
49
 
50
- Step 1: Clone llama.cpp from GitHub.
51
- ```
52
- git clone https://github.com/ggerganov/llama.cpp
53
  ```
54
 
55
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
56
- ```
57
- cd llama.cpp && LLAMA_CURL=1 make
58
- ```
59
 
60
- Step 3: Run inference through the main binary.
61
- ```
62
- ./llama-cli --hf-repo ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF --hf-file persianmind-v1.0-q4_k_m.gguf -p "The meaning to life and the universe is"
63
- ```
64
- or
65
  ```
66
- ./llama-server --hf-repo ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF --hf-file persianmind-v1.0-q4_k_m.gguf -c 2048
67
  ```
 
1
  ---
 
2
  language:
3
  - multilingual
4
  - fa
5
  - en
 
6
  license: cc-by-nc-sa-4.0
7
+ library_name: transformers
8
+ tags:
9
+ - text-generation-inference
10
+ - llama-cpp
11
+ - gguf-my-repo
12
  metrics:
13
  - bleu
14
  - comet
15
  - accuracy
16
  - perplexity
17
  - spearmanr
 
 
 
 
 
18
  inference: false
19
+ pipeline_tag: text-generation
20
  co2_eq_emissions:
21
  emissions: 232380
22
+ source: 'PersianMind: A Cross-Lingual Persian-English Large Language Model. https://arxiv.org/abs/2401.06466'
23
+ training_type: fine-tuning
24
+ hardware_used: 4 RTX3090 24GB GPUs
25
+ geographical_location: Tehran, Iran
26
  ---
27
 
28
  # ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF
29
  This model was converted to GGUF format from [`universitytehran/PersianMind-v1.0`](https://huggingface.co/universitytehran/PersianMind-v1.0) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
30
  Refer to the [original model card](https://huggingface.co/universitytehran/PersianMind-v1.0) for more details on the model.
 
31
  ## Use with llama.cpp
 
32
 
33
+ Install llama.cpp through brew.
 
34
 
35
+ ```bash
36
+ brew install ggerganov/ggerganov/llama.cpp
37
  ```
38
  Invoke the llama.cpp server or the CLI.
39
 
40
+ CLI:
 
 
 
41
 
 
42
  ```bash
43
+ llama-cli --hf-repo ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF --model persianmind-v1.0.Q4_K_M.gguf -p "The meaning to life and the universe is"
44
  ```
45
 
46
+ Server:
47
 
48
+ ```bash
49
+ llama-server --hf-repo ysn-rfd/PersianMind-v1.0-Q4_K_M-GGUF --model persianmind-v1.0.Q4_K_M.gguf -c 2048
 
50
  ```
51
 
52
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
 
 
 
53
 
 
 
 
 
 
54
  ```
55
+ git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m persianmind-v1.0.Q4_K_M.gguf -n 128
56
  ```