Triangle104 commited on
Commit
86db918
·
verified ·
1 Parent(s): e59a798

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +162 -0
README.md ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - long-cot-reasoning
4
+ - transformers
5
+ - mamba2
6
+ - llms
7
+ - chain-of-thought
8
+ - llama-cpp
9
+ - gguf-my-repo
10
+ license: apache-2.0
11
+ language:
12
+ - en
13
+ datasets:
14
+ - Daemontatox/LongCOT-Reason
15
+ - Daemontatox/alpaca_reasoning_COT
16
+ base_model: Daemontatox/Sphinx2.0
17
+ pipeline_tag: text-generation
18
+ library_name: transformers
19
+ model-index:
20
+ - name: Sphinx2.0
21
+ results:
22
+ - task:
23
+ type: text-generation
24
+ name: Text Generation
25
+ dataset:
26
+ name: IFEval (0-Shot)
27
+ type: wis-k/instruction-following-eval
28
+ split: train
29
+ args:
30
+ num_few_shot: 0
31
+ metrics:
32
+ - type: inst_level_strict_acc and prompt_level_strict_acc
33
+ value: 71.23
34
+ name: averaged accuracy
35
+ source:
36
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Daemontatox%2FSphinx2.0
37
+ name: Open LLM Leaderboard
38
+ - task:
39
+ type: text-generation
40
+ name: Text Generation
41
+ dataset:
42
+ name: BBH (3-Shot)
43
+ type: SaylorTwift/bbh
44
+ split: test
45
+ args:
46
+ num_few_shot: 3
47
+ metrics:
48
+ - type: acc_norm
49
+ value: 49.4
50
+ name: normalized accuracy
51
+ source:
52
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Daemontatox%2FSphinx2.0
53
+ name: Open LLM Leaderboard
54
+ - task:
55
+ type: text-generation
56
+ name: Text Generation
57
+ dataset:
58
+ name: MATH Lvl 5 (4-Shot)
59
+ type: lighteval/MATH-Hard
60
+ split: test
61
+ args:
62
+ num_few_shot: 4
63
+ metrics:
64
+ - type: exact_match
65
+ value: 2.72
66
+ name: exact match
67
+ source:
68
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Daemontatox%2FSphinx2.0
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: GPQA (0-shot)
75
+ type: Idavidrein/gpqa
76
+ split: train
77
+ args:
78
+ num_few_shot: 0
79
+ metrics:
80
+ - type: acc_norm
81
+ value: 5.82
82
+ name: acc_norm
83
+ source:
84
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Daemontatox%2FSphinx2.0
85
+ name: Open LLM Leaderboard
86
+ - task:
87
+ type: text-generation
88
+ name: Text Generation
89
+ dataset:
90
+ name: MuSR (0-shot)
91
+ type: TAUR-Lab/MuSR
92
+ args:
93
+ num_few_shot: 0
94
+ metrics:
95
+ - type: acc_norm
96
+ value: 13.05
97
+ name: acc_norm
98
+ source:
99
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Daemontatox%2FSphinx2.0
100
+ name: Open LLM Leaderboard
101
+ - task:
102
+ type: text-generation
103
+ name: Text Generation
104
+ dataset:
105
+ name: MMLU-PRO (5-shot)
106
+ type: TIGER-Lab/MMLU-Pro
107
+ config: main
108
+ split: test
109
+ args:
110
+ num_few_shot: 5
111
+ metrics:
112
+ - type: acc
113
+ value: 46.49
114
+ name: accuracy
115
+ source:
116
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=Daemontatox%2FSphinx2.0
117
+ name: Open LLM Leaderboard
118
+ ---
119
+
120
+ # Triangle104/Sphinx2.0-Q5_K_S-GGUF
121
+ This model was converted to GGUF format from [`Daemontatox/Sphinx2.0`](https://huggingface.co/Daemontatox/Sphinx2.0) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
122
+ Refer to the [original model card](https://huggingface.co/Daemontatox/Sphinx2.0) for more details on the model.
123
+
124
+ ## Use with llama.cpp
125
+ Install llama.cpp through brew (works on Mac and Linux)
126
+
127
+ ```bash
128
+ brew install llama.cpp
129
+
130
+ ```
131
+ Invoke the llama.cpp server or the CLI.
132
+
133
+ ### CLI:
134
+ ```bash
135
+ llama-cli --hf-repo Triangle104/Sphinx2.0-Q5_K_S-GGUF --hf-file sphinx2.0-q5_k_s.gguf -p "The meaning to life and the universe is"
136
+ ```
137
+
138
+ ### Server:
139
+ ```bash
140
+ llama-server --hf-repo Triangle104/Sphinx2.0-Q5_K_S-GGUF --hf-file sphinx2.0-q5_k_s.gguf -c 2048
141
+ ```
142
+
143
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
144
+
145
+ Step 1: Clone llama.cpp from GitHub.
146
+ ```
147
+ git clone https://github.com/ggerganov/llama.cpp
148
+ ```
149
+
150
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
151
+ ```
152
+ cd llama.cpp && LLAMA_CURL=1 make
153
+ ```
154
+
155
+ Step 3: Run inference through the main binary.
156
+ ```
157
+ ./llama-cli --hf-repo Triangle104/Sphinx2.0-Q5_K_S-GGUF --hf-file sphinx2.0-q5_k_s.gguf -p "The meaning to life and the universe is"
158
+ ```
159
+ or
160
+ ```
161
+ ./llama-server --hf-repo Triangle104/Sphinx2.0-Q5_K_S-GGUF --hf-file sphinx2.0-q5_k_s.gguf -c 2048
162
+ ```