DaKitch commited on
Commit
bc3ac92
·
verified ·
1 Parent(s): 709ed31

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +166 -0
README.md ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: cognitivecomputations/Dolphin3.0-Llama3.1-8B
3
+ datasets:
4
+ - OpenCoder-LLM/opc-sft-stage1
5
+ - OpenCoder-LLM/opc-sft-stage2
6
+ - microsoft/orca-agentinstruct-1M-v1
7
+ - microsoft/orca-math-word-problems-200k
8
+ - NousResearch/hermes-function-calling-v1
9
+ - AI-MO/NuminaMath-CoT
10
+ - AI-MO/NuminaMath-TIR
11
+ - allenai/tulu-3-sft-mixture
12
+ - cognitivecomputations/dolphin-coder
13
+ - HuggingFaceTB/smoltalk
14
+ - cognitivecomputations/samantha-data
15
+ - m-a-p/CodeFeedback-Filtered-Instruction
16
+ - m-a-p/Code-Feedback
17
+ language:
18
+ - en
19
+ license: llama3.1
20
+ tags:
21
+ - llama-cpp
22
+ - gguf-my-repo
23
+ model-index:
24
+ - name: Dolphin3.0-Llama3.1-8B
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ name: Text Generation
29
+ dataset:
30
+ name: IFEval (0-Shot)
31
+ type: wis-k/instruction-following-eval
32
+ split: train
33
+ args:
34
+ num_few_shot: 0
35
+ metrics:
36
+ - type: inst_level_strict_acc and prompt_level_strict_acc
37
+ value: 76.21
38
+ name: averaged accuracy
39
+ source:
40
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=cognitivecomputations%2FDolphin3.0-Llama3.1-8B
41
+ name: Open LLM Leaderboard
42
+ - task:
43
+ type: text-generation
44
+ name: Text Generation
45
+ dataset:
46
+ name: BBH (3-Shot)
47
+ type: SaylorTwift/bbh
48
+ split: test
49
+ args:
50
+ num_few_shot: 3
51
+ metrics:
52
+ - type: acc_norm
53
+ value: 27.63
54
+ name: normalized accuracy
55
+ source:
56
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=cognitivecomputations%2FDolphin3.0-Llama3.1-8B
57
+ name: Open LLM Leaderboard
58
+ - task:
59
+ type: text-generation
60
+ name: Text Generation
61
+ dataset:
62
+ name: MATH Lvl 5 (4-Shot)
63
+ type: lighteval/MATH-Hard
64
+ split: test
65
+ args:
66
+ num_few_shot: 4
67
+ metrics:
68
+ - type: exact_match
69
+ value: 10.5
70
+ name: exact match
71
+ source:
72
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=cognitivecomputations%2FDolphin3.0-Llama3.1-8B
73
+ name: Open LLM Leaderboard
74
+ - task:
75
+ type: text-generation
76
+ name: Text Generation
77
+ dataset:
78
+ name: GPQA (0-shot)
79
+ type: Idavidrein/gpqa
80
+ split: train
81
+ args:
82
+ num_few_shot: 0
83
+ metrics:
84
+ - type: acc_norm
85
+ value: 4.36
86
+ name: acc_norm
87
+ source:
88
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=cognitivecomputations%2FDolphin3.0-Llama3.1-8B
89
+ name: Open LLM Leaderboard
90
+ - task:
91
+ type: text-generation
92
+ name: Text Generation
93
+ dataset:
94
+ name: MuSR (0-shot)
95
+ type: TAUR-Lab/MuSR
96
+ args:
97
+ num_few_shot: 0
98
+ metrics:
99
+ - type: acc_norm
100
+ value: 8.97
101
+ name: acc_norm
102
+ source:
103
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=cognitivecomputations%2FDolphin3.0-Llama3.1-8B
104
+ name: Open LLM Leaderboard
105
+ - task:
106
+ type: text-generation
107
+ name: Text Generation
108
+ dataset:
109
+ name: MMLU-PRO (5-shot)
110
+ type: TIGER-Lab/MMLU-Pro
111
+ config: main
112
+ split: test
113
+ args:
114
+ num_few_shot: 5
115
+ metrics:
116
+ - type: acc
117
+ value: 22.13
118
+ name: accuracy
119
+ source:
120
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard#/?search=cognitivecomputations%2FDolphin3.0-Llama3.1-8B
121
+ name: Open LLM Leaderboard
122
+ ---
123
+
124
+ # DaKitch/Dolphin3.0-Llama3.1-8B-Q4_K_M-GGUF
125
+ This model was converted to GGUF format from [`cognitivecomputations/Dolphin3.0-Llama3.1-8B`](https://huggingface.co/cognitivecomputations/Dolphin3.0-Llama3.1-8B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
126
+ Refer to the [original model card](https://huggingface.co/cognitivecomputations/Dolphin3.0-Llama3.1-8B) for more details on the model.
127
+
128
+ ## Use with llama.cpp
129
+ Install llama.cpp through brew (works on Mac and Linux)
130
+
131
+ ```bash
132
+ brew install llama.cpp
133
+
134
+ ```
135
+ Invoke the llama.cpp server or the CLI.
136
+
137
+ ### CLI:
138
+ ```bash
139
+ llama-cli --hf-repo DaKitch/Dolphin3.0-Llama3.1-8B-Q4_K_M-GGUF --hf-file dolphin3.0-llama3.1-8b-q4_k_m.gguf -p "The meaning to life and the universe is"
140
+ ```
141
+
142
+ ### Server:
143
+ ```bash
144
+ llama-server --hf-repo DaKitch/Dolphin3.0-Llama3.1-8B-Q4_K_M-GGUF --hf-file dolphin3.0-llama3.1-8b-q4_k_m.gguf -c 2048
145
+ ```
146
+
147
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
148
+
149
+ Step 1: Clone llama.cpp from GitHub.
150
+ ```
151
+ git clone https://github.com/ggerganov/llama.cpp
152
+ ```
153
+
154
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
155
+ ```
156
+ cd llama.cpp && LLAMA_CURL=1 make
157
+ ```
158
+
159
+ Step 3: Run inference through the main binary.
160
+ ```
161
+ ./llama-cli --hf-repo DaKitch/Dolphin3.0-Llama3.1-8B-Q4_K_M-GGUF --hf-file dolphin3.0-llama3.1-8b-q4_k_m.gguf -p "The meaning to life and the universe is"
162
+ ```
163
+ or
164
+ ```
165
+ ./llama-server --hf-repo DaKitch/Dolphin3.0-Llama3.1-8B-Q4_K_M-GGUF --hf-file dolphin3.0-llama3.1-8b-q4_k_m.gguf -c 2048
166
+ ```