brittlewis12 commited on
Commit
44a5d41
·
verified ·
1 Parent(s): 2cb21b2

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +121 -0
README.md ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/gemma-3-1b-it
3
+ pipeline_tag: text-generation
4
+ inference: true
5
+ language:
6
+ - en
7
+ license: gemma
8
+ model_creator: google
9
+ model_name: google/gemma-3-1b-it
10
+ model_type: gemma3_text
11
+ quantized_by: brittlewis12
12
+ tags:
13
+ - reasoning
14
+ - gemma
15
+ ---
16
+
17
+ # Gemma 3 1B IT GGUF
18
+
19
+ **Original model**: [Gemma 3 1B IT](https://huggingface.co/google/gemma-3-1b-it)
20
+
21
+ **Model creator**: [Google DeepMind](https://huggingface.co/google)
22
+
23
+ > Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
24
+
25
+ > Gemma 3 has a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions. Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone.
26
+
27
+ This repo contains GGUF format model files for Google DeepMind’s Gemma 3 1B IT (instruction-tuned).
28
+
29
+ ### What is GGUF?
30
+
31
+ GGUF is a file format for representing AI models. It is the third version of the format,
32
+ introduced by the llama.cpp team on August 21st 2023.
33
+
34
+ Converted with llama.cpp build b4875 (revision [7841fc7](https://github.com/ggml-org/llama.cpp/commits/7841fc723e059d1fd9640e5c0ef19050fcc7c698)),
35
+ using [autogguf-rs](https://github.com/brittlewis12/autogguf-rs).
36
+
37
+ ### Prompt template: [Gemma Instruct]()
38
+
39
+ ```
40
+ {{system_prompt}}
41
+ <start_of_turn>user
42
+ {{prompt}}<end_of_turn>
43
+ <start_of_turn>model
44
+
45
+
46
+ ```
47
+
48
+ ---
49
+
50
+ ## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac!
51
+
52
+ ![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg)
53
+
54
+ [cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
55
+ - create & save **Characters** with custom system prompts & temperature settings
56
+ - download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
57
+ * or, use an API key with the chat completions-compatible model provider of your choice -- ChatGPT, Claude, Gemini, DeepSeek, & more!
58
+ - make it your own with custom **Theme colors**
59
+ - powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggml-org/llama.cpp), with **haptics** during response streaming!
60
+ - **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
61
+ * if you **already have the app**, download Gemma 3 1B IT now!
62
+ * <cnvrsai:///models/search/hf?id=brittlewis12/gemma-3-1b-it-GGUF>
63
+ - follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date
64
+
65
+ ### Gemma 3 1B IT in cnvrs on macOS
66
+
67
+ ![gemma-3 in cnvrs](https://cdn-uploads.huggingface.co/production/uploads/63b64d7a889aa6707f155cdb/dt0eLAU9pKyKnuu8oCw0D.png)
68
+
69
+ ---
70
+
71
+ ## Original Model Evaluation
72
+
73
+ > These models were evaluated against a large collection of different datasets and
74
+ metrics to cover different aspects of text generation:
75
+
76
+ #### Reasoning and factuality
77
+
78
+ | Benchmark | Metric | Gemma 3 PT 1B | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
79
+ | ------------------------------ |----------------|:--------------:|:-------------:|:--------------:|:--------------:|
80
+ | [HellaSwag][hellaswag] | 10-shot | 62.3 | 77.2 | 84.2 | 85.6 |
81
+ | [BoolQ][boolq] | 0-shot | 63.2 | 72.3 | 78.8 | 82.4 |
82
+ | [PIQA][piqa] | 0-shot | 73.8 | 79.6 | 81.8 | 83.3 |
83
+ | [SocialIQA][socialiqa] | 0-shot | 48.9 | 51.9 | 53.4 | 54.9 |
84
+ | [TriviaQA][triviaqa] | 5-shot | 39.8 | 65.8 | 78.2 | 85.5 |
85
+ | [Natural Questions][naturalq] | 5-shot | 9.48 | 20.0 | 31.4 | 36.1 |
86
+ | [ARC-c][arc] | 25-shot | 38.4 | 56.2 | 68.9 | 70.6 |
87
+ | [ARC-e][arc] | 0-shot | 73.0 | 82.4 | 88.3 | 89.0 |
88
+ | [WinoGrande][winogrande] | 5-shot | 58.2 | 64.7 | 74.3 | 78.8 |
89
+ | [BIG-Bench Hard][bbh] | few-shot | 28.4 | 50.9 | 72.6 | 77.7 |
90
+ | [DROP][drop] | 1-shot | 42.4 | 60.1 | 72.2 | 77.2 |
91
+
92
+ [hellaswag]: https://arxiv.org/abs/1905.07830
93
+ [boolq]: https://arxiv.org/abs/1905.10044
94
+ [piqa]: https://arxiv.org/abs/1911.11641
95
+ [socialiqa]: https://arxiv.org/abs/1904.09728
96
+ [triviaqa]: https://arxiv.org/abs/1705.03551
97
+ [naturalq]: https://github.com/google-research-datasets/natural-questions
98
+ [arc]: https://arxiv.org/abs/1911.01547
99
+ [winogrande]: https://arxiv.org/abs/1907.10641
100
+ [bbh]: https://paperswithcode.com/dataset/bbh
101
+ [drop]: https://arxiv.org/abs/1903.00161
102
+
103
+ #### Multilingual
104
+
105
+ | Benchmark | Gemma 3 PT 1B | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
106
+ | ------------------------------------ |:-------------:|:-------------:|:--------------:|:--------------:|
107
+ | [MGSM][mgsm] | 2.04 | 34.7 | 64.3 | 74.3 |
108
+ | [Global-MMLU-Lite][global-mmlu-lite] | 24.9 | 57.0 | 69.4 | 75.7 |
109
+ | [WMT24++][wmt24pp] (ChrF) | 36.7 | 48.4 | 53.9 | 55.7 |
110
+ | [FloRes][flores] | 29.5 | 39.2 | 46.0 | 48.8 |
111
+ | [XQuAD][xquad] (all) | 43.9 | 68.0 | 74.5 | 76.8 |
112
+ | [ECLeKTic][eclektic] | 4.69 | 11.0 | 17.2 | 24.4 |
113
+ | [IndicGenBench][indicgenbench] | 41.4 | 57.2 | 61.7 | 63.4 |
114
+
115
+ [mgsm]: https://arxiv.org/abs/2210.03057
116
+ [flores]: https://arxiv.org/abs/2106.03193
117
+ [xquad]: https://arxiv.org/abs/1910.11856v3
118
+ [global-mmlu-lite]: https://huggingface.co/datasets/CohereForAI/Global-MMLU-Lite
119
+ [wmt24pp]: https://arxiv.org/abs/2502.12404v1
120
+ [eclektic]: https://arxiv.org/abs/2502.21228
121
+ [indicgenbench]: https://arxiv.org/abs/2404.16816