brittlewis12 commited on
Commit
93f5f3c
·
verified ·
1 Parent(s): e0785d1

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +142 -0
README.md ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/gemma-3-4b-it
3
+ pipeline_tag: text-generation
4
+ inference: true
5
+ language:
6
+ - en
7
+ license: gemma
8
+ model_creator: google
9
+ model_name: gemma-3-4b-it
10
+ model_type: gemma3_text
11
+ quantized_by: brittlewis12
12
+ tags:
13
+ - reasoning
14
+ - gemma
15
+ ---
16
+
17
+ # Gemma 3 4B IT GGUF
18
+
19
+ **Original model**: [Gemma 3 4B IT](https://huggingface.co/google/gemma-3-4b-it)
20
+
21
+ **Model creator**: [Google DeepMind](https://huggingface.co/google)
22
+
23
+ > Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models.
24
+
25
+ > Gemma 3 has a large, 128K context window, multilingual support in over 140 languages, and is available in more sizes than previous versions. Gemma 3 models are well-suited for a variety of text generation and image understanding tasks, including question answering, summarization, and reasoning. Their relatively small size makes it possible to deploy them in environments with limited resources such as laptops, desktops or your own cloud infrastructure, democratizing access to state of the art AI models and helping foster innovation for everyone.
26
+
27
+ This repo contains GGUF format model files for Google DeepMind’s Gemma 3 4B IT (instruction-tuned).
28
+
29
+ ### What is GGUF?
30
+
31
+ GGUF is a file format for representing AI models. It is the third version of the format,
32
+ introduced by the llama.cpp team on August 21st 2023.
33
+
34
+ Converted with llama.cpp build b4875 (revision [7841fc7](https://github.com/ggml-org/llama.cpp/commits/7841fc723e059d1fd9640e5c0ef19050fcc7c698)),
35
+ using [autogguf-rs](https://github.com/brittlewis12/autogguf-rs).
36
+
37
+ ### Prompt template: [Gemma Instruct](https://huggingface.co/google/gemma-3-4b-it/raw/main/tokenizer_config.json)
38
+
39
+ ```
40
+ {{system_prompt}}
41
+ <start_of_turn>user
42
+ {{prompt}}<end_of_turn>
43
+ <start_of_turn>model
44
+
45
+
46
+ ```
47
+
48
+ ---
49
+
50
+ ## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac!
51
+
52
+ ![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg)
53
+
54
+ [cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
55
+ - create & save **Characters** with custom system prompts & temperature settings
56
+ - download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
57
+ * or, use an API key with the chat completions-compatible model provider of your choice -- ChatGPT, Claude, Gemini, DeepSeek, & more!
58
+ - make it your own with custom **Theme colors**
59
+ - powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggml-org/llama.cpp), with **haptics** during response streaming!
60
+ - **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
61
+ * if you **already have the app**, download Gemma 3 4B IT now!
62
+ * <cnvrsai:///models/search/hf?id=brittlewis12/gemma-3-4b-it-GGUF>
63
+ - follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date
64
+
65
+ ### Gemma 3 4B IT in cnvrs on macOS
66
+
67
+ ![gemma-3 in cnvrs](https://cdn-uploads.huggingface.co/production/uploads/63b64d7a889aa6707f155cdb/FVVJ1WtuGglQuhCSwc6kh.png)
68
+
69
+ ---
70
+
71
+ ## Original Model Evaluation
72
+
73
+ > These models were evaluated against a large collection of different datasets and
74
+ metrics to cover different aspects of text generation:
75
+
76
+ #### Reasoning and factuality
77
+
78
+ | Benchmark | Metric | Gemma 3 PT 1B | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
79
+ | ------------------------------ |----------------|:--------------:|:-------------:|:--------------:|:--------------:|
80
+ | [HellaSwag][hellaswag] | 10-shot | 62.3 | 77.2 | 84.2 | 85.6 |
81
+ | [BoolQ][boolq] | 0-shot | 63.2 | 72.3 | 78.8 | 82.4 |
82
+ | [PIQA][piqa] | 0-shot | 73.8 | 79.6 | 81.8 | 83.3 |
83
+ | [SocialIQA][socialiqa] | 0-shot | 48.9 | 51.9 | 53.4 | 54.9 |
84
+ | [TriviaQA][triviaqa] | 5-shot | 39.8 | 65.8 | 78.2 | 85.5 |
85
+ | [Natural Questions][naturalq] | 5-shot | 9.48 | 20.0 | 31.4 | 36.1 |
86
+ | [ARC-c][arc] | 25-shot | 38.4 | 56.2 | 68.9 | 70.6 |
87
+ | [ARC-e][arc] | 0-shot | 73.0 | 82.4 | 88.3 | 89.0 |
88
+ | [WinoGrande][winogrande] | 5-shot | 58.2 | 64.7 | 74.3 | 78.8 |
89
+ | [BIG-Bench Hard][bbh] | few-shot | 28.4 | 50.9 | 72.6 | 77.7 |
90
+ | [DROP][drop] | 1-shot | 42.4 | 60.1 | 72.2 | 77.2 |
91
+
92
+ [hellaswag]: https://arxiv.org/abs/1905.07830
93
+ [boolq]: https://arxiv.org/abs/1905.10044
94
+ [piqa]: https://arxiv.org/abs/1911.11641
95
+ [socialiqa]: https://arxiv.org/abs/1904.09728
96
+ [triviaqa]: https://arxiv.org/abs/1705.03551
97
+ [naturalq]: https://github.com/google-research-datasets/natural-questions
98
+ [arc]: https://arxiv.org/abs/1911.01547
99
+ [winogrande]: https://arxiv.org/abs/1907.10641
100
+ [bbh]: https://paperswithcode.com/dataset/bbh
101
+ [drop]: https://arxiv.org/abs/1903.00161
102
+
103
+ #### STEM and code
104
+
105
+ | Benchmark | Metric | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
106
+ | ------------------------------ |----------------|:-------------:|:--------------:|:--------------:|
107
+ | [MMLU][mmlu] | 5-shot | 59.6 | 74.5 | 78.6 |
108
+ | [MMLU][mmlu] (Pro COT) | 5-shot | 29.2 | 45.3 | 52.2 |
109
+ | [AGIEval][agieval] | 3-5-shot | 42.1 | 57.4 | 66.2 |
110
+ | [MATH][math] | 4-shot | 24.2 | 43.3 | 50.0 |
111
+ | [GSM8K][gsm8k] | 8-shot | 38.4 | 71.0 | 82.6 |
112
+ | [GPQA][gpqa] | 5-shot | 15.0 | 25.4 | 24.3 |
113
+ | [MBPP][mbpp] | 3-shot | 46.0 | 60.4 | 65.6 |
114
+ | [HumanEval][humaneval] | 0-shot | 36.0 | 45.7 | 48.8 |
115
+
116
+ [mmlu]: https://arxiv.org/abs/2009.03300
117
+ [agieval]: https://arxiv.org/abs/2304.06364
118
+ [math]: https://arxiv.org/abs/2103.03874
119
+ [gsm8k]: https://arxiv.org/abs/2110.14168
120
+ [gpqa]: https://arxiv.org/abs/2311.12022
121
+ [mbpp]: https://arxiv.org/abs/2108.07732
122
+ [humaneval]: https://arxiv.org/abs/2107.03374
123
+
124
+ #### Multilingual
125
+
126
+ | Benchmark | Gemma 3 PT 1B | Gemma 3 PT 4B | Gemma 3 PT 12B | Gemma 3 PT 27B |
127
+ | ------------------------------------ |:-------------:|:-------------:|:--------------:|:--------------:|
128
+ | [MGSM][mgsm] | 2.04 | 34.7 | 64.3 | 74.3 |
129
+ | [Global-MMLU-Lite][global-mmlu-lite] | 24.9 | 57.0 | 69.4 | 75.7 |
130
+ | [WMT24++][wmt24pp] (ChrF) | 36.7 | 48.4 | 53.9 | 55.7 |
131
+ | [FloRes][flores] | 29.5 | 39.2 | 46.0 | 48.8 |
132
+ | [XQuAD][xquad] (all) | 43.9 | 68.0 | 74.5 | 76.8 |
133
+ | [ECLeKTic][eclektic] | 4.69 | 11.0 | 17.2 | 24.4 |
134
+ | [IndicGenBench][indicgenbench] | 41.4 | 57.2 | 61.7 | 63.4 |
135
+
136
+ [mgsm]: https://arxiv.org/abs/2210.03057
137
+ [flores]: https://arxiv.org/abs/2106.03193
138
+ [xquad]: https://arxiv.org/abs/1910.11856v3
139
+ [global-mmlu-lite]: https://huggingface.co/datasets/CohereForAI/Global-MMLU-Lite
140
+ [wmt24pp]: https://arxiv.org/abs/2502.12404v1
141
+ [eclektic]: https://arxiv.org/abs/2502.21228
142
+ [indicgenbench]: https://arxiv.org/abs/2404.16816