Spestly commited on
Commit
a782ef2
ยท
verified ยท
1 Parent(s): e0c7aae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +97 -1
README.md CHANGED
@@ -3,4 +3,100 @@ license: mit
3
  base_model:
4
  - rubenroy/Zurich-7B-GCv2-5m
5
  library_name: transformers
6
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  base_model:
4
  - rubenroy/Zurich-7B-GCv2-5m
5
  library_name: transformers
6
+ ---
7
+ # Maverick Model Card
8
+
9
+ ## Model Overview
10
+
11
+ **Maverick** is a 14.7-billion-parameter causal language model fine-tuned from [Ruben Roy's Zurich-14B-GCv2-5m](https://huggingface.co/rubenroy/Zurich-14B-GCv2-5m). The base model, Zurich-14B-GCv2-5m, is itself a fine-tuned version of Alibaba's Qwen 2.5 14B Instruct model, trained on the GammaCorpus v2-5m dataset. Maverick is designed to excel in various STEM fields and general natural language processing tasks, offering enhanced reasoning and instruction-following capabilities.๎ˆ†
12
+
13
+ ## Model Details
14
+
15
+ - **Model Developer:** Aayan Mishra
16
+ - **Model Type:** Causal Language Model
17
+ - **Architecture:** Transformer with Rotary Position Embeddings (RoPE), SwiGLU activation, RMSNorm, and Attention QKV bias
18
+ - **Parameters:** 14.7 billion total (13.1 billion non-embedding)๎ˆ†
19
+ - **Layers:** 48
20
+ - **Attention Heads:** 28 for query and 4 for key-value (Grouped Query Attention)โ€‹
21
+ - **Vocabulary Size:** Approximately 151,646 tokens
22
+ - **Context Length:** Supports up to 131,072 tokens
23
+ - **Languages Supported:** Over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, and Arabic
24
+ - **License:** MIT
25
+
26
+ ## Training Details
27
+
28
+ Maverick was fine-tuned using the Unsloth framework on a single NVIDIA A100 GPU. The fine-tuning process spanned approximately 90 minutes over 60 epochs, utilising a curated dataset focused on instruction-following and STEM-related content. This approach aimed to enhance the model's performance in complex reasoning and academic tasks.๎ˆ†
29
+
30
+ ## Intended Use
31
+
32
+ Maverick is designed for a range of applications, including but not limited to:
33
+
34
+ - **STEM Reasoning:** Assisting with problem-solving and explanations in science, technology, engineering, and mathematics.
35
+ - **Academic Assistance:** Providing support for tutoring, essay composition, and research inquiries.
36
+ - **General NLP Tasks:** Engaging in text completion, summarisation, and question-answering tasks.
37
+ - **Data Analysis:** Offering insights and interpretations of data-centric queries.
38
+
39
+ While Maverick is a powerful tool for various applications, it is not intended for real-time, safety-critical systems or for processing sensitive personal information.๎ˆ†
40
+
41
+ ## How to Use
42
+
43
+ To utilize Maverick, ensure that you have the latest version of the `transformers` library installed:
44
+
45
+ ```bash
46
+ pip install transformers
47
+ ```
48
+
49
+
50
+ Here's an example of how to load the Maverick model and generate a response:
51
+
52
+ ```python
53
+ from transformers import AutoModelForCausalLM, AutoTokenizer
54
+
55
+ model_name = "Spestly/Maverick-1-7B"
56
+
57
+ model = AutoModelForCausalLM.from_pretrained(
58
+ model_name,
59
+ torch_dtype="auto",
60
+ device_map="auto"
61
+ )
62
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
63
+
64
+ prompt = "Explain the concept of entropy in thermodynamics."
65
+ messages = [
66
+ {"role": "system", "content": "You are Maverick, an AI assistant designed to be helpful."},
67
+ {"role": "user", "content": prompt}
68
+ ]
69
+ text = tokenizer.apply_chat_template(
70
+ messages,
71
+ tokenize=False,
72
+ add_generation_prompt=True
73
+ )
74
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
75
+
76
+ generated_ids = model.generate(
77
+ **model_inputs,
78
+ max_new_tokens=512
79
+ )
80
+ generated_ids = [
81
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
82
+ ]
83
+
84
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
85
+ print(response)
86
+ ```
87
+
88
+ ## Limitations
89
+
90
+ Users should be aware of the following limitations:
91
+
92
+ - **Biases:** Maverick may exhibit biases present in its training data. Users should critically assess outputs, especially in sensitive contexts.
93
+ - **Knowledge Cutoff:** The model's knowledge is current up to August 2024. It may not be aware of events or developments occurring after this date.
94
+ - **Language Support:** While primarily trained on English data, performance in other languages may be inconsistent.
95
+
96
+ ## Acknowledgements
97
+
98
+ Maverick builds upon the work of [Ruben Roy](https://huggingface.co/rubenroy), particularly the Zurich-14B-GCv2-5m model, which is a fine-tuned version of Alibaba's Qwen 2.5 14B Instruct model. Gratitude is also extended to the open-source AI community for their contributions to tools and frameworks that facilitated the development of Maverick.
99
+
100
+ ## License
101
+
102
+ Maverick is released under the [MIT License](https://opensource.org/license/mit), permitting wide usage with proper attribution.๎ˆ†