Martico2432 commited on
Commit
145cbdf
·
verified ·
1 Parent(s): b3b6708

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +72 -3
README.md CHANGED
@@ -1,3 +1,72 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - theblackcat102/evol-codealpaca-v1
5
+ language:
6
+ - en
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - code
10
+ ---
11
+ # Lizard
12
+
13
+ A lightweight 17m parameter coding model.
14
+
15
+ ## Model Details
16
+
17
+ ### Model Description
18
+
19
+ - **Developed by:** Martico2432
20
+ - **Model type:** Transformer
21
+ - **Language(s) (NLP):** English
22
+ - **License:** MIT
23
+
24
+ ## Uses
25
+
26
+ - Experimentation with small models
27
+
28
+
29
+ ### Direct Use
30
+
31
+ - Try to get the model to give a coherent code
32
+
33
+
34
+ ### Out-of-Scope Use
35
+
36
+ This model won't work for any malicious goal. It's too dumb
37
+
38
+ ## Bias, Risks, and Limitations
39
+
40
+ 17m parameters is not a lot, and it limits a lot it's usage
41
+
42
+ ### Recommendations
43
+
44
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model
45
+
46
+ ## How to Get Started with the Model
47
+
48
+ Use the code in the repository of this model to get started
49
+
50
+ ## Training Details
51
+
52
+ ### Training Data
53
+
54
+ This model was trained using the **theblackcat102/evol-codealpaca-v1** dataset
55
+
56
+ ### Training Procedure
57
+
58
+
59
+
60
+ #### Training Hyperparameters
61
+
62
+ - **Training regime:** fp32 <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
63
+
64
+ ## Technical Specifications
65
+
66
+ #### Hardware
67
+
68
+ - Any device with 8 gb of memory should be able to run it
69
+
70
+ #### Software
71
+
72
+ - PyTorch