thermal666 commited on
Commit
fcef847
·
verified ·
1 Parent(s): 20a976a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -14
README.md CHANGED
@@ -15,14 +15,13 @@ Nature Language Model (NatureLM) is a sequence-based science foundation model de
15
  - Model type: Sequence-based science foundation model
16
  - Language(s): English
17
  - License: MIT License
 
18
 
19
 
20
  # Model sources
21
  ## Repository:
22
- We provide four repositories for 1B and 8x7B models, including both base versions and instruction-finetuned versions.
23
 
24
- - https://huggingface.co/microsoft/NatureLM-1B
25
- - https://huggingface.co/microsoft/NatureLM-1B-Inst
26
  - https://huggingface.co/microsoft/NatureLM-8x7B
27
  - https://huggingface.co/microsoft/NatureLM-8x7B-Inst
28
 
@@ -43,6 +42,12 @@ NatureLM currently not ready to use in clinical applications, without rigorous e
43
  NatureLM is not a general-purpose language model and is not designed or optimized to perform general tasks like text summarization or Q&A.
44
  ### Use by Non-Experts
45
  NatureLM outputs scientific entities (e.g., molecules, proteins, materials) and requires expert interpretation, validation, and analysis. It is not intended for use by non-experts or individuals without the necessary domain knowledge to evaluate and verify its outputs. Outputs, such as small molecule inhibitors for target proteins, require rigorous validation to ensure safety and efficacy. Misuse by non-experts may lead to the design of inactive or suboptimal compounds, resulting in wasted resources and potentially delaying critical research or development efforts.
 
 
 
 
 
 
46
 
47
  ## Risks and limitations
48
  NatureLM may not always generate compounds or proteins precisely aligned with user instructions. Users are advised to apply their own adaptive filters before proceeding. Users are responsible for verification of model outputs and decision-making.
@@ -60,17 +65,10 @@ Preprocessing
60
  The training procedure involves two stages: Stage 1 focuses on training newly introduced tokens while freezing existing model parameters. Stage 2 involves joint optimization of both new and existing parameters to enhance overall performance.
61
 
62
  ## Training hyperparameters
63
- - Learning Rate:
64
- - 1B model: 1×10<sup>−4</sup>
65
- - 8x7B model: 2×10<sup>−4</sup>
66
- - Batch Size (Sentences):
67
- - 1B model: 4096
68
- - 8x7B model: 1536
69
- - Context Length (Tokens):
70
- - All models: 8192
71
- - GPU Number (H100):
72
- - 1B model: 64
73
- - 8x7B model: 256
74
 
75
  ## Speeds, sizes, times
76
 
 
15
  - Model type: Sequence-based science foundation model
16
  - Language(s): English
17
  - License: MIT License
18
+ - Finetuned from model: one version of the model is finetuned from Mixtral-8x7B-v0.1
19
 
20
 
21
  # Model sources
22
  ## Repository:
23
+ We provide two repositories for 8x7B models, including both base versions and instruction-finetuned versions.
24
 
 
 
25
  - https://huggingface.co/microsoft/NatureLM-8x7B
26
  - https://huggingface.co/microsoft/NatureLM-8x7B-Inst
27
 
 
42
  NatureLM is not a general-purpose language model and is not designed or optimized to perform general tasks like text summarization or Q&A.
43
  ### Use by Non-Experts
44
  NatureLM outputs scientific entities (e.g., molecules, proteins, materials) and requires expert interpretation, validation, and analysis. It is not intended for use by non-experts or individuals without the necessary domain knowledge to evaluate and verify its outputs. Outputs, such as small molecule inhibitors for target proteins, require rigorous validation to ensure safety and efficacy. Misuse by non-experts may lead to the design of inactive or suboptimal compounds, resulting in wasted resources and potentially delaying critical research or development efforts.
45
+ ### CBRN Applications (Chemical, Biological, Radiological, and Nuclear)
46
+ NatureLM is not intended for the design, development, or optimization of agents or materials for harmful purposes, including but not limited to weapons of mass destruction, bioterrorism, or other malicious uses.
47
+ ### Unethical or Harmful Applications
48
+ The use of NatureLM must align with ethical research practices. It is not intended for tasks that could cause harm to individuals, communities, or the environment.
49
+
50
+
51
 
52
  ## Risks and limitations
53
  NatureLM may not always generate compounds or proteins precisely aligned with user instructions. Users are advised to apply their own adaptive filters before proceeding. Users are responsible for verification of model outputs and decision-making.
 
65
  The training procedure involves two stages: Stage 1 focuses on training newly introduced tokens while freezing existing model parameters. Stage 2 involves joint optimization of both new and existing parameters to enhance overall performance.
66
 
67
  ## Training hyperparameters
68
+ - Learning Rate: 2×10<sup>−4</sup>
69
+ - Batch Size (Sentences): 8x7B model: 1536
70
+ - Context Length (Tokens): 8192
71
+ - GPU Number (H100): 8x7B model: 256
 
 
 
 
 
 
 
72
 
73
  ## Speeds, sizes, times
74