Update README.md
Browse files
README.md
CHANGED
@@ -15,14 +15,13 @@ Nature Language Model (NatureLM) is a sequence-based science foundation model de
|
|
15 |
- Model type: Sequence-based science foundation model
|
16 |
- Language(s): English
|
17 |
- License: MIT License
|
|
|
18 |
|
19 |
|
20 |
# Model sources
|
21 |
## Repository:
|
22 |
-
We provide
|
23 |
|
24 |
-
- https://huggingface.co/microsoft/NatureLM-1B
|
25 |
-
- https://huggingface.co/microsoft/NatureLM-1B-Inst
|
26 |
- https://huggingface.co/microsoft/NatureLM-8x7B
|
27 |
- https://huggingface.co/microsoft/NatureLM-8x7B-Inst
|
28 |
|
@@ -43,6 +42,12 @@ NatureLM currently not ready to use in clinical applications, without rigorous e
|
|
43 |
NatureLM is not a general-purpose language model and is not designed or optimized to perform general tasks like text summarization or Q&A.
|
44 |
### Use by Non-Experts
|
45 |
NatureLM outputs scientific entities (e.g., molecules, proteins, materials) and requires expert interpretation, validation, and analysis. It is not intended for use by non-experts or individuals without the necessary domain knowledge to evaluate and verify its outputs. Outputs, such as small molecule inhibitors for target proteins, require rigorous validation to ensure safety and efficacy. Misuse by non-experts may lead to the design of inactive or suboptimal compounds, resulting in wasted resources and potentially delaying critical research or development efforts.
|
|
|
|
|
|
|
|
|
|
|
|
|
46 |
|
47 |
## Risks and limitations
|
48 |
NatureLM may not always generate compounds or proteins precisely aligned with user instructions. Users are advised to apply their own adaptive filters before proceeding. Users are responsible for verification of model outputs and decision-making.
|
@@ -60,17 +65,10 @@ Preprocessing
|
|
60 |
The training procedure involves two stages: Stage 1 focuses on training newly introduced tokens while freezing existing model parameters. Stage 2 involves joint optimization of both new and existing parameters to enhance overall performance.
|
61 |
|
62 |
## Training hyperparameters
|
63 |
-
- Learning Rate:
|
64 |
-
|
65 |
-
|
66 |
-
-
|
67 |
-
- 1B model: 4096
|
68 |
-
- 8x7B model: 1536
|
69 |
-
- Context Length (Tokens):
|
70 |
-
- All models: 8192
|
71 |
-
- GPU Number (H100):
|
72 |
-
- 1B model: 64
|
73 |
-
- 8x7B model: 256
|
74 |
|
75 |
## Speeds, sizes, times
|
76 |
|
|
|
15 |
- Model type: Sequence-based science foundation model
|
16 |
- Language(s): English
|
17 |
- License: MIT License
|
18 |
+
- Finetuned from model: one version of the model is finetuned from Mixtral-8x7B-v0.1
|
19 |
|
20 |
|
21 |
# Model sources
|
22 |
## Repository:
|
23 |
+
We provide two repositories for 8x7B models, including both base versions and instruction-finetuned versions.
|
24 |
|
|
|
|
|
25 |
- https://huggingface.co/microsoft/NatureLM-8x7B
|
26 |
- https://huggingface.co/microsoft/NatureLM-8x7B-Inst
|
27 |
|
|
|
42 |
NatureLM is not a general-purpose language model and is not designed or optimized to perform general tasks like text summarization or Q&A.
|
43 |
### Use by Non-Experts
|
44 |
NatureLM outputs scientific entities (e.g., molecules, proteins, materials) and requires expert interpretation, validation, and analysis. It is not intended for use by non-experts or individuals without the necessary domain knowledge to evaluate and verify its outputs. Outputs, such as small molecule inhibitors for target proteins, require rigorous validation to ensure safety and efficacy. Misuse by non-experts may lead to the design of inactive or suboptimal compounds, resulting in wasted resources and potentially delaying critical research or development efforts.
|
45 |
+
### CBRN Applications (Chemical, Biological, Radiological, and Nuclear)
|
46 |
+
NatureLM is not intended for the design, development, or optimization of agents or materials for harmful purposes, including but not limited to weapons of mass destruction, bioterrorism, or other malicious uses.
|
47 |
+
### Unethical or Harmful Applications
|
48 |
+
The use of NatureLM must align with ethical research practices. It is not intended for tasks that could cause harm to individuals, communities, or the environment.
|
49 |
+
|
50 |
+
|
51 |
|
52 |
## Risks and limitations
|
53 |
NatureLM may not always generate compounds or proteins precisely aligned with user instructions. Users are advised to apply their own adaptive filters before proceeding. Users are responsible for verification of model outputs and decision-making.
|
|
|
65 |
The training procedure involves two stages: Stage 1 focuses on training newly introduced tokens while freezing existing model parameters. Stage 2 involves joint optimization of both new and existing parameters to enhance overall performance.
|
66 |
|
67 |
## Training hyperparameters
|
68 |
+
- Learning Rate: 2×10<sup>−4</sup>
|
69 |
+
- Batch Size (Sentences): 8x7B model: 1536
|
70 |
+
- Context Length (Tokens): 8192
|
71 |
+
- GPU Number (H100): 8x7B model: 256
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
72 |
|
73 |
## Speeds, sizes, times
|
74 |
|