Transformers
English
Mixture of Experts
olmo
flexolmo
akshitab commited on
Commit
49305e0
·
verified ·
1 Parent(s): e69350f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -17,7 +17,7 @@ library_name: transformers
17
  > FlexOlmo-7x7B-1T (without router training) is a Mixture-of-Experts with 33B total parameters, combining independently trained experts on public-mix, news, math, code, academic texts, creative writing, and Reddit data. The public-mix expert is trained on 1T tokens of public data while the other experts are branched from the public-mix expert and trained on 50B tokens of their respective data.
18
 
19
  This information and more can also be found:
20
- - **Paper**: https://allenai.org/papers/FlexOlmo
21
  - **Code**: https://github.com/allenai/FlexOlmo
22
  - **Blog**: https://allenai.org/blog/flexolmo
23
  - **Data and corresponding models**:
 
17
  > FlexOlmo-7x7B-1T (without router training) is a Mixture-of-Experts with 33B total parameters, combining independently trained experts on public-mix, news, math, code, academic texts, creative writing, and Reddit data. The public-mix expert is trained on 1T tokens of public data while the other experts are branched from the public-mix expert and trained on 50B tokens of their respective data.
18
 
19
  This information and more can also be found:
20
+ - **Paper**: https://allenai.org/papers/flexolmo
21
  - **Code**: https://github.com/allenai/FlexOlmo
22
  - **Blog**: https://allenai.org/blog/flexolmo
23
  - **Data and corresponding models**: