Update README.md
Browse files
README.md
CHANGED
@@ -28,7 +28,7 @@ It has state-of-the-art performance among multimodal models with a similar size
|
|
28 |
You can find all models in the Molmo family [here](https://huggingface.co/collections/allenai/molmo-66f379e6fe3b8ef090a8ca19).
|
29 |
**Learn more** about the Molmo family [in our announcement blog post](https://molmo.allenai.org/blog).
|
30 |
|
31 |
-
MolmoE-1B is a multimodal Mixture-of-Experts LLM with 1.5B active and 7.2B total parameters
|
32 |
It nearly matches the performance of GPT-4V on both academic benchmarks and human evaluation, and achieves state-of-the-art performance among similarly-sized open multimodal models.
|
33 |
|
34 |
This checkpoint is a **preview** of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
|
@@ -100,9 +100,9 @@ print(generated_text)
|
|
100 |
| Model | Average Score on 11 Academic Benchmarks | Human Preference Elo Rating |
|
101 |
|-----------------------------|-----------------------------------------|-----------------------------|
|
102 |
| Molmo 72B | 81.2 | 1077 |
|
103 |
-
|
|
104 |
| Molmo 7B-O | 74.6 | 1051 |
|
105 |
-
| MolmoE 1B
|
106 |
| GPT-4o | 78.5 | 1079 |
|
107 |
| GPT-4V | 71.1 | 1041 |
|
108 |
| Gemini 1.5 Pro | 78.3 | 1074 |
|
|
|
28 |
You can find all models in the Molmo family [here](https://huggingface.co/collections/allenai/molmo-66f379e6fe3b8ef090a8ca19).
|
29 |
**Learn more** about the Molmo family [in our announcement blog post](https://molmo.allenai.org/blog).
|
30 |
|
31 |
+
MolmoE-1B is a multimodal Mixture-of-Experts LLM with 1.5B active and 7.2B total parameters based on [OLMoE-1B-7B-0924](https://huggingface.co/allenai/OLMoE-1B-7B-0924).
|
32 |
It nearly matches the performance of GPT-4V on both academic benchmarks and human evaluation, and achieves state-of-the-art performance among similarly-sized open multimodal models.
|
33 |
|
34 |
This checkpoint is a **preview** of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.
|
|
|
100 |
| Model | Average Score on 11 Academic Benchmarks | Human Preference Elo Rating |
|
101 |
|-----------------------------|-----------------------------------------|-----------------------------|
|
102 |
| Molmo 72B | 81.2 | 1077 |
|
103 |
+
| Molmo 7B-D | 77.3 | 1056 |
|
104 |
| Molmo 7B-O | 74.6 | 1051 |
|
105 |
+
| **MolmoE 1B (this model)** | **68.6** | **1032** |
|
106 |
| GPT-4o | 78.5 | 1079 |
|
107 |
| GPT-4V | 71.1 | 1041 |
|
108 |
| Gemini 1.5 Pro | 78.3 | 1074 |
|