Update README.md
Browse files
README.md
CHANGED
@@ -10,25 +10,25 @@ base_model:
|
|
10 |
|
11 |
|
12 |
|
13 |
-
#
|
14 |
|
15 |
[](https://github.com/oumi-ai/oumi)
|
16 |
|
17 |
|
18 |
|
19 |
## Model Description
|
20 |
-
**
|
21 |
|
22 |
-
|
23 |
|
24 |
## Model Sources
|
25 |
|
26 |
<!-- Provide the basic links for the model. -->
|
27 |
|
28 |
- π **Paper:** https://arxiv.org/abs/2502.08820
|
29 |
-
- π **Project Page:** https://emrecanacikgoz.github.io/
|
30 |
- π» **Repository:** https://github.com/oumi-ai/oumi/tree/main/configs/projects/calm
|
31 |
-
- π **Dataset:** https://huggingface.co/datasets/uiuc-convai/
|
32 |
|
33 |
|
34 |
|
@@ -36,11 +36,11 @@ CALM-8B is trained on a **multi-task dataset** covering dialogue state tracking,
|
|
36 |
---
|
37 |
## Model Details
|
38 |
|
39 |
-
- **Model Name:**
|
40 |
- **Developed by:** Colloboration of UIUC Conversational AI LAB and Oumi
|
41 |
- **License:** cc-by-nc-4.0
|
42 |
- **Architecture:** Fine-tuned **Llama 3.1 8B Instruct**
|
43 |
-
- **Training Data:**
|
44 |
- **Fine-tuning Framework:** [Oumi](https://github.com/oumi-ai/oumi)
|
45 |
- **Training Hardware:** 8 NVIDIA H100 GPUs
|
46 |
- **Training Duration:** ~8 hours
|
@@ -79,7 +79,7 @@ CALM-8B is trained on a **multi-task dataset** covering dialogue state tracking,
|
|
79 |
- **Gradient Accumulation Steps:** 1
|
80 |
|
81 |
---
|
82 |
-
## π‘
|
83 |
<img src="table.png" alt="CALM-IT Dataset Statistics" width="800"/>
|
84 |
|
85 |
|
@@ -95,8 +95,8 @@ CALM-8B is trained on a **multi-task dataset** covering dialogue state tracking,
|
|
95 |
```python
|
96 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
97 |
|
98 |
-
tokenizer = AutoTokenizer.from_pretrained("uiuc-convai/
|
99 |
-
model = AutoModelForCausalLM.from_pretrained("uiuc-convai/
|
100 |
```
|
101 |
|
102 |
### π Example Oumi Inference
|
@@ -116,8 +116,8 @@ oumi train -c ./oumi_train.yaml
|
|
116 |
```
|
117 |
|
118 |
---
|
119 |
-
- **Task-Specific Calibration:** While
|
120 |
-
- **Scalability to Larger Models:** Future iterations (
|
121 |
- **Open-Source Expansion:** All datasets, training scripts, and model checkpoints are publicly available to foster further research.
|
122 |
|
123 |
## Acknowledgements
|
@@ -132,7 +132,7 @@ This model is licensed under [Creative Commons NonCommercial (CC BY-NC 4.0)](htt
|
|
132 |
If you use **CALM-8B** in your research, please cite:
|
133 |
```
|
134 |
@misc{acikgoz2025singlemodelmastermultiturn,
|
135 |
-
title={Can a Single Model Master Both Multi-turn Conversations and Tool Use?
|
136 |
author={Emre Can Acikgoz and Jeremiah Greer and Akul Datta and Ze Yang and William Zeng and Oussama Elachqar and Emmanouil Koukoumidis and Dilek Hakkani-TΓΌr and Gokhan Tur},
|
137 |
year={2025},
|
138 |
eprint={2502.08820},
|
|
|
10 |
|
11 |
|
12 |
|
13 |
+
# CoALM-8B: Conversational Agentic Language Model
|
14 |
|
15 |
[](https://github.com/oumi-ai/oumi)
|
16 |
|
17 |
|
18 |
|
19 |
## Model Description
|
20 |
+
**CoALM-8B** is the smallest open-source model of **CoALM** (Conversational Agentic Language Model) series, designed to integrate both **Task-Oriented Dialogue (TOD) capabilities** and **Language Agent (LA) functionalities** into a unified system. By fine-tuning on **CoALM-IT**, a novel dataset that interleaves multi-turn ReAct-based reasoning with complex API usage, CoALM-8B achieves promising results on TOD and function-calling benchmarks.
|
21 |
|
22 |
+
CoALM-8B is trained on a **multi-task dataset** covering dialogue state tracking, function calling, and multi-turn reasoning. The model outperforms top domain-specific models on key evaluation benchmarks: **MultiWOZ 2.4 (TOD), BFCL V3 (LA), and API-Bank (LA).**
|
23 |
|
24 |
## Model Sources
|
25 |
|
26 |
<!-- Provide the basic links for the model. -->
|
27 |
|
28 |
- π **Paper:** https://arxiv.org/abs/2502.08820
|
29 |
+
- π **Project Page:** https://emrecanacikgoz.github.io/CoALM/
|
30 |
- π» **Repository:** https://github.com/oumi-ai/oumi/tree/main/configs/projects/calm
|
31 |
+
- π **Dataset:** https://huggingface.co/datasets/uiuc-convai/CoALM-IT
|
32 |
|
33 |
|
34 |
|
|
|
36 |
---
|
37 |
## Model Details
|
38 |
|
39 |
+
- **Model Name:** CoALM-8B
|
40 |
- **Developed by:** Colloboration of UIUC Conversational AI LAB and Oumi
|
41 |
- **License:** cc-by-nc-4.0
|
42 |
- **Architecture:** Fine-tuned **Llama 3.1 8B Instruct**
|
43 |
+
- **Training Data:** CoALM-IT dataset
|
44 |
- **Fine-tuning Framework:** [Oumi](https://github.com/oumi-ai/oumi)
|
45 |
- **Training Hardware:** 8 NVIDIA H100 GPUs
|
46 |
- **Training Duration:** ~8 hours
|
|
|
79 |
- **Gradient Accumulation Steps:** 1
|
80 |
|
81 |
---
|
82 |
+
## π‘ CoALM-IT Dataset
|
83 |
<img src="table.png" alt="CALM-IT Dataset Statistics" width="800"/>
|
84 |
|
85 |
|
|
|
95 |
```python
|
96 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
97 |
|
98 |
+
tokenizer = AutoTokenizer.from_pretrained("uiuc-convai/CoALM-8B")
|
99 |
+
model = AutoModelForCausalLM.from_pretrained("uiuc-convai/CoALM-8B")
|
100 |
```
|
101 |
|
102 |
### π Example Oumi Inference
|
|
|
116 |
```
|
117 |
|
118 |
---
|
119 |
+
- **Task-Specific Calibration:** While CoALM-8B generalizes well across tasks, performance can improve with domain-specific fine-tuning.
|
120 |
+
- **Scalability to Larger Models:** Future iterations (CoALM-70B, CoALM-405B) extend capabilities to larger-scale agentic conversations.
|
121 |
- **Open-Source Expansion:** All datasets, training scripts, and model checkpoints are publicly available to foster further research.
|
122 |
|
123 |
## Acknowledgements
|
|
|
132 |
If you use **CALM-8B** in your research, please cite:
|
133 |
```
|
134 |
@misc{acikgoz2025singlemodelmastermultiturn,
|
135 |
+
title={Can a Single Model Master Both Multi-turn Conversations and Tool Use? CoALM: A Unified Conversational Agentic Language Model},
|
136 |
author={Emre Can Acikgoz and Jeremiah Greer and Akul Datta and Ze Yang and William Zeng and Oussama Elachqar and Emmanouil Koukoumidis and Dilek Hakkani-TΓΌr and Gokhan Tur},
|
137 |
year={2025},
|
138 |
eprint={2502.08820},
|