Safetensors
English
llama
emrecanacikgoz commited on
Commit
9387781
Β·
verified Β·
1 Parent(s): 5c9c794

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -13
README.md CHANGED
@@ -10,25 +10,25 @@ base_model:
10
 
11
 
12
 
13
- # CALM-8B: Conversational Agentic Language Model
14
 
15
  [![Made with Oumi](https://badgen.net/badge/Made%20with/Oumi/%23085CFF?icon=https%3A%2F%2Foumi.ai%2Flogo_dark.svg)](https://github.com/oumi-ai/oumi)
16
 
17
 
18
 
19
  ## Model Description
20
- **CALM-8B** is the smallest open-source model of **CALM** (Conversational Agentic Language Model) series, designed to integrate both **Task-Oriented Dialogue (TOD) capabilities** and **Language Agent (LA) functionalities** into a unified system. By fine-tuning on **CALM-IT**, a novel dataset that interleaves multi-turn ReAct-based reasoning with complex API usage, CALM-8B achieves promising results on TOD and function-calling benchmarks.
21
 
22
- CALM-8B is trained on a **multi-task dataset** covering dialogue state tracking, function calling, and multi-turn reasoning. The model outperforms top domain-specific models on key evaluation benchmarks: **MultiWOZ 2.4 (TOD), BFCL V3 (LA), and API-Bank (LA).**
23
 
24
  ## Model Sources
25
 
26
  <!-- Provide the basic links for the model. -->
27
 
28
  - πŸ“ **Paper:** https://arxiv.org/abs/2502.08820
29
- - 🌐 **Project Page:** https://emrecanacikgoz.github.io/CALM/
30
  - πŸ’» **Repository:** https://github.com/oumi-ai/oumi/tree/main/configs/projects/calm
31
- - πŸ’Ž **Dataset:** https://huggingface.co/datasets/uiuc-convai/CALM-IT
32
 
33
 
34
 
@@ -36,11 +36,11 @@ CALM-8B is trained on a **multi-task dataset** covering dialogue state tracking,
36
  ---
37
  ## Model Details
38
 
39
- - **Model Name:** CALM-8B
40
  - **Developed by:** Colloboration of UIUC Conversational AI LAB and Oumi
41
  - **License:** cc-by-nc-4.0
42
  - **Architecture:** Fine-tuned **Llama 3.1 8B Instruct**
43
- - **Training Data:** CALM-IT dataset
44
  - **Fine-tuning Framework:** [Oumi](https://github.com/oumi-ai/oumi)
45
  - **Training Hardware:** 8 NVIDIA H100 GPUs
46
  - **Training Duration:** ~8 hours
@@ -79,7 +79,7 @@ CALM-8B is trained on a **multi-task dataset** covering dialogue state tracking,
79
  - **Gradient Accumulation Steps:** 1
80
 
81
  ---
82
- ## πŸ’‘ CALM-IT Dataset
83
  <img src="table.png" alt="CALM-IT Dataset Statistics" width="800"/>
84
 
85
 
@@ -95,8 +95,8 @@ CALM-8B is trained on a **multi-task dataset** covering dialogue state tracking,
95
  ```python
96
  from transformers import AutoModelForCausalLM, AutoTokenizer
97
 
98
- tokenizer = AutoTokenizer.from_pretrained("uiuc-convai/CALM-8B")
99
- model = AutoModelForCausalLM.from_pretrained("uiuc-convai/CALM-8B")
100
  ```
101
 
102
  ### πŸ›  Example Oumi Inference
@@ -116,8 +116,8 @@ oumi train -c ./oumi_train.yaml
116
  ```
117
 
118
  ---
119
- - **Task-Specific Calibration:** While CALM-8B generalizes well across tasks, performance can improve with domain-specific fine-tuning.
120
- - **Scalability to Larger Models:** Future iterations (CALM-70B, CALM-405B) extend capabilities to larger-scale agentic conversations.
121
  - **Open-Source Expansion:** All datasets, training scripts, and model checkpoints are publicly available to foster further research.
122
 
123
  ## Acknowledgements
@@ -132,7 +132,7 @@ This model is licensed under [Creative Commons NonCommercial (CC BY-NC 4.0)](htt
132
  If you use **CALM-8B** in your research, please cite:
133
  ```
134
  @misc{acikgoz2025singlemodelmastermultiturn,
135
- title={Can a Single Model Master Both Multi-turn Conversations and Tool Use? CALM: A Unified Conversational Agentic Language Model},
136
  author={Emre Can Acikgoz and Jeremiah Greer and Akul Datta and Ze Yang and William Zeng and Oussama Elachqar and Emmanouil Koukoumidis and Dilek Hakkani-TΓΌr and Gokhan Tur},
137
  year={2025},
138
  eprint={2502.08820},
 
10
 
11
 
12
 
13
+ # CoALM-8B: Conversational Agentic Language Model
14
 
15
  [![Made with Oumi](https://badgen.net/badge/Made%20with/Oumi/%23085CFF?icon=https%3A%2F%2Foumi.ai%2Flogo_dark.svg)](https://github.com/oumi-ai/oumi)
16
 
17
 
18
 
19
  ## Model Description
20
+ **CoALM-8B** is the smallest open-source model of **CoALM** (Conversational Agentic Language Model) series, designed to integrate both **Task-Oriented Dialogue (TOD) capabilities** and **Language Agent (LA) functionalities** into a unified system. By fine-tuning on **CoALM-IT**, a novel dataset that interleaves multi-turn ReAct-based reasoning with complex API usage, CoALM-8B achieves promising results on TOD and function-calling benchmarks.
21
 
22
+ CoALM-8B is trained on a **multi-task dataset** covering dialogue state tracking, function calling, and multi-turn reasoning. The model outperforms top domain-specific models on key evaluation benchmarks: **MultiWOZ 2.4 (TOD), BFCL V3 (LA), and API-Bank (LA).**
23
 
24
  ## Model Sources
25
 
26
  <!-- Provide the basic links for the model. -->
27
 
28
  - πŸ“ **Paper:** https://arxiv.org/abs/2502.08820
29
+ - 🌐 **Project Page:** https://emrecanacikgoz.github.io/CoALM/
30
  - πŸ’» **Repository:** https://github.com/oumi-ai/oumi/tree/main/configs/projects/calm
31
+ - πŸ’Ž **Dataset:** https://huggingface.co/datasets/uiuc-convai/CoALM-IT
32
 
33
 
34
 
 
36
  ---
37
  ## Model Details
38
 
39
+ - **Model Name:** CoALM-8B
40
  - **Developed by:** Colloboration of UIUC Conversational AI LAB and Oumi
41
  - **License:** cc-by-nc-4.0
42
  - **Architecture:** Fine-tuned **Llama 3.1 8B Instruct**
43
+ - **Training Data:** CoALM-IT dataset
44
  - **Fine-tuning Framework:** [Oumi](https://github.com/oumi-ai/oumi)
45
  - **Training Hardware:** 8 NVIDIA H100 GPUs
46
  - **Training Duration:** ~8 hours
 
79
  - **Gradient Accumulation Steps:** 1
80
 
81
  ---
82
+ ## πŸ’‘ CoALM-IT Dataset
83
  <img src="table.png" alt="CALM-IT Dataset Statistics" width="800"/>
84
 
85
 
 
95
  ```python
96
  from transformers import AutoModelForCausalLM, AutoTokenizer
97
 
98
+ tokenizer = AutoTokenizer.from_pretrained("uiuc-convai/CoALM-8B")
99
+ model = AutoModelForCausalLM.from_pretrained("uiuc-convai/CoALM-8B")
100
  ```
101
 
102
  ### πŸ›  Example Oumi Inference
 
116
  ```
117
 
118
  ---
119
+ - **Task-Specific Calibration:** While CoALM-8B generalizes well across tasks, performance can improve with domain-specific fine-tuning.
120
+ - **Scalability to Larger Models:** Future iterations (CoALM-70B, CoALM-405B) extend capabilities to larger-scale agentic conversations.
121
  - **Open-Source Expansion:** All datasets, training scripts, and model checkpoints are publicly available to foster further research.
122
 
123
  ## Acknowledgements
 
132
  If you use **CALM-8B** in your research, please cite:
133
  ```
134
  @misc{acikgoz2025singlemodelmastermultiturn,
135
+ title={Can a Single Model Master Both Multi-turn Conversations and Tool Use? CoALM: A Unified Conversational Agentic Language Model},
136
  author={Emre Can Acikgoz and Jeremiah Greer and Akul Datta and Ze Yang and William Zeng and Oussama Elachqar and Emmanouil Koukoumidis and Dilek Hakkani-TΓΌr and Gokhan Tur},
137
  year={2025},
138
  eprint={2502.08820},