Update README.md
Browse files
README.md
CHANGED
@@ -10,13 +10,13 @@ pipeline_tag: text-generation
|
|
10 |
library_name: transformers
|
11 |
tags:
|
12 |
- tools
|
13 |
-
-
|
14 |
---
|
15 |
# Qwen2.5-Coder-32B-Glaive-ToolCall
|
16 |
|
17 |
## Model Description
|
18 |
|
19 |
-
This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) specifically enhanced for tool calling capabilities. The model has been trained using the `
|
20 |
|
21 |
## Model Details
|
22 |
|
@@ -123,11 +123,11 @@ The following chart shows the training loss progression during the fine-tuning p
|
|
123 |
|
124 |

|
125 |
|
126 |
-
*Training loss curve demonstrating stable convergence over 3 epochs with the
|
127 |
|
128 |
## Limitations
|
129 |
|
130 |
-
- The model's tool calling capabilities are primarily trained on the patterns present in the
|
131 |
- Performance may vary for highly specialized or domain-specific tools not represented in the training data
|
132 |
- Like all LLMs, the model may occasionally generate plausible-sounding but incorrect tool calls
|
133 |
- The model requires careful prompt engineering for optimal tool calling performance
|
|
|
10 |
library_name: transformers
|
11 |
tags:
|
12 |
- tools
|
13 |
+
- functions
|
14 |
---
|
15 |
# Qwen2.5-Coder-32B-Glaive-ToolCall
|
16 |
|
17 |
## Model Description
|
18 |
|
19 |
+
This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) specifically enhanced for tool calling capabilities. The model has been trained using the **Glaive Function Calling v2** dataset (`glaiveai/glaive-function-calling-v2`) to significantly improve its ability to understand, generate, and execute function calls in various programming and automation contexts.
|
20 |
|
21 |
## Model Details
|
22 |
|
|
|
123 |
|
124 |

|
125 |
|
126 |
+
*Training loss curve demonstrating stable convergence over 3 epochs with the Glaive Function Calling v2 dataset.*
|
127 |
|
128 |
## Limitations
|
129 |
|
130 |
+
- The model's tool calling capabilities are primarily trained on the patterns present in the Glaive Function Calling v2 dataset
|
131 |
- Performance may vary for highly specialized or domain-specific tools not represented in the training data
|
132 |
- Like all LLMs, the model may occasionally generate plausible-sounding but incorrect tool calls
|
133 |
- The model requires careful prompt engineering for optimal tool calling performance
|