sccastillo commited on
Commit
f9c45d2
·
verified ·
1 Parent(s): 915fb93

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -3
README.md CHANGED
@@ -11,12 +11,20 @@ tags:
11
  base_model: unsloth/llama-3-8b-bnb-4bit
12
  ---
13
 
14
- # Uploaded model
 
 
 
 
 
 
15
 
 
 
 
 
16
  - **Developed by:** sccastillo
17
  - **License:** apache-2.0
18
  - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
19
-
20
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
11
  base_model: unsloth/llama-3-8b-bnb-4bit
12
  ---
13
 
14
+ # LLama 3 for router module in RAG
15
+
16
+ While developing complex RAG applications, I found a common need for router functionality to map user queries to different system workflows (and APIs). The router acts as a dispatcher that can enhance responsiveness and accuracy by choosing the best workflow or API based on the query context. This implies that we need to produce structured output from unstructured input text.
17
+
18
+ To this end, and to avoid external dependencies for this part of the system, I will undertake a simple exercise to fine-tune the new Llama 3 model to process text input and generate JSON-like output. My hope is that we can seamlessly integrate various models to reinforce complex applications in production settings, building a robust critical infraestructures for the semantical modules.
19
+
20
+ For training, we will use structured data from [azizshaw](https://huggingface.co/azizshaw/text_to_json). The dataset has 485 rows and contains 'input', 'output' and 'instruction' columns.
21
 
22
+ For a quick evaluation, let's use another dataset for text-to-JSON, the **Diverse Restricted JSON Data Extraction**, curated by: The paraloq analytics team ([here](https://huggingface.co/datasets/paraloq/json_data_extraction))
23
+
24
+
25
+ # Uploaded model
26
  - **Developed by:** sccastillo
27
  - **License:** apache-2.0
28
  - **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
 
29
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
30