license: apache-2.0 tags:
text-generation
conversational
mistral
fine-tuned
chatbot
bangkit widget:
text: "Hello! I'm looking for recommendations on agricultural products."
Fine-Tuned Mistral Model for Agrease Application This repository contains a fine-tuned version of a Mistral Large Language Model, specifically adapted for the "Agrease" application. The model was developed as part of a capstone project for Bangkit Academy 2024 Batch 2.
The primary goal of this project was to create a specialized conversational AI capable of assisting users within the Agrease application, likely by providing information and recommendations based on data from various marketplaces.
Model Description Base Model: The model is a fine-tuned version of a Mistral v3 Large Language Model.
Fine-tuning Task: The model was fine-tuned for conversational question-answering and recommendations.
Training Data: The training data was collected by scraping various online marketplaces using Python libraries such as BeautifulSoup and Scrapy.
Performance: The fine-tuning process achieved a final training loss of 11%.
Intended Use This model is intended to be used as a chatbot or a conversational agent within a larger application. It can answer user queries, provide product recommendations, and engage in domain-specific conversations related to the "Agrease" application's scope.
How to Use You can use this model with the transformers library for text generation.
from transformers import pipeline
Load the text generation pipeline from the Hugging Face Hub
Replace "your-username/model-name" with the actual model path
generator = pipeline('text-generation', model='your-username/model-name')
Example prompt
prompt = "What are the best fertilizers for rice paddies in a tropical climate?"
Generate a response
response = generator(prompt, max_length=150, num_return_sequences=1)
print(response[0]['generated_text'])
Training Data The dataset used for fine-tuning was created by scraping publicly available data from various e-commerce and marketplace websites. The scraping was performed using custom Python scripts with BeautifulSoup and Scrapy. The collected data was then processed and formatted into a conversational format suitable for training a large language model.
Training Procedure The fine-tuning was performed using the PyTorch framework on the collected dataset. The training focused on minimizing the cross-entropy loss to improve the model's ability to generate relevant and coherent responses in a conversational context. The final model achieved a training loss of 0.11
- Downloads last month
- 45
4-bit