YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

license: apache-2.0 tags:

text-generation

conversational

mistral

fine-tuned

chatbot

bangkit widget:

text: "Hello! I'm looking for recommendations on agricultural products."

Fine-Tuned Mistral Model for Agrease Application This repository contains a fine-tuned version of a Mistral Large Language Model, specifically adapted for the "Agrease" application. The model was developed as part of a capstone project for Bangkit Academy 2024 Batch 2.

The primary goal of this project was to create a specialized conversational AI capable of assisting users within the Agrease application, likely by providing information and recommendations based on data from various marketplaces.

Model Description Base Model: The model is a fine-tuned version of a Mistral v3 Large Language Model.

Fine-tuning Task: The model was fine-tuned for conversational question-answering and recommendations.

Training Data: The training data was collected by scraping various online marketplaces using Python libraries such as BeautifulSoup and Scrapy.

Performance: The fine-tuning process achieved a final training loss of 11%.

Intended Use This model is intended to be used as a chatbot or a conversational agent within a larger application. It can answer user queries, provide product recommendations, and engage in domain-specific conversations related to the "Agrease" application's scope.

How to Use You can use this model with the transformers library for text generation.

from transformers import pipeline

Load the text generation pipeline from the Hugging Face Hub

Replace "your-username/model-name" with the actual model path

generator = pipeline('text-generation', model='your-username/model-name')

Example prompt

prompt = "What are the best fertilizers for rice paddies in a tropical climate?"

Generate a response

response = generator(prompt, max_length=150, num_return_sequences=1)

print(response[0]['generated_text'])

Training Data The dataset used for fine-tuning was created by scraping publicly available data from various e-commerce and marketplace websites. The scraping was performed using custom Python scripts with BeautifulSoup and Scrapy. The collected data was then processed and formatted into a conversational format suitable for training a large language model.

Training Procedure The fine-tuning was performed using the PyTorch framework on the collected dataset. The training focused on minimizing the cross-entropy loss to improve the model's ability to generate relevant and coherent responses in a conversational context. The final model achieved a training loss of 0.11

Downloads last month
45
GGUF
Model size
7.25B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support