Agrease-Chatbot / README.md
cakra84's picture
Update README.md
198c9ac verified
license: apache-2.0
tags:
text-generation
conversational
mistral
fine-tuned
chatbot
bangkit
widget:
text: "Hello! I'm looking for recommendations on agricultural products."
Fine-Tuned Mistral Model for Agrease Application
This repository contains a fine-tuned version of a Mistral Large Language Model, specifically adapted for the "Agrease" application. The model was developed as part of a capstone project for Bangkit Academy 2024 Batch 2.
The primary goal of this project was to create a specialized conversational AI capable of assisting users within the Agrease application, likely by providing information and recommendations based on data from various marketplaces.
Model Description
Base Model: The model is a fine-tuned version of a Mistral v3 Large Language Model.
Fine-tuning Task: The model was fine-tuned for conversational question-answering and recommendations.
Training Data: The training data was collected by scraping various online marketplaces using Python libraries such as BeautifulSoup and Scrapy.
Performance: The fine-tuning process achieved a final training loss of 11%.
Intended Use
This model is intended to be used as a chatbot or a conversational agent within a larger application. It can answer user queries, provide product recommendations, and engage in domain-specific conversations related to the "Agrease" application's scope.
How to Use
You can use this model with the transformers library for text generation.
from transformers import pipeline
# Load the text generation pipeline from the Hugging Face Hub
# Replace "your-username/model-name" with the actual model path
generator = pipeline('text-generation', model='your-username/model-name')
# Example prompt
prompt = "What are the best fertilizers for rice paddies in a tropical climate?"
# Generate a response
response = generator(prompt, max_length=150, num_return_sequences=1)
print(response[0]['generated_text'])
Training Data
The dataset used for fine-tuning was created by scraping publicly available data from various e-commerce and marketplace websites. The scraping was performed using custom Python scripts with BeautifulSoup and Scrapy. The collected data was then processed and formatted into a conversational format suitable for training a large language model.
Training Procedure
The fine-tuning was performed using the PyTorch framework on the collected dataset. The training focused on minimizing the cross-entropy loss to improve the model's ability to generate relevant and coherent responses in a conversational context. The final model achieved a training loss of 0.11