This repository contains the code and resources for building an E-Commerce FAQ Chatbot using Parameter Efficient Fine Tuning with LoRA Technique. The project aims to develop a chatbot for an E-Commerce site by leveraging Large Language Models (LLMs) and adopting a fine-tuning approach using the Falcon-7B model. The fine-tuning is performed with Parameter Efficient Fine Tuning (PEFT) and the LoRA (Low-Rank Adaptation) Technique to enhance the model's performance.
In the fast-paced world of e-commerce, handling customer queries efficiently is crucial. This project introduces a chatbot solution leveraging advanced language models to automate responses to frequently asked questions. The fine-tuned model, Falcon-7B, is trained on a custom dataset extracted from Kaggle, addressing common user queries in the e-commerce domain.
The project builds upon pre-trained models, including OpenAI's GPT models and META's LLAMA models. It also explores existing chatbots like IBM Watson Assistant and Ada Healthcare Chatbot. The comparison between RAG (Retrieval Augmented Generation) and fine-tuned models is discussed.
The dataset, sourced from Kaggle, comprises 79 samples with questions and corresponding answers. The split includes 67 samples for training (85%) and 12 samples for testing (15%).
Link to Dataset on Kaggle
The methodology involves using the FALCON-7B model, fine-tuning with PEFT and LoRA Technique, and leveraging key libraries such as Transformers and Peft by Hugging Face. The process includes dataset preprocessing, LoRA adapters configuration, and model training.
The model demonstrates decreasing loss values across epochs, indicating positive training trends. The Bleu score, used for evaluation, showcases the model's proficiency in generating responses aligned with expected results.
The project contributes to enhancing customer support in e-commerce through advanced language models. While the achieved results are promising, further experiments with larger datasets and continuous model refinement are recommended.
Acknowledgments are extended to the developers of the FALCON-7B model, the Hugging Face community, Kaggle for hosting the dataset, and the faculty at the University of New Haven.
A Streamlit application has been developed for local use, requiring a GPU with at least 16 gigabytes of video RAM (vRAM) for optimal performance. the app in this repository. checkout in streamlit-app dir
Feel free to explore, experiment, and contribute to further improvements.