--- language: en license: mit model_name: gemma-3b-chat-support tags: - chat-support - customer-service - gemma-3b - fine-tuned - schema-guided-dialogue base_model: - google/gemma-3-1b-it metrics: - bleu ---

Gemma 3B Chat Support Assistant

Image

## Model Details This model is a fine-tuned version of [google/gemma-3-1b-it](https://huggingface.co/google/gemma-3-1b-it). It has been trained using [TRL](https://github.com/huggingface/trl). Specifically optimized for customer support conversations across multiple domains. It leverages the Schema-Guided Dialogue (SGD) dataset to provide helpful, contextual responses. ## Capabilities This model can help with: * **Conversational Support** - Handles multi-turn dialogues with context retention across various customer inquiries * **Information Retrieval** - Provides relevant information based on user requests across multiple domains * **Context Management** - Maintains conversation history to provide coherent and contextually appropriate responses * **Multi-domain Assistance** - Flexibly switches between different domains and topics as the conversation evolves * **User Engagement** - Creates dynamic, personalized responses that adapt to the changing context of a conversation ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="iprajwaal/gemma-3b-chat-support", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Advanced Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline model_id = "iprajwaal/gemma-3b-chat-support" model = AutoModelForCausalLM.from_pretrained(model_id) tokenizer = AutoTokenizer.from_pretrained(model_id) pipe = pipeline("text-generation", model=model, tokenizer=tokenizer) messages = [ {"role": "system", "content": "You are a helpful customer support assistant."}, {"role": "user", "content": "I need to book a hotel in New York for next weekend."} ] response = pipe(messages, max_new_tokens=100) print(response[0]["generated_text"]) ``` ## Model Details * **Base Model**: google/gemma-3-1b-it * **Fine-tuning Method**: LoRA with 4 epochs * **Architecture**: Gemma 3B (instruction-tuned variant) * **Language**: English * **License**: MIT ## Training Data This model was fine-tuned on a carefully filtered subset of the [Schema-Guided Dialogue dataset](https://github.com/google-research-datasets/dstc8-schema-guided-dialogue) (SGD). The SGD dataset consists of over 20,000 annotated multi-domain, task-oriented conversations between users and virtual assistants spanning 20 domains, including banking, events, media, calendar, travel, weather, and accommodation. For fine-tuning purposes, we extracted only the user messages and assistant responses, filtering out the complex annotations while preserving the natural conversational flow. This approach allowed us to create a clean dataset focused on high-quality customer support interactions. ## Intended Uses This model is designed for: * Customer support automation * Virtual assistants * Chat interfaces * Conversational AI applications * Multi-domain dialogue systems ## Limitations * May not have knowledge of events after its training data cutoff (May 2025) * Performance may vary on domains not well-represented in the training data * The model may occasionally generate incorrect information when uncertain * As with all language models, outputs should be reviewed for accuracy in critical applications ### Framework versions * TRL: 0.15.2 * Transformers: 4.51.3 * Pytorch: 2.6.0+cu124 * Datasets: [Schema-Guided Dialogue dataset](https://github.com/google-research-datasets/dstc8-schema-guided-dialogue) (SGD) * Tokenizers: 0.21.1 ## Citations Cite TRL as: ```bibtex @misc{iprajwaal, title = {{Gemma-3b-chat-support: A Fine-tuned Customer Support Assistant}}, author = {Prajwal Kumbar}, year = 2025, note = {Fine-tuned using the Schema-Guided Dialogue dataset and the TRL library}, publisher = {Hugging Face}, howpublished = {\url{https://huggingface.co/iprajwaal/gemma-3b-chat-support}} } ```