You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

๐Ÿง  Qwen3-4B-Instruct Coaching Customer Support Model

Base Model: unsloth/Qwen3-4B-Instruct-2507 License: Apache-2.0
Libraries Used: unsloth, trl, transformers, torch, datasets, peft, accelerate, bitsandbytes


๐Ÿ“˜ Model Overview

This model is a fine-tuned version of Qwen3-4B-Instruct, specialized for educational support and institute FAQs.
It helps provide structured and factual answers about JEE and NEET coaching programs, batches, fees, and facilities.

Key Abilities:

  • Answers academic and course-related questions
  • Explains coaching programs and structures
  • Shares institute details (fees, modes, schedules)
  • Simulates realistic studentโ€“counselor chat conversations

๐Ÿงฉ Fine-tuning Details

  • Framework: PyTorch

  • Libraries:

    • ๐Ÿค— Transformers
    • PEFT (Parameter-Efficient Fine-Tuning)
    • Unsloth (for faster fine-tuning)
    • BitsAndBytes (4-bit/8-bit quantization)
    • Accelerate
    • Datasets
  • Training Objective: Instruction-following and educational support conversations

  • Input Format: Alpaca/ShareGPT style

  • Output Format: Chat-style assistant response


๐Ÿง  Intended Use

Use Case Description
๐Ÿ“š Educational Assistant Answers student queries about JEE/NEET batches and programs.
๐Ÿ’ฌ Chatbot Integration Can be integrated into coaching institute websites or portals.
๐Ÿงพ FAQ Automation Ideal for handling structured questionโ€“answer interactions..

โš™๏ธ Example Usage

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

model_name = "startelelogic/Qwen3-4B-Instruct-2507-Customer-Support"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")

prompt = "List all NEET programs offered by EduQuest Academy."
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=300)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

๐Ÿ“ Citation

@misc {qwen3technicalreport, title={Qwen3 Technical Report}, author={Qwen Team}, year={2025}, eprint={2505.09388}, archivePrefix={arXiv}, primaryClass={cs.CL}, url={https://arxiv.org/abs/2505.09388}, }

Downloads last month
-
Safetensors
Model size
4B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for startelelogic/Qwen3-4B-Instruct-2507-Customer-Support

Finetuned
(392)
this model

Paper for startelelogic/Qwen3-4B-Instruct-2507-Customer-Support