Instructions to use Shobhank-iiitdwd/NLP-Medical-Intent-Detector with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use Shobhank-iiitdwd/NLP-Medical-Intent-Detector with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-classification", model="Shobhank-iiitdwd/NLP-Medical-Intent-Detector")# Load model directly from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("Shobhank-iiitdwd/NLP-Medical-Intent-Detector") model = AutoModelForSequenceClassification.from_pretrained("Shobhank-iiitdwd/NLP-Medical-Intent-Detector") - Notebooks
- Google Colab
- Kaggle
Model Card for Model ID
Model Details
Model Description
This model is designed for classifying medical intents from text using a specialized variant of ClinicalBERT. It was trained to understand and classify various medical intents based on patient dialogues and medical records. The model aims to enhance the understanding of medical phrases and intents in clinical settings.
Pretraining Data
The Clinical Intent Model was trained on a diverse dataset consisting of:
Training Data: A large multicenter dataset including a variety of medical dialogues and intents.
Model Pretraining
Pretraining Procedures Base Model: The model is initialized from the ClinicalBERT base model. Training Objective: The model is fine-tuned on the classification task of predicting medical intents. During fine-tuning, the model was exposed to a labeled dataset where phrases were categorized into predefined intents.
Pretraining Hyperparameters
- Batch Size: 32
- Maximum Sequence Length: 256
- Learning Rate: 2e-5
- Epochs: 20
- Optimizer: AdamW
- Scheduler: Linear scheduler with warm-up
- Downloads last month
- 2