DistilBERT NLU Intent Classification

Fine-tuned DistilBERT model for intent classification in Natural Language Understanding (NLU) systems.

Model Details

  • Base Model: distilbert-base-uncased
  • Task: Intent Classification (Sequence Classification)
  • Number of Labels: 8
  • Framework: PyTorch + Transformers

Supported Intents

ID Intent Description
0 BILLING_ISSUE Problems with bills or charges
1 CANCEL_SUBSCRIPTION Cancel service requests
2 CHECK_BALANCE Balance inquiry
3 GOODBYE Farewell messages
4 GREETING Hello/welcome messages
5 MODIFY_PROFILE Update account details
6 ROAMING_ACTIVATION Enable roaming
7 ROAMING_DEACTIVATION Disable roaming

Usage

from transformers import pipeline

classifier = pipeline("text-classification", model="sidde/distilbert-nlu-intent-classification")

# Single prediction
result = classifier("I want to check my balance")
print(result)
# [{"label": "CHECK_BALANCE", "score": 0.98}]

Training Details

  • Dataset: 772 examples (custom intent dataset)
  • Train/Eval Split: 80/20 with stratification
  • Epochs: 10
  • Batch Size: 16
  • Learning Rate: 2e-5
  • Hardware: NVIDIA L4 GPU on OpenShift AI

Deployment

This model is deployed on OpenShift AI using KServe.

License

Apache 2.0

Downloads last month
4
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support