xlm-emotion

This model is a fine-tuned version of xlm-roberta-base for Emotion Classification (28 labels). It is specifically trained to recognize emotional nuances in Midwest Emo and Math Rock lyrical styles.

To ensure high reliability and prevent the model from being biased toward majority classes (like Grief), the training was conducted on a balanced version of the anggars/mbti-emotion dataset using undersampling techniques.

Model description

  • Model Type: XLM-RoBERTa Base
  • Labels: 28 Emotion Categories
  • Dataset: Balanced anggars/mbti-emotion (max 500 samples per class combo)
  • Language: English & Indonesian

Training results

The following results were achieved on the evaluation set during the 3-epoch training process:

Training Loss Epoch Step Validation Loss Accuracy
0.5297 1.0 5618 0.4876 0.8480
0.3340 2.0 11236 0.4059 0.8732
0.2116 3.0 16854 0.3954 0.8876

Intended uses & limitations

This model is designed for sentiment and emotion analysis in creative writing, specifically lyrics. Limitations: Since it's trained on a specific subgenre context, it may perform differently on general conversational text or technical documents.

Training procedure

Training hyperparameters

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: AdamW with betas=(0.9,0.999) and epsilon=1e-08
  • num_epochs: 3

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.5.1+cu124
  • Datasets 3.1.0
  • Tokenizers 0.20.3
Downloads last month
107
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 1 Ask for provider support

Model tree for anggars/xlm-emotion

Finetuned
(3856)
this model

Dataset used to train anggars/xlm-emotion

Spaces using anggars/xlm-emotion 2

Evaluation results