| | --- |
| | license: mit |
| | datasets: |
| | - SemEvalWorkshop/sem_eval_2018_task_1 |
| | language: |
| | - en |
| | - ar |
| | base_model: |
| | - FacebookAI/xlm-roberta-base |
| | pipeline_tag: text-classification |
| | --- |
| | π XLM-R Multi-Emotion Classifier π |
| |
|
| | π Mission Statement |
| |
|
| | The XLM-R Multi-Emotion Classifier is built to understand human emotions across multiple languages, helping researchers, developers, and businesses analyze sentiment in text at scale. |
| |
|
| | From social media monitoring to mental health insights, this model is designed to decode emotions with accuracy and fairness. |
| |
|
| | π― Vision |
| |
|
| | Our goal is to create an AI-powered emotion recognition model that: |
| | β’ π Understands emotions across cultures and languages |
| | β’ π€ Bridges the gap between AI and human psychology |
| | β’ π‘ Empowers businesses, researchers, and developers to extract valuable insights from text |
| | |
| | π Model Overview |
| |
|
| | Model Name: msgfrom96/xlm_emo_multi |
| | Architecture: XLM-RoBERTa (Multi-Lingual Transformer) |
| | Task: Multi-label Emotion Classification |
| | Languages: English, Arabic |
| | Dataset: SemEval-2018 Task 1: Affect in Tweets |
| |
|
| | The model predicts multiple emotions per text using multi-label classification. It can recognize emotions like: |
| | β’ π Anger, Anticipation, Disgust, Fear, Joy, Sadness, Surprise, Trust, Love, Optimism, Pessimism |
| | |
| | π¦ How to Use |
| |
|
| | Load Model and Tokenizer |
| |
|
| | from transformers import AutoModelForSequenceClassification, AutoTokenizer |
| |
|
| | model_name = "msgfrom96/xlm_emo_multi" |
| | |
| | # Load model and tokenizer |
| | model = AutoModelForSequenceClassification.from_pretrained(model_name) |
| | tokenizer = AutoTokenizer.from_pretrained(model_name) |
| | |
| | # Example text |
| | text = "I can't believe how amazing this is! So happy and excited!" |
| | |
| | # Tokenize input |
| | inputs = tokenizer(text, return_tensors="pt", padding=True, truncation=True) |
| |
|
| | # Get model predictions |
| | outputs = model(**inputs) |
| | print(outputs.logits) # Raw emotion scores |
| | |
| | Interpreting Results |
| | |
| | The model outputs logits (raw scores) for each emotion. Apply a sigmoid activation to convert these into probabilities: |
| | |
| | import torch |
| | |
| | probs = torch.sigmoid(outputs.logits) |
| | print(probs) |
| | |
| | Each score represents the probability of an emotion being present in the text. |
| | |
| | β‘ Training & Fine-Tuning Details |
| | β’ Base Model: XLM-RoBERTa (xlm-roberta-base) π |
| | β’ Dataset: SemEval-2018 (English & Arabic Tweets) π |
| | β’ Training Strategy: Multi-label classification π₯ |
| | β’ Optimizer: AdamW βοΈ |
| | β’ Batch Size: 16 ποΈββοΈ |
| | β’ Learning Rate: 2e-5 π― |
| | β’ Hardware: Trained on AWS SageMaker with CUDA GPU support π |
| | β’ Evaluation Metric: Macro-F1 & Micro-F1 π |
| | β’ Best Model Selection: Auto-selected via load_best_model_at_end=True β
|
| | |
| | π Citations & References |
| | |
| | If you use this model, please cite the following sources: |
| | |
| | π SemEval-2018 Dataset |
| | Mohammad, S., Bravo-Marquez, F., Salameh, M., & Kiritchenko, S. (2018). βSemEval-2018 Task 1: Affect in Tweets.β Proceedings of SemEval-2018. |
| | π Paper Link |
| | |
| | π XLM-RoBERTa |
| | Conneau, A., Khandelwal, K., Goyal, N., Chaudhary, V., Wenzek, G., GuzmΓ‘n, F., Grave, E., Ott, M., Zettlemoyer, L., & Stoyanov, V. (2020). βUnsupervised Cross-lingual Representation Learning at Scale.β Proceedings of ACL 2020. |
| | π Paper Link |
| | |
| | π Transformers Library |
| | Hugging Face (2020). βπ€ Transformers: State-of-the-art Natural Language Processing for Pytorch and TensorFlow 2.0.β |
| | π Library Docs |
| | |
| | π€ Contributing |
| | |
| | Want to improve the model? Feel free to: |
| | β’ Train it on more languages π |
| | β’ Optimize for low-resource devices π₯ |
| | β’ Integrate it into real-world applications π‘ |
| | β’ Submit pull requests or discussions π |
| | |
| | π Acknowledgments |
| | |
| | Special thanks to the Hugging Face team, SemEval organizers, and the NLP research community for providing the tools and datasets that made this model possible. π |
| | |
| | π Connect & Feedback |
| | |
| | π¬ Questions? Issues? Create a discussion on the Hugging Face Model Hub |
| | π§ Email: gleiser2@hotmail.com |
| | |
| | --- |
| | license: mit |
| | --- |