Giter VIP home page Giter VIP logo

semeval2025-task11's Introduction

Google Group | Create an Issue | Join Discord | Contact Us

Bridging the Gap in Text-Based Emotion Detection

Emotions are simultaneously familiar and mysterious. On the one hand, we all express and manage our emotions every day. Yet, on the other hand, emotions are complex, nuanced, and sometimes hard to articulate. We also use language in subtle and complex ways to express emotion (Wiebe et al. 2005, Mohammad and Kiritcheko 2018, Mohammad et al. 2018). Further, people are highly variable in how they perceive and express emotions (even within the same culture or social group). Thus, we can never truly identify how one is feeling based on something that they have said with absolute certainty.

Emotion recognition is not one task but an umbrella term for several tasks such as detecting the emotions of the speaker, identifying what emotion a piece of text is conveying and detecting emotions evoked in a reader (Mohammad 2021, Mohammad 2023).

This task is on perceived emotion and focuses on:

  • Determining what emotion most people will think the speaker may be feeling given a sentence or short text snippet uttered by the speaker.

The task is not about:

  • The emotion evoked in the reader.
  • The emotion of someone else mentioned in the text.
  • Or even the true emotion of the speaker (which cannot be definitively known from just a short text snippet).

We acknowledge the importance of this distinction as perceived emotions can differ from actual emotions due to various factors such as cultural context, individual differences in emotional expression, and the limitations of text-based communication (Van Woensel and Nevil 2019, Wakerfield 2021).

Languages

We include a large number of languages with many predominantly spoken in regions characterised by a relatively limited availability of NLP resources (e.g., Africa, Asia, Eastern Europe and Latin America):

Tracks

Participants can participate in one or more of the following tracks:

Track A: Multi-label Emotion Detection

Given a target text snippet, predict the perceived emotion(s) of the speaker. Specifically, select whether each of the following emotions apply: Joy, Sadness, Fear, Anger, Surprise, or Disgust. In other words, label the text snippet with: joy (1) or no joy (0), sadness (1) or no sadness (0), anger (1) or no anger (0), surprise (1) or no surprise (0), and disgust (1) or no disgust (0).

A training dataset with gold emotion labels will be provided for this track.

Track B: Emotion Intensity

Given a target text and a target emotion, predict the intensity for each of the target emotion classes.

The set of target emotions includes: Joy, Sadness, Fear, Anger, Surprise, or Disgust.

The set of ordinal classes of intensity includes: 0 for no emotion 1 for a low degree of emotion 2 for a moderate degree of emotion 3 for a high degree of emotion.

Track C: Cross-lingual Emotion Detection

Given a labeled training set in one of the languages given above, predict the emotion labels of a new text instance in a different target language using the set of six emotion classes:Joy, Sadness, Fear, Anger, Surprise, or Disgust.

Dataset and Download Links

For each track, we provide the sample, training, and evaluation datasets. Find the links to download the datasets for each track below:

Track Download Link File ID
Track A Track A Dataset 1Pvptx6XDsfLcR0IGvGUV4ZDD1qezyzUo
Track B Track B Dataset 1OCzDN5RuWdos47P3TvIzIqVcspf1_4yZ
Track C Track C Dataset ``
All Tracks Complete Dataset 1vYggpyd0O0FNL99OvHHUp_wsCKVOcZWn

Download the dataset using gdown:

  1. Install gdown if you haven't already:

    pip install gdown

  2. Use the following commands to download the datasets using provided IDs:

    gdown --folder https://drive.google.com/drive/folders/<file_id>

The table below shows the languages of the different datasets, their sizes and the release status of their pilot samples, training, and development (dev) sets. Note that ✓ means released and can be found in the data folder.

No. Language Code Pilot Data Training Dev Size
1 Afrikaans AFR
2 Algerian Arabic ARQ
3 Amharic AMH
4 Arabic ARB
5 Brazilian Portuguese PTB
6 Chinese ZHO
7 English ENG
8 German DEU
9 Hausa HAU
10 Hindi HIN
11 Igbo IBO
12 Indonesian IND
13 isiXhosa XHO
14 isiZulu ZUL
15 Javanese JAV
16 Kinyarwanda KIN
17 Marathi MAR
18 Moroccan Arabic ARY
19 Mozambican Portuguese PTM
20 Nigerian-Pidgin PCM
21 Oromo ORM
22 Romanian RON
23 Russian RUS
24 Setswana TSN
25 Somali SOM
26 Spanish SPA
27 Swahili SWA
28 Swedish SWE
29 Tatar TAT
30 Tigrinya TIR
31 Ukrainian UKR
32 Xitsonga TSO
33 Yoruba YOR

Evaluation

The performance of each submission will be evaluated using F1-score based on the predicted labels and the gold ones. Participants will be provided with an evaluation script and a starter kit that includes a simple baseline.

Important Dates and Phases of the Task

Description Deadline
Sample Data Ready 15 July 2024
Training Data Ready 02 September 2024
Evaluation Start 10 January 2025
Evaluation End 31 January 2025
System Description Paper Due 28 February 2025
Notification to authors 31 March 2025
Camera ready due 21 April 2025
SemEval workshop 2025 (co-located with a major NLP conference)

The task will be divided into three phases: Development, Evaluation, and Post-Evaluation. The following summarize the phases and their timelines.

Development Phase: Codalab submission link coming soon
  • This phase runs from 02 September to 10 January 2024.
  • Train (with gold labels) and dev data (without gold labels) will be released for this phase.
  • Train and evaluate your model on the dev set via CodaLab.
  • Up to 999 submissions are allowed, and the leaderboard is open for you to view your results and those of others.
Evaluation Phase: Codalab submission link coming soon
  • This phase runs from around 10 January to 31 January 2024 (tentative).
  • Test data will be released (without gold labels).
  • Participants will have the opportunity to evaluate their models on the test data.
  • Each team is allowed only one submission. This single submission will be considered your official entry for the competition.
  • The leaderboard is disabled and will only be published after the submission deadline.
Post-Evaluation Phase: Codalab submission link coming soon
  • Starts around 31 January 2024 and never ends.
  • In this phase, you can still submit and test your system even after the official competition ends. This way, you can keep improving your work.
  • We will make the leaderboard public again so you can see how you are doing compared to others.
  • You can use CodaLab to talk with other participants, share ideas, and learn how to make your system better.

How to Participate

  1. Register: Sign up on the Codalab competition platform (link to be provided).
  2. Track: Decide on the track(s) you want to participate in (Track A, B and/or C)
  3. Download: Access to the datasets for each track will be provided in this repository.
  4. Develop: Build your models using the provided data.
  5. Submit: Submit your predictions on the Codalab competition platform (link to be provided).

Competition Rules and Terms

1. Consent to Public Release of Scores
  • By submitting results, you consent to the public release of your scores on:
    • the competition website,
    • at the designated workshop,
    • in associated proceedings.
  • Task organizers have discretion over the release and choice of metrics.
  • Scores may include:
    • automatic and manual quantitative judgments,
    • qualitative judgments,
    • other metrics as deemed appropriate.
2. Score Release and Validity
  • Task organizers reserve the right to withhold scores for:
    • incomplete submissions,
    • erroneous submissions,
    • deceptive submissions,
    • rule-violating submissions.
  • Inclusion of a submission's scores does not constitute endorsement.
3. Team Participation Rules
  • Participants may be involved in only one team.
  • Exceptions may be granted with prior approval from organizers.
4. Account Management
  • Each team must create and use exactly one account on the designated platform.
5. Team Constitution
  • Team membership cannot be changed after the evaluation period begins.
6. Development Period Rules
  • Teams can submit up to 999 submissions.
  • Results are visible only to the submitting team.
  • Leaderboard is disabled.
  • Warnings and errors are visible for each submission.
7. Evaluation Period Rules
  • Teams are limited to 3 submissions.
  • Only the final submission is considered official.
  • Warnings and errors are visible for each submission.
8. Post-Competition
  • Gold labels will be released after the competition.
  • Teams are encouraged to report results on all system variants in their description paper.
  • Official submission results must be clearly indicated.
9. Public Release of Submissions
  • Final team submissions may be made public after the evaluation period.
10. Disclaimer about the Datasets
  • Organizers and affiliated institutions provide no warranties on dataset correctness or completeness.
  • They are not liable for dataset access or usage.
11. Peer Review Process
  • Each participant will review another team's system description paper.
12. Dataset Usage Restrictions
  • Datasets should only be used for scientific or research purposes.
  • Any other use is explicitly prohibited.
  • Datasets must not be redistributed or shared with third parties.
  • Interested parties should be directed to the official website.
13. Final ranking
  • To be included in the official task ranking, you **MUST** submit a system description paper.

Dataset paper

We will soon release a dataset paper that describes the data collection, annotation process, and baseline experiments. This paper will provide additional details and information that will be useful for the task participants.

Communication

  • Join our Discord Channel to ask questions and receive updates (coming soon).
  • If you have any questions or issues, please feel free to create an issue.
  • Contact organizers at: emotion-semeval-2025-organisers[at]googlegroups[dot]com

FAQs

Do I have to participate in all languages for a given track?
  • No, you can participate in one or more languages.
How will you verify my submitted model?
  • To be included in the final team rankings of our shared task, it is mandatory for participants to submit a system description paper describing their approaches and methodologies in detail, therefore ensuring scientific integrity.
When will you release the gold labels?
  • For the dev set, the gold labels will be released when the evaluation phase starts and the gold labels for the different test sets will be released after the competition is over.
Can I use LLMs in the different tracks?
  • Yes.
Will I be included in the final ranking if I do not write a system description paper?
  • No. You MUST write a system description paper to be included in the final ranking.
I have never written a system description paper. How can I write one?
  • We will have an online writing tutorial and share resources to help you write a system description paper.
Do I need to pay conference registration fees and/or attend SemEval for my paper to be published?
  • It is not required to attend the SemEval workshop for the paper to be published. You do not have to pay any registration fees if you do not attend the workshop. However, if you want to attend the workshop, you need to pay for attendance.
## Our system did not perform very well, should I still write a system description paper?
  • We want to hear from **all** of you even if you did not outperform other systems!Write about the details of your system. (**Yes we want your insights from any negative results!**)

Resources

  1. SemEval 2025 Shared Tasks
  2. Frequently Asked Questions about SemEval
  3. Paper Submission Requirements
  4. Guidelines for Writing Papers
  5. Paper style files
  6. Previous shared-tasks on emotion detection (to be added)
  7. Resources for Beginners (to be added)
  8. Paper submission link (to be added)

References

Janyce Wiebe, Theresa Wilson, and Claire Cardie. "Annotating expressions of opinions and emotions in language." Language resources and evaluation 39 (2005): 165-210.

Saif M. Mohammad,, and Svetlana Kiritchenko. "Understanding Emotions: A Dataset of Tweets to Study Interactions between Affect Categories." Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). 2018.

Saif M. Mohammad, Felipe Bravo-Marquez, Mohammad Salameh, and Svetlana Kiritchenko. In Proceedings of the International Workshop on Semantic Evaluation (SemEval-2018), New Orleans, LA, USA, June 2018.

Lieve Van Woensel and Nissy Nevil. 2019. What if your emotions were tracked to spy on you? European Parliamentary Research Service, PE 634.415. https://www.europarl.europa.eu/RegData/etudes/ATAG/2019/634415/EPRS_ATA(2019)634415_EN.pdf.

Jane Wakefield. 2021. AI emotion-detection software tested on Uyghurs. BBC. https://www.bbc.com/news/technology-57101248.

Saif M. Mohammad "Ethics sheet for automatic emotion recognition and sentiment analysis." Computational Linguistics 48.2 (2022): 239-278.

Saif M. Mohammad "Best Practices in the Creation and Use of Emotion Lexicons." Findings of the Association for Computational Linguistics: EACL 2023. 2023.

Organizers

Shamsuddeen Hassan Muhammad, Seid Muhie Yimam , Nedjma Ousidhoum, Idris Abdulmumin, Ibrahim Said Ahmad, Alham Fikri Aji, Vladimir Araujo, Abinew Ali Ayele, Tadesse Destaw Belay, David Ifeoluwa Adelani, Daniela Teodorescu, Jan Philip Wahle, Terry Ruas, Nirmal Surange, Yi Zhou, Alexander Panchenko, Vukosi Marivate, Saif M. Mohammad

semeval2025-task11's People

Contributors

shmuhammadd avatar seyyaw avatar nedjmaou avatar vgaraujov avatar jodiechou avatar

Stargazers

Jan Philip Wahle avatar  avatar JianxiangTian avatar Taha avatar Lukman Aliyu Jibril avatar Devendra Kayande avatar  avatar Abdullahi Yunus avatar

Forkers

jodiechou

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.