Multilingual Tweet Intimacy Analysis

Organized by pedropei - Current server time: March 30, 2025, 2:58 a.m. UTC

First phase

Development
Sept. 1, 2022, 11 p.m. UTC

End

Competition Ends
Jan. 31, 2023, 11:59 p.m. UTC

Multilingual Tweet Intimacy Analysis

Intimacy is a fundamental social aspect of language. This SemEval shared task focuses on predicting the intimacy of tweets in 10+ languages. This task is co-organized by University of Michigan and Snap Inc.

[ALERT] You might see offensive or sexual content in the dataset.

Please email pedropei@umich.edu if you need the test set!

 

Full ranking and detailed scores are available in this google sheet. We encourage you to check out your detailed ranking on each language as the overall ranking may not fully reflect your effort on certain languages.  You may check our baseline performances in this task paper and please fill in this very short participation survey to access the test data and evaluation script.

 

 

Final rankings:

 Top 10 teams on predicting overall intimacy score

  1.    lazybob, Ohio State University

  2.    UZH_CLyp, University of Zurich

  3.    opi, OPI

  4.    tmn, University of Tyumen

  5.    OPD, Interactive Entertainment Group of Netease Inc

  6.    lottery, Individual Participant

  7.    DUTH, Democritus University of Thrace

  8.    Zhegu, University of Electronic Science and Technology of China

  9.    arizonans, University of Arizona, University of Arizona, University of Arizona

  10.    irel, International Institute of Information Technology, Hyderabad

 

Top 10 teams on predicting languages in the training data

  1.    king001, Ping An Insurance (Group) Company

  2.    lazybob, Ohio State University

  3.    opi, OPI

  4.    cyclejs, sunwoda

  5.    OPD, Interactive Entertainment Group of Netease Inc

  6.    lottery, Individual Participant

  7.    ODA_SRIB, Samsung Research and Development Institute Bangalore

  8.    UZH_CLyp, University of Zurich

  9.    tmn, University of Tyumen

  10.    water, Individual Participant

 

Top 10 teams on predicting languages not in the training data

  1.    lazybob, Ohio State University

  2.    DUTH, Democritus University of Thrace

  3.    irel, International Institute of Information Technology, Hyderabad

  4.    kean_nlp, Wen Zhou Kean University

  5.    opi, OPI

  6.    UZH_CLyp, University of Zurich

  7.    Zhegu, University of Electronic Science and Technology of China

  8.    arizonans, University of Arizona, University of Arizona, University of Arizona

  9.    tmn, University of Tyumen

  10.    ZBL2W, Dalian University of Technology

 

Task description

  • The goal of this task is to predict the intimacy of tweets in 10 languages. You are given a set of tweets in six languages (English, Spanish, Italian, Portuguese, French, and Chinese) annotated with intimacy scores ranging from 1-5 to train your model.

  • You are encouraged (but not required) to also use the question intimacy dataset (Pei and Jurgens, 2020) which contains 2247 English questions from Reddit as well as another 150 questions from Books, Movies, and Twitter. Please note that the intimacy scores in this dataset range from -1 to 1 so you might need to consider data augmentation methods or other methods mapping the intimacy scores to the 1-5 range in the current task. Please check out the paper for more details about this question intimacy dataset.

  • The model performance will be evaluated on the test set in the given 6 languages as well as an external test set with 4 languages not in the training data (Hindi, Arabic, Dutch and Korean).

  • We will use Pearson's r as the evaluation metric.

 

References

[1] Jiaxin Pei and David Jurgens. 2020. Quantifying Intimacy in Language. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 5307–5326, Online. Association for Computational Linguistics.

Evaluation

The goal is to predict the intimacy of tweets with a range from 1 (not intimate at all) to 5 (very intimate). The performance of the model will be evaluated based on the overall Pearson's r over the test set. 


There are 2 phases:

  • Phase 1: development phase. We provide you with labeled training data for both tasks.
  • Phase 2: final phase. You do not need to do anything. Your last submission of phase 1 will be automatically forwarded. Your performance on the test set will appear on the leaderboard when the organizers finish checking the submissions.

This sample competition allows you to submit either:

  • Only prediction results (no code).

Rules

Submissions must be made before the end of phase 1. You may submit 5 submissions every day and 100 in total.

Development

Start: Sept. 1, 2022, 11 p.m.

Description: Development phase: create models

Evaluation

Start: Jan. 9, 2023, 11 p.m.

Description: The test set will be released and you can submit your predictions in the system

Competition Ends

Jan. 31, 2023, 11:59 p.m.

You must be logged in to participate in competitions.

Sign In