Intimacy is a fundamental social aspect of language. This SemEval shared task focuses on predicting the intimacy of tweets in 10+ languages. This task is co-organized by University of Michigan and Snap Inc.
[ALERT] You might see offensive or sexual content in the dataset.
Please email pedropei@umich.edu if you need the test set!
Full ranking and detailed scores are available in this google sheet. We encourage you to check out your detailed ranking on each language as the overall ranking may not fully reflect your effort on certain languages. You may check our baseline performances in this task paper and please fill in this very short participation survey to access the test data and evaluation script.
Final rankings:
Top 10 teams on predicting overall intimacy score
lazybob, Ohio State University
UZH_CLyp, University of Zurich
opi, OPI
tmn, University of Tyumen
OPD, Interactive Entertainment Group of Netease Inc
lottery, Individual Participant
DUTH, Democritus University of Thrace
Zhegu, University of Electronic Science and Technology of China
arizonans, University of Arizona, University of Arizona, University of Arizona
irel, International Institute of Information Technology, Hyderabad
Top 10 teams on predicting languages in the training data
king001, Ping An Insurance (Group) Company
lazybob, Ohio State University
opi, OPI
cyclejs, sunwoda
OPD, Interactive Entertainment Group of Netease Inc
lottery, Individual Participant
ODA_SRIB, Samsung Research and Development Institute Bangalore
UZH_CLyp, University of Zurich
tmn, University of Tyumen
water, Individual Participant
Top 10 teams on predicting languages not in the training data
lazybob, Ohio State University
DUTH, Democritus University of Thrace
irel, International Institute of Information Technology, Hyderabad
kean_nlp, Wen Zhou Kean University
opi, OPI
UZH_CLyp, University of Zurich
Zhegu, University of Electronic Science and Technology of China
arizonans, University of Arizona, University of Arizona, University of Arizona
tmn, University of Tyumen
ZBL2W, Dalian University of Technology
Task description
The goal of this task is to predict the intimacy of tweets in 10 languages. You are given a set of tweets in six languages (English, Spanish, Italian, Portuguese, French, and Chinese) annotated with intimacy scores ranging from 1-5 to train your model.
You are encouraged (but not required) to also use the question intimacy dataset (Pei and Jurgens, 2020) which contains 2247 English questions from Reddit as well as another 150 questions from Books, Movies, and Twitter. Please note that the intimacy scores in this dataset range from -1 to 1 so you might need to consider data augmentation methods or other methods mapping the intimacy scores to the 1-5 range in the current task. Please check out the paper for more details about this question intimacy dataset.
The model performance will be evaluated on the test set in the given 6 languages as well as an external test set with 4 languages not in the training data (Hindi, Arabic, Dutch and Korean).
We will use Pearson's r as the evaluation metric.
References
[1] Jiaxin Pei and David Jurgens. 2020. Quantifying Intimacy in Language. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 5307–5326, Online. Association for Computational Linguistics.
The goal is to predict the intimacy of tweets with a range from 1 (not intimate at all) to 5 (very intimate). The performance of the model will be evaluated based on the overall Pearson's r over the test set.
There are 2 phases:
This sample competition allows you to submit either:
Submissions must be made before the end of phase 1. You may submit 5 submissions every day and 100 in total.
Start: Sept. 1, 2022, 11 p.m.
Description: Development phase: create models
Start: Jan. 9, 2023, 11 p.m.
Description: The test set will be released and you can submit your predictions in the system
Jan. 31, 2023, 11:59 p.m.
You must be logged in to participate in competitions.
Sign In