LongEval 2023 Classification

Organized by CLEF-LongEval-Classification - Current server time: March 30, 2025, 6:53 a.m. UTC

Previous

Evaluation
May 4, 2023, midnight UTC

Current

Post-Evaluation
May 23, 2023, midnight UTC

End

Competition Ends
Never

Welcome to LongEval Classification task:

In this first edition of the LongEval Lab, we look at the temporal persistence of the systems’ performance. In order to include the feature of temporal persistence as an additional quality for models proposed, participants are asked to suggest temporal IR systems (Task 1) and longitudinal text classifiers (Task 2) that generalize well beyond a train set generated within a limited time frame.

We consider two types of temporal persistence tasks: temporal information retrieval and longitudinal text classification. For each task, we look at a short-term and a long-term performance persistence. We aim to answer a high level question:

Given a longitudinal evolving benchmark for a typical NLP task, what types of models offer better temporal persistence over a short term and a long term?

Objective and scope:

Evaluation Criteria

The submissions will be ranked by *macro F1-score* separately for within, short and long term datasets, though winner systems are the one ranked top using averaged results over all testing sets. The metrics will be computed as follows:

from sklearn.metrics import f1_score
f1_score(y_true, y_pred, average='macro')

image.png

def calculate_RPD (temporal_fscore, within_fscore):
    RPD = (temporal_fscore -  within_fscore) / within_fscore
    return RPD

This it to quantify drop by comparing [within with short] and [within with long] to see if performance of 'within time' is a good quantifier for model's robustness evaluation.

  • Average macro f1-score for all three submissions [within, short and long]

Terms and Conditions

By participating in this task you agree to the following terms and conditions.

(In very specific instances, we might make some exceptions. Please contact the task organisers if you have any further queries.)

  1. By submitting results to this competition, you explicitly consent to the public release of your scores on this website and at LongEval 2023 and in the associated proceedings, at the task organizers' discretion. 
  2. Scores may include, but are not limited to, automatic and manual quantitative judgements, qualitative judgements, and such other metrics as the task organizers see fit. You accept that the ultimate decision of metric choice and score value is that of the task organizers.
  3. You further agree that the task organizers are under no obligation to release scores and that scores may be withheld if it is the task organizers' judgement that the submission was incomplete, erroneous, deceptive, or violated the letter or spirit of the competition's rules. 
  4. Inclusion of a submission's scores is not an endorsement of a team or individual's submission, system, or science.
  5. Each participant can be part of exactly one team.
  6. Each team must create and use exactly one CodaLab account.
  7. The members of a team cannot be changed after the start of the evaluation period. 
  8. During the evaluation period, each team can submit 5 submissions. However, only the final submission will be considered as the official submission to the competition. Also, you will not be able to see the results of your submission on the test set. Until after the evaluation period is complete.
  9. The final submissions of each team will be made public at the end of the competition. 
  10. The organisers and their affiliated institutions make no warranties regarding the dataset provided for this task  including but not limited to correctness or completeness. They cannot be held liable for providing access to the datasets or the usage of the datasets.
  11. The dataset should only be used for scientific or research purposes. Any other use is explicitly prohibited.
  12. The dataset must not be redistributed or shared in part or full with any third party. Redirect interested parties to this website.

Practice

Start: April 25, 2023, midnight

Evaluation

Start: May 4, 2023, midnight

Post-Evaluation

Start: May 23, 2023, midnight

Competition Ends

Never

You must be logged in to participate in competitions.

Sign In