News: The LongEval workshop at CLEF 2023 is on Tuesday, 19 September, afternoon. All times are Thessaloniki times.
|14:00 - 15:30||LongEval Classifiaction task: Overview and Participants Presentations|
|15:30 - 16:00||Coffee Break|
|16:00 - 17:30||LongEval Retrieval Task: Overview and Participant Presentatons|
News: Train data for both tasks were released. Please see the Data section for details.
News: Test data for the retrieval task was released.
News: Test data for the classification task was released.
News: Relevance judgements for the retrieval task were released.
News: Classification golden labels for the classification task were released.
In this page we present the first CLEF 2023 shared task evaluating the temporal persistence of information retrieval (IR) systems and text classifiers. The task is motivated by recent research showing that the performance of these models drops as the test data becomes more distant in time from the training data. LongEval differs from traditional IR and classification shared task with special considerations on evaluating models that mitigate performance drop over time. We envisage that this task will bring more attention from the NLP community to the problem of temporal generalisability of models, what enables or prevents it, potential solutions and limitations.
The CLEF 2023 LongEval Lab encourages participants to develop temporal information retrieval systems and longitudinal text classifiers that survive through dynamic temporal text changes, introducing time as a new dimension for ranking models performance.
For Task 1. LongEval-Retrieval: firstname.lastname@example.org
For Task 2. LongEval-Classification: Rabab Alkhalifa
Join our slack channel for any question.