|
--- |
|
license: cc-by-nc-sa-4.0 |
|
task_categories: |
|
- text-classification |
|
language: |
|
- en |
|
- ar |
|
- bg |
|
- de |
|
- el |
|
- it |
|
- pl |
|
- ro |
|
- uk |
|
tags: |
|
- subjectivity-detection |
|
- news-articles |
|
viewer: true |
|
pretty_name: 'CLEF 2025 CheckThat! Lab - Task 1: Subjectivity in News Articles' |
|
size_categories: |
|
- 1K<n<10K |
|
configs: |
|
- config_name: arabic |
|
data_files: |
|
- split: train |
|
path: |
|
- "data/arabic/train_ar.tsv" |
|
- split: dev |
|
path: |
|
- "data/arabic/dev_ar.tsv" |
|
- split: dev_test |
|
path: |
|
- "data/arabic/dev_test_ar.tsv" |
|
- split: test |
|
path: |
|
- "data/arabic/test_ar_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: bulgarian |
|
data_files: |
|
- split: train |
|
path: |
|
- "data/bulgarian/train_bg.tsv" |
|
- split: dev |
|
path: |
|
- "data/bulgarian/dev_bg.tsv" |
|
- split: dev_test |
|
path: |
|
- "data/bulgarian/dev_test_bg.tsv" |
|
sep: "\t" |
|
- config_name: english |
|
data_files: |
|
- split: train |
|
path: |
|
- "data/english/train_en.tsv" |
|
- split: dev |
|
path: |
|
- "data/english/dev_en.tsv" |
|
- split: dev_test |
|
path: |
|
- "data/english/dev_test_en.tsv" |
|
- split: test |
|
path: |
|
- "data/english/test_en_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: german |
|
data_files: |
|
- split: train |
|
path: |
|
- "data/german/train_de.tsv" |
|
- split: dev |
|
path: |
|
- "data/german/dev_de.tsv" |
|
- split: dev_test |
|
path: |
|
- "data/german/dev_test_de.tsv" |
|
- split: test |
|
path: |
|
- "data/german/test_de_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: greek |
|
data_files: |
|
- split: test |
|
path: |
|
- "data/greek/test_gr_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: italian |
|
data_files: |
|
- split: train |
|
path: |
|
- "data/italian/train_it.tsv" |
|
- split: dev |
|
path: |
|
- "data/italian/dev_it.tsv" |
|
- split: dev_test |
|
path: |
|
- "data/italian/dev_test_it.tsv" |
|
- split: test |
|
path: |
|
- "data/italian/test_it_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: multilingual |
|
data_files: |
|
- split: dev_test |
|
path: |
|
- "data/multilingual/dev_test_multilingual.tsv" |
|
- split: test |
|
path: |
|
- "data/multilingual/test_multilingual_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: polish |
|
data_files: |
|
- split: test |
|
path: |
|
- "data/polish/test_pol_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: romanian |
|
data_files: |
|
- split: test |
|
path: |
|
- "data/romanian/test_ro_unlabeled.tsv" |
|
sep: "\t" |
|
- config_name: ukrainian |
|
data_files: |
|
- split: test |
|
path: |
|
- "data/ukrainian/test_ukr_unlabeled.tsv" |
|
sep: "\t" |
|
--- |
|
|
|
# CLEF‑2025 CheckThat! Lab Task 1: Subjectivity in News Articles |
|
|
|
Systems are challenged to distinguish whether a sentence from a news article expresses the subjective view of the author behind it or presents an objective view on the covered topic instead. |
|
|
|
This is a binary classification tasks in which systems have to identify whether a text sequence (a sentence or a paragraph) is subjective (**SUBJ**) or objective (**OBJ**). |
|
|
|
The task comprises three settings: |
|
- **Monolingual**: train and test on data in a given language L |
|
- **Multilingual**: train and test on data comprising several languages |
|
- **Zero-shot**: train on several languages and test on unseen languages |
|
|
|
## Datasets statistics |
|
|
|
* **English** |
|
- train: 830 sentences, 532 OBJ, 298 SUBJ |
|
- dev: 462 sentences, 222 OBJ, 240 SUBJ |
|
- dev-test: 484 sentences, 362 OBJ, 122 SUBJ |
|
* **Italian** |
|
- train: 1613 sentences, 1231 OBJ, 382 SUBJ |
|
- dev: 667 sentences, 490 OBJ, 177 SUBJ |
|
- dev-test - 513 sentences, 377 OBJ, 136 SUBJ |
|
* **German** |
|
- train: 800 sentences, 492 OBJ, 308 SUBJ |
|
- dev: 491 sentences, 317 OBJ, 174 SUBJ |
|
- dev-test - 337 sentences, 226 OBJ, 111 SUBJ |
|
* **Bulgarian** |
|
- train: 729 sentences, 406 OBJ, 323 SUBJ |
|
- dev: 467 sentences, 175 OBJ, 139 SUBJ |
|
- dev-test - 250 sentences, 143 OBJ, 107 SUBJ |
|
- test: TBA |
|
* **Arabic** |
|
- train: 2,446 sentences, 1391 OBJ, 1055 SUBJ |
|
- dev: 742 sentences, 266 OBJ, 201 SUBJ |
|
- dev-test - 748 sentences, 425 OBJ, 323 SUBJ |
|
|
|
## Input Data Format |
|
|
|
The data will be provided as a TSV file with three columns: |
|
> sentence_id <TAB> sentence <TAB> label |
|
|
|
Where: <br> |
|
* sentence_id: sentence id for a given sentence in a news article<br/> |
|
* sentence: sentence's text <br/> |
|
* label: *OBJ* and *SUBJ* |
|
|
|
**Note:** For English, the training and development (validation) sets will also include a fourth column, "solved_conflict", whose boolean value reflects whether the annotators had a strong disagreement. |
|
|
|
**Examples:** |
|
|
|
> b9e1635a-72aa-467f-86d6-f56ef09f62c3 Gone are the days when they led the world in recession-busting SUBJ |
|
> |
|
> f99b5143-70d2-494a-a2f5-c68f10d09d0a The trend is expected to reverse as soon as next month. OBJ |
|
|
|
## Output Data Format |
|
|
|
The output must be a TSV format with two columns: sentence_id and label. |
|
|
|
## Evaluation Metrics |
|
|
|
This task is evaluated as a classification task using F1-macro measure. Other metrics include Precision, Recall, and F1 of the SUBJ class and the macro-averaged scores. |
|
|
|
## Scorers |
|
|
|
The code base with the scorer script is available on the original GitLab repository - [clef2025-checkthat-lab-task1](https://gitlab.com/checkthat_lab/clef2025-checkthat-lab/-/tree/main/task1). |
|
|
|
To evaluate the output of your model which should be in the output format required, please run the script below: |
|
|
|
> python evaluate.py -g dev_truth.tsv -p dev_predicted.tsv |
|
|
|
where dev_predicted.tsv is the output of your model on the dev set, and dev_truth.tsv is the golden label file provided by authors. |
|
|
|
The file can be used also to validate the format of the submission, simply use the provided test file as gold data. |
|
|
|
## Baselines |
|
|
|
The code base with the script to train the baseline model is provided in the original GitLab repository - [clef2025-checkthat-lab-task1](https://gitlab.com/checkthat_lab/clef2025-checkthat-lab/-/tree/main/task1). |
|
The script can be run as follow: |
|
|
|
> python baseline.py -trp train_data.tsv -ttp dev_data.tsv |
|
|
|
where train_data.tsv is the file to be used for training and dev_data.tsv is the file on which doing the prediction. |
|
|
|
The baseline is a logistic regressor trained on a Sentence-BERT multilingual representation of the data. |
|
|
|
## Leaderboard |
|
|
|
The leaderboard is available in the original GitLab repository - [clef2025-checkthat-lab-task1](https://gitlab.com/checkthat_lab/clef2025-checkthat-lab/-/tree/main/task1). |
|
|
|
## Related Work |
|
|
|
The dataset was used in [AI Wizards at CheckThat! 2025: Enhancing Transformer-Based Embeddings with Sentiment for Subjectivity Detection in News Articles](https://huggingface.co/papers/2507.11764). |
|
|
|
Information regarding the annotation guidelines can be found in the following papers: |
|
|
|
> Federico Ruggeri, Francesco Antici, Andrea Galassi, aikaterini Korre, Arianna Muti, Alberto Barron, _[On the Definition of Prescriptive Annotation Guidelines for Language-Agnostic Subjectivity Detection](https://ceur-ws.org/Vol-3370/paper10.pdf)_, in: Proceedings of Text2Story — Sixth Workshop on Narrative Extraction From Texts, CEUR-WS.org, 2023, Vol 3370, pp. 103 - 111 |
|
|
|
> Francesco Antici, Andrea Galassi, Federico Ruggeri, Katerina Korre, Arianna Muti, Alessandra Bardi, Alice Fedotova, Alberto Barrón-Cedeño, _[A Corpus for Sentence-level Subjectivity Detection on English News Articles](https://arxiv.org/abs/2305.18034)_, in: Proceedings of Joint International Conference on Computational Linguistics, Language Resources and Evaluation (COLING-LREC), 2024 |
|
|
|
> Suwaileh, Reem, Maram Hasanain, Fatema Hubail, Wajdi Zaghouani, and Firoj Alam. "ThatiAR: Subjectivity Detection in Arabic News Sentences." arXiv preprint arXiv:2406.05559 (2024). |
|
> |
|
|
|
## Credits |
|
|
|
### ECIR 2025 |
|
|
|
Alam, F. et al. (2025). The CLEF-2025 CheckThat! Lab: Subjectivity, Fact-Checking, Claim Normalization, and Retrieval. In: Hauff, C., et al. Advances in Information Retrieval. ECIR 2025. Lecture Notes in Computer Science, vol 15576. Springer, Cham. https://doi.org/10.1007/978-3-031-88720-8_68 |
|
|
|
```bibtex |
|
@InProceedings{10.1007/978-3-031-88720-8_68, |
|
author="Alam, Firoj |
|
and Stru{\ss}, Julia Maria |
|
and Chakraborty, Tanmoy |
|
and Dietze, Stefan |
|
and Hafid, Salim |
|
and Korre, Katerina |
|
and Muti, Arianna |
|
and Nakov, Preslav |
|
and Ruggeri, Federico |
|
and Schellhammer, Sebastian |
|
and Setty, Vinay |
|
and Sundriyal, Megha |
|
and Todorov, Konstantin |
|
and V., Venktesh", |
|
editor="Hauff, Claudia |
|
and Macdonald, Craig |
|
and Jannach, Dietmar |
|
and Kazai, Gabriella |
|
and Nardini, Franco Maria |
|
and Pinelli, Fabio |
|
and Silvestri, Fabrizio |
|
and Tonellotto, Nicola", |
|
title="The CLEF-2025 CheckThat! Lab: Subjectivity, Fact-Checking, Claim Normalization, and Retrieval", |
|
booktitle="Advances in Information Retrieval", |
|
year="2025", |
|
publisher="Springer Nature Switzerland", |
|
address="Cham", |
|
pages="467--478", |
|
isbn="978-3-031-88720-8", |
|
} |
|
``` |
|
|
|
### CLEF 2025 LNCS |
|
|
|
```bibtex |
|
@InProceedings{clef-checkthat:2025-lncs, |
|
author = { |
|
Alam, Firoj |
|
and Struß, Julia Maria |
|
and Chakraborty, Tanmoy |
|
and Dietze, Stefan |
|
and Hafid, Salim |
|
and Korre, Katerina |
|
and Muti, Arianna |
|
and Nakov, Preslav |
|
and Ruggeri, Federico |
|
and Schellhammer, Sebastian |
|
and Setty, Vinay |
|
and Sundriyal, Megha |
|
and Todorov, Konstantin |
|
and Venktesh, V |
|
}, |
|
title = {Overview of the {CLEF}-2025 {CheckThat! Lab}: Subjectivity, Fact-Checking, Claim Normalization, and Retrieval}, |
|
editor = { |
|
Carrillo-de-Albornoz, Jorge and |
|
Gonzalo, Julio and |
|
Plaza, Laura and |
|
García Seco de Herrera, Alba and |
|
Mothe, Josiane and |
|
Piroi, Florina and |
|
Rosso, Paolo and |
|
Spina, Damiano and |
|
Faggioli, Guglielmo and |
|
Ferro, Nicola |
|
}, |
|
booktitle = {Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF 2025)}, |
|
year = {2025} |
|
} |
|
``` |
|
|
|
### CLEF 2025 CEUR papers |
|
|
|
```bibtex |
|
@proceedings{clef2025-workingnotes, |
|
editor = "Faggioli, Guglielmo and |
|
Ferro, Nicola and |
|
Rosso, Paolo and |
|
Spina, Damiano", |
|
title = "Working Notes of CLEF 2025 - Conference and Labs of the Evaluation Forum", |
|
booktitle = "Working Notes of CLEF 2025 - Conference and Labs of the Evaluation Forum", |
|
series = "CLEF~2025", |
|
address = "Madrid, Spain", |
|
year = 2025 |
|
} |
|
``` |
|
|
|
### Task 1 overview paper |
|
|
|
```bibtex |
|
@inproceedings{clef-checkthat:2025:task1, |
|
title = {Overview of the {CLEF-2025 CheckThat!} Lab Task 1 on Subjectivity in News Article}, |
|
author = { |
|
Ruggeri, Federico and |
|
Muti, Arianna and |
|
Korre, Katerina and |
|
Stru{\ss}, Julia Maria and |
|
Siegel, Melanie and |
|
Wiegand, Michael and |
|
Alam, Firoj and |
|
Biswas, Rafiul and |
|
Zaghouani, Wajdi and |
|
Nawrocka, Maria and |
|
Ivasiuk, Bogdan and |
|
Razvan, Gogu and |
|
Mihail, Andreiana |
|
}, |
|
crossref = {clef2025-workingnotes} |
|
} |
|
``` |