KUCST at CheckThat 2023: How good can we be with a generic model?
Publikation: Working paper › Preprint › Forskning
Standard
KUCST at CheckThat 2023 : How good can we be with a generic model? / Agirrezabal, Manex.
2023.Publikation: Working paper › Preprint › Forskning
Harvard
APA
Vancouver
Author
Bibtex
}
RIS
TY - UNPB
T1 - KUCST at CheckThat 2023
T2 - How good can we be with a generic model?
AU - Agirrezabal, Manex
PY - 2023
Y1 - 2023
N2 - In this paper we present our method for tasks 2 and 3A at the CheckThat2023 shared task. We make use of a generic approach that has been used to tackle a diverse set of tasks, inspired by authorship attribution and profiling. We train a number of Machine Learning models and our results show that Gradient Boosting performs the best for both tasks. Based on the official ranking provided by the shared task organizers, our model shows an average performance compared to other teams.
AB - In this paper we present our method for tasks 2 and 3A at the CheckThat2023 shared task. We make use of a generic approach that has been used to tackle a diverse set of tasks, inspired by authorship attribution and profiling. We train a number of Machine Learning models and our results show that Gradient Boosting performs the best for both tasks. Based on the official ranking provided by the shared task organizers, our model shows an average performance compared to other teams.
KW - cs.CL
M3 - Preprint
BT - KUCST at CheckThat 2023
ER -
ID: 374968865