Evaluating Annotation Consistency in Offensive Language Detection: A Data Analytics Approach on the TweetEval Dataset
Abstract
Most machine learning models are not only highly
dependent on difficult datasets but also on the quality of labeled
data they are trained on, especially for offensive content detection.
In this paper, we study the TweetEval dataset to provide a
comparison of its ground truth with manually annotated labels;
inter-annotator agreements are applied here as a metric for
assessing the consistency of annotation. Cohen’s Kappa coefficient
is used to quantify how much each pair of annotators agreed and
where they differed. In-depth examination of missed classifications
demonstrates other difficulties with manual labelling: subjective
interpretation, context dependency, and annotator bias. The in-
sights gathered demonstrate how manual annotation can have
positive and negative effects on further model training practices,
highlighting the importance of standardized annotation guidelines.
In their actions, the findings contribute to enhancing offensive
content detection models by advocating dataset reliability and the
reduction of inconsistencies in labeling.
Keywords:
—TweetEval Dataset, Annotation Consistency, Inter- Annotator Agreement,Cohen’s Kappa,, Offensive Language Detection, Hybrid Models,Annotator BiasPublished
Issue
Section
License
Copyright (c) 2025 International Journal on Emerging Research Areas

This work is licensed under a Creative Commons Attribution 4.0 International License.
All published work in this journal is licensed under the Creative Commons Attribution 4.0 International License (CC BY 4.0). This license permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
How to Cite
Similar Articles
- Devasangeeth A J, Athul MS, Madhav K Vinod, Basil Byju, Seon saju, Amarnadh K S, Angelo joseph, Rohith PM, Hima AU, SMART VEHICLE RENTAL SYSTEM , International Journal on Emerging Research Areas: Vol. 5 No. 1 (2025): IJERA
- Ethen Biju, Chris Mathew, Alina Ann Joseph, Diya Kalyan, Ria Mathews, DaceStudio: AI-Driven Code Editing for Next-Gen Software Development , International Journal on Emerging Research Areas: Vol. 5 No. 1 (2025): IJERA
- NITHYA M V, ADIL SIYAD K.M, AFINSHA P.B, GAUTHAM T.S, ABHIJITH K.P, SALIH SUDHEER, ARJUN SANKAR R.S, C.S ADHITHYAN, JEWELLERY SHOPPING WITH FACIAL RECOGNITION , International Journal on Emerging Research Areas: Vol. 5 No. 1 (2025): IJERA
- Rema M K, Muhamed Ajmal K R, Deepak T G, Roshini M, Muhammed Bazir, INTERACTIVE TOY , International Journal on Emerging Research Areas: Vol. 3 No. 1 (2023): IJERA
- Rince Joseph AS , Rinil Johns , Rinku Theres Jose, Riya Ann Sojan, Siju John , Interview Preparation System: A Smart Platform for Technical and Behavioral Readiness , International Journal on Emerging Research Areas: Vol. 5 No. 1 (2025): IJERA
- Dr. Sinciya P.O, AN EFFECT OF DISTANCE MEASURES IN CLASSIFYING LARGE DATASETS , International Journal on Emerging Research Areas: Vol. 5 No. 1 (2025): IJERA
- Muhammed Aqeel Haroon, Niyas, Muhammed Sajid Nizar, Muzaid Musthafa, Lamer.Ind: A Smart and Interactive Online Textile Platform , International Journal on Emerging Research Areas: Vol. 5 No. 1 (2025): IJERA
- Nihal Anil, Ms. Nighila Abhish, Jesila Joy , Noora Sajil , P R Vishnuraj, Augmented Neat Algorithm For Enhanced Cognitive Interaction (NEAT-X) , International Journal on Emerging Research Areas: Vol. 4 No. 1 (2024): IJERA
- Ashish George, Fida Fathima N, Aswin Kumar A, Nishok Perumal A , Lini Ickappan, GITSHUB - A COMPREHENSIVE PLATFORM FOR ACADEMIC NETWORKING, MENTORSHIP, AND CAREER DEVELOPMENT , International Journal on Emerging Research Areas: Vol. 5 No. 1 (2025): IJERA
You may also start an advanced similarity search for this article.
