
International Journal For Multidisciplinary Research
E-ISSN: 2582-2160
•
Impact Factor: 9.24
A Widely Indexed Open Access Peer Reviewed Multidisciplinary Bi-monthly Scholarly International Journal
Home
Research Paper
Submit Research Paper
Publication Guidelines
Publication Charges
Upload Documents
Track Status / Pay Fees / Download Publication Certi.
Editors & Reviewers
View All
Join as a Reviewer
Get Membership Certificate
Current Issue
Publication Archive
Conference
Publishing Conf. with IJFMR
Upcoming Conference(s) ↓
WSMCDD-2025
GSMCDD-2025
Conferences Published ↓
RBS:RH-COVID-19 (2023)
ICMRS'23
PIPRDA-2023
Contact Us
Plagiarism is checked by the leading plagiarism checker
Call for Paper
Volume 7 Issue 2
March-April 2025
Indexing Partners



















Contextual Understanding and Reprocessing in Sarcasm Detection: a Study of BERT Vs Logistic Regression
Author(s) | Dharmendra Bhadauria, Kalyani Singh, Arun Bhardwaj, S M Aqdas Hashmi |
---|---|
Country | India |
Abstract | Detecting sarcasm in natural language presents distinct challenges, primarily due to its reliance on contextual clues and the use of subtle, contradictory expressions. This research examines how the BERT model and Logistic Regression (LR) perform in detecting sarcasm, emphasizing the importance of contextual understanding and various preprocessing methods. Using a social media dataset, we evaluate both models on sentiment analysis tasks after applying different preprocessing methods including tokenization, elimination of stop words, and noise reduction. We assess the models' performance using evaluation metrics like F1 scores and confusion matrices to determine their accuracy in recognizing sarcastic expressions. Our findings suggest that although BERT's bidirectional architecture provides a deeper understanding of linguistic context, it does not always outperform the simpler LR model when noise reduction preprocessing is applied. This indicates that preprocessing techniques can affect model performance in varying ways, and selecting NLP models should align with the unique needs of the sentiment analysis objective. This study provides meaningful perspectives on the role of contextual models and preprocessing methods in improving the precision of sarcasm detection. |
Keywords | Sarcasm Detection, Sentiment Analysis, BERT, Logistic Regression, Tokenization, Preprocessing Techniques |
Field | Computer > Artificial Intelligence / Simulation / Virtual Reality |
Published In | Volume 6, Issue 6, November-December 2024 |
Published On | 2024-12-28 |
DOI | https://doi.org/10.36948/ijfmr.2024.v06i06.32271 |
Short DOI | https://doi.org/g82gkp |
Share this

E-ISSN 2582-2160

CrossRef DOI is assigned to each research paper published in our journal.
IJFMR DOI prefix is
10.36948/ijfmr
Downloads
All research papers published on this website are licensed under Creative Commons Attribution-ShareAlike 4.0 International License, and all rights belong to their respective authors/researchers.
