Back to Search Start Over

Exploring Transformer Models and Domain Adaptation for Detecting Opinion Spam in Reviews

Authors :
Christopher G Harris
Source :
Proceedings of the XXth Conference of Open Innovations Association FRUCT, Vol 36, Iss 1, Pp 249-255 (2024)
Publication Year :
2024
Publisher :
FRUCT, 2024.

Abstract

As online reviews play a crucial role in purchasing decisions, businesses are increasingly incentivized to generate positive reviews, sometimes resorting to fake reviews or opinion spam. Detecting opinion spam requires well-trained models, but obtaining annotated training data in the same domain (e.g., hotels) can be challenging. Transfer learning addresses this by leveraging training data from a similar domain (e.g., restaurants). This paper examines three popular transformer models—BERT, RoBERTa, and DistilBERT—to evaluate how training data from different domains, including imbalanced datasets, impacts Transformer model performance. Notably, our evaluation of hotel opinion spam detection achieved an AUC of 0.927 using RoBERTa trained on YelpChi restaurant data.

Details

Language :
English
ISSN :
23057254 and 23430737
Volume :
36
Issue :
1
Database :
Directory of Open Access Journals
Journal :
Proceedings of the XXth Conference of Open Innovations Association FRUCT
Publication Type :
Academic Journal
Accession number :
edsdoj.10b8b403c781431393b7dbc4b3e719b7
Document Type :
article
Full Text :
https://doi.org/10.23919/FRUCT64283.2024.10749897