Effective Opinion Spam Detection: A Study on Review Metadata Versus Content
, and
May 20, 2020
About this article
Article Category: Research Paper
Published Online: May 20, 2020
Page range: 76 - 110
Received: Jan 21, 2020
Accepted: Apr 23, 2020
DOI: https://doi.org/10.2478/jdis-2020-0013
Keywords
© 2020 Ajay Rastogi et al., published by Sciendo
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.
Figure 1

Figure 2

Figure 3

Figure 4

Figure 5

Figure 6

Figure 7

Figure 8

Figure 9

Figure 10

Classifiers performance on YelpNYC dataset using both behavioral and textual features over all three settings_
SVM | LR | MLP | NB | |||||
---|---|---|---|---|---|---|---|---|
Behavioral | Textual | Behavioral | Textual | Behavioral | Textual | Behavioral | Textual | |
AP | 0.6866 | 0.6999 | 0.7029 | 0.6860 | ||||
Recall | 0.5945 | 0.7011 | 0.6878 | 0.3103 | ||||
F1 (Macro) | 0.6250 | 0.6495 | 0.6472 | 0.5462 | ||||
F1 (Micro) | 0.6259 | 0.6517 | 0.6491 | 0.5759 | ||||
AP | 0.6566 | 0.6444 | 0.6708 | 0.6640 | ||||
Recall | 0.5270 | 0.5073 | 0.6089 | 0.3611 | ||||
F1 (Macro) | 0.6052 | 0.5897 | 0.6173 | 0.5655 | ||||
F1 (Micro) | 0.6069 | 0.5917 | 0.6179 | 0.5851 | ||||
AP | 0.8345 | 0.8367 | 0.8345 | 0.8357 | ||||
Recall | 0.3396 | 0.6844 | 0.6770 | 0.7006 | ||||
F1 (Macro) | 0.6177 | 0.7369 | 0.7048 | 0.7332 | ||||
F1 (Micro) | 0.6601 | 0.7385 | 0.7109 | 0.7344 |
Brief summary of features used by comparing methods under reviewer-centric, review-centric and product-centric settings_
Rayana & Akoglu (2015) Features | |||||
---|---|---|---|---|---|
– | |||||
– | |||||
– | |||||
– | |||||
– | |||||
Dataset statistics after preprocessing (for YelpZip and YelpNYC)_
Dataset | # Reviews (spam%) | # Reviewers (spammer%) | # Products (restaurants) |
---|---|---|---|
YelpZip (Preprocessed) | 356,766 (4.66%) | 49,841 (9.21%) | 3,975 |
YelpNYC (Preprocessed) | 90,906 (7.58%) | 15,351 (10.67%) | 873 |
Classifiers performance on YelpZip dataset using both behavioral and textual features over all three settings_
SVM | LR | MLP | NB | |||||
---|---|---|---|---|---|---|---|---|
Behavioral | Textual | Behavioral | Textual | Behavioral | Textual | Behavioral | ||
AP | 0.6682 | 0.6717 | 0.6783 | 0.6558 | ||||
Recall | 0.6063 | 0.6537 | 0.6497 | 0.3140 | ||||
F1 (Macro) | 0.6260 | 0.6340 | 0.6343 | 0.5491 | ||||
F1 (Micro) | 0.6268 | 0.6343 | 0.6353 | 0.5808 | ||||
AP | 0.6461 | 0.6232 | 0.6581 | 0.6478 | ||||
Recall | 0.4348 | 0.3775 | 0.5947 | 0.3612 | ||||
F1 (Macro) | 0.5888 | 0.5655 | 0.6180 | 0.5663 | ||||
F1 (Micro) | 0.5998 | 0.5830 | 0.6187 | 0.5876 | ||||
AP | 0.8440 | 0.8421 | 0.8499 | 0.8432 | ||||
Recall | 0.7795 | 0.7774 | 0.7731 | 1.0000 | ||||
F1 (Macro) | 0.7279 | 0.7293 | 0.7347 | 0.3537 | ||||
F1 (Micro) | 0.7321 | 0.7333 | 0.7384 | 0.5472 |
Algorithm for balancing the feature set_
|
|
Dataset statistics (for YelpZip and YelpNYC)_
Dataset | # Reviews (spam %) | # Reviewers (spammer %) | # Products (restaurants) |
---|---|---|---|
YelpZip | 608,598 (13.22%) | 260,277 (23.91%) | 5,044 |
YelpNYC | 359,052 (10.2 7%) | 160,225 (17.79%) | 923 |
Brief description of behavioral and textual features employed under reviewer-centric, review-centric and product-centric settings_
Setting | Featuretype | Feature | Description |
---|---|---|---|
Reviewer-centric and Product-centric | Behavioral | ARD | Average rating deviation ( |
WRD | Weighted rating deviation ( | ||
MRD | Maximum rating deviation | ||
BST | Burstiness ( | ||
ERR | Early review ratio | ||
MNR | Maximum number of reviews ( | ||
RPR | Ratio of positive reviews ( | ||
RNR | Ratio of negative reviews ( | ||
FRR | First review ratio ( | ||
EXRR | Extreme rating ratio | ||
TRRR | Top ranked reviews ratio | ||
BRRR | Bottom ranked reviews ratio | ||
Textual | MCS | Maximum content similarity ( | |
ACS | Average content similarity ( | ||
AFPP | Average first-person pronouns ratio | ||
ASPP | Average second-person pronouns ratio | ||
AFTAPP | Average first-and-third-person to all-person pronouns ratio | ||
ASAPP | Average second-person to all-person pronouns ratio | ||
ASW | Average subjective words ratio | ||
AOW | Average objective words ratio | ||
AInW | Average informative words ratio | ||
AImW | Average imaginative words ratio | ||
ARL | Average review length ( | ||
Review-centric | Behavioral | RD | Rating deviation ( |
ERD | Early rating deviation | ||
ETF | Early time frame ( | ||
EXT | Extreme rating ( | ||
TRR | Top ranked review | ||
BRR | Bottom ranked review | ||
RR | Review rank ( | ||
RL | Review length ( | ||
Textual | RPW | Ratio of positive words ( | |
RNW | Ratio of negative words ( | ||
RFPP | Ratio of first-person pronouns ( | ||
RSPP | Ratio of second-person pronouns ( | ||
RFTAPP | Ratio of first-and-third-person to all-person pronouns | ||
RSAPP | Ratio of second-person to all-person pronouns | ||
RSW | Ratio of subjective words ( | ||
ROW | Ratio of objective words ( | ||
RInW | Ratio of informative words ( | ||
RImW | Ratio of imaginative words ( |
Statistical significance of results obtained on behavioral and textual features using Z-test analysis_
Reviewer-centric | Review-centric | Product-centric | |||||
---|---|---|---|---|---|---|---|
Z-test statistic | P-value | Z-test statistic | P-value | Z-test statistic | P-value | ||
YelpZip | ROC-AUC | 30.03 | ~ 0.0 | 53.40 | 0.0 | 3.14 | 0.0016 |
Avg. Precision | 27.69 | ~ 0.0 | 37.58 | ~ 0.0 | 3.31 | 0.0009 | |
F1-Score (micro) | 20.88 | ~ 0.0 | 48.91 | 0.0 | 2.07 | 0.0377 | |
YelpNYC | ROC-AUC | 23.02 | ~ 0.0 | 47.44 | 0.0 | 4.59 | ~ 0.0 |
Avg. Precision | 23.35 | ~ 0.0 | 33.48 | ~ 0.0 | 3.86 | 0.0001 | |
F1-Score (micro) | 22.41 | ~ 0.0 | 30.17 | ~ 0.0 | 8.73 | ~ 0.0 |