Skip to main content

Generating Sentiment-Preserving Fake Online Reviews Using Neural Language Models and Their Human- and Machine-Based Detection

  • Conference paper
  • First Online:
Advanced Information Networking and Applications (AINA 2020)

Abstract

Advanced neural language models (NLMs) are widely used in sequence generation tasks because they are able to produce fluent and meaningful sentences. They can also be used to generate fake reviews, which can then be used to attack online review systems and influence the buying decisions of online shoppers. To perform such attacks, it is necessary for experts to train a tailored LM for a specific topic. In this work, we show that a low-skilled threat model can be built just by combining publicly available LMs and show that the produced fake reviews can fool both humans and machines. In particular, we use the GPT-2 NLM to generate a large number of high-quality reviews based on a review with the desired sentiment and then using a BERT based text classifier (with accuracy of 96%) to filter out reviews with undesired sentiments. Because none of the words in the review are modified, fluent samples like the training data can be generated from the learned distribution. A subjective evaluation with 80 participants demonstrated that this simple method can produce reviews that are as fluent as those written by people. It also showed that the participants tended to distinguish fake reviews randomly. Three countermeasures, Grover, GLTR, and OpenAI GPT-2 detector, were found to be difficult to accurately detect fake review.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

Notes

  1. 1.

    https://github.com/openai/gpt-2.

  2. 2.

    https://github.com/nshepperd/gpt-2.

  3. 3.

    https://github.com/NVIDIA/sentiment-discovery.

  4. 4.

    An image of the interface is available at https://nii-yamagishilab.github.io/fakereview_interface/.

References

  1. Bengio, Y., Ducharme, R., Vincent, P., Jauvin, C.: A neural probabilistic language model. JMLR 3(Feb), 1137–1155 (2003)

    Google Scholar 

  2. Lu, S., Zhu, Y., Zhang, W., Wang, J., Yu, Y.: Neural text generation: past, present and beyond, CoRR, abs/1803.07133 (2018)

    Google Scholar 

  3. Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: ICLR (2015)

    Google Scholar 

  4. Vinyals, O., Toshev, A., Bengio, S., Erhan, D.: Show and tell: a neural image caption generator. In: CVPR (2015)

    Google Scholar 

  5. See, A., Liu, P.J., Manning, C.D.: Get to the point: Summarization with pointer-generator networks. In: ACL, pp. 1073–1083 (2017)

    Google Scholar 

  6. Serban, I.V., Sordoni, A., Bengio, Y., Courville, A., Pineau, J.: Building end-to-end dialogue systems using generative hierarchical neural network models. In: AAAI (2016)

    Google Scholar 

  7. Sundermeyer, M., Schlüter, R., Ney, H.: LSTM neural networks for language modeling. In: INTERSPEECH (2012)

    Google Scholar 

  8. Luca, M., Zervas, G.: Fake it till you make it: Reputation, competition, and Yelp review fraud. Manage. Sci. 62, 3412–3427 (2016)

    Article  Google Scholar 

  9. Yao, Y., Viswanath, B., Cryan, J., Zheng, H., Zhao, B.Y.: Automated crowdturfing attacks and defenses in online review systems. In: CCS, pp. 1143–1158 (2017)

    Google Scholar 

  10. Juuti, M., Sun, B., Mori, T., Asokan, N.: Stay on-topic: Generating context-specific fake restaurant reviews. In: ESORICS (2018)

    Google Scholar 

  11. Radford, A., Wu, J., Child, R., Luan, D.: Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners (2019)

    Google Scholar 

  12. Devlin, J., Chang, M.-W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding, arXiv preprint arXiv:1810.04805 (2018)

  13. He, R., McAuley, J.: Ups and downs: modeling the visual evolution of fashion trends with one-class collaborative filtering. In: WWW, pp. 507–517 (2016)

    Google Scholar 

  14. Zhang, X., Zhao, J., LeCun, Y.: Character-level convolutional networks for text classification. In: NeuRIPS, pp. 649–657 (2015)

    Google Scholar 

  15. Zellers, R., Holtzman, A., Rashkin, H., Bisk, Y., Farhadi, A., Roesner, F., Choi, Y.: Defending against neural fake news, arXiv preprint arXiv:1905.12616 (2019)

  16. Gehrmann, S., Strobelt, H., Rush, A.M.: GLTR: statistical detection and visualization of generated text. In: ACL (2019)

    Google Scholar 

  17. Solaiman, I., Brundage, M., Clark, J., Askell, A., Herbert-Voss, A., Wu, J., Radford, A., Krueger, G., Kim, J.W., Kreps, S., McCain, M., Newhouse, A., Blazakis, J., McGuffie, K., Wang, J.: Release strategies and the social impacts of language models (2019)

    Google Scholar 

  18. Lee, K., Webb, S., Ge, H.: The dark side of micro-task marketplaces: characterizing Fiverr and automatically detecting crowdturfing. In: ICWSM (2014)

    Google Scholar 

  19. Motoyama, M., McCoy, D., Levchenko, K., Savage, S., Voelker, G.M.: Dirty jobs: the role of freelance labor in web service abuse. In: USENIX Security Symposium (2011)

    Google Scholar 

  20. Liang, B., Li, H., Su, M., Bian, P., Li, X., Shi, W.: Deep text classification can be fooled. In: IJCAI, pp. 4208–4215 (2018)

    Google Scholar 

  21. Ebrahimi, J., Rao, A., Lowd, D., Dou, D.: HotFlip: white-box adversarial examples for text classification. In: ACL (2018)

    Google Scholar 

  22. Bengio, Y., Ducharme, R., Vincent, P., Janvin, C.: A neural probabilistic language model. JMLR 3, 1137–1155 (2003)

    MATH  Google Scholar 

  23. Kim, Y., Jernite, Y., Sontag, D., Rush, A.M.: Character-aware neural language models. In: AAAI, pp. 2741–2749 (2016)

    Google Scholar 

  24. Sennrich, R., Haddow, B., Birch, A.: Neural machine translation of rare words with subword units. In: NeuRIPS, pp. 1715–1725 (2016)

    Google Scholar 

  25. Verwimp, L., Pelemans, J., Hamme, H.V., Wambacq, P.: Character-word LSTM language models. CoRR, abs/1704.02813 (2017)

    Google Scholar 

  26. Mikolov, T., Karafiát, M., Burget, L., Cernocký, J., Khudanpur, S.: Recurrent neural network based language model. In: INTERSPEECH, pp. 1045–1048 (2010)

    Google Scholar 

  27. Shen, X., Oualil, Y., Greenberg, C., Singh, M., Klakow, D.: Estimation of gap between current language models and human performance. In: INTERSPEECH (2017)

    Google Scholar 

  28. Sundermeyer, M., Schlueter, R., Ney, H.: LSTM neural networks for language modeling. In: INTERSPEECH, pp. 194–197 (2012)

    Google Scholar 

  29. Krause, B., Lu, L., Murray, I., Renals, S.: Multiplicative LSTM for sequence modelling, CoRR abs/1609.07959 (2016)

    Google Scholar 

  30. Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L.U., Polosukhin, I.: Attention is all you need. In: NeuRIPS, pp. 5998–6008 (2017)

    Google Scholar 

  31. Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)

    Google Scholar 

  32. Radford, A., Józefowicz, R., Sutskever, I.: Learning to generate reviews and discovering sentiment, CoRR, abs/1704.01444 (2017)

    Google Scholar 

  33. Puri, R., Kirby, R., Yakovenko, N., Catanzaro, B.: Large scale language modeling: converging on 40GB of text in four hours. In: SBAC-PAD, pp. 290–297 (2018)

    Google Scholar 

  34. Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D., Levy, O., Lewis, M., Zettlemoyer, L., Stoyanov, V.: RoBERTA: a robustly optimized BERT pretraining approach, arXiv preprint arXiv:1907.11692 (2019)

  35. Sriram, A., Jun, H., Satheesh, S., Coates, A.: Cold fusion: training seq2seq models together with language models. In: Interspeech (2017)

    Google Scholar 

  36. Stahlberg, F., Cross, J., Stoyanov, V.: Simple fusion: return of the language model, arXiv preprint arXiv:1809.00125 (2018)

Download references

Acknowledgments

This research was carried out when the first and second authors were at the National Institute of Informatics (NII) of Japan in 2018 and 2019 as part of the NII International Internship Program. This work was partially supported by a JST CREST Grant (JPMJCR18A6) (VoicePersonae Project), Japan, and by MEXT KAKENHI Grants (16H06302, 17H04687, 18H04120, 18H04112, 18KT0051), Japan.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to David Ifeoluwa Adelani .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Adelani, D.I., Mai, H., Fang, F., Nguyen, H.H., Yamagishi, J., Echizen, I. (2020). Generating Sentiment-Preserving Fake Online Reviews Using Neural Language Models and Their Human- and Machine-Based Detection. In: Barolli, L., Amato, F., Moscato, F., Enokido, T., Takizawa, M. (eds) Advanced Information Networking and Applications. AINA 2020. Advances in Intelligent Systems and Computing, vol 1151. Springer, Cham. https://doi.org/10.1007/978-3-030-44041-1_114

Download citation

Publish with us

Policies and ethics