Tired of Fake Restaurant Reviews? This AI Can Help
Was this restaurant review written by a machine or a person? Not so easy, is it? (It's a fake created by the researchers' AI model, made to imitate the review layout on Yelp. No fakes were posted online in the study.) Credit: Aalto University
Sites like TripAdvisor, Yelp and Amazon display user reviews of products and services. Consumers take heed: nine out of ten people read these peer reviews and trust what they see. In fact, up to 40% of users decide to make a purchase based on only a couple of reviews, and great reviews make people spend 30% more on their purchases.
Yet not all reviews are legitimate. Fake reviews written by real people are already common on review sites, but the amount of fakes generated by machines is likely to increase substantially.
According to doctoral student Mika Juuti at Aalto University, fake reviews based on algorithms are nowadays easy, accurate and fast to generate. Most of the time, people are unable to tell the difference between genuine and machine-generated fake reviews.
‘Misbehaving companies can either try to boost their sales by creating a positive brand image artificially or by generating fake negative reviews about a competitor. The motivation is, of course, money: online reviews are a big business for travel destinations, hotels, service providers and consumer products,’ says Mika Juuti.
In 2017, researchers from the University of Chicago described a method for training a machine learning model, a deep neural network, using a dataset of three million real restaurant ratings on Yelp. After the training, the model generated fake restaurant reviews character by character.
There was a slight hiccup in the method, however; it had a hard time staying on topic. For a review of a Japanese restaurant in Las Vegas, the model could make references to an Italian restaurant in Baltimore. These kinds of errors are, of course, easily spotted by readers.
To help the review generator stay on the mark, Juuti and his team used a technique called neural machine translation to give the model a sense of context. Using a text sequence of ‘review rating, restaurant name, city, state, and food tags’, they started to obtain believable results.
‘In the user study we conducted, we showed participants real reviews written by humans and fake machine-generated reviews and asked them to identify the fakes. Up to 60% of the fake reviews were mistakenly thought to be real,’ says Juuti.
Juuti and his colleagues then devised a classifier that would be able to spot the fakes. The classifier turned out to perform well, particularly in cases where human evaluators had the most difficulties in telling whether a review is real or not.
The study was conducted in collaboration with Aalto University’s Secure Systems research group and researchers from Waseda University in Japan. It was presented at the 2018 European Symposium on Research in Computer Security in September.
This article has been republished from materials provided by Aalto University. Note: material may have been edited for length and content. For further information, please contact the cited source.
Reference: Mika Juuti, Bo Sun, Tatsuya Mori, N. Asokan: Stay On-Topic: Generating Context-specific Fake Restaurant Reviews. https://arxiv.org/abs/1805.02400
Computer bits are binary, with a value of 0 or 1. By contrast, neurons in the brain can have all kinds of different internal states, depending on the input that they received. This allows the brain to process information in a more energy-efficient manner than a computer. A new study hopes to bring the two closer together.
MIT researchers have developed a cryptographic system that could help neural networks identify promising drug candidates in massive pharmacological datasets, while keeping the data private. Secure computation done at such a massive scale could enable broad pooling of sensitive pharmacological data for predictive drug discovery.