Products -> Semantic Analysis for All Accommodation Reviews
Semantic Analysis for All Accommodation Reviews consist of 3-10 different engines optimized for reviews of different accommodation (hotel, hostel, etc.). Each engine consists of 120+ Semantic Models. Each Semantic model was specially designed, built, tested, and re-tested on hundreds of thousands of accommodation reviews from 10 different sources. All presented Semantic Models work with an unparalleled precision of 90-95%.
Below products is a part of this package (Semantic Analysis for All Accommodation Reviews). Depending on your needs, we prepare 3-10 different engines for different types of accommodations (hotel, hostel, apartment, b&b, cabins, etc.).
To see more of what each engine consists of and what this technology can do, choose the product from the list.
First, you process all reviews of a given hotel and extract important information with high accuracy. Then, it is easy to compare these results with other hotels, and valuable patterns emerge.
We analyzed all 4423 of your reviews, compared with 450 000 reviews of similar businesses and extracted 14 Recommendations/Tips how you can improve your business:
Compare your parameters with similar businesses
Here, we focused on showing classified negative opinions critical to hotel owners and grouped them semantically.
By clicking on a specific semantic category (here: Noisy rooms) we get specific sentences/segments and "not the whole review" where people mentioned noisy rooms. We detect all the information that is semanticaly correlated to noisy rooms (e.g. "cleaning service did not their job"; "it was far from clean"), not just related keywords (e.g. dirty, messy). This is the advantage if you use Cognitive Learning instead of keywords/machine learning models.
Below is a simple comparison made only using 200 reviews and limited categories/areas. You have 149 Semantic Models at your disposal which can be used to make comparisons in general or in very specific areas.
Marriott, San Francisco (8.5 on booking.com)vs. Hilton, San Francisco (7.8 on booking.com)
bathroom clean 1 dirty 10
cleaning service good 4 poor 4
bed dirty 1
Positive: 30, Negative: 16
bathroom clean 1 dirty 7
cleaning service good 5 poor 16
bed dirty 6
Positive: 33, Negative: 43
helpful, friendly 94
unhelpful, unfriendly 9
check-in fast 8 slow 1
Positive: 102, Negative: 10
helpful, friendly 57
unhelpful, unfriendly 35
check-in fast 22 slow 8
Positive: 79, Negative: 43
Addons comparison (comparing only negative opinions):
not free 20
parking not free 12 small 1
not free 9
parking not free 16
Clients Opinion comparison:
overall good, satisfied 6
will return 4
will recommend 2
overall disappointed 2
will not return 3
Positive: 12, Negative: 5
overall good, satisfied 4
will return 5
will recommend 4
overall disappointed 14
will not return 7
not recommend 3
Positive: 13, Negative: 24
Marriott, San Francisco wins (Positive: 144, Negative: 68)
over Hilton, San Francisco (Positive: 105, Negative: 141).
On Booking.com Marriott also got a higher rate (8.5) compared to Hilton rate (7.5).
Search based on your individual preferences matched with opinions of similar people to you and data from reviews
What is important to you? Personalize your search
We remember these parameters while reading millions of reviews and find you the best possible place
We analyzed 3 million reviews from Berlin to match your personal needs:
Meet the Alchemist Traveler. He has been everywhere and he has read all reviews in the world (and he remembers them all).
He is patient and he can spare a moment to plan your perfect vacation.
Instead of reading thousands of reviews individually, you can see a semantic summary and get a feeling for what is contained within them in 20 seconds. You can click for specific information and see extracted sentences from reviews.
Hotelhoteloutdated 23 smelly 22 clean 15 not as advertised 4
Roomroomclean 34 spacious 22 well equiped 15 sth broken 4
Servicestaffhelpful, friendly 23 unhelpful, unfriendly 15 check-in slow 12
Foodbreakfastrich, plenty options 13 not free as promised 9 great 6
Pricehoteloverpriced 23 money waste 15 adequate price 4
Locationclose tohighway 7 city center 5 public transportation 4
Addonswififree 11 problem 9 slow 7
Otherdeal breakerssth stolen 4bugs 2
The automatic summary made from reviews - after processing by Semantic Analysis for Hotel Reviews (no additional NLP/NLU/ML or post-processing needed).
The first no-block title (Hotel, Price etc.) is a Category from our API, blue blocks (Bathroom, Payment etc.) are Aspects, green and red blocks (outdated, well equipped etc.) is the Feature and Polarity (Feature is text, Polarity is color).
See the live demo of semantic summary made from hotel reviews (powered by our technology).
All displayed results make sense. By clicking on specific information, you can display actual sentences (not whole reviews).
We analyzed 4765 Reviews and get Top Strengths & Weaknesses of this place
(so you do not have to read hundreds of reviews)
|Top 5 Strengths (People were the most satisfied with):|
|Top 3 Weaknesses (People complained about):|
|Problem with Refund||
|Problem with WiFi||
|show me more...|
Client loyalty analysis
Overall good/satisfied vs. Overall dissapointed
Will come back vs. Will not come back
Will recommend vs. Will not recommend
Information extracted by 149 Semantic Models is precise enough to create automatic text summary.
Natural Language Generation Summary from Reviews could look like this:
We analyzed all your reviews for the past 3 months, combined them with reviews of similar businesses and automatically created this summary:
Most of people (88%) was satisfied with their stay, a lot of people (47 mentions) wrote that they would come back and will recommend this apartment (24 m.).
Reviewers state that rooms in this apartment are clean (34 mentions) and spacious (22 m.) and well-equipped (15 m.). Unfortunately, reviewers also stated that the apartment was outdated (23 m.) and smelly (22 m.). The majority found the staff helpful and friendly (23 m.) but some stated that they were unprofessional (19 m.) and the check-in process (12 m.) was slow.
There were also some issues with the Pool (23m.) and the Gym (15m.)
Generally, people say it is a good value for money (42m.).
|Semantic distribution summary out of text:|
|Would come back||
|Issues with Pool||
|Issues with Gym||
|Good value for money||
If you would like to talk more about these solutions, or discover what is possible
We present it on a hotel review (Semantic Analysis for Hotel Reviews):
Simplified output after processing by Semantic Analysis for Hotel Reviews:
|1.||The hotel was clean and renovated, service was friendly too||
staff helpful, friendly +2
hotel renovated +1
|2.||But thats it|
|3.||The bathroom was dirty, Shower got poor pressure||
shower problem -2
bathroom dirty -2
|4.||I think there were fleas in our bed||
|5.||The pool was closed during our stay and staff couldn't tell us where it will be open and does not seem to care||
staff unfriendly, unhelpful -2
|6.||The breakfast was a joke, the portions was small and we were told that we could pay to get more, so it wasn't free as promised||
poor, limited -1
breakfast not free as promised -2
breakfast bad -1
|7.||Wanted to get my money back but they refused, not coming back||
will not return -2
payment refound problem -1
We specialize in creating dedicated Language Understanding APIs for specific reviews or other user-generated content. We focused in travel, food, apps, surveys, profanity & toxicity detection and prepared a set of Language Understanding APIs for these domains. We also develop private custom-made Language Understanding APIs for any kind of text (reviews, comments, or other user-generated content). Contact Us to discuss what is possible, or Send Us Your Dataset to see how it works on your data.
We stand by the statement that we detect more information than Watson or than any other Deep Learning solution and we are more accurate than Google in a specific domain. We are also much more cost-effective. And we allow the processing of as much data as you want with no additional cost (0 cents / text).
This is possible because we created the next-generaion process of developing NLU systems in which we can build Semantic Models 10 x faster. It is designed from the ground up to process reviews and other user-generated content, not proper english texts. It is the result of 13 years of experience in NLP/NLU, 8 years in processing reviews, and more than 50 implementations (2007-2020) in companies, corporations, and startups of NLP/NLU systems and Language Understanding APIs to process reviews and other user-generated content.
We are passionate and proud of our technology and aim to provide the optimal technology for reviews and other user-generated content. We will be happy to help you implement your brilliant ideas and discover what is possible.
|"Breakfast was tasty"||=>||Breakfast: positive|
|"Breakfast was huge"||=>||Breakfast: positive|
|"Breakfast was not included"||=>||Breakfast: negative|
|"Breakfast was very tasty but limited choices"||=>||Breakfast: neutral|
|"Breakfast was delicious but you had to pay extra"||=>||Breakfast: neutral|
|"Breakfast was tasty"||=>||Breakfast: tasty|
|"Breakfast was huge"||=>||Breakfast: plenty options|
|"Breakfast was not included"||=>||Breakfast: not included|
|"Breakfast was very tasty but limited choices"||=>||Breakfast: tasty | poor, limited|
|"Breakfast was delicious but you had to pay extra"||=>||Breakfast: tasty | not included|
Language has colors. Do not reduce it to black & white.
Technological differences of Unicorn NLP Cognitive Learning solution (compared to other NLP/NLU/ML/DeepLearning solutions):
The semantic model is the newest approach to information extraction models where boundaries of those models are not determined by the structure of a language (or grammar, or used words), but the type of information you are detecting. It does not matter how the user writes an opinion about a specific object, e.g. “breakfast was a joke”, “Muffin and cold coffee is not a breakfast”, “Breakfast wasn’t included as written”, our semantic model still detects it and provides structured-data. With "breakfast" example, it is relatively easy to provide a shallow analysis (e.g. Sentiment Analysis) with the current NLP techniques. If you are looking for more sophisticated information (e.g. is it safe?, is it handicap friendly?, will someone come back?, Is it sustainable? Is it pet-friendly?) the situation gets more complicated. Current techniques provide a part of this information and their accuracy relies on the keywords used in reviews. Users are very creative in the reviews and they are not using a keyword approach. Let’s take one example. If we want to answer the question “is it safe?”, you need to capture a lot of information from reviews without common keywords safe/dangerous. In practice, they write this in tens or hundreds of ways: e.g. “there was a lot of drunken people outside”, “I did not feel good because of the neighbors”, “someone yelled in the night and the front desk lady did nothing”. We believe that if the human can understand what is written in the review, then technology should detect it as well. That is why we call them Semantic Models, not information extraction models.
Our Unicorn NLP Cognitive Learning Environment is a place where we design and precisely craft all the semantic models. It is a perfect combination of human and computer intelligence. It is the result of 13 years of experience in the NLP/NLU field and 50+ implementation in 2007-2020. It is the secret sauce of our technology, and it allows us to deliver the best technology in the world for reviews and other user-generated content. Statistical algorithms provide the data based on the processing of hundreds of thousands of reviews, but the humans are making all the crucial decisions. Human does not only annotate the data (like in machine learning approach), but also create semantic boundaries, semantic rules, domain-specific dictionaries, describe common domain-specific misspellings and grammar errors, and whatever it takes to achieve precision=90-95%. Given the pace of AI development, this is a significant advantage in information extraction (in NLP/NLU), and it will be appropriate for the next decade. Think about our environment as a way of creating tens of thousands of very precise, hand-crafted semantic rules but with an AI as an assistant rigorously validating new ideas and marking parts of algorithms and semantic code that needs improvement. AI assistant provides new ideas based on constant simulations on real data and tens of statistical tools.
We designed a unique semantic programming language (QL4Reviews) to extract information from reviews and other user-generated content. With this programming language, you build semantic models ten times faster than in other current technologies. Every Semantic Model is programmed in QL4Reviews and consists of 50-2000 lines of semantic code. Heavily optimized engine allows processing through 150+ Semantic Models on 1 medium-CPU Amazon AWS instance - 1 million reviews a day. The speed of the core engine was essential for us from the beginning. It helps us daily speed the process of development and make several iterations and validation on data a day.
Our technology is designed from the ground up to process reviews and other user-generated content, not proper-grammar texts. It's resistance to errors allows processing also machine-translated reviews/texts (via Google Translate or Microsoft Translate). This approach provides almost as good results as processing English reviews but there is no need to rewrite and maintain semantic models into other languages. Consequently, it is a much more cost-effective solution, and it is easier to maintain and scale (using Google Translation API allows you to process 105 languages on day one).
Read more & See demo: Processing machine-translated non-English Travel Reviews
This is not just another limited black box that you still need to configure to your data, and learn how to use it. You get dedicated Language Understanding API with a human-like accuracy, which is train&tested on your data. We redesigned every layer of technology and we wanted to address the biggest challenge in the NLP/NLU industry. We wanted to change the way you use and think about Natural Language Understanding API. In our opinion, if you are using NLP API, you should not get your hands dirty and spend inordinate time thinking about how to best implement this API into your application. We wanted to hide all the NLP complexity inside the system and provide you ready-to-use data. We believe that NLP API should be simple and easy to use without linguistic, machine learning or other expertise. You do not need to spend weeks on improving the accuracy of "domain-independent" NLP API, to adjust to Your domain or to make decisions about complex output parameters. Getting output parameters like confidence (e.g. 0.7852) or polarity (e.g. 0.4352) or keywords detected (e.g. friendly, 0.6786; disappointed, -0.7564) gives every developer problems and frankly, they are not standard or predictable. They are rather a consequence of used technology inside, and putting these parameters for the output in our opinion, is the way to leave configuration and responsibility for decision-making to You. We believe in simple human-like output where a human can understand what this API detected. We believe that we have developed excellent semantic models for specific domains, so you can get tailored results and focus on your business. We position ourselves as a read-to-use technology, not a standard API that is a tool that you need to configure and train to your data.
If you want to get custom-made private API, or receive more info about the custom On-Premise version (with 0 cents/text),
or want to discuss what is possible
If you want to test it or see how it works on your data, send us a dataset
(we do not collect your data)