In the modern day, public discussion and critiquing of products and services occurs beyond dedicated mediums, and now also takes place in the realm of social media, too.
Online Hilton Hotel reviews are currently found on tripadvisor, trustpilot, and expedia. The majority of reviewers gave a score between 3 and 5, so if a new customer browses online reviews on any of the previously mentioned review sites, they may consider booking a room at the Hilton.
What if they already made up there mind from hearing what a friend had to say? Potential customers, could have their hotel choice be influenced by a tweet. Opinions are shared constantly on social media platforms, and are read by their followers. The knowledge, of what these followers think about our hotel, from reading these online posts, could help us better understand the general public's perception of our hotel.
By using sentiment analysis, on existing hotel reviews from Tripadvisor.com, I created a model that can quantify on a scale of 1-5, how the author of a tweet on twitter, or a post on a reddit thread, feels about our hotel, and as a result, also how the readers think about us. If a review classifies to be less than a score of 3, this post/tweet could be looked into, find out why they had a negative opinion of our hotel, and in return fix the problem.
Here is a demo application of the review classifier: https://hilton-hotel-app.herokuapp.com/
Try it out by inputting a review that you found online, or wrote yourself, and it will clasify the review with a score between 1-5.
Email: [email protected]
LinkedIn: www.linkedin.com/in/ahilan-srivishnumohan/
Medium: www.medium.com/@candyahs
Show/Hide
- Data : folder containing all data files
- 1.tripadvisor_scraped_hotel_reviews.csv: webscraped data before any changes
- 2.hotel_reviews_structured.csv: data after balancing and cleaning
- 3.x_train_data.csv: training data with x values from preprocessed dataset
- 3.y_train_data.csv: training data with y values from preprocessed dataset
- 4.x_test_data.csv: test data with x values from preprocessed dataset
- 4.y_test_data.csv: test data with y values from preprocessed dataset
- Images : folder containing images used for README and presentation pdf
- Models : folder containing trained models saved with pickle
- Adabooost.pkl, Decision Tree.pkl, KNN.pkl, Logistic Regression.pkl, Naive Bayes.pkl, Neural Network.pkl, Random Forest.pkl, Stacking.pkl, SVM.pkl, Voting.pkl, XGBoost.pkl
- Tripadvisor_Webscrape : folder containing all webscraping files
- Tripadvisor: folder containing .py files and spiders used
- spiders: folder containing spider files and datasets
- hotels.py: main spider .py file for scraping hotel reviews from Tripadvisor
- tripadvisor_scraped_hotel_reviews.csv: csv file with data to be used for project
- init.py, items.py, middlewares.py, pipelines.py, settings.py: default scrapy files used for webscrape
- spiders: folder containing spider files and datasets
- scrapy.cfg: scrap config file
- Tripadvisor: folder containing .py files and spiders used
- 1.Webscraping_Early_EDA_and_Cleaning.ipynb : notebook with early data exploration and data manipulation
- 2.Further_EDA_and_Preprocessing.ipynb : notebook with feature engineering and nlp preprocessing
- 3.Modelling_and_Hyperparameter_Tuning.ipynb : notebook with all the models created
- 4.Evaluation : notebook with final model selection, testing, and model evaluation
- 5.Neural_Network_Modelling : notebook with an extra model training using neural networks
- 6.Revaluation_and_Deployment : notebook comparing old best model and NN model, and deployment
- Classification.py : contains classes with classifcation methods
- Ensemble.py : contains classes with ensemble methods
- Helpers_NN.py : contains methods used to help with neural network processes
- Hilton_Hotel_App.py : contains script to run app
- Hilton__Hotel_Presentation.pdf : presentation summarising project case, processese, and findings
- Procfile : file supporting Heroku application
- requirements.txt : dependencies for heroku application
Show/Hide
- Python
- Pandas
- Numpy
- Matplotlib
- Seaborn
- NLTK
- Scrapy
- Scikit-Learn
- Keras
- Tensorflow
- Streamlit
- Heroku
Show/Hide
-
Early EDA and Cleaning
- 1.1 Webscrape Process Explained
- 1.2 Imports
- 1.3 Checking for Nulls
- 1.4 Converting Score Column
- 1.5 Adjusting Class Imbalance for Scores
- 1.6 Joining Review Part 1 with Review Part 2 in New Column Review
- 1.7 Removing Review Part 1 and Review Part 2 Columns
- 1.8 Saving Structured Dataset as a CSV
-
Further EDA and Preprocessing
- 2.1 Imports
- 2.2 Checking Frequency of Words and Phrases in Review Summaries
- 2.3 Checking Frequency of Words and Phrases in Reviews
- 2.4 Stemming and Lemming
- 2.5 Train Test Split
- 2.6 TF-IDF Vectorisation for Reviews
- 2.7 TF-IDF Vectorisation for Review Summaries
- 2.8 Joining Reviews With Review Summaries
- 2.9 Saving Preprocessed Dataset as CSVs
-
Modelling and Hyperparameter Tuning
- 3.1 Imports
- 3.2 Train and Validation Split
- 3.3 Decision Tree (Baseline)
- 3.4 Random Forest
- 3.5 Logistic Regression
- 3.6 Support Vector Machines
- 3.7 Guassian Naive Bayes
- 3.8 KNN
- 3.9 Adaboost (Logistic Regression)
- 3.10 XGBoost (Logistic Regression)
- 3.11 Voting
- 3.12 Stacking
- 3.13 All Models Compared
- 3.14 Best Model (Logistic Regression) - Deeper Look
- 3.15 Saving Best Model
-
Evaluation
- 4.1 Imports
- 4.2 Best Model Selection
- 4.3 Best Model Tested
- 4.4 Deeper Diver Into Best Model
- 4.5 Application Deployability
-
Neural Network Modelling
- 5.1 Imports
- 5.2 One Hot Encoding Score Column
- 5.3 Train Test Split
- 5.4 Add Suffix to the Review Summary to Distinguish the Difference
- 5.5 Removing Punctuation and Tokenizing Review Column
- 5.6 Creating a Dictionary With Words That Appear in Reviews and an Index
- 5.7 Indexing Words in Reviews Using Dictionary
- 5.8 Combining Indexed Review Summary and Indexed Review Into a Single Column Called All Preprocessed Review
- 5.9 Modelling
- 5.10 Testing Model
- 5.11 Test Confusion Matrix
- 5.12 Saving Model
-
Revaluation and Deployment
- 6.1 Imports
- 6.2 Comparing Stacking Model with Neural Network Model
- 6.3 Deployment
Show/Hide
I set a goal of a minimum of 5000 reviews to scrape, before choosing the specific hotels. I then chose the 5 Hilton hotels with the highest number of reviews, to scrape; London Gatwick Airport, London Metropole, London Euston, London Croydon, and London - West End. Between these 5 hotels there were 17538 reviews, I had plenty room to filter or drop reviews and retain at least my minimum of 5000.
The structure of each review consisted of a 1-5 scale score rating in bubble form, a review summary, and a detailed review split into p1 and p2 (depending on if there was a read more option). Each page on tripadvisor had 5 reviews per page, so I had to navigate between pages using tripadvisor's next page function.
The root URL I used was 'www.tripadvisor.co.uk'
The 5 starting URL extensions I used were:
- '/Hotel_Review-g187051-d239658-Reviews-Hotel_Hilton_London_Gatwick_Airport-Crawley_West_Sussex_England.html/'
- '/Hotel_Review-g186338-d193089-Reviews-Hilton_London_Metropole-London_England.html/'
- '/Hotel_Review-g186338-d192048-Reviews-Hilton_London_Euston-London_England.html/'
- '/Hotel_Review-g186338-d193102-Reviews-DoubleTree_by_Hilton_Hotel_London_West_End-London_England.html/'
- '/Hotel_Review-g504167-d192599-Reviews-Hilton_London_Croydon-Croydon_Greater_London_England.html'
From these pages I chose to extract 5 different features:
- hotel_name
- review_summary
- review_p1
- review_p2
- score
I used a scrapy spider to crawl the website to scrape the requested data. Scrapy proved the be efficient and fast at extracting the data. I ran the spider script (hotels.py) for around 20 minutes, on the 13th May 2020.
The initial shape of the dataset was (35078,5). The 5 columns was as expected, but there were double the number of rows as the number of reviews scraped. There were null rows with only hotel_name and no other values, so I removed those rows, bringing us back to the expected 17538.
This project entailed the use of classification models, and for reliable results, I had to remove reviews to undo class imbalance. Using this visualisation I saw that were much less reviews with a score of 1 compared to reviews with a score of 3, 4, and 5. To combat this imbalance, I randomly removed reviews with scores of 2, 3, 4, and 5, to match with 1 (1881 reviews).
I combined the review p1 and review p2 column into one to make future vectorisation much easier, then I saved the cleaned dataset as a csv, for the next stage.
Show/Hide
The cleaned dataset had a shape of (9405,4). I started with some analysis on the text columns; review and review summary.
Using the FreqDist function in the ntlk library I plotted a graph with the most frequent words and phrases in both columns. Stopwords were removed to capture the more meaningful words.
Distribution Plot of Frequent Words and Phrases in Text ( Review Summary (Left) and Review (Right) )
I had noticed a lot of the most frequent words in the review text happened to be words with no sentimental impact, so I iteratively removed unmeaningful words such as 'room', 'hotel', 'hilton' etc. I did this as a precaution, as some of these words may impact my model accuracies.
World Cloud of Frequent Words and Phrases in Text After Removing Unmeaningful Words ( Review Summary (Left) and Review (Right) )
To narrow down the feature words I applied stemmation and lemmitisation to both the reviews and review summaries.
Stemmation had broken down some words into words that don't exist, whereas lemmitisation had simplified adjectives and verbs to their root form. I chose to continue with the lemmitised version of the texts for further processing.
Prior to vectorising the current dataset, I did a train, test split to save the test data for after modelling.
Using the lemmed texts for review and review summary I used TF-IDF vectorisation with an ngram range of 2, leaving me with a vectorised dataset with 138 words and phrases (112 from reviews and 26 from review summaries). I then saved the x and y train data in separate csv files for modelling.
Show/Hide
I have created .py files; Classifiction.py and Ensemble.py with classes, that contain functions to simplify the modelling process, and to neaten up the modelling notebook.
I did another data split into Train and Validation data in preparation for using GridSearch Cross Validation. I also chose Stratified 5-fold has a my choice for cross validating.
For the majority of models I created, I applied hyperparameter tuning, where I started with a broad range of hyperparameters, and tuned for optimal train accuracy and validation accuracy.
Initially, I thought the validation accuracy was low for most of the models I created, but when considering these models were attempting to classify for 5 different classes, 0.45 and greater seems very reasonable (where 0.2 = randomly guessing correctly).
I have saved all the models using the pickle library's dump function and stored them in the Models folder.
Show/Hide
I focused on 3 factors of defining a good model:
- Good Validation Accuracy
- Good Training Accuracy
- Small Difference between Training and Validation Accuracy
I chose the Stacking ensemble model ( (Adaboost with log_reg_2) stacked with log_reg_2 ) as my best model, because it has the highest validation accuracy with only around 3.5% drop from train to validation in accuracy. I wanted to minimise overfitting and make the model as reusable as possible. Stacking achieved a reasonable training accuracy as well, although it did not reach the level of some of the other ensemble techniques.
I next tested the best model with the earlier saved test data. The model managed to get a high test accuracy, similar to the validation data from the model training stage. This is very good, proving that prioritising a high validation score, and minimising the difference between train and validation accuracy, has helped it classify new review texts very well.
Looking at the precision, recall, and f1 score, I also noticed the scores were higher around scores of 1 and 5, lower for 2, 3, and 4. This shows that the models performs well on more extreme opinions on reviews than mixed opinions.
Looking into different metrics and deeper into my best model; Stacking, I learnt that most the False Postives came from close misses (e.g. predicting a score of 4 for a true score of 5). This is best shown by these two confusion matrixes (validation and test).
The adjacent squares of the diagonal going across the confusion matrix, shows that the model's second highest prediction for a given class (review score) is always a review score that is +-1 the true score. Very few reviews that have a score of 5, have been predicted to have a score of 1 or 2. This is very relieving to know, the majority of the error for the model, is no different to the error a human may make classifying a review to a score with a scale of 1-5.
- most errors were near misses (e.g. 5 predicted as 4)
- extreme scores (1 and 5) were relatively accurate
- comparable to human prediction
- reusable and consistent
Given the classifcation problem is 5 way multi-class one and the adjacent classes can have overlap in the english language even to humans, this model I have created can be deployed.
Applying this model will address the problem of not having a full understanding of public opinion of our hotel. We can apply this to new sources for opinions on our hotel and yield more feedback then we did had before.
- Model using neural networks - see if better accuracy can be achieved
- Create a working application to test new reviews written by people
- Try a different pre-processing approach and see if model performances change
- Bring in new sources of data to see if there are significant differences on frequent words used
Show/Hide
I experimented with different classifcation and ensemble methods to help classify hotel review scores. Some performed well, but there was definitely room for improvement, so I wanted to explore a deep learning approach.
- Input Layer: 17317 Nodes (one for each word in training data + 4 extra for padding, unknown words, start of review, and unused words)
- Embedding Layer: takes 17317 unique items and maps them into a 16 dimensional vector space
- Global Average 1D Pooling Layer: scales down 16 dimensional layer
- Dense Hidden Layer: 16 Nodes (using relu activation function)
- Dense Output Layer: 5 nodes for each score (using sigmoid activation function)
Show/Hide
I tested the neural network model using the test data and achieved an accuracy of 0.5710 which is better than the stacking model accuracy of 0.5077, by over 5%.
I wanted to look at the confusion matrix, as this gives a better idea of how the model is performing over all 5 classes.
The error is more contained within adjacent scores with the neural network model. Almost zero confusion between extreme scores 1 and 5, and minimal confusion with scores 2 and 4. Although a score of 3 can be harder to predict, there is definitely an improvement from the Stacking model.
After seeing the improvements from the Stacking model, I was more confident about deploying the model for actionable use.
I planned on future improvements being the addition of the neural network model and then creating an application for the model, so as a next step I decided to make a working application to test out new reviews using streamlit. I have deployed the app using Heroku: https://hilton-hotel-app.herokuapp.com/.
Using this model, we will learn more about our new and old customers, then we can improve Hilton Hotel's guest satisfaction, and as a result increase customer retention and bring in new travelers.
- Create a webscraper spider for twitter, reddit, etc for further model assessment