Next Article in Journal
From XS to XL Urban Nature: Examining Access to Different Types of Green Space Using a ‘Just Sustainabilities’ Framework
Next Article in Special Issue
Significant Labels in Sentiment Analysis of Online Customer Reviews of Airlines
Previous Article in Journal
Life Cycle Assessment of Wood Pellet Production in Thailand
Previous Article in Special Issue
Online Company Reputation—A Thorny Problem for Optimizing Corporate Sustainability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Bayesian Network to Predict Online Review Helpfulness

1
College of Business Administration, Sejong University, Seoul 05006, Korea
2
Department of Global Administration, SKK Business School/SAIHST (Samsung Advanced Institute for Health Sciences & Technology), Sungkyunkwan University, Seoul 03063, Korea
3
Department of Software, Sejong University, Seoul 05006, Korea
*
Author to whom correspondence should be addressed.
Sustainability 2020, 12(17), 6997; https://doi.org/10.3390/su12176997
Submission received: 2 August 2020 / Revised: 17 August 2020 / Accepted: 22 August 2020 / Published: 27 August 2020
(This article belongs to the Special Issue Online Reputation and Sustainability)

Abstract

:
The enormous volume and largely varying quality of available reviews provide a great obstacle to seek out the most helpful reviews. While Naive Bayesian Network (NBN) is one of the matured artificial intelligence approaches for business decision support, the usage of NBN to predict the helpfulness of online reviews is lacking. This study intends to suggest HPNBN (a helpfulness prediction model using NBN), which adopts NBN for helpfulness prediction. This study crawled sample data from Amazon website and 8699 reviews comprise the final sample. Twenty-one predictors represent reviewer and textual traits as well as product traits of the reviews. We investigate how the expanded list of predictors including product, reviewer, and textual characteristics of eWOM (online word-of-mouth) has an effect on helpfulness by suggesting conditional probabilities of the binned determinants. The prediction accuracy of NBN outperformed that of the k-nearest neighbor (kNN) method and the neural network (NN) model. The results of this study can support determining helpfulness and support website design to induce review helpfulness. This study will help decision-makers predict the helpfulness of the review comments posted to their websites and manage more effective customer satisfaction strategies. When prospect customers feel such review helpfulness, they will have a stronger intention to pay a regular visit to the target website.

1. Introduction

eWOM (online word-of-mouth) plays an increasingly enormous role for consumers to obtain information on purchase decisions and as one of the WOM communications, reviews represent the bulk of user-generated content [1]. One of the critical challenges for market-driven product design to satisfy customer needs is to analyze vast amounts of online reviews to determine helpful reviews accurately [2]. Online reviews are described as “peer-generated product evaluations posted on company or third party websites” [3] or “a type of product information created by users based on personal usage experience” [4]. Online reviews increase consumer awareness and represent a credible source of information regarding product and service quality [5]. They can support manufacturers in predicting consumer reaction to their products for improving the products [6].
Recently, researchers have shown that customers, in creating online reviews, are under social influence from others [7,8]. The importance of explaining or predicting review helpfulness can be understood from its role in attracting customers’ behaviors. Review helpfulness can be treated as enabling the promotion of purchasing behavior [9]. The intention of consumers to online purchase is greatly influenced by perceived review usefulness [10]. The usefulness of eWOM information has a significant influence on customer trust [10]. Consumers will provide greater trust in review contents if reviews are valuable to satisfy their information needs [11], which makes a rigorous analytics method crucial to suggest the prediction of online review helpfulness. Thus, this so-called “social navigation” community-based voting technique [12] is much adopted to support readers in sorting out useful reviews, especially for products with a huge amount of online reviews. Online reviews can bring diagnostic value, representing their helpfulness for several stages of the decision process in purchasing. Major websites like Amazon.com present a “Most Helpful First” option in sorting reviews according to the “usefulness” dimension, using the count of readers who responded affirmatively to the question asking whether the review is useful.
This study has the following objectives. First, this study examines the previous literature on predicting helpfulness [13,14,15,16], offers an expanded list of predictors, including product, reviewer, and textual characteristics in order to explain helpfulness, and through NBN, suggests the importance of each determinant using conditional probability of the binned determinants. This may be especially insightful because review helpfulness can be affected by the characteristics of the product, reviewer, and textual traits. For the class 1 of helpfulness class, if the conditional probability of the bin with greater value exceeds that of other bins and that for the class 0 of helpfulness class, this shows that the determinant is positively affecting helpfulness. Further, through comparing the conditional probability of different determinants, we can suggest which has the greater effect on helpfulness than the other determinants. This will improve understanding of the impact of product, reviewer, and textual traits on online review usefulness in a consumer-oriented mechanism where reviews are highlighted according to their expected helpfulness.
Second, while there are several studies of using business intelligence (BI) in online reviews, the studies using NBN to predict helpfulness are rare. This study intends to overcome this gap by suggesting and validating a data mining model using NBN for determining helpfulness of reviews. Because of its simplicity and high computational efficiency, Naive Bayesian Network (NBN) is a good machine learning tool providing high prediction performance, especially appropriate for samples with high dimensions. NBN is a machine learning algorithm which is computationally data-driven and produces good performance on noisy datasets. NBN is posited to have simplicity and computational efficiency because NBN is a nonparametric method and NBN does not require assumption on the distribution of data and estimating or optimizing parameters or weights like parametric methods such as regression or neural network. Given that an enormous number of review records are required to be processed, this study focuses on investigating the performance of NBN by comparing it with other methodologies based on the studies regarding the factors affecting helpfulness [3,14,15,16,17,18]. It is necessary to use NBN in predicting helpfulness as the application a Naive Bayesian network requires less complexity and computational resources than other methods, given the large volume of review data to be processed.
The sample data crawled from Amazon.com are used to predict review helpfulness based on the traits of product, review, and text, which is represented as HPNBN (a helpfulness prediction model using NBN), where review helpfulness is transformed into a categorical predictor representing high or low value. Our study enables the automatically analyzing of review content and improvement in the ability of BI by identifying targeted customer opinions. In order to validate our model, the predictive performance of NBN is contrasted with that of multivariate discriminant analysis (MDA) and k-nearest neighbor (kNN) and the neural networks (NN) method in terms of prediction error.

2. Research Background

Online review classification becomes a critical task because of the proliferated availability of online reviews due to the advent of big data analytics technologies and the increasing need to access them in order to suggest marketing strategy. In the past several years, for text classification, many methods based on machine learning and statistical theory have been suggested to overcome information overload problems in predicting helpfulness of reviews. An enormous number of product reviews on online sites exist to attract customers’ attention. Using data analytics becomes much more necessary as it becomes crucial for website management to choose and improve helpfulness of reviews to lower the likelihood that consumers stop vising their websites and to improve the attractiveness of their websites [9]. While researchers have suggested effective solutions in the field of text categorization [19], studies on data mining for determining the online review helpfulness need to be expanded in order to provide coherent and cohesive insights [20].
Recent studies using data mining techniques in various online product review analyses include applying data mining models to predict review helpfulness [16,17,21,22,23]. Various BI methods are employed such as k-nearest neighbors (kNN) [24], inductive learning based on rough set theory [25], ordinal logistic regression analysis [21], backpropagation neural networks [26], a neural deep learning model [27,28,29,30], support vector regression [31], and content analysis to assess text comments using logistic regression [18].
Although there exist many studies using BI in online reviews, the studies using NBN to predict helpfulness are scarce. It is required to adopt NBN in predicting helpfulness as the application of a Naive Bayesian network demands less complexity and computational resources than other methods.

3. Determinants of Helpfulness

The quality assessment becomes crucial as a great number of reviews that are different in quality are posted everyday [32]. Helpfulness can be described as the ratio of positive votes out of total votes a review receives [33]. As helpfulness is often treated as a binary helpfulness measure according to whether the original percentage is greater than a threshold value [20,23], our study considers helpfulness as a binary value showing “helpful” or “unhelpful” after applying threshold point of 0.5, which is generally suggested as a well adopted threshold in the previous literature [21,24]; when helpfulness is greater than 0.5, the review becomes “helpful”, otherwise, the review becomes “unhelpful”.
While there exist a large number of studies on review helpfulness prediction, the rationale regarding specific selection of features is still vague and needs further research [34,35,36,37]. As this study intends to encompass various aspects of reviews beyond text, our study categorizes various factors affecting review usefulness according to Baek et al. [38], Filieri [11], Filieri et al. [39], Ghose and Ipeirotis [17], and Lee and Cheoh [26]. Our study classifies determinants to helpfulness as product, reviewer, and text-related features as suggested in Figure 1, denoting the specific previous studies supporting their effects.
For product traits, the review usefulness of a specific product is affected by the review activity level of that product [17,26,41]. Chen and Tseng [40] discovered that reviews with a number of comments on multiple features of products turned out to be high-quality reviews. Consumers demand different information to support purchase decisions on search or experience goods [3]. For a search good, it is easier for customers to obtain objective and comparable information about product quality before purchasing. An experience good is having subjective attributes which are hard to compare and to obtain information regarding product quality before purchasing.
Based on the previous literature which covers product, reviewer, and textual characteristics separately, this study integrates these previous models and suggests the model of determinants for review helpfulness which are classified into product, reviewer, and review as in Figure 1. The product data encompass (1) type of product (experience goods = 1, search goods = 0), (2) sales rank, (3) product’s total number of reviews, (4) product’s review rating, and (5) retail price of product.
For reviewer traits, reviewer rank, disclosure of reviewer name, visibility of source, expertise, and trustworthiness affect individuals’ perception of product, message, and reviewer [9,14,43,44,45]. O’Mahony and Smyth [23] suggest the categories of the features for helpfulness as user reputation, social, and sentiment. The authorship of a product review is positively associated with the review persuasiveness [9]. A reviewer’s reputation represented by reviewer helpfulness is a crucial intrinsic quality for review helpfulness [48] and is directly related to review trustfulness [6]. Consumers are encouraged to contribute to post reviews to maintain reputation. Reviewer ranking is crucial for review helpfulness of experience goods [38]. Reviewer’s expertise and disclosure of reviewer name turned out to be important predictors for the helpfulness from meta-analysis of the eWOM literature [13]. Reviewer engagement features include reputation, commitment, and current activity (the total number of reviews by reviewer) [31].
Reviewer attributes also have an effect on consumer perceptions of reviews [18]. The reviewer traits include (6) reviewer ranking, (7) reviewer’s total helpfulness votes, (8) reviewer’s total votes, (9) proportion of all past helpful votes (among the total count of votes) which is the reviewer history macro, (10) reviewer’s total number of reviews, and (11) reviewer name disclosure.
For textual traits, review depth and review rating are important determinants to helpfulness [46]. The review content-related signals are the most important predictors for helpfulness [15,22,46]. Data filtering with sentiment analysis is likely to increase the evaluation performance in determining useful reviews [49]. Reviews with extreme arguments tend to be considered as more helpful [21]. Review depth (the number of arguments) is positively related to agreements by providing readers more to think about [38], and can improve diagnosticity and helpfulness [3] and perceived value of the reviews. The linguistic style cues (such as counts of wh-words like where and which, sentences, and words) exert a positive influence on helpfulness.
The textual traits represent the stylistic traits which can have an effect on helpfulness [17,21]. The textual traits include (12) elapsed days after launching until posting, (13) elapsed days since posting until parsing, (14) number of words in a review, (15) number of misspelt words, (16) number of one letter long words, (17) number of words 2–9 letters long, (18) number of words 10+ letters long, (19) review rating, (20) review extremity indicating the absolute deviation of the reviewers’ ratings from the average ratings, and (21) review’s total helpfulness votes.
We suggest how product, reviewer, and textual characteristics of eWOM have an effect on helpfulness by indicating conditional probabilities of variables. For the class 1 of helpfulness class, if the conditional probability of the bin with greater value is greater, this shows that the determinant is positively affecting helpfulness. For instance, the last bin of the product’s review rating has the greatest probability among all bins for the class 1 of helpfulness class, and provides greater probability than for class 0 of the helpfulness class, and this shows that product’s review rating is positively affecting helpfulness. Further, by comparing the conditional probability of different determinants, we can suppose which has the greatest effect on helpfulness than the other determinants.

4. Naive Bayesian Network (NBN) in This Study

The Bayes classifier is based on three principles [50]: (1) Search out the other cases where the predictor values are identical to new case; (2) Decide what classes they all belong to or which class is majority class; (3) Regard that class as the prediction class for the new case. The Bayesian network, which is a directed and acyclic graph, is widely used where training time is linear in both the number of instances and attributes. The nodes and arcs of the Bayesian network are attributes and attribute dependencies, respectively. The conditional probabilities pertaining to each node show the attribute dependencies. The nodes in a Bayesian network are represented by the set of attributes { X 1 , , X p } , and the   attribute   values   of   an   instance   E i   are   ( x 1 , , x p ) . The top node belonging to a Bayesian network is represented by the class variable C , and the value which C has is c. The following represents Bayesian network classifier
C ( E )   =   arg   max c C   P ( c ) P ( x 1 , , x p | c )
The Bayes modification includes the following procedures:(1) For class 1, produce the individual probabilities that each predictor value in the case to be classified ( x 1 , , x p ) occurs in class 1; (2) Multiply these probabilities each other, then multiplies the proportion of cases belonging to class 1; (3) Repeat steps 1 and 2 for all classes; (4) Calculate a probability for class i by taking the value produced in step 2 for class i and dividing it by the sum of such values for all classes; (5) For the set of predictor values, the prediction class is determined as the class with the highest probability value. Given that all attributes are conditionally independent, the following is the Bayes formula for the probability that a case has outcome class C i :
P n b ( C i | x 1 , , x p )   =   P ( x 1 , , x p | C i ) P ( C i )   P ( x 1 , , x p | C 1 ) P ( C 1 ) + + P ( x 1 , , x p | C m ) P ( C m )
The Naive Bayes formula to calculate the probability that a case with a given set of predictor values x 1 , , x p belong to class 1 among m classes is as follows:
P n b ( C 1 | x 1 , , x p )   =   P ( C 1 ) [ P ( x 1 | C 1 ) P ( x 2 | C 1 ) P ( x p | C 1 ) ] P ( C 1 ) [ P ( x 1 | C 1 ) P ( x 2 | C 1 ) P ( x p | C 1 ) ] + + P ( C m ) [ P ( x 1 | C m ) P ( x 2 | C m ) P ( x p | C m ) ]
Due to the computational simplicity, it is easy to construct a classifier for Naive Bayes. Its prediction performance is close to that of perceptron networks or decision trees [51]. The assumption of conditional independence leads to its computation efficiency and simplicity of NBN. As the simple form for the Bayesian network classifier, the classifier for Naive Bayes is as follows:
c ( E ) =   arg   max c C   P ( c ) i   =   1 p P ( x i | c )
Although there exist many studies of using BI in online reviews, the studies adopting NBN to predict helpfulness are scarce. This study purports to investigate this gap by investigating and testing a data mining model for predicting helpfulness of reviews. Given the enormous number of review records to be analyzed, NBN is competitive in predicting helpfulness because of requiring less complex and computational resources than other methods.
Our model based on Naive Bayesian Network has advantages in that firstly, it employs crucial determinants for helpfulness, as suggested in Figure 1, and secondly, our model can allow no underlying assumptions of determinants, providing simplicity and computational efficiency by adopting a simplistic cutoff probability method which establishes a cutoff probability (mostly 0.5) for the helpfulness class. While the conditional independence assumption results in biased posterior probabilities, it is easy to build a Naive Bayes classifier due to the simplicity of computing the probability that a case has outcome class C i . Our model has important merit in selecting crucial determinants affecting online review helpfulness. In order to accomplish this merit, this study crawled sample data from Amazon.com and 8699 reviews comprise the final sample, and 21 predictors represent reviewer and textual traits as well as product traits of the reviews. In June 2014, the sample was crawled and the year date of reviews range from 1998 to 2013 (see Table 1 and Table A1 (Appendix A) for the distributions of sample and variables). All the reviews are in Korean. We crawled the site using software aid. The software used to crawl the sample includes Python and Beautifulsoup. We include all the records once they were crawled from Amazon.
The types of products are shown in Table 1. Only product type and the disclosure of reviewer name among the independent variables are categorical variables. In order to experiment with the Bayesian classifier, we discretize continuous variables into multiple bins (see last column in Table A1). The number of bins for each variable is determined such that the count of values pertaining to each bin is same. The value for each bin is given by the mean of the values belonging to each bin. In order to examine the importance for each variable for helpfulness, the mean difference for each of the 21 variables between the helpful group and non-helpful group is contrasted based on the training sample in Partition ID of 1 (Table 2). Except the number of words one letter long, all variables show significant difference between these groups. This study continues to use the number of words one letter long as it is theoretically posited to be important variable in previous studies such as Cao et al. [21] and Ghose and Ipeirotis [17].

5. Results

To validate our model performance, this study compares the predictive performance of the NBN model with that of MDA, kNN, and NN, which are widely used statistical and artificial intelligence (AI) analytics models for forecasting categorical outcomes. kNN (best k method) or case-based reasoning (CBR) are for experience-based problem solving and learning by retrieving past solutions and by adapting past solutions to solve new problems [50]. MDA is chosen as a reference method as this is a highly popular and powerful statistical classification method used as a benchmarking method like regression [50,52]. When we suppose a linear relationship among variables, MDA provides computationally fast results, classifying a large number of records. kNN is chosen as a reference method because in order to compare NBN with another AI method, our study chooses another nonparametric method like kNN in order to compare the predictive capability of these two nonparametric AI methods. kNN selects from one k-nearest neighbor and determines the best k which results in the lowest error rate in test set. Our study compares NBN with NN as NN is a popular method with high tolerance to noisy data and the ability to capture a highly complicated relationship among variables [53], which are the traits that non-parametric methods like kNN or NBN do not provide. NN is chosen to compare NBN with a parametric method based on optimized weights of networks. This study uses one hidden layer with the same number of nodes in the hidden layer as the number of nodes in the input layer.
To investigate the prediction performance of NBN, MDA, kNN, and NN, the study sample is divided into two subsamples—a training sample and a test sample, which are composed of 8449 and 250 records, respectively. The n-fold cross validation of samples is utilized to show the stability of the comparison results. A total of 8699 records are divided into 35 subsets, each of which has 250 records, respectively. When each of these subsets is used as a test sample one by one, the records left over are used as a training sample (Table 3). That is, for comparing the prediction performance among data mining methods, 35 pairs of training and test samples are built, each having 8449 and 250 records, respectively. Based on each training sample, BI methods create a prediction result for the target variables in the corresponding test sample. The prediction error is computed using a classification rate for helpfulness class. The classification rates are then averaged across a number of test samples. Further, in order to reduce the possibility of overfitting and collinearity, our study reduces the number of variables into 12 variables after considering the theoretical significance in their relation with helpfulness and the number of bins (variables with a larger number of bins are avoided as they can cause overfitting in NBN): (1) type of product (experience goods = 1, search goods = 0), (2) product’s total number of reviews, (3) product’s review rating, (4) reviewer name disclosure, (5) elapsed days after launching until posting, (6) elapsed days since posting until parsing, (7) number of misspelling words, (8) number of length one words, (9) number of length 2–9 words, (10) number of length 10 words, (11) review rating, and (12) review extremity indicating the absolute deviation of the reviewers’ ratings from the average ratings. From 35 test samples, the predictive ability of the NBN, MDA, kNN, and NN can be suggested from their misclassification ratios (Table 3).
The prediction of helpfulness class in the corresponding test sample is produced using each trained NBN, MDA, kNN, and NN model. In order to examine the extent that there exists a major difference among errors in three models, a paired t-test is used for the average misclassification ratios from 35 test samples. NBN is comparable to MDA as the t-value (t = −0.023) is not so large. NBN largely outperformed kNN and NN, as the t-value is very large and significant (t = −5.319 p < 0.000; t = −7.244, p < 0.000). We used a smaller number (12) of input variables for comparison of prediction performance and thus, overfitting or multicollinearity due to the number of variables can be a smaller issue in classification rate. The significant difference in classification rate between NBN and kNN or NN can be due to the binned input variables, which probably lower the prediction performance of kNN or NN.
We investigate how the past traits of eWOM have an effect on helpfulness by suggesting conditional probabilities of variables. To provide the conditional probabilities of NBN, MDA, the 8699 cases are randomly divided into training sample (60%) and test sample (40%), each having 5219 and 3480 cases, respectively. The conditional probabilities are suggested in Table 4. The greater the conditional probability for some specific value of predictors, the greater the importance of that variable with that specific value. For instance, the conditional probability for binned proportion of review rating having 5, given that output class (helpfulness class) equal to 1 is 0.66, is greater than 0.51 in the case of the helpfulness class being equivalent to 0. This shows that review rating contributes to helpfulness class.
Table 5 shows one example of prediction performance of training and test samples using NBN and the accuracy of test data is not much lower than that of the training sample, showing that the prediction model is trained appropriately for effective predicting.
The lift chart in Figure 2 is to validate performance of the NBN model against “no model, pick randomly”, which classifies the class for all cases in the test sample as the most frequent class and is usually adopted as the benchmark [50]. The large gap between the two lines in the lift curve indicates the ability of the NBN model to predict the important class (i.e., helpfulness class = 1), relative to its average prevalence.
By using product, reviewer, and textual traits of eWOM and using the crawled sample of 8699 reviews and 21 predictors of eWOM, we explain how the past characteristics of eWOM exert an influence on helpfulness by suggesting conditional probabilities of determinants. The greater conditional probability for some specific value of determinants represents the greater effect of that determinant with that specific value. The classification accuracy of NBN outperformed that of the k-nearest neighbor (kNN) method, which indicates the main premise of our method. This shows that NBN can be utilized to search out helpful reviews, providing an effective marketing solution. This indicates better marketing practice in online reviews because NBN can decrease the processing loads required to classify helpful reviews from a large number of review records.

6. Implications and Conclusions

6.1. Implications for Researchers

This study provides several implications for researchers. First, this study is one of a few research works to use NBN in predicting helpfulness. Given the enormous volume of review records to be processed, this study focuses on examining the performance of NBN by comparing with other methodologies. It turns out that the NBN model is comparable to conventional statistical analysis i.e., MDA, and largely outperformed the other BI method, i.e., kNN, in prediction performance, indicating that the proposed NBN method has merits in especially high dimensional data, as in this sample of online reviews where the relationships between independent and dependent variables (helpfulness) have complexity. NBN is one of the machine learning algorithms which is computationally based, data-driven, and unlike statistical models, does not require assumptions on data distribution, and performs well on noisy datasets [54,55], and is a rigorous machine learning method particularly suitable for samples with high dimensions [56]. NBN in this study can help overcome the prediction difficulty caused by the large variation exhibited by review content and quality, and enables the analyzing of a large number of online consumer reviews, which await full analysis of their characteristics. NBN is well suited for learning from nonlinear relationships among variables which do not demand prior knowledge of specific functional form. Our study tested a smaller number of determinants for comparison of prediction performance and this shows that overfitting or multicollinearity from the number of determinants can be less of an issue in predicting helpfulness. The significant difference between NBN and kNN or NN can be explained as that the binned input variables possibly decrease the prediction performance of kNN or NN.
Second, this study develops the previous literature on predicting helpfulness to provide an expanded list of predictors including product, reviewer, and textual data in order to explain helpfulness. This may be especially insightful because review helpfulness can be affected by the characteristics of the product, reviewer, and textual characteristics. For the class 1 of helpfulness class, if the conditional probability of the bin with greater value turns out to be greater than for the class 0 of helpfulness class, this indicates that the determinant is positively affecting helpfulness. For example, the last bin of product’s review rating has the greatest probability among all bins for the class 1 of helpfulness class, and provides greater probability than for the class 0 of helpfulness class; this shows that product’s review rating is positively affecting helpfulness. Further, by comparing the conditional probability of different determinants, we can suppose which determinant has the greatest effect on helpfulness than the other determinants. This will improve understanding of the impact of product, reviewer, and textual traits on online review helpfulness in a consumer-oriented mechanism where reviews are highlighted according to their expected helpfulness.
There are limitations and future research issues. Our study is data-driven and thus, the results can be different with another dataset. When using a different dataset, the selection of variables can be adjusted or reduced to encompass more crucial variables, and the number of data in this study can be expanded to provide more stable results. The determinants composed in this study can also be expandable to include more recent data and future studies can consider other variables with significant relationships with helpfulness. The relationship between determinants and helpfulness should be further investigated on a theoretical basis of eWOM research. Further, the NBN method used in this study can be expanded further to accommodate more unique hybrid methods in predicting helpfulness of online review comments. For instance, deep learning methods like LSTM (Long Short-Term Memory) and CNN (Convolutional Neural Network) can be integrated with NBN to create a more differentiated approach.

6.2. Implications for Practitioners

This study provides implications for practice. NBN can suggest a promising method to rate and highlight very recently posted reviews which readers have not spent enough time on rating. With the selection of variables having more a crucial relationship with helpfulness, the expanded number of recent data than our current sample can provide more useful results than other methods in terms of online marketing business practice. This can help in finding helpful reviews better and e-commerce and social media websites can improve their existing filtering and rating process. This study can provide advantages to social network sites by supporting them to evaluate new reviews rapidly and adjust their ranking for presentation in an efficient way. Our model supports buyers to create better reviews and thus, assist other buyers in determining their buying decisions. The results of this study can offer guidelines for building helpful reviews and for implementation of consumer online review sites to facilitate helpfulness votes. The study provides several clues about online review site design by indicating the impact of eWOM traits on review helpfulness. The effective prediction of helpfulness is important because helpful reviews can greatly enhance attitudes toward purchasing online and potential benefits to companies in terms of sales [14]. The review sites with the crucial eWOM traits for helpfulness can contribute to the sales and market share of companies.
This study can be utilized to concentrate on the factors enhancing helpfulness like review extremity, reviewer ranking, and reviewer name disclosure to adjust the design of review websites and rank reviews according to these traits. Marketers should evaluate important aspects of eWOM for helpfulness in building e-marketing strategies. Using the prediction results, an online review site can rank the new reviews rapidly so that the more helpful reviews can be displayed prominently. Those helpful reviews tend to receive the consumer attention for that product. The manufacturer could evaluate helpful reviews in building an improved manufacturing and marketing strategy.

6.3. Conclusions

Applying a data mining approach to eWOM is invaluable to recommend helpful reviews [22] because eWOM is a crucial way to understand and predict consumer needs and purchase decisions. This paper intends to predict review helpfulness with NBN by using various characteristics of eWOM and using a crawled sample of 8699 reviews and 21 predictors of eWOM. We investigate how the past characteristics of eWOM have an effect on helpfulness by suggesting conditional probabilities of variables. The greater conditional probability for some specific value of predictors indicates the greater importance of that variable with that specific value. The prediction accuracy of NBN outperformed that of the k-nearest neighbor (kNN) method, which shows the validity of our method. Our study can be used to find helpful reviews, which can turn out to be a useful marketing tool. Our study can support practice in online review marketing as NBN would reduce the efforts necessary to predict review helpfulness from an enormous volume of eWOM data.

Author Contributions

Conceptualization, S.L.; methodology, S.L., J.Y.C.; software, S.L., J.Y.C.; validation, J.Y.C.; formal analysis, S.L., J.Y.C.; resources, K.C.L.; data curation, S.L. and J.Y.C.; writing—original draft preparation, S.L.; writing—review and editing, K.C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Distribution of research variables (NA = not applicable).
Table A1. Distribution of research variables (NA = not applicable).
CategoryVariablesMinMaxMeanStandard DeviationNumber of Bins
ProductType of Product010.410.49NA
Sales Rank131,183,28030,481.5048,288.011394
Product’s Total Number of Reviews138537.2940.275124
Product’s Review Rating1.505.004.370.4669
Retail Price of Product0.0182.9911.759.19149
ReviewerReviewer Ranking09,712,9591,572,362.553,448,559.04664
Reviewer’s Total Helpfulness Votes074,1513755.598203.176232
Reviewer’s Total Votes081,3804884.289675.629187
Proportion of all Past Helpful Votes0.001.000.550.4311
Reviewer’s Total Number of Reviews111,351902.981495.14135
Reviewer Name Disclosure010.350.48NA
ReviewElapsed Days after Launching until Posting050132243.511231.30327
Elapsed Days since Posting until Parsing150471960.871150.1630
Number of Words in a Review61361140.76130.36111
Number of Misspelt Words1262.642.04129
Number of Words One Letter Long1121.570.77114
Number of Words 2–9 Letters Long135145.0126.29117
Number of Words 10 Letters Long1262.562.02165
Review Rating154.341.0142
Review Extremity0.003.800.660.6376
Reviews’ Total Helpfulness Votes02053.6710.178703
Output VariableHelpfulness Class0.001.000.550.43NA

References

  1. Kroenke, D.M. Experiencing MIS, 3rd ed.; Prentice Hall: Upper Saddle River, NJ, USA, 2012. [Google Scholar]
  2. Liu, Y.; Jin, J.; Ji, P.; Harding, J.A.; Fung, R.Y.K. Identifying helpful online reviews: A product designer’s perspective. Comput. Aided Des. 2013, 45, 180–194. [Google Scholar] [CrossRef]
  3. Mudambi, S.M.; Schuff, D. What makes a helpful online review? A study of customer reviews on Amazon.com. MIS Q. 2010, 34, 185–200. [Google Scholar] [CrossRef] [Green Version]
  4. Chen, Y.; Xie, J. Online consumer review: Word-of-mouth as a new element of marketing communication mix. Manag. Sci. 2008, 54, 477–491. [Google Scholar] [CrossRef] [Green Version]
  5. Li, X.; Hitt, L.M. Self-selection and information role of online product reviews. Inf. Syst. Res. 2008, 19, 456–474. [Google Scholar] [CrossRef] [Green Version]
  6. Ku, Y.-C.; Wei, C.-P.; Hsiao, H.-W. To whom should I listen? Finding reputable reviewers in opinion-sharing communities. Decis. Support Syst. 2012, 53, 534–542. [Google Scholar] [CrossRef]
  7. Kunja, S.R.; Gvrk, A. Examining the effect of eWOM on the customer purchase intention through value co-creation (VCC) in social networking sites (SNSs): A study of select Facebook fan pages of smartphone brands in India. Manag. Res. Rev. 2018, 43, 245–269. [Google Scholar] [CrossRef]
  8. Wang, A.; Zhang, M.; Hann, I.H. Socially nudged: A quasi-experimental study of friends’ social influence in online product ratings. Inf. Syst. Res. 2018, 29, 641–655. [Google Scholar] [CrossRef] [Green Version]
  9. Li, M.; Huang, L.; Tan, C.-H.; Wei, K.-K. Helpfulness of online product reviews as seen by consumers: Source and content features. Int. J. Electron. Commer. 2013, 17, 101–136. [Google Scholar] [CrossRef] [Green Version]
  10. Elwalda, A.; Lü, K.; Ali, M. Perceived derived attributes of online customer reviews. Comput. Hum. Behav. 2016, 56, 306–319. [Google Scholar] [CrossRef] [Green Version]
  11. Filieri, R. What makes online reviews helpful? A diagnosticity-adoption framework to explain informational and normative influences in e-WOM. J. Bus. Res. 2015, 68, 1261–1270. [Google Scholar] [CrossRef]
  12. Gilbert, E.; Karahalios, K. Understanding deja reviewers. In Proceedings of the CSCW 2010: The 2010 ACM Conference on Computer Supported Cooperative Work, Savannah, GA, USA, 6–10 February 2010; pp. 225–228. [Google Scholar]
  13. Hong, H.; Xua, D.; Wang, G.A.; Fan, W. Understanding the determinants of online review helpfulness: A meta-analytic investigation. Decis. Support Syst. 2017, 102, 1–11. [Google Scholar] [CrossRef]
  14. Lee, S.; Choeh, J.Y. The determinants of helpfulness of online reviews. Behav. Inf. Technol. 2016, 35, 853–863. [Google Scholar] [CrossRef]
  15. Siering, M.; Muntermann, J.; Rajagopalan, B. Explaining and predicting online review helpfulness: The role of content and reviewer-related signals. Decis. Support Syst. 2018, 108, 1–12. [Google Scholar] [CrossRef]
  16. Singh, J.P.; Irani, S.; Rana, N.P.; Dwivedi, Y.K.; Saumya, S.; Roy, P.K. Predicting the “helpfulness” of online consumer reviews. J. Bus. Res. 2017, 70, 346–355. [Google Scholar] [CrossRef]
  17. Ghose, A.; Ipeirotis, P.G. Estimating the helpfulness and economic impact of product reviews: Mining text and reviewer characteristics. IEEE Trans. Knowl. Data Eng. 2011, 23, 1498–1512. [Google Scholar] [CrossRef] [Green Version]
  18. Weathers, D.; Swain, S.D.; Grover, V. Can online product reviews be more helpful? Examining characteristics of information content by product type. Decis. Support Syst. 2016, 79, 12–23. [Google Scholar] [CrossRef]
  19. Bai, X. Predicting consumer sentiments from online text. Decis. Support Syst. 2011, 50, 732–742. [Google Scholar] [CrossRef]
  20. Pan, Y.; Zhang, J.Q. Born unequal: A study of the helpfulness of user-generated product reviews. J. Retail. 2011, 87, 598–612. [Google Scholar] [CrossRef]
  21. Cao, Q.; Duan, W.; Gan, Q. Exploring determinants of voting for the ‘helpfulness’ of online user reviews: A text mining approach. Decis. Support Syst. 2011, 50, 511–521. [Google Scholar] [CrossRef]
  22. Krishnamoorthy, S. Linguistic features for review helpfulness prediction. Expert Syst. Appl. 2015, 42, 3751–3759. [Google Scholar] [CrossRef]
  23. O’Mahony, M.P.; Smyth, B. A classification-based review recommender. Knowl. Based Syst. 2010, 23, 323–329. [Google Scholar] [CrossRef] [Green Version]
  24. Lee, S.; Choeh, J.Y. Exploring the determinants of and predicting the helpfulness of online user reviews using decision trees. Manag. Decis. 2017, 55, 681–700. [Google Scholar] [CrossRef]
  25. Chung, W.; Tseng, T.-L. Discovering business intelligence from online product reviews: A rule-induction framework. Expert Syst. Appl. 2012, 39, 11870–11879. [Google Scholar] [CrossRef]
  26. Lee, S.; Choeh, J.Y. Predicting the Helpfulness of Online Reviews using Multilayer Perceptron Neural Networks. Expert Syst. Appl. 2014, 41, 3041–3046. [Google Scholar] [CrossRef]
  27. Du, J.; Rong, J.; Michalska, S.; Wang, H.; Zhang, Y. Feature selection for helpfulness prediction of online product reviews: An empirical study. PLoS ONE 2019, 14, e0226902. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Fan, M.; Feng, C.; Guo, L.; Sun, M.; Li, P. Product-Aware Helpfulness Prediction of Online Reviews. In Proceedings of the WWW’19: The World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2715–2721. [Google Scholar] [CrossRef]
  29. Malik, M.S.I. Predicting users’ review helpfulness: The role of significant review and reviewer characteristics. Soft Comput. 2020, 24, 13913–13928. [Google Scholar] [CrossRef]
  30. Olatunji, I.E.; Li, X.; Wai Lam, W. Context-aware helpfulness prediction for online product reviews. In Proceedings of the 15th Asia Information Retrieval Societies Conference, AIRS 2019, Hong Kong, China, 7–9 November 2019. [Google Scholar] [CrossRef]
  31. Ngo-Ye, T.L.; Sinha, A.P. The influence of reviewer engagement characteristics on online review helpfulness: A text regression model. Decis. Support Syst. 2014, 61, 47–58. [Google Scholar] [CrossRef]
  32. Min, H.-J.; Park, J.C. Identifying helpful reviews based on customer’s mentions about experiences. Expert Syst. Appl. 2012, 39, 11830–11838. [Google Scholar] [CrossRef]
  33. Shannon, K.M. Predicting Amazon review helpfulness ratio. Tech. Rep. 2017. [Google Scholar] [CrossRef]
  34. Almutairi, Y.; Abdullah, M.; Alahmadi, D. Review helpfulness prediction: Survey. Period. Eng. Nat. Sci. 2019, 7, 420–432. [Google Scholar] [CrossRef]
  35. Du, J.; Rong, J.; Wang, H.; Zhang, Y. Helpfulness prediction for online reviews with explicit content-rating interaction. In Web Information Systems Engineering—WISE 2019; Cheng, R., Mamoulis, N., Sun, Y., Huang, X., Eds.; Springer: Cham, Switzerland, 2019; Volume 11881. [Google Scholar]
  36. Chiriatti, G.; Brunato, D.; Dell’Orletta, F.; Venturi, G. What Makes a Review Helpful? Predicting the Helpfulness of Italian TripAdvisor Reviews; Istituto di Linguistica Computazionale “Antonio Zampolli” (ILC–CNR): Pisa, Italy, 2019. [Google Scholar]
  37. Haque, M.E.; Tozal, M.E.; Islam, A. Helpfulness prediction of online product reviews. In Proceedings of the 18th ACM Symposium on Document Engineering 2018, Halifax, NS, Canada, 28–31 August 2018. [Google Scholar] [CrossRef]
  38. Baek, H.; Ahn, J.; Choi, Y. Helpfulness of online consumer reviews: Readers’ objectives and review cues. Int. J. Electron. Commer. 2012, 17, 99–126. [Google Scholar] [CrossRef]
  39. Filieri, R.; Raguseo, E.; Vitari, C. When are extreme ratings more helpful? Empirical evidence on the moderating effects of review characteristics and product type. Comput. Hum. Behav. 2018, 88, 134–142. [Google Scholar] [CrossRef]
  40. Chen, C.C.; Tseng, Y.-D. Quality evaluation of product reviews using an information quality framework. Decis. Support Syst. 2011, 50, 755–768. [Google Scholar] [CrossRef]
  41. Kwon, O.; Sung, Y. Shifting selves and product reviews: How the effects of product reviews vary depending on the self-views and self-regulatory goals of consumers. Int. J. Electron. Commer. 2012, 17, 59–82. [Google Scholar] [CrossRef]
  42. Cheung, C.M.; Lee, M.K. What drives consumers to spread electronic word of mouth in online consumer-opinion platforms. Decis. Support Syst. 2012, 53, 218–225. [Google Scholar] [CrossRef]
  43. Dou, X.; Walden, J.A.; Lee, S.; Lee, J.Y. Does source matter? Examining source effects in online product reviews. Comput. Hum. Behav. 2012, 28, 1555–1563. [Google Scholar] [CrossRef]
  44. Filieri, R.; Raguseo, E.; Vitari, C. What moderates the influence of extremely negative ratings? The role of review and reviewer characteristics. Int. J. Hosp. Manag. 2019, 77, 333–341. [Google Scholar] [CrossRef]
  45. Huang, A.H.; Chen, K.; Yen, D.C.; Tran, T.P. A study of factors that contribute to online review helpfulness. Comput. Hum. Behav. 2015, 48, 17–27. [Google Scholar] [CrossRef]
  46. Korfiatis, N.; García-Bariocanal, E.; Sánchez-Alonso, S. Evaluating content quality and helpfulness of online product reviews: The interplay of review helpfulness vs. review content. Electron. Commer. Res. Appl. 2012, 11, 205–217. [Google Scholar] [CrossRef]
  47. Schindler, R.M.; Bickart, B. Perceived helpfulness of online consumer reviews: The role of message content and style. J. Consum. Behav. 2012, 11, 234–243. [Google Scholar] [CrossRef] [Green Version]
  48. Lee, S.; Choeh, J.Y. The interactive impact of online word-of-mouth and review helpfulness on box office revenue. Manag. Decis. 2018, 56, 849–866. [Google Scholar] [CrossRef]
  49. Lubis, F.F.; Rosmansyah, Y.; Supangkat, S.H. Improving course review helpfulness prediction through sentiment analysis. In Proceedings of the 2017 International Conference on ICT for Smart Society (ICISS), Tangerang, Indonesia, 18–19 September 2017. [Google Scholar] [CrossRef]
  50. Shmueli, G.; Patel, N.R.; Bruce, P.C. Data Mining for Business Analytics, 3rd ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2018. [Google Scholar]
  51. Koc, L.; Mazzuchi, T.A.; Sarkani, S. A network intrusion detection system based on a Hidden Naïve Bayes multiclass classifier. Expert Syst. Appl. 2012, 39, 13492–13500. [Google Scholar] [CrossRef]
  52. Olson, D.; Shi, Y. Introduction to Data Mining; McGraw-Hill International Edition: New York, NY, USA, 2007. [Google Scholar]
  53. Bogaert, M.; Ballings, M.; Hosten, M.; Van den Poel, D. Identifying soccer players on Facebook through predictive analytics. Decis. Anal. 2017, 14, 274–297. [Google Scholar] [CrossRef] [Green Version]
  54. Farid, D.M.; Zhang, L.; Rahman, C.M.; Hossain, M.A.; Strachan, R. Hybrid decision tree and naïve Bayes classifiers for multi-class classification tasks. Expert Syst. Appl. 2014, 41, 1937–1946. [Google Scholar] [CrossRef]
  55. Gao, Y.; Xu, A.; Hu, P.J.-H.; Cheng, T.-H. Incorporating association rule networks in feature category-weighted Naive Bayes model to support weaning decision making. Decis. Support Syst. 2017, 96, 27–38. [Google Scholar] [CrossRef]
  56. Hernández-González, J.; Inza, I.; Lozano, J.A. Learning Bayesian network classifiers from label proportions. Pattern Recognit. 2013, 46, 3425–3440. [Google Scholar]
Figure 1. The model of determinants for review helpfulness (See [3,6,9,13,14,15,17,21,22,23,24,26,31,38,40,41,42,43,44,45,46,47]).
Figure 1. The model of determinants for review helpfulness (See [3,6,9,13,14,15,17,21,22,23,24,26,31,38,40,41,42,43,44,45,46,47]).
Sustainability 12 06997 g001
Figure 2. Lift chart for test sample (Partition ID = 1).
Figure 2. Lift chart for test sample (Partition ID = 1).
Sustainability 12 06997 g002aSustainability 12 06997 g002b
Table 1. The sample of the study.
Table 1. The sample of the study.
Items FrequencyPercent
Disclosure of Reviewer Name Product ClassYes305035.1
No564964.9
Baby Product10.0
Book790.9
DVD329637.9
Music502657.8
Pet Products10.0
Single detail page Miscellaneous40.0
Sports10.0
Toy440.5
Video2382.7
Video Games70.1
Wireless Devices20.0
Product TypeExperience Goods358541.2
Search Goods511458.8
Helpfulness Class1 (helpfulness greater than 0.5)465453.5
0404546.5
Total 8699100
Table 2. Mean difference between helpfulness classes (based on training sample in Partition ID of 1).
Table 2. Mean difference between helpfulness classes (based on training sample in Partition ID of 1).
Helpfulness ClassMeanStandard DeviationMean Difference tdfSignificance
(2-Tailed)
Type of Product10.390.490.053.38 5218.00.001
00.430.50
Sales Rank133,218.5450,360.34
027,343.5241,932.33−5875.02−4.545218.00.000
Product’s Total Number of Reviews133.6533.996.956.235218.00.000
040.6146.29
Products’ Review Rating14.410.44−0.08−6.195218.00.000
04.330.47
Retail Price of Product112.099.10−0.79−3.165218.00.002
011.298.99
Reviewer Ranking1894,193.682,695,149.471,558,694.8216.505218.00.000
02,452,888.504,067,635.66
Reviewer’s Total Helpfulness Votes15089.6110,331.94−2705.20−11.645218.00.000
02384.415353.88
Reviewer’s Total Votes16226.6311,737.64−2642.58−9.595218.00.000
03584.057369.19
Proportion of all Past Helpful Votes10.920.12−0.81−182.955218.00.000
00.110.19
Reviewer’s Total Number of Reviews1981.071538.58−121.45−2.835218.00.005
0859.621560.28
Reviewer Name Disclosure10.340.470.031.895218.00.059
00.370.48
Elapsed Days after Launching until Posting11995.591186.81482.3014.285218.00.000
02477.891251.35
Elapsed Days since Posting until Parsing12206.001087.19−504.27
01701.731184.53−504.27−16.035218.00.000
Number of Words in a Review1158.17130.14−44.83−13.155218.00.000
0113.34114.08
Number of Misspelt Words12.681.71−0.10−2.005218.00.045
02.581.76
Number of Words One Letter Long11.560.700.010.765218.00.449
01.570.72
Number of Words 2–9 Letters Long148.3725.77−8.63−12.445218.00.000
039.7424.10
Number of Words 10+ Letters Long12.621.72−0.18−4.085218.00.000
02.441.44
Review Rating14.530.79−0.42−15.455218.00.000
04.111.15
Review Extremity10.530.450.2514.685218.00.000
00.780.75
Review’s Total Helpfulness Votes16.3613.95
00.651.49−5.72−20.155218.00.000
Table 3. Comparison of misclassification rate (t-value for NBN-MDA = −0.023, significance = 0.982; t-value for NBN-kNN = −5.319, significance = 0.000, t-value for NBN-NN = −7.244, significance = 0.000).
Table 3. Comparison of misclassification rate (t-value for NBN-MDA = −0.023, significance = 0.982; t-value for NBN-kNN = −5.319, significance = 0.000, t-value for NBN-NN = −7.244, significance = 0.000).
Partition ID.NBNMultivariate Discriminant Analysis (MDA)K-Nearest Neighbor (kNN)Neural Network (NN)
129.628.433.239.2
226.827.636.836.8
337.639.241.642.4
43230.83635.2
524.824.827.626.8
641.239.240.844
725.224.828.826.4
831.63233.237.6
93030.83234.4
10343437.641.6
1134.831.235.236.4
1239.239.642.446.8
13363841.238
1430.43235.638
1532.43032.435.6
1635.233.238.440.4
1733.236.832.433.2
1841.644.436.842.8
1931.233.231.632
2030.824.833.629.2
213231.635.242.4
2233.636.838.438.8
2331.22833.234.4
2431.232.431.628
25343237.242.8
2634.43637.236
273234.441.237.6
283842.846.448.4
2927.628.431.630
3038.837.242.841.2
3139.63837.241.2
3231.23235.239.2
3337.237.636.439.2
3435.234.435.639.6
3533.731.235.238.7
Min24.824.827.626.4
Max41.644.446.448.4
Mean33.3533.3636.0537.55
Standard deviation4.164.864.125.23
Table 4. Conditional probabilities for example variables (using the first random partition).
Table 4. Conditional probabilities for example variables (using the first random partition).
Classes
Input Variables10
ValueProbabilitiesValueProbabilities
Product type0.000.610.000.56
1.000.391.000.44
Reviewer name disclosure0.000.650.000.64
1.000.351.000.36
Binned proportion of all past helpful votes0.000.000.000.69
0.180.000.180.09
0.460.000.460.21
0.680.170.680.00
0.850.140.850.00
1.000.681.000.00
Binned number of words in a review1.000.521.000.48
1.580.051.580.10
2.000.372.000.36
3.000.053.000.04
4.000.014.000.01
6.110.006.110.01
Binned Number of length 10 words1.000.251.000.25
2.000.182.000.15
2.560.302.560.37
3.000.103.000.10
4.000.064.000.05
5.000.045.000.03
6.000.026.000.02
7.000.017.000.01
8.000.018.000.01
9.000.019.000.00
10.000.0010.000.00
13.200.0113.200.00
Binned review rating1.000.011.000.06
2.000.022.000.06
3.000.073.000.12
4.000.244.000.25
5.000.665.000.51
Table 5. Classification matrix for predicting helpfulness class (using the first test sample in n-fold cross validation of samples).
Table 5. Classification matrix for predicting helpfulness class (using the first test sample in n-fold cross validation of samples).
5.1. Training Sample
Class# Cases# Errors% Error
13953148537.57
04496117826.20
Overall8449266331.52
5.2. Test Sample
Class# Cases# Errors% Error
1923436.96
01584025.32
Overall2507429.6

Share and Cite

MDPI and ACS Style

Lee, S.; Lee, K.C.; Choeh, J.Y. Using Bayesian Network to Predict Online Review Helpfulness. Sustainability 2020, 12, 6997. https://doi.org/10.3390/su12176997

AMA Style

Lee S, Lee KC, Choeh JY. Using Bayesian Network to Predict Online Review Helpfulness. Sustainability. 2020; 12(17):6997. https://doi.org/10.3390/su12176997

Chicago/Turabian Style

Lee, Sangjae, Kun Chang Lee, and Joon Yeon Choeh. 2020. "Using Bayesian Network to Predict Online Review Helpfulness" Sustainability 12, no. 17: 6997. https://doi.org/10.3390/su12176997

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop