1 GAZING AT THE STARS IS NOT ENOUGH, LOOK AT THE WORD CONSTRUALS TOO ABSTRACT Online reviews affect purchase decisions by providing information posted online by previous consumers. Past research on review helpfulness focused on its numerical star ratings and on the number of words included in its textual portion. This study proposes also including the qualitative information in those reviews to achieve a more comprehensive understanding of the factors that impact review helpfulness. The paper does so by looking at novel versus established products. To enable this addition the study proposes a method to classify the words/terms in online reviews into low- and high-level construal words by applying Latent Semantic Analysis (LSA). Low-level construal terms are words dealing with the experience of other consumers. High-level construal terms are words dealing with characteristics and features of the product itself. Results show that for established products, numerical ratings, such as the number of stars, and low-level construal strongly increase review helpfulness. Adding high-level construal terms actually reduced helpfulness. For novel products, the number of stars, the number of previous assessments of review helpfulness, and the verification that the reviewer bought the product add to perceived helpfulness. That is not the case of high- and low-level construal types of information. Online consumers may lack the knowledge to evaluate what textual information is relevant for novel products. Both academics and practitioners can benefit from a more comprehensive understanding of what increases the perceived helpfulness of online product reviews. KEYWORDS Online consumer reviews, e-WOM, e-commerce, review helpfulness, latent semantic analysis, construal terms 2 INTRODUCTION Online consumer reviews are an increasingly important source of information in purchase decisions, complementing and sometimes substituting other forms of business-to-consumer and consumer-to-consumer information (Chevalier and Mayzlin 2006). Online reviews provide information based on the personal experience of previous buyers of a specific product working as free "sales assistants" to help other consumers identify the product that best matches their needs and tastes (Chen and Xie 2008). The development of websites whose primary focus is to provide purchase information through consumer reviews – e.g. Consumeraffairs.com or Angieslist.com – exemplifies their growing popularity as relevant product information providers (Anderson and Simester 2014). In fact, a 2012 survey conducted by Nielsen among 28,000 Internet users from 56 countries reported that 70% of respondents considered online consumer reviews a trustworthy source of information, ranking them as the second most trusted form of advertising among 19 different choices (Nielsen 2012). Online product reviews impact sales (Chevalier and Mayzlin 2006; Chintagunta et al. 2010; Floyd et al. 2014; Liu 2006; Zhu and Zhang 2010), especially helpful reviews (Chen et al. 2008). There is, therefore, a growing interest in understanding the importance of online product reviews, as evidenced by the abundance of available literature (see Appendix B for a list of some of the most relevant papers on online consumer reviews). This study contributes to the expanding literature on what makes an online review helpful by including in the analysis the high- and low-level construal terms in the written portion of the reviews. By doing so, this paper presents an integrative approach that includes what people talk about in the textual part of each of those online reviews. 3 The degree of helpfulness that potential customers assign to a review is a common measure to assess the value of the information included in it (Mudambi and Schuff 2010). On many sites, this helpfulness measure is the summation of the positive and negative votes of the specific review by potential buyers. Review helpfulness can also represent a public endorsement of a particular review message by a peer consumer (Metzger et al. 2010). More ‘helpful’ product reviews can provide a competitive advantage due to their value as a differentiation tool and can increase the time spent on an e-commerce site (Mudambi and Schuff 2010). The importance of review helpfulness ratings has raised practical and academic interest in identifying the elements that increase the perceived helpfulness of these product reviews. Past research into what makes online consumers reviews helpful has concentrated on the easily accessible numeric characteristics, such as the number of stars and on counting the number of words included in the textual portion (e.g. Mudambi and Schuff 2010; Pan and Zhang 2011). This study introduces a new method in marketing that allows for an assessment of the qualitative content of those reviews too. The paper automatically integrates the construal meaning of words in the product review to determine how the content of those reviews, in addition to the variables employed in previous research, affects subsequent helpfulness evaluations. This study proposes a methodology to do so based on applying Latent Semantic Analysis (LSA) and Construal Level Theory (CLT) to identify low- and high-level construal words in the reviews. Subsequently, these two variables are tested along with other variables employed in previous literature to assess their importance in determining the helpfulness of the review. Rather than identify what specific words potential customers seek for a specific product, this method has the potential to classify the words in any review of any product into low- and high-level construal words applying CLT and then suggest when low- or high-level construal words, or both, also contribute in forming 4 helpfulness assessments. This paper tests its hypotheses and the revised theoretical model over product reviews relating to 16 types of light bulbs (LEDs and incandescent light bulb) available for purchase at Amazon.com. The contributions of this study are in: (1) proposing a new methodology based on LSA to identify key terms in online reviews, (2) showing how CLT can be applied to classify those key terms into low- and high-level construal terms, (3) developing a framework to assess the impact of the informational elements available at the individual review level on its assessed helpfulness and (4) showing that the relative importance of the variables differs depending on whether the product is novel or established. The textual part of consumer reviews has been the focus of limited research in marketing, possibly because of a lack of an appropriate methodology (Chevalier and Mayzlin 2006), leaving previous attempts to process text with noisy results (Godes and Mayzlin 2004). Adding a textual analysis of the kind proposed in this study is important because research suggests that the textual part of product reviews may allow for richer descriptions of the interaction of buyers with the product (King et al. 2014). The remainder of the paper is organized as follows. In the second section, we develop the theoretical background, leading to the research hypotheses. The third section introduces the conceptual model. The fourth section explains the methodology. The fifth section is the data analysis. The sixth section discusses the findings, limitations, and avenues for future research. THEORETICAL DEVELOPMENT The Context. Electronic Word-of-Mouth and Review Helpfulness 5 Online consumer reviews are a form of word-of-mouth (WOM). WOM is a term that includes different types of informal communications among consumers about products, services, or companies (Liu 2006). WOM is a source of purchasing information that influences potential customers’ behavior (Brown and Reingen 1987; McFadden and Train 1996). It is often considered by consumers to be a reliable and valid source of information that can reduce purchase risks (Bickart and Schindler 2001; Godes and Mayzlin 2004; Mayzlin 2006). Possibly, this is because of its non-commercial nature and because consumers generally trust other peer consumers more than they trust marketers or advertisers (Sen and Lerman 2007). The electronic counterpart of WOM (e-WOM) is defined as “any positive or negative statement made by potential, actual or former consumers about a product or company, which is made available to a multitude of people and institutions via the Internet” (p.39, Hennig-Thurau et al. 2004). Typically, e-WOM interactions take place among people who have no previous connection or relationship. This constitutes a major difference compared to traditional WOM (Dellarocas 2003). e-WOM includes a broad range of elements in addition to online reviews such as blogs, social networking sites, online forums, and electronic bulletin boards (Floyd et al. 2014). Although there is no standard e-WOM structure, online consumer reviews usually include several informational elements. These elements include: (1) review valence (frequently referred to as ‘number of stars’); (2) review text (where reviewers provide further qualitative information); (3) the verification that the reviewer actually purchased the product or service; and (4) an overall summary helpfulness score of the review. In the case of Amazon.com this helpfulness score is implemented by answering the question “was this review helpful to you?” with a dichotomous “yes” or “no” answer (thumb up/thumb down). The helpfulness assessment 6 is reported by people other than the person who posted elements (1) and (2). This helpfulness evaluation is the dependent variable in the conceptual model of this study. Review helpfulness is “the extent to which consumers perceive the product review as being capable of facilitating judgment or purchase decisions” (p. 103, Li et al. 2013). Helpful product reviews provide potential buyers with pertinent information from previous consumers with the purpose of informing the purchasing decision process (Mudambi and Schuff 2010). Review helpfulness can be understood as the principal tool to assess how potential customers evaluate the quality of the information provided in each individual review. Information quality can reduce purchase uncertainty, which can be crucial within the e-commerce context (Mudambi and Schuff 2010). This is important because helpfulness ratings can impact sales (Chen et al. 2008; Chen and Xie 2008; Ghose et al. 2012), presumably by certifying to the value of the information they provide (Ghose and Ipeirotis 2011). As a prominent example of the relevance of the helpfulness ratings, Spool (2009) estimated that Amazon.com added $2.7 billion to its annual revenue by requesting that potential customers assess the perceived helpfulness of the review, and, through these evaluations, making the more ‘helpful’ reviews more visible to potential customers. Additionally, helpfulness ratings allow consumers to filter the overwhelming amount of information available on popular e-commerce sites. This may allow potential customers to make purchase decisions even in situations of information overload (Cao et al. 2011). The availability of product reviews can also increase the time that potential buyers spend on a particular website, thanks to the so called “stickiness” effect of product reviews (Mudambi and Schuff 2010). This time spent on a site can be increased even more if the information provided actually aids consumers in their decision making, i.e. if the information provided is in fact ‘helpful.’ 7 Theory Base and Hypotheses Allowing for this evidenced importance of e-WOM, the objective of this study is to suggest a theory base and a related methodology that also accounts for the qualitative portion of online reviews. This proposed theory base and related methodology should be applicable in a consistent manner across products and service types. This need for consistent applicability is highlighted because the specific wording that makes a review of a certain product or service helpful might not readily apply to another product or service. In an attempt to create such an overarching theory and supporting methodology, this section introduces CLT as a classifier of eWOM information into low- and high-level construal terms and LSA as a methodology to identify those terms. Being able to apply such a theory and methodology, and achieve a more comprehensive approach to the informational elements that add to helpfulness is arguably necessary because: (1) the web economy lowers search costs for most products (Alba et al. 1997; Lynch Jr and Ariely 2000); (2) the likelihood of information overload is greater due to the higher number of alternatives and to the lower searching costs, especially for the most popular e-commerce sites (Alba et al. 1997; Peterson et al. 1997; Shapiro and Varian 1999); (3) the entry costs are very low in the online marketplaces for sellers, making it very difficult for customers to differentiate among legitimate vendors and “fly-by-night” sellers (Tan 1999); (4) the spatial difference with the online retailer makes it infeasible to physically examine the product prior to purchase (Tan 1999); and (5) there is a greater assortment of products on e-commerce sites, including versioned and customized possibilities (Peterson et al. 1997; Shapiro and Varian 1999). The textual part of online product reviews can provide at least a partial solution to those issues. However, little research has looked at how to assess the value of the textual portions, 8 even in the form of subjective helpfulness assessments. This scarce attention may be due to a lack of appropriate methodologies to analyze large amounts of textual data (Chevalier and Mayzlin 2006) and due to costly analyses and noisy results (Godes and Mayzlin 2004). Although the numerical part of product reviews conveys valuable information, the textual part could arguably allow for richer description (King et al. 2014). The text may also convey additional helpful information that may have nothing to do with the actual interaction with the product such as packaging and customer service. Moreover, analyzing a numeric value rather than analyzing its related textual review ignores the possibility that the description may deal with more than one aspect of the quality construct or the uniformity of consumer preferences (Archak et al. 2011). The application of CLT with LSA provides such a methodology to assess the text portion of online reviews and add it to the existing research on what makes a review helpful. CLT and the Difference between Low- and High-Level Construal Terms A limiting issue in assessing the textual part of online product reviews is that specific terms that might make a posting helpful about one product or service may not be as helpful with another. For example, a term such as “courteous” may be important across services, but “luxurious” may be more important in contexts relating to opulence. And this is where CLT as the theory base is advantageous. CLT provides a basis for distinguishing between two different types of information: high-level and low-level construal terms (Irmak et al. 2013; Liberman and Trope 1998). High-level construal terms deal with the desirability of a product or service based on its essential features (i.e. why would customers want to buy this product or service). Considering the desirability of an object, event, or course of action, high-level construal information puts more importance on the end results (Trope 2012). High-level construal terms are abstract, simple, structured and coherent, decontextualized, focused on the primary core, 9 superordinate, and goal relevant (Trope and Liberman 2003). As such, in an e-commerce context, high-level construal terms could be understood as comprising the type of information that is typically provided by manufacturers and retailers as they describe the product itself and its relevant features. Examples of this type of information are technical details of RAM memory or the name of a concert performer. In the specific case of the database of this study, the technical details of a light bulb, such as its color temperature, constitute high-level construal information. In contrast, low-level construal terms provide information about the feasibility or specific ‘how’ nuances of the product or service. This could include details about how other customers used a product, their experience with it, and secondary features that are not as salient as those included in the main desirability core. Low-level construal terms are relatively more concrete, complex, unstructured, contextualized, focused on secondary features, subordinate, and goal irrelevant (Trope and Liberman 2003). Consumer reviews provide also such low-level construal terms (De Maeyer 2012). Previous research suggests that potential buyers value low-level construal information in some cases as much as they do high-levels construal information (Irmak et al. 2013). As Trope states, high-level construal information is a way of thinking in a bigpicture way, whereas low-level construal information is more detail-oriented (Trope 2012). Specific examples of this type of low-level construal terms might include testimonials about how pleasant the light produced by an LED is and how it allowed the person writing the review to see nuanced colors in paintings better. Low-level and high-level construal terms complement each other. Low-level construal terms provide information that allow potential customers to learn from the experiences of previous buyers. High-level construal terms reported in the textual portion of a review can support retailers’ claims. High- and Low-Level Construal Information 10 Economics of Information (EoI) provides a theoretical framework for the necessity of gathering pre-purchase information by consumers. The basic premise of EoI is that there exists an information asymmetry between buyers and sellers, i.e. potential customers are not perfectly informed about all the available product or service choices in the market (Biswas 2004). The classic EoI literature suggests that a potential buyer, operating within this asymmetric information market, would acquire information about the product until the point at which the marginal cost of further information acquisition equals or exceeds its marginal benefit (Stigler 1961). Research shows that in online settings consumers actively search for purchase information before buying (e.g. Punj and Moore 2009) and that their decisions are affected by the recommendations of others (Li and Zhan 2011; Moe and Trusov 2011). Both low- and high-level construal terms embedded in the text provided in an online consumer review could provide people reading it with such valuable information about the experiences of others, in addition to the other sources available. Such terms may carry meaning, regardless of the rest of the text. Indeed, some words such as “horrible”, a low-level construal term, and “efficient”, a high-level construal term, carry strong messages when applied in the context of light, and may do so even regardless of the rest of the sentence. It is true that meaningchanging prefixes, such as adding “not” before “good”, may confuse such an approach. Nonetheless, such prefixes are sufficiently rare in English that analyzing individual words across many documents is enough for LSA to successfully factor together words to identify synonyms (Gomez et al. 2012; Islam et al. 2012; Valle-Lisboa and Mizraji 2007), answer multiple choice exam questions as successfully as students do in an introduction to psychology course (Landauer et al. 1998b), and score comparably to nonnative speakers on TOEFL (Landauer and Dumais 1997). In the specific data analyzed in this study, meaning-changing prefixes were very rare. 11 Before discussing the hypotheses, some background discussion of LSA is necessary to clarify operationally what constitutes low-level and high-level construal terms in this context. What LSA does, more on that in the next sections, is to identify through singular value decomposition (SVD), a process somewhat equivalent to a two-way principal components analysis (PCA), what terms (e.g. words) and what documents (in this case, the online product reviews) factor together in a document-to-terms frequency matrix (DTM). In analyzing terms, LSA supports both the analysis of single words as well as phrases. (Typically, and also in this study, terms are only single words.) Terms that factor together are assumed to carry some latent shared meaning, much as factors are assumed to do in a PCA. And, as in a PCA, only those factors that explain high degrees of the variance are typically retained. The resulting retained factors identify the words that carry the most weight in explaining the variance in the DTM. In the specific analysis run in this study some of those factors contained words that were the same words as the manufacturers’ websites used to describe the products. The words associated with those factors were determined to be high-level construal terms. Note that LSA drops commonly used words such as "the", so the retained words in this category are primarily technical in nature. Words associated with the other factors were treated as low-level construal terms. In other words, LSA allowed the classification of commonly occurring high eigenvalue words in the online reviews as either replicating words that appear in the manufacturers’ websites or other information that the authors of those online reviews posted. The words that repeated the manufacturers’ websites technical information were treated as high-level construal terms. The other words were treated as low-level construal terms, relating to information other than the technical experience of the consumers. As a result of this process, it was possible to identify how many high-level construal terms and how many low-level construal terms appeared 12 in each online product review. Conceptually, this process could be thought of as the equivalent of asking human raters to identify the most influential words in the online product reviews and count the number of times each review has what seems to them subjectively as low- or as highlevel construal terms. LSA does this automatically and objectively. The Importance of Low-Level Construal Information Exploring product information allows consumers to make better purchase decisions (Kohli et al. 2004). Low-level construal terms provide information about the experience of consumers beyond what the website details. That experience can come from the actual interaction with the product to other related experiences such as product packaging and shipping. This low-level evidence is generally not easy to access through other means different from eWOM. Presumably, the more low-level construal words there are, the more helpful the review should be as a response to the need of a more informed purchase decision. More terms mean either more information, in the case of different terms, or emphasizing a point, in the case of a repetition of a term. More low-level construal words may also be the result of adding words that load on other factors, adding additional facets of information. Rationally, if people reading the post are seeking information, and if, as LSA contends, terms are facets of that information, then the more such information is provided, the more helpful the online review should be. Consistent with this, the first hypothesis posits: H1: The more low-level construal terms included in a review, the more its perceived helpfulness The Importance of High-Level Construal Information High-level construal terms add to the helpfulness of the review in an equivalent, but different, way. Consumers lack physical cues to evaluate products in electronic marketplaces. This lack of cues creates risk and uncertainty since customers cannot anticipate if the purchase 13 will result in the expected benefits. To reduce this uncertainty and risk, individuals seek to verify from other sources the claims that manufacturers make (Racherla et al. 2012; Ramirez et al. 2002). Online reviews can be one such source of verification and vice versa, they may also question claims manufacturers make (Lee et al. 2015). Research suggests that consumers largely trust other consumers more than they trust claims made by sellers (Sen and Lerman 2007). And so, if the terms a manufacturer uses in its website to describe the product, such as describing a bulb as efficient or its light as warm, are repeated in the online review then it would at least somewhat confirm what the manufacture claims. Applying the same logic as in the previous hypothesis, the more these terms are used, the stronger the message of confirmation with the claims the manufacturer makes should be. The helpfulness of an online review should be higher, the more such confirmation it provides. Following this rationale, the second hypothesis of this study is as follows: H2: The more high-level construal terms included in a review, the more its perceived helpfulness Other Sources of Information in Individual Consumer Reviews Customer ratings, such as the number of stars, are known important quantitative informational elements of e-WOM, as they can assist potential customers in learning about the quality of a product (Filieri 2015). However, existing literature has reported contrasting results regarding how customer ratings affect the decision-making process. Several papers found that extremely negative ratings are more influential, while others found that the extremely positive ones are the most relevant for this process (e.g. Hu et al. 2009; Kuan et al. 2015; Pan and Zhang 2011; Sen and Lerman 2007; Wu 2013). Under the specific conditions of consumers evaluating e-WOM helpfulness, Pan and Zhang (2011) found that positive ratings are more influential than negative ones and are rated as more helpful. In other words, these authors suggest that there is a 14 positive correlation between the number of stars and review helpfulness. If that is the case, the number of stars might be considered as a surrogate recommendation (Alex and Prendergast 2009). Furthermore, from a practical standpoint, the existence of a market for artificially inflating ratings and fake reviews (e.g. News 2013) also suggests that there should be a positive relationship between the number of stars and perceived helpfulness. Due to the strongly contrasting results reported in the literature and to achieve a comprehensive approach to the informational elements available in online reviews, this paper tests the positive relationship reported previously by Pan and Zhang (2011): H3: The more stars a review has, the more its perceived helpfulness Potential customers are not perfectly informed about all the available product or service choices in the market (Biswas 2004). Previous literature suggests that this problem can be lessened by the provision of product information by third-parties (Faulhaber and Yao 1989). Product reviews are suitable to mitigate this lack of market knowledge if their information is pertinent, but only if their source is believable (Forman et al. 2008). Filieri (2015) found that if the information source in e-WOM is assessed as credible, it will be perceived as more helpful. One such aspect of credibility is that the person posting the online product review actually bought the product. Conversely, if that person did not buy the product, the credibility in making claims about it may be questionable. Perhaps indicating how strong that indication can be, many sites – such as Amazon.com – include a verification that the reviewer actually purchased the product. Consequently, it might be expected therefore that a verification that the review writer bought the product should increase the perceived helpfulness of the review: H4: The confirmation that the reviewer purchased the product will increase the perceived helpfulness of the review 15 Accounting for Product Novelty The need for information gathering could vary substantially depending on product novelty. Novel products are relatively unknown to potential consumers, and so consumers might seek more information about those products to reduce their perceived risk in purchase (Dowling and Staelin 1994; Hawkins and Mothersbaugh 2010; Wasson 1978; Wind and Robertson 1983). Because customers seek more information in the case of novel products, it follows that when such information is available, reviews will be more helpful. Logically, it is expected that information about what the buyers do not know much about, as with a novel product, should be more helpful than telling the buyers what they presumably already know, as with an established one. Previous research agrees that this is likely the case (Berger 2013; Wu 2013). The fifth hypothesis tests if novelty impacts positively the helpfulness evaluation of a review: H5: The perceived helpfulness of a review will be higher among novel products CONCEPTUAL MODEL Summarizing the hypotheses, the conceptual model (Figure 1) expands previous research about what determines the perceived helpfulness of a review (e.g. Mudambi and Schuff 2010; Pan and Zhang 2011) by integrating into the model also the qualitative information available in the review. [Insert Figure 1 about here] A control variable, the number of previous helpfulness evaluations (Number of Evaluations) that each review received, was added to Figure 1. This variable is included in the theoretical model for two main reasons. First, the algorithm employed by Amazon.com to rank 16 the most helpful reviews is based on the aggregated number of evaluations that each review received. This feature has an “anchoring effect,” leading to more helpful reviews receiving more votes since they are made more salient for potential buyers (Cao et al. 2011; Wan and Nakayama 2014). Second, previous literature shows that consumers rely on the recommendations of other online consumers when making purchasing decisions (e.g. Forman et al. 2008; Li and Zhan 2011). This social persuasion can also impact the helpfulness assessment of a review. The number of previous helpfulness assessments can have an impact on the subsequent evaluations, as potential customers may follow the recommendations of others regarding the helpfulness perception of the review (Cialdini and Goldstein 2004; Deutsch and Gerard 1955). Previous literature also controlled for the number of helpfulness evaluations (Mudambi and Schuff 2010). The model also includes a weighting criterion, the total number of words included in the textual portion (Number of Words). Previous research (e.g. Chevalier and Mayzlin 2006) suggests that the length of the text also affects consumers. Various measures of text length have been employed, such as the number of characters, the number of words, and the number of sentences (e.g. Pan and Zhang 2011; Salehan and Kim 2016). The literature on online reviews also suggests that a positive relationship exists between the length of the textual part and review helpfulness (Pan and Zhang 2011). This finding would call for the inclusion of this variable in our model as an independent variable. (This was tested. Adding the number of words as a control variable instead of as a weight did not significantly affect perceived helpfulness and was not selected for entry in a stepwise regression. This is important, since it highlights that the high- and low-level construal words may have a more relevant impact on review helpfulness than the raw number of words.) In addition to this, the inclusion of the number of words as an independent variable may bias the analysis, since some e-commerce sites such as Amazon.com do not provide 17 reviewers with the ability to leave the textual portion blank, forcing them to add at least the number of stars in a textual format –i.e. “five stars.” Figure 2 shows the variables included in the conceptual model as they were extracted from each of the product reviews in the database: [Insert Figure 2 about here] METHODOLOGY Analyzing the text portion of the online reviews to count the number of high-level and the number of low-level construal terms in each review was performed with LSA. LSA is a fully automatic statistical method that determines the similarity of meanings across words (named “terms” in LSA) and documents (in this case online consumer reviews) by analyzing large bodies of text. Because LSA applied SVD, terms can be associated with each other even if they do not co-occur together. It is enough for two terms to both co-occur with a third term to make the former two terms close in meaning (Landauer et al. 1998a). LSA has been argued to approximate some aspects of human learning from vast amounts of text, as well as human discrimination of meaning similarity between words (Landauer et al. 2013). Those characteristics make it especially suitable for the kind of analysis performed in this study. Another advantage of the usage of LSA is that it allows for the analysis of a large number of documents, in this case product reviews, with minimal human intervention. This ability to analyze large amounts of textual data has been highlighted in previous literature as a relevant area for further research (e.g. Archak et al. 2011; Ghose et al. 2012). The application of LSA in this study was to identify factors that best describe high- and low-construal terms. 18 The dataset analyzed consisted of 759 product reviews relating to 16 types of light bulbs available for purchase at Amazon.com (8 incandescent light bulb packs and 8 LED light bulbs). Those reviews were published online from the time the product was released until February 11, 2015. The selection of this specific category is based on three main reasons: (1) lack of seasonality; (2) lack of heavy marketing activity, e.g. advertising, which reduces potential noise during the analysis; and (3) fewer reviews than other very popular products, which reduces the potential impact of other variables associated with the review – such as its posting order. The two types of products, LED and incandescent light bulbs, differ in that LEDs were relatively new in the period of the data collection. They were introduced in the market only after 2008 (Energy 2013). All incandescent light bulbs are established, mostly functionally equivalent products. For each of these 16 products the variables collected were product ID, review ID, review valence, review text, number of words included in the review text, verified purchase, number of helpfulness evaluations, and the helpfulness rating of the review. At the time of the data collection, these 759 reviews were assessed by 1,632 potential buyers on their perceived helpfulness. Additionally, a helpfulness score was calculated for each review by summing the number of positive votes minus the number of negative votes. For instance, a review with “1 of 3 people found this helpful” received a score of -1 (+1 for one potential buyer assessing the review as helpful, -2 for two potential buyers assessing the review as non-helpful). See the Review Helpfulness clause in Figure 2. Product description provided by the seller/manufacturer was analyzed to identify the high-level type of information included in the textual portion of reviews (as shown in figure 3). [Insert Figure 3 about here] 19 Stop words (such as “the” or “a”), numbers, and punctuation were removed from the body of the text. Then, a DTM was constructed from the remaining words in the product description through a standard LSA function in R. Instead of working with raw frequencies, LSA employs a transformation called “weighting.” The goal of this transformation is to significantly reduce the influence of very frequent words that carry little meaning and give more weight to those that convey more information (Landauer et al. 2013). The default most common weighting approach was applied (Dumais 2004). This approach entails the multiplication of the log of the local term frequencies (localWeight) by the inverse of the entropy of the word in the text body (globalWeight): ; Where: i = term number j = document number wordLocalFrequency = number of appearances of term i in document j wordGlobalFrequency = total number of appearances of term i in all the documents ncontexts = total number of documents From this transformed DTM, a pool of potential high-level words were obtained for selection following the criterion described in Appendix A. After applying this selection process, a total of 33 words were ascribed to the high-level representations category. ‘High-Level 20 Construals’ in the conceptual model is the sum of the number of times that any of these 33 words appeared in the textual portion of each review. To identify the low-level construal terms, the same transformed DTM was calculated over the textual portion of the reviews – as the first step of the process in Appendix A. After applying this process, a total number of 131 terms were selected. ‘Low-Level Construals’ in the conceptual model is the number of times that any of those 131 terms appeared in the textual portion of the review. The following GLM was run to test the research hypotheses: (1) Where: y = Helpfulness score x1 = Number of terms ascribed to low-level construals in the review x2 = Number of terms ascribed to high-level construals in the review x3 = Number of stars x4 = Verification of purchase: yes = 1/ no = 0. This is a classification variable x5 = Product novelty: LED = 1 / incandescent = 0. This is a classification variable x6 = Number of Evaluations. This is a control variable weight = Number of words in the textual portion of the review ANALYSIS 21 Descriptive statistics are shown in Table 1. Model 1 shows analysis results for all the data, Model 2 for incandescent bulbs only, and Model 3 for LED only. Table 1 shows that LED reviews were more helpful, had more words, and had both more high- and low-level construal terms, but fewer stars. LED, however, had fewer confirmed purchases. [Insert Table 1 about here] The research model in Figure 1 was tested with GLM, adding the number of words as a weight and specifying that product novelty and purchase verification are classification variables (i.e. are nominal rather than real numbers). Results are shown in Table 2 as they refer to the three models in Table 1. To verify the analyses, the models were run in a linear regression model too, producing equivalent results. The Durbin Watson statistic at 1.983 was insignificant. Significant coefficients in Table 2 are emphasized in bold. The results of the linear regression with standardized betas (to enable comparison among the independent variables) is shown as Model 4 in Table 2. Note that the standardized betas of the number of low-level construal terms (standardized β=.09), high-level construal terms (standardized β=-.12), and the number of stars (standardized β=.12) are close in size, suggesting that construal terms are as important as the number of stars. (The interactions of the verification of purchase and product novelty with lowlevel construals, high-level construals, and number of stars were also tested. None of these interactions were significant, and none were selected automatically for entry by stepwise regression.) [Insert Table 2 about here] The GLM analyses shows a nuanced picture. The more low-level construal terms there were in the review, the more helpful the review was, supporting H1. But this is the case only thanks to the incandescent bulb category. This suggests that potential buyers reading the reviews 22 about novel products are not very interested about the experience of other buyers with those products, but they are interested when it concerns established products. H2, that having more high-level construal terms in the report will increase its helpfulness score, was not supported. The coefficient is significant and very significant in the incandescent category but with a negative rather than positive sign. Contrary to H2, the data indicates that the more the number of high-level construal terms there are, the less helpful the review is. It seems that buyers reading the review are not interested in other buyers confirming what the manufacturer says, especially in the case of established products. Buyers reading the review are behaving as if they are treating the inclusions of such high-level construal terms as a waste of their time, reducing review helpfulness. H3, that more stars will increase helpfulness, was supported in both categories. H4, that verification of purchase will increase helpfulness, was only supported in the LED category. Buyers reading the review are behaving as if to say that they know people buy incandescent light bulbs. Therefore, that verification in the case of incandescent light bulbs adds no new information and is insignificant. Overall, reviews are more helpful if the product being reviewed is an LED as shown in Table 1. According to Table 2, H5 is not statistically supported. As expected, reviews with more evaluations (the control variable) were assessed as more helpful. A post hoc analysis was then run to verify the implication of the study that it is the number of high- and of low-level construal terms, rather than the numbers of words altogether, that determines helpfulness. In that post hoc analysis a stepwise linear regression was run with all the terms in Model 1 and including beyond those also the number of words. The stepwise regression was significant (F7,751 =133.1, p-value<.0001, R2=.55). Using the default entry and retaining values of .05, the stepwise procedure included only the number of stars (β=.44, std. error=.08, p-value<.0001), supporting H3, the control number of evaluations (β=.63, std. 23 error=.02, p-value<.0001), and the number of low-level construal terms (β=.08, std. error=.03, pvalue=.005), supporting H1. The variable number of words was not chosen by the stepwise procedure to be entered. DISCUSSION Online reviews are clearly important to consumers and their helpfulness score does increase as their number of stars increases as Pan and Zhang (2011) suggested. Verification of purchase is also an important contributor to making the review more helpful, but apparently in our data only with regard to the novel category of LED bulbs. Adding beyond previous studies of online reviews, this study shows that also the written component of the online reviews can contribute to their helpfulness, but in a perhaps less than obvious manner. Classifying the written portion of the reviews into low- and high-level construal terms (through terms that reflect how other customers employed a product and their experiences with it versus terms that replicate what the manufacturers say, respectively), shows that low-level construal terms contribute to the assessed helpfulness of the review but only with the established product category. High-level construal terms, in contrast, seem to be detrimental in making the review helpful, and significantly so with the established product category. Buyers assessing the helpfulness of the review are behaving as if online reviews that replicate terms that the manufacturers use to describe the product is counterproductive to making the reviews helpful, possibly indicating that such detail may even be a waste of their time. The importance of the number of low-level construal terms, rather than the overall number of words, as shown in the post hoc analysis, suggests that it is not the length of the 24 review that makes it helpful but the type of terms used in it. Moreover, the classification of the written portion of the reviews into low- and high-level construal terms adds insight into what increases helpfulness, but notably, also into what detracts from making a review helpful. Apparently, some kinds of information, in this case identified as high-level construal terms, are at least in these data, of no value and potentially even be detrimental to making the review helpful. This suggests that encouraging reviewers to describe specific experiences during the whole purchase process should be considered. People access online reviews because they want to learn from other consumers, but, apparently, selectively. Therefore, telling people who access online reviews about consumer experiences is being helpful. Telling them what they already know, or at least what they can learn directly from the manufacturer, is, at least in these data, counterproductive. It makes the review less helpful. Contribution to Theory Online reviews are clearly helpful. Discovering what exactly makes them helpful is the topic of this study. Specifically, what is the contribution of high- and of low-level construal terms embedded in the review to make it helpful, and presumably how do they influence the decision-making process of other interested customers? And, relatively, how important are the high- and low-level construal terms in the textual part of the review relative to its numerical stars rating? The approach this study took, analyzing low- and high-level construal terms rather than number of words as done by previous research, sheds new light on these questions. Putting this approach into context, past research that looked at the written portion of online reviews examined text length (Pan and Zhang 2011), the impact of linguistic styles (Ludwig et al. 2013), the role of emotions (e.g. Felbermayr and Nanopoulos 2016; Li and Zhan 2011), and the influence of very specific features of the product included in the textual part of the review 25 (Archak et al. 2011). Also splitting review texts into a consumer experience category and a product features category has been done before (e.g. Li et al. 2013). What is new in the approach taken in this study is that this categorization is centered on construal terms, and it is objective and numeric. This duality of the textual portion fits with CLT. While the application of CLT is not novel in marketing – especially within consumer behavior research (e.g. Irmak et al. 2013; Kim et al. 2009; Trope et al. 2007) – this theory could possibly by applied to shed light also to ecommerce research. Indeed, among the online review studies listed in the Appendix B, only Jin et al. (2014) included CLT in their theoretical framework. Using CLT to classify terms into lowlevel construal terms (related to the usage experience) and into high-level construal terms (related to the features of the product), and assessing essentially equivalent functionality over novel and established products, adds new insight into what makes reviews helpful. For example, that, at least in these data, novel products have almost double the number of low-level construal terms than the established product category (see Table 1) and almost triple the number of highlevel construal terms. The even higher ratio of high-level construal terms in novel products may be because information about features and characteristics may be easier to convey with fewer terms, due to their more objective nature (such as the color temperature of the light being '2700K'). Information on how previous buyers used the product, the description of the usage context, and other additional experiences, such as how long it took the package to be delivered, will be more varied and hence demand more words. This new approach to assessing review helpfulness expands previous research on the topic. The results suggest that is not only the “raw” number of words, i.e. how much is written, that may increase the helpfulness perception of the review – an approach common in the early literature (e.g. Chevalier and Mayzlin 2006) – but, rather, what is written as reflected by the 26 construal terms. It is not that previous research did not recognize the importance of what versus how much, but rather that apparently previous research did not have the tools to objectively do so. Indeed, Filieri (2015) suggested that potential buyers are more influenced by the quality of the information (information quality being defined by information depth and breadth, relevance, credibility, and factuality) than by information quantity. Mudambi and Schuff (2010) also suggested that information depth is an important predictor of review helpfulness. Cao et al. (2011) found that semantic characteristics (related to the ‘substance’ of the text) are the most important in determining helpfulness votes. The addition of high- and low-level construal terms allows the analysis of the text portion through perhaps more nuanced lenses. This new perspective might also apply to the emerging literature in marketing that diverts from employing just review ratings – such as the number of stars – in the analyses (Wang et al. 2015; Wu et al. 2015). Review ratings have been widely employed in previous literature – for a meta-analysis on the impact of the number of stars and the volume of reviews on sales see Floyd et al. (2014). Nevertheless, recent literature reports important shortcomings of relying solely on these reviews ratings (De Langhe et al. 2016). The analyses presented suggest that an intrinsic attribute of the product, its novelty, does not, by itself, increase the assessed helpfulness of the review as suggested. The novelty of the product, however, does determine what specific informational elements online buyers rely on when forming an idea of the performance and quality of a product. For established products, low-level construal terms and the number of stars increase helpfulness. For novel products, the number of stars, the verification of purchase, and the number of previous evaluations are more indicative than the remaining variables. Possibly, consumers of novel products are less certain what to look for in the written section of online reviews. 27 Previous research recognized the importance of providing product information for novel products (e.g. Wasson 1978; Wind and Robertson 1983) in order to overcome the perceived risk that is due to the lack of knowledge about the product (Hawkins and Mothersbaugh 2010). Research also acknowledges that consumers’ interest switches from different aspects of information depending on whether the product is a novel or an established one (Park and Kim 2008). To the best of our knowledge, this is the first attempt to address the inquiry from King et al. (2014) on how the textual content relates to numerical ratings. In our data, for established products, more stars and more terms that describe the experiences of previous customers increases the helpfulness ratings of the review. Both variables move in the same direction. However, more stars and fewer terms describing features and characteristics significantly increase helpfulness evaluations. These two variables move in opposing directions. This is an interesting finding. High-level type of information has been traditionally provided by vendors and manufacturers through advertising or other forms of business-to-consumers communications. The inclusion of these terms, referring essential features and characteristics, may not be helpful in corroborating claims made by manufacturers as suggested. Conversely, the inclusion of this high-level type of information may signal some kind of manipulation of the review from the vendor or manufacturer (Hu et al. 2011a; Hu et al. 2011b). Potential customers may expect reviewers to express their personal experiences with the product (Hu et al. 2012) rather than report features, as vendors and manufacturers usually do. In the case of novel products, neither low-level nor high-level construal terms contributed to increase review helpfulness. A potential explanation of this finding is that the lack of knowledge about the product may hinder the ability to assess what information is helpful in making purchase decisions. Potential buyers are exposed to concepts and episodes through product reviews. These concepts and episodes gain depth of 28 meaning by becoming associated with additional concepts and episodes. The process of building knowledge structures around a product depends on the exposure of the reader to enough information. Therefore, acquiring the ability to assess the helpfulness of the information may require them to build enough knowledge about the product (Hawkins and Mothersbaugh 2010; Piaget 2001). As Table 2 suggests, in the case of novel products, potential customers rely more on numerical information such as stars or the number of previous helpfulness evaluations. In the case of established products, online customers value more complex information, such as lowlevel construals probably due to a better knowledge of what messages can be considered relevant. The novelty of the product is also relevant when assessing the impact of other informational elements of individual reviews, such as the verification that the reviewer purchased the product. The literature on e-WOM has found contrasting results regarding the influence of different aspects of source credibility. Filieri (2015) suggested that the review will be perceived as more helpful if the information source is considered credible. Willemsen et al. (2011) found a weak positive relationship between the perception of helpfulness of the information and the claims of expertise of reviewers. Other research found no impact of source credibility over perceived information helpfulness (Cheung et al. 2008). The inclusion of novelty, may help explain those contrasting results. The analysis section shows that the helpfulness scores of the novel category improved significantly with the verification that the reviewer purchased the product. Potential customers may rely more heavily on the information provided by verified reviewers. They probably assess such information as more credible. Contribution to Methodology 29 LSA is a natural language processing tool that can factor statistically influential terms across large bodies of text. CLT is a popular theory in the social sciences that classifies information into high- and low-construal terms. Adding CLT to LSA can, as done in this study, guide the classification of terms into terms dealing with the product and terms dealing with other experience. This methodological framework might have potential beyond the context of online consumer reviews and e-commerce. Social scientists may be able to apply this methodology to study considerably larger bodies of text than the traditional qualitative approaches, especially as it requires no human coders or raters. The application of LSA as well as other natural language processing techniques for text analysis offers new avenues of research within online contexts. A similar methodology as the one presented in this paper could be applied to analyze other sources of online data, such as social media or specialized blogs data. Likewise, LSA can provide new insights about the learning process from textual information online. LSA has been argued to approximate how people learn from text and how they discern the similarity of meaning among words (Landauer et al. 2013). LSA may be adequate to simulate ‘what’ potential buyers can learn from the textual data available online, e.g. from product reviews. This research avenue differs from previous studies in marketing that focused more on ‘how’ consumers learn from online reviews (Zhao et al. 2013). As a potential application of this simulation, LSA could provide additional insights on how to develop better ways to overcome information overload in online settings. Limitations and Avenues for Future Research The data analyzed contained 16 products from two different categories that were deliberately chosen because they are essentially functionally equivalent: both product categories are used for the same purpose of producing artificial light. As research suggests that product type 30 may play an important role in determining the perceived helpfulness of online reviews (e.g. Moore 2015; Mudambi and Schuff 2010; Pan and Zhang 2011), replicating the study with a broader scope of products may reveal new insight. This may be especially relevant in the case of low-level construal terms, which are highly contextualized (Trope and Liberman 2003). What is perceived as helpful within a specific context may not be perceived as such within a different one. In this sense, marketers may be interested in characterizing and making more accessible those reviews that are considered more helpful. Marketers could also be interested in making less helpful reviews less accessible or even deleting them. A potential way to achieve this goal is through an extensive characterization of how the informational elements impact the perceived helpfulness of a review across different product categories. Conclusions New methodological approaches are needed to achieve a comprehensive understanding of the role of the different informational elements available online, especially in the case of textual data that, nowadays, is ubiquitous for online users. Textual data is readily available to influence the purchase intentions of potential customers anytime. In the U.S. just mobile connections to the Internet generate more than 18 million megabytes of data every minute (James 2016). This research proposes and illustrates how LSA, in connection with a theoretical framework, provided in this case by CLT, can advance the current marketing methodological approaches to achieve that comprehensive approach. The method proposed in this study can identify some aspects of what information from reviews is considered important and when it is important. This could open new avenues into the study of online reviews, determining what information is important and when. 31 REFERENCES Alba, J., J. Lynch, B. Weitz, C. Janiszewski, and et al. (1997), "Interactive home shopping: Consumer, retailer, and manufacturer incentives to participate in electronic marketplaces," Journal of Marketing, 61 (3), 38-53. Alex, S.L.T. and G. Prendergast (2009), "Is a "star" worth a thousand words?," European Journal of Marketing, 43 (11/12), 1269-80. Anderson, E. T. and D. I. Simester (2014), "Reviews Without a Purchase: Low Ratings, Loyal Customers, and Deception," Journal of Marketing Research, 51 (3), 249-69. Archak, N., A. Ghose, and P.G. Ipeirotis (2011), "Deriving the Pricing Power of Product Features by Mining Consumer Reviews," Management Science, 57 (8), 1485-509. Berger, J. (2013), Contagious. New York, NY: Simon & Schuster. Bickart, B. and R.M. Schindler (2001), "Internet forums as influential sources of consumer information," Journal of Interactive Marketing, 15 (3), 31-40. Biswas, D. (2004), "Economics of information in the Web economy: Towards a new theory?," Journal of Business Research, 57 (7), 724-33. Brown, J.J. and P.H. Reingen (1987), "Social Ties and Word-of-Mouth Referral Behavior," Journal of Consumer Research, 14 (3), 350-62. Cao, Q., W. Duan, and Q. Gan (2011), "Exploring determinants of voting for the “helpfulness” of online user reviews: A text mining approach," Decision Support Systems, 50 (2), 511-21. Chen, P.-Y., S. Dhanasobhon, and M.D. Smith (2008), "All reviews are not created equal: The disaggregate impact of reviews and reviewers at amazon. com," Available at SSRN: http://ssrn.com/abstract=918083. Chen, Y. and J. Xie (2008), "Online Consumer Review: Word-of-Mouth as a New Element of Marketing Communication Mix," Management Science, 54 (3), 477-91. Cheung, C. M. K., M. K.O. Lee, and N. Rabjohn (2008), "The impact of electronic word‐of‐mouth: The adoption of online opinions in online customer communities," Internet Research, 18 (3), 229-47. 32 Chevalier, J.A. and D. Mayzlin (2006), "The Effect of Word of Mouth on Sales: Online Book Reviews," Journal of Marketing Research, 43 (3), 345-54. Chintagunta, Pradeep K., Shyam Gopinath, and Sriram Venkataraman (2010), "The effects of online user reviews on movie box office performance: accounting for sequential rollout and aggregation across local markets," Marketing Science, 29, 944+. Cialdini, R.B. and N.J. Goldstein (2004), "Social influence: Compliance and conformity," Annu. Rev. Psychol., 55, 591-621. De Langhe, B., P. M. Fernbach, and D. R. Lichtenstein (2016), "Navigating by the Stars: Investigating the Actual and Perceived Validity of Online User Ratings," Journal of Consumer Research, 42 (6), 817-33. De Maeyer, P. (2012), "Impact of online consumer reviews on sales and price strategies: a review and directions for future research," Journal of Product & Brand Management, 21 (2), 132-39. Dellarocas, C. (2003), "The Digitization of Word of Mouth: Promise and Challenges of Online Feedback Mechanisms," Management Science, 49 (10), 1407-24. Deutsch, M. and H.B. Gerard (1955), "A study of normative and informational social influences upon individual judgment," The journal of abnormal and social psychology, 51 (3), 629. Dowling, G.R. and R. Staelin (1994), "A Model of Perceived Risk and Intended Risk-Handling Activity," Journal of Consumer Research, 21 (1), 119-34. Dumais, S.T. (2004), "Latent semantic analysis," Annual Review of Information Science and Technology, 38 (1), 188-230. Energy, US Department of (2013), "The History of the Light Bulb." Faulhaber, G.R. and D.A. Yao (1989), ""Fly-by-Night" Firms and the Market for Product Reviews," The Journal of Industrial Economics, 38 (1), 65-77. Felbermayr, A. and A. Nanopoulos (2016), "The Role of Emotions for the Perceived Usefulness in Online Customer Reviews," Journal of Interactive Marketing, 36, 60-76. Filieri, R. (2015), "What makes online reviews helpful? A diagnosticity-adoption framework to explain informational and normative influences in e-WOM," Journal of Business Research, 68 (6), 1261-70. 33 Floyd, K., R. Freling, S. Alhoqail, H. Y. Cho, and T. Freling (2014), "How Online Product Reviews Affect Retail Sales: A Meta-analysis," Journal of Retailing, 90 (2), 217-32. Forman, C., A. Ghose, and B. Wiesenfeld (2008), "Examining the Relationship Between Reviews and Sales: The Role of Reviewer Identity Disclosure in Electronic Markets," Information Systems Research, 19 (3), 291-313,93-95. Ghose, A. and P.G. Ipeirotis (2011), "Estimating the Helpfulness and Economic Impact of Product Reviews: Mining Text and Reviewer Characteristics," IEEE Transactions on Knowledge and Data Engineering 23 (10), 1498-512. Ghose, A., P.G. Ipeirotis, and B. Li (2012), "Designing Ranking Systems for Hotels on Travel Search Engines by Mining User-Generated and Crowdsourced Content," Marketing Science, 31 (3), 493-520. Godes, D. and D. Mayzlin (2004), "Using online conversations to study word-of-mouth communication," Marketing Science, 23, 545+. Gomez, Juan Carlos, Erik Boiy, and Marie-Francine Moens (2012), "Highly Discriminative Statistical Features for Email Classification," Knowlege Information Systems, 31, 23-53. Hawkins, D.I. and D.L. Mothersbaugh (2010), Consumer Behavior. Building Marketing Strategy (11th ed.). New York, NY: McGraw-Hill/Irwin. Hennig-Thurau, T., K.P. Gwinner, G. Walsh, and D.D. Gremler (2004), "Electronic word-of-mouth via consumer-opinion platforms: What motivates consumers to articulate themselves on the Internet?," Journal of Interactive Marketing, 18 (1), 38. Hu, N., I. Bose, Y. Gao, and L. Liu (2011a), "Manipulation in digital word-of-mouth: A reality check for book reviews," Vol. 50: Elsevier B.V. Hu, N., I. Bose, N.S. Koh, and L. Liu (2012), "Manipulation of online reviews: An analysis of ratings, readability, and sentiments," Decision Support Systems, 52 (3), 674-84. Hu, N., L. Liu, and V. Sambamurthy (2011b), "Fraud detection in online consumer reviews," Decision Support Systems, 50 (3), 614-26. Hu, N., J. Zhang, and P.A. Pavlou (2009), "Overcoming the J-shaped distribution of product reviews," Commun. ACM, 52 (10), 144-47. Irmak, C., C.J. Wakslak, and Y. Trope (2013), "Selling the Forest, Buying the Trees: The Effect of Construal Level on Seller- Buyer Price Discrepancy," Journal of Consumer Research, 40 (2), 284-97. 34 Islam, Aminul, Evangelos Milios, and Vlado Keselj (2012), "Text Similarity using Google Tri-grams," in 25th Canadian Conference on Artificial Intelligence, L. Kosseim and D. Inkpen (Eds.) Vol. LNAI 7310. Toronto, Canada: Springer. James, J. (2016), "Data Never Sleeps 4.0," (accessed 09/05/2016, 2016), [available at https://www.domo.com/blog/2016/06/data-never-sleeps-4-0/]. Jin, Liyin, Bingyan Hu, and Yanqun He (2014), "The Recent versus The Out-Dated: An Experimental Examination of the Time-Variant Effects of Online Consumer Reviews," Journal of Retailing, 90 (4), 55266. Kim, Y.-J., J. Park, and R. Wyer (2009), "Effects of Temporal Distance and Memory on Consumer Judgments," Journal of Consumer Research, 36 (4), 634-45. King, R.A., P. Racherla, and V.D. Bush (2014), "What We Know and Don't Know About Online Word-ofMouth: A Review and Synthesis of the Literature," Journal of Interactive Marketing, 28 (3), 167-83. Kohli, R., S. Devaraj, and M. A. Mahmood (2004), "Understanding Determinants of Online Consumer Satisfaction: A Decision Process Perspective," Journal of Management Information Systems, 21 (1), 11535. Kuan, K.K.Y., K.-L. Hui, P. Prasarnphanich, and H.-Y. Lai (2015), "What Makes a Review Voted? An Empirical Investigation of Review Voting in Online Review Systems," Journal of the Association for Information Systems, 16 (1), 48-71. Landauer, T.K., P.W. Foltz, and D. Laham (1998a), "An introduction to latent semantic analysis," Discourse Proceses, 25 (2-3), 259-84. Landauer, T.K., D.S. McNamara, S. Dennis, and W. Kintsch (2013), Handbook of latent semantic analysis. New York, NY: Psychology Press. Landauer, Thomas K. and Susan T. Dumais (1997), "A Solution to Plato’s Problem: The Latent Semantic Analysis Theory of Acquisition, Induction, and Representation of Knowledge," Psychological Review, 104 (2), 211-40. Landauer, Thomas K., Peter W. Foltz, and D. Laham (1998b), "An Introduction to Latent Semantic Analysis," Discourse Processes, 25 (2&3), 259-84. Lee, J., J.-N. Lee, and B.C.Y. Tan (2015), "Antecedents of cognitive trust and affective distrust and their mediating roles in building customer loyalty," Information Systems Frontiers, 17 (1), 159-75. 35 Li, J. and L. Zhan (2011), "Online Persuasion: How the Written Word Drives WOM," Journal of Advertising Research, 51 (1), 239-57. Li, M., L. Huang, C-H Tan, and K-K Wei (2013), "Helpfulness of Online Product Reviews as Seen by Consumers: Source and Content Features," International Journal of Electronic Commerce, 17 (4), 101-36. Liberman, N. and Y. Trope (1998), "The role of feasibility and desirability considerations in near and distant future decisions: A test of temporal construal theory," Journal of Personality and Social Psychology, 75 (1), 5-18. Liu, Y. (2006), "Word of Mouth for Movies: Its Dynamics and Impact on Box Office Revenue," Journal of Marketing, 70 (3), 74-89. Ludwig, S., K. de Ruyter, M. Friedman, E. C. Brüggen, M. Wetzels, and G. Pfann (2013), "More Than Words: The Influence of Affective Content and Linguistic Style Matches in Online Reviews on Conversion Rates," Journal of Marketing, 77 (1), 87-103. Lynch Jr, J.G. and D. Ariely (2000), "Wine online: Search costs affect competition on price, quality, and distribution," Marketing Science, 19 (1), 83-103. Mayzlin, D. (2006), "Promotional Chat on the Internet," Marketing Science, 25 (2), 155-63,201. McFadden, D.L. and K.E. Train (1996), "Consumers' Evaluation of New Products: Learning from Self and Others," Journal of Political Economy, 104 (4), 683-703. Metzger, M.J., A.J. Flanagin, and R.B. Medders (2010), "Social and Heuristic Approaches to Credibility Evaluation Online," Journal of Communication, 60 (3), 413-39. Moe, W. W. and M. Trusov (2011), "The Value of Social Dynamics in Online Product Ratings Forums," Journal of Marketing Research 48 (3), 444-56. Moore, S. G. (2015), "Attitude Predictability and Helpfulness in Online Reviews: The Role of Explained Actions and Reactions," Journal of Consumer Research, 42 (1), 30-44. Mudambi, S.M. and D. Schuff (2010), "What Makes a Helpful Online Review? A Study of Customer Reviews on Amazon.com," MIS Quarterly, 34 (1), 185-200. News, BBC (2013), "Samsung probed in Taiwan over 'fake web reviews'," in BBC News Technology. 36 Nielsen (2012), "Consumer Trust in Online, Social and Mobile Advertising Grows," Nielsen. Pan, Y. and J.Q. Zhang (2011), "Born Unequal: A Study of the Helpfulness of User-Generated Product Reviews," Journal of Retailing, 87 (4), 598-612. Park, D.-H. and S. Kim (2008), "The effects of consumer knowledge on message processing of electronic word-of-mouth via online consumer reviews," Electronic Commerce Research and Applications, 7 (4), 399-410. Peterson, R.A. , S. Balasubramanian, and B.J. Bronnenberg (1997), "Exploring the Implications of the Internet for Consumer Marketing " Journal of the Academy of Marketing Science, 25 (4), 329-46. Piaget, J. (2001), Studies in Reflecting Abstraction. New York, NY: Psychology Press. Punj, G. and R. Moore (2009), "Information search and consideration set formation in a web-based store environment," Journal of Business Research, 62 (6), 644-50. Racherla, P., M. Mandviwalla, and D.J. Connolly (2012), "Factors affecting consumers' trust in online product reviews," Journal of Consumer Behaviour, 11 (2), 94-104. Ramirez, A., J.B. Walther, J.K. Burgoon, and M. Sunnafrank (2002), "Information Seeking Strategies, Uncertainty, and Computer-Mediated Communication. Toward a Conceptual Model," Human Communication Research, 28 (2), 213-28. Salehan, M. and D.J. Kim (2016), "Predicting the performance of online consumer reviews: A sentiment mining approach to big data analytics," Decision Support Systems, 81, 30-40. Sen, S. and D. Lerman (2007), "Why are you telling me this? An examination into negative consumer reviews on the Web," Journal of Interactive Marketing, 21 (4), 76-94. Shapiro, C. and H.R. Varian (1999), Information Rules. Boston, MA: Harvard Business School Publications. Spool, J.M. (2009), "The Magic Behind Amazon's 2.7 Billion Dollar Question," in User Interface Engineering: User Interface Engineering. Stigler, G.J. (1961), "The Economics of Information," Journal of Political Economy, 69 (3), 213-25. Tan, S.J. (1999), "Strategies for reducing consumers’ risk aversion in Internet shopping," Journal of Consumer Marketing, 16 (2), 163-80. 37 Trope, Y. and N. Liberman (2003), "Temporal construal," Psychological Review, 110 (3), 403-21. Trope, Y., N. Liberman, and C. J. Wakslak (2007), "Construal Levels and Psychological Distance: Effects on Representation, Prediction, Evaluation, and Behavior," Journal of Consumer Psychology, 17 (2), 83-95. Trope, Y.L. (2012), "Construal Level Theory," in Handbook of Theories of Social Psychology, P. K. Van Lange, ed. Washington DC: Sage Publications Ltd. . Valle-Lisboa, Juan C. and Eduardo Mizraji (2007), "The Uncovering of Hidden Structures by Latent Semantic Analysis," Information Sciences, 177 (19), 4122–47. Wan, Y. and M. Nakayama (2014), "The reliability of online review helpfulness," Journal of Electronic Commerce Research, 15 (3), 179. Wang, F., X. Liu, and E. Fang (2015), "User Reviews Variance, Critic Reviews Variance, and Product Sales: An Exploration of Customer Breadth and Depth Effects," Journal of Retailing, 91 (3), 372-89. Wasson, C.R. (1978), Dynamic competitive strategy & product life cycles: Austin Press. Willemsen, L. M., P. C. Neijens, F. Bronner, and J. A. de Ridder (2011), "“Highly Recommended!” The Content Characteristics and Perceived Usefulness of Online Consumer Reviews," Journal of ComputerMediated Communication, 17 (1), 19-38. Wind, Y. and T.S Robertson (1983), "Marketing strategy: New directions for theory and research," The Journal of Marketing, 12-25. Wu, C., H. Che, T.Y. Chan, and X. Lu (2015), "The Economic Value of Online Reviews," Marketing Science, 34 (5), 739-54. Wu, P.F. (2013), "In Search of Negativity Bias: An Empirical Study of Perceived Helpfulness of Online Reviews," Psychology & Marketing, 30 (11), 971-84. Zhao, Y., S. Yang, V. Narayan, and Y. Zhao (2013), "Modeling Consumer Learning from Online Product Reviews," Marketing Science, 32 (1), 153-69. Zhu, F. and X. Zhang (2010), "Impact of Online Consumer Reviews on Sales: The Moderating Role of Product and Consumer Characteristics," Journal of Marketing, 74 (2), 133-48. 38 TABLES Mean Std Helpfulness Score N Mean Std Number of Words N Mean Number of Stars in Std the Review N Mean Number of HighStd Level Construal Terms in the Review N Mean Number of LowLevel Construal Std Terms in the Review N Verification of Purchase Mean Std N Review is about Model 1 Model 2 Established Both Established Product and Novel Category Product (Incandescent Categories bulbs) only 1.28 0.57 4.51 3.31 759 431 55.18 33.57 77.8 28.08 759 431 4.09 4.26 1.39 1.28 759 431 7.11 4.13 9.92 3.43 759 431 11.57 8.32 12.6 5.19 759 431 0.85 0.35 759 0.96 0.21 431 Model 3 Novel Product Category (LED) only 2.2 5.58 328 83.59 107.56 328 3.87 1.5 328 11.02 13.63 328 15.84 17.33 328 0.72 0.45 328 Significance levels *** at .001, ** at .01, * at .05, and # at .1 Table 1. Descriptive statistics of the complete dataset T test comparing Incandescent bulbs with LED -4.70*** -8.21*** 3.74*** -8.93*** -7.61*** 8.93*** 39 Model 1 (all the data) Hypothesis H1 H2 H3 H4 H5 Contr ol Intercept Number of LowLevel Construal Terms Number of High-Level Construal Terms Number of Stars Verification of Purchase (Bought = 1) Product Novelty (LED = 1) Number of Evaluations R square -2.24(.41)*** Model 2 Model 3 (only (only LED incandescent bulbs) bulbs) Estimate (Std. Err) -11.02(3.66)** -2.70(.86)** Model 4 Standardized Linear Regression .02(.01)* 1.25(.13)*** .01(.03) .09* -.04(.01)** -1.03(.24)*** -.05(.04) -.12** .57(.08) *** 3.33(.52)*** 1.21(.20)*** .12*** .68(.28)* 4.34(3.21) 2.85(.65)*** .05* .40(.26) .03 .77(.01)*** -.16(.07)* .82(.03)*** .91** .83 .55 .87 .83 Significance levels *** at .001, ** at .01, * at .05, and # at .1 Table 2. GLM and Standardized Linear Regression Analyses Results 40 FIGURES Figure 1. Conceptual model 41 Figure 2. Variables extracted from reviews included in the conceptual model 42 Figure 3. Operationalization of high-level construal terms 43 APPENDIX A The high- and low-level construal word selection criterion is based on a tradeoff between the number of words selected and the average importance of those words. The construal words selection follows an 11-step process. Step (1), in addition to the transformed DTM (DTM1) already constructed to identify potential high-level words, a new transformed DTM (DTM2) was constructed over the textual portion of the 759 reviews following the same methodology as in the case of the potential highlevel construal words. Those words that are found in both matrices (DTM1 and DTM2) become the pool of potential high-level words. All others in the review DTM2 are the pool of low-level terms. Step (2), a reduced-rank SVD was calculated over DTM2. SVD retains the k-largest singular values of this matrix and the remainders are set to 0. This results in a reduced-dimension SVD representation, which is the best k-dimensional approximation to the original matrix employing least-squares. After applying SVD, each document and each term is represented as a k-dimensional vector in a semantic space (specifically in this research, as a vector of 149 dimensions). Step (3), similarities and associations were then computed on this reduced-dimension semantic space. Since both documents and terms are represented as vectors in this space, document-document, term-term, and document-term similarities and associations can be calculated (Dumais 2004). Step (4), from this semantic space, cosine measures of “closeness in meaning” were computed by collecting the 20 terms most close to every one of the original 759 reviews. 1,497 terms were obtained – with their corresponding appearance frequencies in each 44 review. Step (5), the total frequency for each word i is calculated over each document j for a total of r documents: Where: x = frequency of word i in document j Step (6), the i rows are ordered in descending order by Fi. Step (7), to obtain the average importance of each word across the 759 reviews R and in each of the 16 product descriptions D, we calculated the average importance of word k in matrix M as: Where: y = estimate of how many times a word would occur in each of an infinite sample of reviews (Landauer et al. 1998a) (each cell value in the semantic space) k = word in the semantic space (row in the semantic space) l = documents in the semantic space (column in the semantic space). They vary between 1 and t t = total number of documents in the semantic space (r in M = R, d in M = D) M = semantic space matrix (R = reviews, D = descriptions) Choosing words for inclusion in the construal words set is based on the average cumulative importance of these words, following the frequency ordering obtained in step (6). Step (8), to choose a cutoff point for the average of words to retain given the importance of the words, we inspect a graph of the cumulative average importance over the full set of potential construal words. 45 Cumulative Average Importance Cumulative Average Term Importance 2 1.5 1 0.5 0 0 200 400 600 800 1000 1200 1400 1600 Number of Terms in Construal Set Upon inspection of the previous figure, a cutoff point of .5 is chosen. We seek to include only those words with an average cumulative importance greater than .5. Step (9), maintaining the ordering found in step (6), we calculate the average cumulative importance at each unique frequency level in the set of potential construal terms. Step (10), for each frequency level, we calculate the cumulative number of words included in the construal set and the cumulative average importance for the two matrices. Step (11), we take the average importance at each unique frequency level. Our inclusion pool is based on the average of the cumulative average importance for the unique frequency levels from matrix R and the cumulative average importance for the matching words found in matrix D. All words that have a unique frequency level where this average is greater than or equal to .5 are included. In this case, we find that the minimum frequency of the included review words is 25, the resulting number of words is 164 and the breakdown of highlevel and low-level construal words is 33 and 131 respectively. 46 APPENDIX B Authors (year) Vol. (Issue) Key Findings Sample Journal of Marketing Research 51 (3) Deceptive reviews may not be limited to the strategic actions of firms. Instead, the phenomenon may be far more prevalent, extending to individual customers who have no financial incentive to influence product ratings 330,975 reviews from a prominent retailer Management Science 57 (8) This paper demonstrates how textual data can be used to learn consumers’ relative preferences for different product features and also how text can be used for predictive modeling of future changes in sales Sales data and review data for digital cameras and camcorders from Amazon Establishing Trust in Electronic Commerce through Online Word of Mouth: An Examination across Genders Finding disseminators via electronic word of mouth message for effective marketing communications Exploring determinants of voting for the “helpfulness” of online user reviews: A text mining approach Journal of Management Information Systems 24 (4) The effect of trust on intention to shop online is stronger for women than for men. Men value their ability to post content online, whereas women value the responsive participation of other consumers to the content they have posted. Online word-of-mouth quality affects online trust differently across genders Decision Support Systems 67 The authors identify three types of opinion leaders through e-WOM using a message-based approach. They demonstrate that e-WOM of opinion leaders drives product sales due to their product experience and knowledge background Decision Support Systems 50 (2) The findings show that the semantic characteristics are more influential than other characteristics in affecting how many helpfulness votes reviews receive. The findings also suggest that reviews with extreme opinions receive more helpfulness votes than those with mixed or neutral opinions Chen, C. C., & Tseng, Y.-D. (2011) Quality evaluation of product reviews using an information quality framework Decision Support Systems 50 (4) The authors treat the evaluation of review quality as a classification problem and employ an effective information quality framework to extract representative review features Chen, Y., & Xie, J. (2008) Online Consumer Review: Word-of-Mouth as a New Element of Marketing Communication Mix Management Science 54 (3) The results reveal that if the review information is sufficiently informative, the seller-created product attribute information and buyer-created review information will interact with each other Survey focused on asynchronous online forums, 1,561 usable responses Amazon review dataset. Sample of 350,122 book, music, video and DVD titles Data from CNET Download.com for software programs in Enterprise Computing The first 150 reviews were collected for ten digital cameras and ten mp3 players on Amazon Third-party reviews from CNET.com, consumer reviews from Amazon.com, product attributes from CNET.com Decision Support Systems 53 (1) The authors focus on the factors that drive consumers to spread positive eWOM in online consumeropinion platforms. They identified a number of key motives of consumers' eWOM intention and developed an associated model 203 members of a consumer review community Decision Support Systems 54 (1) The authors used the social communication framework to summarize and classify prior eWOM studies. They identified key factors related to the major elements of the social communication literature and built an integrative framework explaining the impact of eWOM communication on consumer behavior No empirical test Journal of Marketing Research 43 (3) Reviews are overwhelmingly positive; an improvement in a book's reviews leads to an increase in relative sales at that site; the impact of one-star reviews is greater than the impact of five-star reviews; and customers read review text rather than relying only on summary statistics Book characteristics and user reviews from Amazon.com and bn.com Anderson, E. T., & Simester, D. I. (2014) Archak, N., Ghose, A., & Ipeirotis, P. G. (2011) Awad, N. F., & Ragowsky, A. (2008) Bao, T., & Chang, T. S. (2014) Cao, Q., Duan, W., & Gan, Q. (2011) Cheung, C. M. K., & Lee, M. K. O. (2012) Cheung, C. M. K., & Thadani, D. R. (2012) Chevalier, J. A., & Mayzlin, D. (2006) Title Journal Reviews Without a Purchase: Low Ratings, Loyal Customers, and Deception Deriving the Pricing Power of Product Features by Mining Consumer Reviews What drives consumers to spread electronic word of mouth in online consumeropinion platforms The impact of electronic word-of-mouth communication: A literature analysis and integrative model The Effect of Word of Mouth on Sales: Online Book Reviews 47 Chintagunta, P. K., Gopinath, S., & Venkataraman, S. (2010) Clemons, E. K., Gao, G. G., & Hitt, L. M. (2006) De Langhe, B., Fernbach, P. M., & Lichtenstein, D. R. (2016) de Matos, C., & Rossi, C. (2008) Dellarocas, C. (2003) Dellarocas, C. (2006) Dellarocas, C., Gao, G., & Narayan, R. (2010) Duan, W., Gu, B., Whinston, A.B. (2008) Duhan, D. F., Johnson, S. D., Wilcox, J. B., & Harrell, G. D. (1997) Floyd, K., Freling, R., Alhoqail, S., Cho, H. Y., & Freling, T. (2014) Ghose, A., Ipeirotis, P. G., & Li, B. (2012) The effects of online user reviews on movie box office performance: accounting for sequential rollout and aggregation across local markets Marketing Science When Online Reviews Meet Hyperdifferentiation: A Study of the Craft Beer Industry Navigating by the Stars: Investigating the Actual and Perceived Validity of Online User Ratings Word-of-mouth communications in marketing: a meta-analytic review of the antecedents and moderators The Digitization of Word of Mouth: Promise and Challenges of Online Feedback Mechanisms Strategic Manipulation of Internet Opinion Forums: Implications for Consumers and Firms Are Consumers More Likely to Contribute Online Reviews for Hit or Niche Products? Do online reviews matter? — An empirical investigation of panel data 29 (5) The authors find that it is the valence of reviews that seems to matter and not the volume. When the authors carry out the analysis with aggregated national data, they obtain the same results as previous studies that volume matters but not the valence. Using various market-level controls in the national data model, the authors attempt to identify the source of this difference Daily box office ticket sales data for 148 movies along with user review data from Yahoo! Movies Journal of Management Information Systems 23 (2) They find that the variance of ratings and the strength of the most positive quartile of reviews play a significant role in determining which new products grow fastest in the market-place 281,868 ratings for 1,159 U.S. craft brewers from 6,212 reviewers from Ratebeer.com Journal of Consumer Research 42 (6) Journal of the Academy of Marketing Science 36 (4) Management Science 49 (10) Management Science 52 (10) Journal of Management Information Systems Decision Support Systems Average user ratings (1) lack convergence with Consumer Reports scores, (2) are often based on insufficient sample sizes which limits their informativeness, (3) do not predict resale prices in the used-product marketplace, and (4) are higher for more expensive products and premium brands, controlling for Consumer Reports scores WOM valence is a significant moderator, cross-sectional studies show a stronger influence of satisfaction and loyalty on WOM activity than longitudinal studies, studies of WOM behavior show a weaker link between loyalty and WOM activity than studies of WOM intentions, and satisfaction has a stronger relationship with positive WOM than loyalty, whereas (dis)loyalty has a stronger relationship with negative WOM than does (dis)satisfaction The paper provides an overview of relevant work in game theory and economics on the topic of reputation. It discusses how this body of work is being extended and combined with insights from computer science, management science, sociology, and psychology to take into consideration the special properties of online environments The authors' result implies the existence of settings where online forum manipulation benefits consumers. If the precision of honest consumer opinions that firms manipulate is sufficiently high, firms of all types, as well as society, would be strictly better off if manipulation of online forums was not possible 27 (2) The authors' findings suggest that online forum designers who wish to increase the contribution of user reviews for lesser-known products should make information about the volume of previously posted reviews a less-prominent feature of their sites 45 (4) The result shows that online user reviews have little persuasive effect on consumer purchase decisions. Nevertheless, the authors find that box office sales are significantly influenced by the volume of online posting, suggesting the importance of awareness effect 1272 products across 120 vertically differentiated product categories 127 studies, which produced 162 independent samples, 348 effect sizes, and a cumulated N of 64,364 subjects eBay's Case Study Econometric Model Reviews posted on Yahoo! Movies, production and weekly box office Data from Yahoo! Movies, Variety.com, and BoxOfficeMojo.com Influences on consumer use of word-of-mouth recommendation sources Journal of the Academy of Marketing Science 25 (4) The authors present a model that proposes two routes of influence on the choice of recommendation sources. The model proposes that the prior knowledge level of the consumer, the perceived level of task difficulty, and the type of evaluative cues sought by consumers influence their choice of recommendation sources 245 questionnaires How Online Product Reviews Affect Retail Sales: A Metaanalysis Journal of Retailing 90 (2) Online product reviews have a significant impact on sales elasticity. Our research also provides interesting insights about important variables which augment or diminish the influence that online product reviews exert on retailer performance Meta-analysis of 26 empirical studies yielding 443 sales elasticities Designing Ranking Systems for Hotels on Travel Search Engines by Mining UserGenerated and Crowdsourced Content Marketing Science 31 (3) The authors propose to generate a ranking system that recommends products that provide, on average, the best value for the consumer’s money U.S. hotel reservations through Travelocity, supplemented with data from social media sources 48 Godes, D., & Mayzlin, D. (2004) Using online conversations to study word-of-mouth communication Marketing Science 23 (4) Godes, D., & Silva, J. C. (2012) Sequential and Temporal Dynamics of Online Opinion Marketing Science 31 (3) Decision Support Systems Hu, N., Bose, I., Koh, N. S., & Liu, L. (2012) Hu, N., Koh, N. S., & Reddy, S. K. (2014) Huang, L., Tan, C.-H., Ke, W., & Wei, K.-K. (2013) Jin, L., Hu, B., & He, Y. (2014) Khare, A., Labrecque, L. I., & Asare, A. K. (2011) Manipulation of online reviews: An analysis of ratings, readability, and sentiments Ratings lead you to the product, reviews help you clinch it? The mediating role of online review sentiments on product sales Comprehension and Assessment of Product Reviews: A Review-Product Congruity Proposition The Recent versus The OutDated: An Experimental Examination of the TimeVariant Effects of Online Consumer Reviews The Assimilative and Contrastive Effects of Wordof-Mouth Volume: An Experimental Examination of Online Consumer Ratings Online conversations may offer an easy and cost-effective opportunity to measure word of mouth. The authors show that a measure of the dispersion of conversations across communities has explanatory power in a dynamic model of TV ratings The authors find that, once controlling for calendar date, the residual average temporal pattern is increasing. With respect to sequential dynamics, they find that ratings decrease: the nth rating is, on average, lower than the n−1th when controlling for time, reviewer effects, and book effects 44 TV US shows, data from Nielsen ratings, and Usenet newsgroup Book review data gathered from Amazon.com. Reviews for each of the top 350 selling titles 52 (3) The authors discover that around 10.3% of the products are subject to online reviews manipulation. Consumers are only able to detect manipulation taking place through ratings, but not through sentiments The dataset consisted of 4490 books, with 610,713 online reviews Decision Support Systems 57 Ratings do not have a significant direct impact on sales but have an indirect impact through sentiments. Sentiments, however, have a direct significant impact on sales. The two most accessible types of reviews – most helpful and most recent – play a significant role in determining sales Panel of data on over 4000 books from Amazon.com Journal of Management Information Systems 30 (3) The two matching conditions, (1) attribute-based reviews describing a search product and (2) experience-based reviews describing an experience product, could lead consumers to perceive higher review helpfulness and lower cognitive effort to comprehend the reviews Experimental settings Journal of Retailing 90 (4) Although recent online reviews are more influential in shifting consumer preferences towards nearfuture consumption decisions, the relative influence of out-dated online reviews in shifting consumer preferences increases when consumers are making distant-future consumption decisions Experimental Settings Journal of Retailing 87 (1) High volume accentuates or assimilates perceptions of positivity or negativity of WOM targets. Depending upon how high volume interacts with WOM consensus and consumer decision precommitment, it can contrast preference away from the valence of a target also. Consumers differ in their susceptibility to the influence of high volume Experimental Settings Data from 83 blogs, representing approximately 1,376,000 words and 6722 postings Kozinets, R. V., de Valck, K., Wojnicki, A. C., & Wilner, S. J. S. (2010) Networked Narratives: Understanding Word-ofMouth Marketing in Online Communities Journal of Marketing 74 (2) The findings indicate that the network of communications offers four social media communication strategies—evaluation, embracing, endorsement, and explanation. Each is influenced by character narrative, communications forum, communal norms, and the nature of the marketing promotion Liu, Y. (2006) Word of Mouth for Movies: Its Dynamics and Impact on Box Office Revenue Journal of Marketing 70 (3) WOM activities are the most active during a movie's prerelease and opening week and that movie audiences tend to hold relatively high expectations before release but become more critical in the opening week. WOM information offers significant explanatory power for box office revenue, most comes from the volume of WOM and not from its valence WOM data collected from the Yahoo Movies message board Ludwig, S., de Ruyter, K., Friedman, M., Brüggen, E. C., Wetzels, M., & Pfann, G. (2013) More Than Words: The Influence of Affective Content and Linguistic Style Matches in Online Reviews on Conversion Rates Journal of Marketing 77 (1) The influence of positive affective content on conversion rates is asymmetrical. No such taperingoff effect occurs for changes in negative affective content in reviews. Furthermore, positive changes in affective cues and increasing congruence with the product interest group’s typical linguistic style directly and conjointly increase conversion rates Sample of 591 books available for sale on Amazon.com with 18,682 customer reviews Ma, X., Khansa, L., Deng, Y., & Kim, S. S. (2013) Impact of Prior Reviews on the Subsequent Review Process in Reputation Systems Journal of Management Information Systems 30 (3) Male reviewers or those who lack experience, geographic mobility, or social connectedness are more prone to being influenced by prior reviews. The authors also found that longer and more frequent reviews can reduce online reviews’ biases Panel data set of 744 individual consumers collected from Yelp 49 The authors find a unique equilibrium where online word of mouth is persuasive despite the promotional chat activity by competing firms. In this equilibrium, firms spend more resources promoting inferior products, in striking contrast to existing advertising literature Across individuals, the results show that positive ratings environments increase posting incidence, whereas negative ratings environments discourage posting. The authors' results also indicate important differences across individuals in how they respond to previously posted ratings. These dynamics affect the evolution of online product opinions The results show that although ratings behavior is significantly influenced by previously posted ratings and can directly improve sales, the effects are relatively short lived once indirect effects are considered Mayzlin, D. (2006) Promotional Chat on the Internet Marketing Science 25 (2) Moe, W. W., & Schweidel, D. A. (2012) Online Product Opinions: Incidence, Evaluation, and Evolution Marketing Science 31 (3) Moe, W. W., & Trusov, M. (2011) The Value of Social Dynamics in Online Product Ratings Forums Attitude Predictability and Helpfulness in Online Reviews: The Role of Explained Actions and Reactions What Makes a Helpful Online Review? A Study of Customer Reviews on Amazon.com Born Unequal: A Study of the Helpfulness of UserGenerated Product Reviews Predicting the performance of online consumer reviews: A sentiment mining approach to big data analytics The influence of online product recommendations on consumers’ online choices How Does the Variance of Product Ratings Matter? From valence to emotions: Exploring the distribution of emotions in online product reviews The Impact of MarketingInduced versus Word-ofMouth Customer Acquisition on Customer Equity Growth User Reviews Variance, Critic Reviews Variance, and Product Sales: An Exploration of Customer Breadth and Depth Effects Saliency effects of online reviews embedded in the description on sales: Moderating role of reputation Can online product reviews be more helpful? Examining Journal of Marketing Research 48 (3) Journal of Consumer Research 42(1) Results show that review writers explain their actions more than their reactions for utilitarian products, but they explain their reactions more than their actions for hedonic products Experimental Settings MIS Quarterly 34 (1) Review extremity, review depth, and product type affect the perceived helpfulness of the review. Product type moderates the effect of review extremity on the helpfulness of the review 1,587 reviews from Amazon.com across six products Journal of Retailing 87 (4) Decision Support Systems 81 Journal of Retailing 80 (2) Results indicate that subjects who consulted product recommendations selected recommended products twice as often as subjects who did not consult recommendations Experimental Settings Management Science 58 (4) A high average rating indicates a high product quality, whereas a high variance of ratings is associated with a niche product, one that some consumers love and others hate 667 books available from Amazon and BN Decision Support Systems 81 The authors find that there is a difference in the emotional content of reviews across search and experience goods in the early stages of product launch. However, these differences disappear over time as the addition of reviews reduces the information asymmetry gap 15,849 online customer reviews from Amazon.com 45 (1) Marketing-induced customers add more short-term value, but word-of-mouth customers add nearly twice as much long-term value to the firm. Internet firm provided free Web hosting to registered users during a 70-week observation period Journal of Retailing 91 (3) After recognizing a high variance in user reviews, many potential buyers may simply exclude the focal product from their consideration sets for fear that it does not match their needs and preferences. High user reviews variance, in combination with high critic reviews variance, can elicit a sense of uniqueness and enhance purchase intentions of some consumers. Quality signals can strengthen the positive customer depth effect Secondary data and experimental settings combined Decision Support Systems 87 This study explores the effects of online reviews embedded in the product description (OED) on sales. Results indicate that OED has a positive effect on sales, and a high reputation strengthens the impact of OED on sales Secondary data and experimental settings combined Decision Support 79 The authors find that, while review content affects helpfulness in complex ways, these effects are well explained by the proposed framework. The data suggest that review writers who explicitly Over 8000 helpfulness ratings from reviews Moore, S. G. (2015) Mudambi, S. M., & Schuff, D. (2010) Pan, Y., & Zhang, J. Q. (2011) Salehan, M., & Kim, D. J. (2016) Senecal, S., & Nantel, J. (2004) Sun, M. (2012) Ullah, R., Amblee, N., Kim, W., & Lee, H. (2016) Villanueva, J., Yoo, S., & Hanssens, D. M. (2008) Wang, F., Liu, X., & Fang, E. (2015) Wang, Z., Li, H., Ye, Q., & Law, R. (2016) Weathers, D., Swain, S. D., & Journal of Marketing Research Both review valence and length have positive effects on review helpfulness, but the product type moderates these effects. A curvilinear relationship exists between expressed reviewer innovativeness and review helpfulness The authors' findings show that reviews with higher levels of positive sentiment in the title receive more readerships. Sentimental reviews with neutral polarity in the text are also perceived to be more helpful. The length and longevity of a review positively influence both its readership and helpfulness Econometric Model Data from BazaarVoice. 10,460 ratings across 1,811 products Sample of 500 beauty products with 3801 ratings posted Experimental Settings 35,000 online reviews of 20 different products from Amazon.com 50 Grover, V. (2015) characteristics of information content by product type Systems Wu, C., Che, H., Chan, T. Y., & Lu, X. (2015) The Economic Value of Online Reviews Marketing Science Xu, K., Liao, S. S., Li, J., & Song, Y. (2011) Mining comparative opinions from customer reviews for Competitive Intelligence Decision Support Systems Ya, Y., Vadakkepatt, G. G., & Joshi, A. M. (2015) A Meta-Analysis of Electronic Word-of-Mouth Elasticity Journal of Marketing Yang, J., Sarathy, R., & Lee, J. (2016) Yin, D., Bond, S. D., & Zhang, H. (2014) The effect of product review balance and volume on online Shoppers' risk perception and purchase intention Anxious or Angry? Effects of Discrete Emotions on the Perceived Helpfulness of Online Reviews 34 (5) attempt to enhance review diagnosticity or credibility are often ineffective or systematically unhelpful The authors conduct a series of counterfactual experiments and find that the value from Dianping is about 7 CNY for each user, and about 8.6 CNY from each user for the reviewed restaurants. The majority of the value comes from reviews on restaurant quality, and contextual comments are more valuable than numerical ratings in reviews posted on Amazon.com User browsing and user dining choices collected from Dianping.com 50 (4) The authors proposed a novel graphical model to extract and visualize comparative relations between products from customer reviews, with the interdependencies among relations taken into consideration, to help enterprises discover potential risks and further design new products and marketing strategies Data from: 1) online shopping sites; 2) review sites; 3) blogs; 4) social network sites; and 5) emails in CRM 79 (2) The analysis reveals that electronic word-of-mouth volume (valence) elasticity is .236 (.417). Volume and valence elasticities are studied under different contexts 51 studies Decision Support Systems 89 The four proposed risk concerns are good predictors of online shoppers' overall risk in e-commerce; perceived risk is a major determinant of online shoppers' attitude toward purchasing, which in turn determines their purchase intention. No significant causal effect between perceived uncertainty and purchase intention was found 302 questionnaires MIS Quarterly 38 (2) The findings demonstrate the importance of examining discrete emotions in online word-of-mouth, and they carry important practical implications for consumers and online retailers Experimental Settings 72,916 customers' numerical ratings for 1628 different movies from EachMovie Ying, Y., Feinberg, F., & Wedel, M. (2006) Leveraging Missing Ratings to Improve Online Recommendation Systems Journal of Marketing Research 43 (3) Recommendation quality is improved substantially by jointly modeling "selection" and "ratings," both whether and how an item is rated. Accounting for missing ratings and various sources of heterogeneity offers a rich portrait of which items are rated well, which are rated at all, and how these processes are intertwined Yubo, C., & Jinhong, X. (2008) Online Consumer Review: Word-of-Mouth as a New Element of Marketing Communication Mix Management Science 54 (3) The authors show when and how the seller should adjust its own marketing communication strategy in response to consumer reviews. The authors also identify product/market conditions under which the seller benefits from facilitating such buyer-created information The sample includes 120 digital camera models reviewed by CNET.com Examining the influence of online reviews on consumers' decision-making: A heuristic– systematic model Decision Support Systems 67 Argument quality of online reviews, which is characterized by perceived informativeness and perceived persuasiveness, has a significant effect on consumers' purchase intention. In addition, the authors find that source credibility and perceived quantity of reviews have direct impacts on purchase intention The model is empirically tested with 191 users of an existing online review site Modeling Consumer Learning from Online Product Reviews Marketing Science 32 (1) Consumers learn more from online reviews of book titles than from their own experience with other books of the same genre Panel data set of 1,919 book purchases by 243 consumers Impact of Online Consumer Reviews on Sales: The Moderating Role of Product and Consumer Characteristics Journal of Marketing 74 (2) Online reviews are more influential for less popular games and games whose players have greater Internet experience. The article shows differential impact of consumer reviews across products in the same product category and suggests that firms' online marketing strategies should be contingent on product and consumer characteristics Data on console sales and game sales from NPD. Review data from GameSpot.com Zhang, K. Z. K., Zhao, S. J., Cheung, C. M. K., & Lee, M. K. O. (2014) Zhao, Y., Yang, S., Narayan, V., & Zhao, Y. (2013) Zhu, F., & Zhang, X. (2010) Table 6. Most relevant literature on online consumer reviews (key findings as reported by the authors)
© Copyright 2025 Paperzz