This Just In: Fake News Packs a Lot in Title, Uses Simpler, Repetitive Content in Text Body, More Similar to Satire than Real News Benjamin D. Horne and Sibel Adalı Rensselaer Polytechnic Institute 110 8th Street, Troy, New York, USA {horneb, adalis}@rpi.edu Abstract The problem of fake news has gained a lot of attention as it is claimed to have had a significant impact on 2016 US Presidential Elections. Fake news is not a new problem and its spread in social networks is well-studied. Often an underlying assumption in fake news discussion is that it is written to look like real news, fooling the reader who does not check for reliability of the sources or the arguments in its content. Through a unique study of three data sets and features that capture the style and the language of articles, we show that this assumption is not true. Fake news in most cases is more similar to satire than to real news, leading us to conclude that persuasion in fake news is achieved through heuristics rather than the strength of arguments. We show overall title structure and the use of proper nouns in titles are very significant in differentiating fake from real. This leads us to conclude that fake news is targeted for audiences who are not likely to read beyond titles and is aimed at creating mental associations between entities and claims. Introduction The main problem we address in this paper is the following: Is there any systematic stylistic and other content differences between fake and real news? News informs and influences almost all of our everyday decisions. Today, the news is networked, abundant, and fast-flowing through social networks. The sheer number of news stories together with duplication across social contacts is overwhelming. The overload caused by this abundance can force us to use quick heuristics to gain information and make a decision on whether to trust its veracity. These heuristics can come in many forms. In deciding that content is believable by using reader’s own judgment, readers may simply skim an article to understand the main claims instead of reading carefully the arguments and deciding whether the claim is well-supported. In some cases, other heuristics may rely on the trust for the source producing the information or for the social contact who shared it. Often the trust decision is a combination of these heuristics, content, source, and social network all playing a role. This trust decision strucure is supported by the well-studied notion of echo-chambers, in which the sharing of information often conforms to one’s beliefs and is impacted by c 2017, Association for the Advancement of Artificial Copyright Intelligence (www.aaai.org). All rights reserved. homophily (Bakshy, Messing, and Adamic 2015). Furthermore, misleading or wrong information have a higher potential to become viral (Bessi et al. 2015)and lead to negative discussions (Zollo et al. 2015b). Even more troubling, results show resistance by individuals to information challenging established beliefs and that attempts to debunk conspiracy theories are largely ineffective (Zollo et al. 2015a). As an informed public is crucial to any operating democracy, incorrect information is especially dangerous in political news. As many widely discredited claims have become viral and distributed widely in social networks during the 2016 elections (Silverman 2016), the topic of “fake news” and effective methods for debunking it has gained renewed attention. While there is a great deal of work on the social factors leading to dissemination of misinformation in networks (Bessi et al. 2014), there is relatively little work on understanding how fake news content differs from real news content. We would like to understand whether fake news differs systematically from real news in style and language use. While some have argued that the distinction between fake and real news can be a rather arbitrary one, as wellestablished news organizations have been known to disseminate incorrect information on occasion, such organizations operate with a degree of transparency that is not found in fake news sites and risk their established credibility if their news stories are shown to be false. More importantly, their articles adhere to a journalistic style in making and presenting claims. To conduct our study of “fake news”, we study three separate data sets. Two of these data sets are novel: one has been featured by Buzzfeed through their analysis of real and fake news items from 2016 US Elections (Silverman 2016). The second data set, collected by us, contains news articles on US politics from real, fake, and satire news sources. Finally, we look at a third data set containing real and satire articles from a past study (Burfoot and Baldwin 2009). We include satire as a type of fake news that relies on absurdity rather than sound arguments to make claims, but explicitly identifies itself as satire. Fake news in contrast has the intention to deceive, making the reader believe it is correct. We study similarities between fake news and satire to understand different heuristics that they both employ to persuade their readers. The inclusion of satire as a third category of news is a unique contribution of our work. Through a statis- tical study of these three data sets, we show that fake news articles tend to be shorter in terms of content, but use repetitive language and fewer punctuation. Fake news articles differ much more in their titles. Fake titles are longer, use few stop words, and fewer nouns but more proper nouns. Furthermore, we find that fake news is more similar to satire than real news. When fake news is different than satire, the distinction simply exaggerates satire’s differences with real news further. Fake news packs the main claim of the article into its title, which often is about a specific person and entity, allowing the reader to skip reading the article, which tends to be short, repetitive, and less informative. Given the arguments in the article are less of a factor, the persuasion likely relies on heuristics such as conformance of the information to one’s beliefs. Lastly, we illustrate the predictive power of our features by utilizing linear kernel SVMs on small feature subsets. We hope that our study and data sets lead to further study of stylistic conventions used in persuading audiences with limited attention and effective methods to counter them. In the least, it suggests that more attention needs to paid to titles of such articles. Related Work Fake news is certainly not a new phenomenon, and has been well studied in both the fields of journalism and computer science. In particular, it has been studied in two ways: (1) analyzing the spread of fake news and (2) analyzing the content of fake news. Rubin et al. work towards a news verification system using Rhetorical Structure Theory (RST) on content data from NPR’s “Bluff the Listener” (Rubin, Conroy, and Chen 2015), achieving a 63% prediction accuracy over a 56% baseline. Given its data source, this study aims at identifying subtle differences in narrative between different stories, which is not well-suited for news articles. Similarly, Burfoot and Baldwin use SVMs on lexical and semantic features to automatically classify true news content from satire news content. The content features specifically geared towards satire significantly outperform the baseline and achieve high classification precision. (Burfoot and Baldwin 2009). Some of our features and one of our data sets are common with this study, though we use a much larger feature set that also captures the readability of the text and its word usage. Along the same lines, Rubin et al. propose an SVM-based algorithm with 5 language features (Rubin et al. 2016). They achieve 90% accuracy in detecting satire news from real news. Our features and classification method has some overlap with this work. However, these studies do not explicitly target fake news. Yet, as demonstrated by the many events of the 2016 US Election (Silverman 2016), fake and satirical news are clearly different in motivation. Conversely, fake news is motivated by deceiving its readers into thinking a completely false story is real, many times with malicious intent (Sydell 2016). This distinction between the content of three classes of news: satire, fake, and real, is the key contribution of our work. The spread of misinformation in networks has also been studied. Specifically, Bessi et al. study the attention given to misinformation on Facebook. They show that users who often interact with alternative media are more prone to interact with intentional false claims (Bessi et al. 2014). Very recently, Shao et al. launched a platform for tracking online misinformation called Hoaxy (Shao et al. 2016). Hoaxy gathers social news shares and fact-checking through a mix of web scraping, web syndication, and social network APIs. The goal of Hoaxy is to track both truthful and not truthful online information automatically. However, Hoaxy does not do any fact-checking of its own, rather relying on the efforts of fact-checkers such as snopes.com. In our work, we concentrate on content analysis to study fake news for several important reasons. Readers’ assessment of information play a significant role in decisions to disseminate it. Even though the abundance of information leads to limited attention given to each article, users engage with social media with the intention to find and share information. When controlling for attention and individual differences in the need to engage with information, the relevant arguments and message characteristics become the determinant in persuasion and attitude change (OKeefe 2008). Therefore, it is helpful to understand whether there are specific the message characteristics that accompany fake news articles being produced and widely shared. As, textual analysis is well studied in computer science and has been used for highly accurate document classification, it may prove quite helpful in stopping the spread of fake news (Sebastiani 2002). Furthermore, we hope that this type of analysis can be of benefit to grass-root fact-checkers by notifying them of potential fake articles earlier. This benefit in turn can provide more fact-checked content, ultimately helping systems like Hoaxy. Methodology To begin exploring the content of fake news, we develop strict definitions of what real, fake, and satire news stories are. Specifically, real news stories are stories that are known to be true and from well trusted news sources. Fake news stories are stories that are known to be false and are from well known fake news websites that are intentionally trying to spread misinformation. Satire news stories are stories that are from news sources that explicitly state they are satirical and do not intentionally spread misinformation. Satire news is explicitly produced for entertainment. Data sets With these definitions in mind, we use three independent data sets. Data set 1: Buzzfeed election data set First, we collected the news stories found in Buzzfeed’s 2016 article on fake election news on Facebook (Silverman 2016). Buzzfeed gathered this data using the content analysis tool BuzzSumo by first searching for real and fake stories getting the highest engagement on Facebook using various methods during the 9 months before the 2016 US Presidential Election, divided into three 3-month segments. For fake stories, they targeted articles with the highest engagement for key election terms and filtered these by known fake news sources. For real stories, they found the stories getting the highest engagement from well-known news organizations in the same time period. The URL and Facebook engagement statistics of the chosen fake and real stories, 60 each, were made available. To keep the ground truth of real/fake within our definitions, we filtered out stories that are opinion based or were explicitly satire; leaving us with 36 real news stories and 35 fake news stories. Other than this filtering, we took the ground truth as is; thus, it is important to keep in mind the limitations of this data set. First, we do not know if there was any selection bias in collecting the data, which could impact our results. Second, while we can say these were news stories with high user engagement, we cannot say anything about the actual traffic the stories generated. Despite these limitations, this data set provides reasonable ground truth labels and we know all stories were highly shared on social media. Data set 2: Our political news data set Given data set 1 contains political news only, we created our own political news data set to strengthen our analysis and control for the limitations of the first data set. Our data set contains 75 stories from each of the three defined categories of news: real, fake, and satire. We collected this data by first gathering known real, fake, and satire new sources, which can be found in Table 1. The fake news sources were collected from Zimdars’ list of fake and misleading news websites (Zimdars 2016) and have had at least 1 story show up as false on a fact-checking website like snopes.com in the past. The real sources come from Business Insider’s “MostTrusted” list (Engel 2014), and are well established news media companies. The satire sources are sites that explicitly state they are satirical on the front page of their website. Once the sources were chosen, we then randomly selected political stories from each of the sources. Each of these stories must be a “hard” news story, and not an opinion piece. The sources used in this data set have some overlap with data set 1, but all collected stories are different than those in the first data set. While we cannot say anything about the engagement of the articles in data set 2, it avoids the possible limitations of the Buzzfeed data set described above. Furthermore, data set 2 allows us to explicitly analyze all three defined categories of news by having both satire and fake news stories in the same data set. Both data sets 1 and 2 will be publicly available together with this paper at https://github.com/rpitrust/ fakenewsdata1. Data set 3: Burfoot and Baldwin data set Finally, we use a data set from (Burfoot and Baldwin 2009), which consists of 233 satire news stories and 4000 real news stories used in a classification task between these two types of news stories using lexical and semantic features. The authors collect the real news stories using newswire documents sampled from the English Gigaword Corpus. To select satire documents, they hand select satire stories that are closely related in topic to the real stories collected. They manually filter out “non-newsy” satire, similar to the method we use to construct our data set. While Burfoot and Baldwin do control for topic when matching articles, they do not limit themselves to only political topics; thus, this data may not be directly comparable to our other two data sets with respect to some of our more topic-driven features. We include this data set to strengthen our categorical comparisons between satire and real stories. A separate limitation of Burfoot and Baldwin’s data set is that we do not explicitly know the sources of each story and cannot verify that the definition of real news sources by the authors corresponds to ours. Real sources Wall Street Journal The Economist BBC NPR ABC CBS USA Today The Guardian NBC Washington Post Fake sources Ending the Fed True Pundit abcnews.com.co DC Gazette libertywritersnews Before its News Infowars Real News Right Now Satire sources The Onion Huff Post Satire Borowitz Report The Beaverton SatireWire Faking News Table 1: Data set 2 sources Features To study these different articles, we compute many content based features on each data set and categorize them into 3 broad categories: stylistic, complexity, and psychological. Stylistic Features The stylistic features are based on natural language processing to understand the syntax, text style, and grammatical elements of each article content and title. To test for differences in syntax, we use the Python Natural Language Toolkit (Bird 2006) part of speech (POS) tagger and keep a count of how many times each tag appears in the article. Along with this, we keep track of the number of stopwords, punctuation, quotes, negations (no, never, not), informal/swear words, interrogatives (how, when, what, why), and words that appear in all capital letters. For features that need word dictionaries, such as the number of informal words, we use the 2015 Linguistic Inquiry and Word Count (LIWC) dictionaries (Tausczik and Pennebaker 2010). Complexity Features The complexity features are based on deeper natural language processing computations to capture the overall intricacy of an article or title. We look at two levels of intricacy: the sentence level and the word level. To capture the sentence level complexity, we compute the number of words per sentence and each sentence’s syntax tree depth, noun phrase syntax tree depth, and verb phrase syntax tree depth using the Stanford Parser (de Marneffe, MacCartney, and Manning 2006). We expect that more words per sentence and deeper syntax trees mean the average sentence structure complexity is high. To capture the word level complexity, we use several key features. First, we compute the readability of each document using three different grade level readability indexes: Gunning Fog, SMOG Grade, and Flesh-Kincaid grade level index. Each measure computes a grade level reading score based on the number of syllables in words. A higher score means a document takes a higher education level to read. Second, we compute what is called the Type-Token Ratio Data set ID 1 2 3 Metadata description filtered Buzzfeed 2016 election data set our 3 class political news data set Burfoot and Baldwins satire data set # real news 36 75 4000 # fake news 35 75 0 # satire news 0 75 233 Table 2: Ground truth counts and ID number of each data set used in this study. Abbr. GI SMOG FK med depth med np depth med vp depth flu coca c flu coca d TTR avg wlen Description Gunning Fog Grade Readability Index SMOG Readability Index Flesh-Kincaid Grade Readability Index median depth of syntax tree median depth of noun phrase tree median depth of verb phrase tree avg. frequency of least common 3 words using all of the coca corpus avg. frequency of words in each document using all of the coca corpus Type-Token Ratio (lexical diversity) avg. length of each word (a) Complexity Features analytic insight cause discrep tentat certain differ affil power reward risk personal tone affect str neg str pos number of analytic words number of insightful words number of causal words number of discrepancy words number of tentative words number of certainty words number of differentiation words number of affiliation words number of power words number of reward words number of risk words number of personal concern words (work, leisure, religion, money) number of emotional tone words number of emotion words (anger, sad, etc.) strength of negative emotion using SentiStrength strength of positive emotion using SentiStrength (b) Psychology Features Abbr. WC WPS NN NNP PRP PRP$ WP DT WDT CD RB UH VB JJ VBD VBG VBN VBP VBZ focuspast focusfuture i we you shehe quant compare exclaim negate swear netspeak interrog all caps per stop allPunc quotes #vps Description word count words per sentence number of nouns number of proper nouns number of personal pronouns number of possessive pronouns Wh-pronoun number of determinants number of Wh-determinants number of cardinal numbers number of adverbs number of interjections number verbs Adjective number of past tense verbs Verb, gerund or present participle Verb, past participle Verb, non-3rd person singular present Verb, 3rd person singular present number of past tense words number of future tense words number of I pronouns (similar to PRP) number of we pronouns (similar to PRP) number of you pronouns (similar to PRP) number of she or he pronouns (similar to PRP) number of quantifying words number of comparison words number of exclamation marks number of negations (no, never, not) number of swear words number of online slang terms (lol, brb) number of interrogatives (how, what, why) number of word that appear in all capital letters percent of stop words (the, is, on) number of punctuation number of quotes number of verb phrases (c) Stylistic Features Table 3: Different features used in our study (TTR) of a document as the number of unique words divided by the total number of words in the document. TTR is meant to capture the lexical diversity of the vocabulary in a document. A low TTR means a document has more word redundancy and a high TTR means a document has more word diversity (Dillard and Pfau 2002). Third, we compute a word level metric called fluency, used in (Horne et al. 2016). Fluency is meant to capture how common or specialized the vocabulary of a document is. We would say that a common term is more fluent, and easier to interpret by others; while a less common term would be less fluent and more tech- nical. This idea is captured by computing how frequently a term in a document is found in a large English corpus. We use both the Corpus of Contemporary American English (COCA) (Davies 2008 ) corpus to compute this feature. Psychological Features The psychological features are based on well studied word counts that are correlated with different psychological processes, and basic sentiment analysis. We use Linguistic Inquiry and Word Count (LIWC) dictionaries (Tausczik and Pennebaker 2010) to measure cognitive processes, drives, and personal concerns. Along with this, we use LIWC to measure basic bag-of-words sen- timent. We then use SentiStrength (Thelwall et al. 2010) to measure the intensity of positive and negative emotion in each document. SentiStrength is a sentiment analysis tool that reports a negative sentiment integer score between -1 and -5 and a positive sentiment integer score between 1 and 5, where -5 is the most negative and 5 is the most positive. Statistical Tests and Classification Due to our use of small data sets and large number of features, we choose to first approach the problem using two well known hypothesis testing methods, the one-way ANOVA test and the Wilcoxon rank sum test, to find which features differ between the different categories of news. A one-way ANOVA test is used to compare the means of different groups on a dependent variable. ANOVA uses the ratio of the treatment and residual variances to determine if the difference in group means is due to random variation or the treatment. In our case, the treatment is grouping news as fake, real, or satire. ANOVA assumes that the variables are approximately normally distributed. While these assumptions are true for most of our features, we cannot assume it is true for all of them. Thus, we also utilize the Wilcoxon rank sum test to compare two distributions that are not normal. Specifically, for each feature that passes a normality test, we use the one-way ANOVA, otherwise, we will use Wilcoxon rank sum. In both cases, we are looking for a large F-value and a significance level of at least 0.05. These statistical test cannot say anything about predicting classes in the data, as a machine learning classifier would. However, these test can illustrate a shift in the use of a linguistic feature based on the category the article falls in. We would expect that the more significant the difference in a feature, the higher chance a machine learning classifier would be able to separate the data. To better test the predictive power of our features, we will use a linear classifier on a small subset of our features. We select the top 4 features from our hypothesis testing methods for both the body text and title text of the articles. With these 4 features, we will run a Support Vector Machine (SVM) model with a linear kernel and 5-fold cross-validation. Since we are only using a small number of features and our model is a simple linear model on balanced classes, we expect to avoid the over-fitting that comes with small data. Results In this section, we present the most significant features in separating news, referring to each data set by its ID number found in Table 2. The complete results can be found in Tables 4 and 5. Classification results can be found in Table 6. The content of fake and real news articles is substantially different. Our results show there is a significant difference in the content of real and fake news articles. Consistently between data sets 1 and 2, we find that real news articles are significantly longer than fake news articles and that fake news articles use fewer technical words, smaller words, fewer punctuation, fewer quotes, and more lexical redundancy. Along with this, in data set 2, fake articles need Feature WC flu coca c flu coca d flu acad c avg wlen quote allPunc GI FK analytic all caps NN PRP PRP$ DT WDT RB i we you shehe CD compare swear TTR avg negstr avg posstr med depth med np d med vp d Data set 1 Real > Fake Fake > Real Real > Fake Real > Fake Real > Fake Fake > Real Fake > Real Real > Fake Fake > Real Fake > Real Fake > Real Data set 2 Real > Fake > Satire Satire = Fake > Real Satire = Fake > Real Satire > Fake = Real Real > Fake = Satire Real > Fake = Satire Real > Fake = Satire Real = Satire > Fake Real = Satire > Fake Real > Fake = Satire Fake = Satire > Real Real > Fake = Satire Satire > Fake > Real Satire > Fake = Real Fake = Real > Satire Fake = Real > Satire Satire = Fake > Real Satire > Fake = Real Fake > Real = Satire Satire > Fake > Real Satire > Real = Fake Real = Fake > Satire Real > Fake = Satire Satire > Real = Fake Satire > Fake > Real Data set 3 Satire > Real Real > Satire Satire > Real Satire > Real Satire > Real Satire > Real Satire > Real Satire > Real Satire > Real Real > Satire Satire > Real Satire > Real Table 4: Features that differ in the body of the news content (bolded results correspond to p values of 0.00 or less, all other results have p values of at least less than 0.05) Feature WPS flu coca c avg wlen all caps GI FK NNP NN PRP PRP$ DT CD per stop exclaim focuspast analytic #vps Data set 1 Fake > Real Fake > Real Fake > Real Real > Fake Real > Fake Fake > Real Real > Fake Real > Fake Real > Fake Real > Fake Real > Fake Data set 2 Fake > Real = Satire Satire > Fake = Real Real > Fake = Satire Satire > Fake > Real Real > Satire = Fake Real > Satire = Fake Fake = Satire > Real Real > Satire > Fake Data set 3 Satire > Real Real > Satire Real > Satire Satire > Real Satire > Real Satire > Real Real > Satire = Fake Fake > Real = Satire Fake > Satire = Real Real > Satire Real > Satire Fake > Real Fake > Real = Satire Table 5: Features that differ in the title of the news content (bolded results correspond to p values of 0.00 or less, all other results have p values of at least less than 0.05) a slightly lower education level to read, use fewer analytic words, have significantly more personal pronouns, and use fewer nouns and more adverbs. These differences illustrate a strong divergence in both the complexity and the style of content. Fake news articles seem to be filled with less substantial information demonstrated by having a high amount of redundancy, more adverbs, fewer nouns, fewer analytic words, and fewer quotes. Our results also suggest that fake news may be more personal and more self-referential, using words like we, you, and us more often. However, this result is not consistent between data sets and is less significant. This stark difference between real and fake news content is further strengthened by our SVM classification results on the content of fake and real articles in Dataset 2. To classify the content, we use the top 4 features from our statistical analysis: number of nouns, lexical redundancy (TTR), word count, and number of quotes. We achieve a 71% crossvalidation accuracy over a 50% baseline when separating the body texts of real and fake news articles. Similarly, when classifying fake from real content in Data set 1 using the same 4 features, we achieve a 77% accuracy over a 57% baseline. These results are shown in Table 6. Body Title Baseline 50% 50% Fake vs Real 71% 78% Satire vs Real 91% 75% Satire vs Fake 67% 55% Table 6: Linear kernel SVM classification results using the top 4 features for the body and the title texts in Data set 2. Accuracy is the mean of 5-fold cross-validation. news is closely related to Rubin et al’s finding in the quotations of satirical articles (Rubin et al. 2016) where a large number of clauses are packed into a sentence. Saliently, this title structure is not as similar to clickbait as one might expect. Clickbait titles have been shown to have many more function words, more stopwords, more hyperbolic words (extremely positive), more internet slangs, and more possestive nouns rather than proper nouns (Chakraborty et al. 2016). Once again, these differences are further strengthened by our SVM classification results on Data set 2. To classify fake and real news articles by their title, we use the top 4 features from our statistical analysis on titles: the percent of stopwords, number of nouns, average word length, and FKE readability. We achieve a 78% cross-validation accuracy over a 50% baseline, demonstrating the importance of the title in predicting fake and real news. These results are shown in Table 6. In addition, when classifying fake from real titles in Data set 1 using the same 4 features, we achieve a 71% accuracy over a 49% baseline. Table 7 shows the distributions of select features that are significant and consistent between data sets. Data set 1 Data set 2 Titles are a strong differentiating factor between fake and real news. When looking at just the titles of fake and real news articles, we find an even stronger dissimilarity between the two, with high consistency between the data sets and high statistical significance in the differences. Precisely, we find that fake news titles are longer than real news titles and contain simpler words in both length and technicality. Fake titles also used more all capitalized words, significantly more proper nouns, but fewer nouns overall, and fewer stopwords. Interestingly, we also find that in data set 1 fake titles use significantly more analytical words and in data set 2 fake titles use significantly more verb phrases and significantly more past tense words. Overall, these results suggest that the writers of fake news are attempting to squeeze as much substance into the titles as possible by skipping stop-words and nouns to increase proper nouns and verb phrases. Looking at an example from our data will solidify this notion: 1. FAKE TITLE: ”BREAKING BOMBSHELL: NYPD Blows Whistle on New Hillary Emails: Money Laundering, Sex Crimes with Children, Child Exploitation, Pay to Play, Perjury” 2. REAL TITLE: Preexisting Conditions and Republican Plans to Replace Obamacare There is a clear difference in the number of claims being made in each title. The fake title uses many verb phrases and name entities to get many points across, while the real title opts for a brief and general summary statement. This broad pattern shown in the example is consistent across the first two data sets. Interestingly, this finding in the titles of fake Table 7: 95% Confidence plots of all caps, NNP, and per stop. These features were found to be significant across both data sets 1 and 2 for fake and real titles. Top: all caps Middle: NNP Bottom: per stop Fake content is more closely related to satire than to real. When adding in satire articles to the analysis, we find that the majority of our features distributions are common between satire and fake. Specifically, both satire and fake use smaller, fewer technical, and fewer analytic words, as well as, fewer quotes, fewer punctuation, more adverbs, and fewer nouns than real articles. Further, fake and satire use significantly more lexical redundancy than real articles. These claims are further supported by SVM classification results in Table 6. When using the number of nouns, lexical redundancy (TTR), word count, and number of quotes for the body text, we achieve a 91% cross-validation accuracy over a 50% baseline on separating satire from real articles. On the other hand, when separating satire from fake articles, we only acheive a 67% accuracy over a 50% baseline. Similarly, when classifying the titles of articles using the percent of stopwords, number of nouns, average word length, and FKE readability, we achieve a 75% cross-validation accuracy when separating satire titles from real titles, but only achieve a 55% accuracy when separating satire titles from fake titles. While this accuracy is a reasonable improvement over baseline, it is not nearly as high as big of an improvement as separating satire from real or real from fake. Overall, these results paint an interesting picture where satire and fake news articles are written in a less investigative way. This conclusion is in sync with what we know of satire (Randolph 1942) (Burfoot and Baldwin 2009). Satire news articles do not have the goal of creating sound arguments, but often make absurd and eccentric claims, whereas real news articles must back up the information they are providing with direct quotes, domain specific knowledge, and reasonable analysis. Real news persuades through arguments, while fake news persuades through heuristics. To better explain our findings, we look at the Elaboration Likelihood Model (ELM) of persuasion, well-studied in communications. According to ELM, people are persuaded through two different routes: central and peripheral (Petty and Cacioppo 1986). The central route of persuasion results from the attentive examination of the arguments and message characteristics presented, involving a high amount of energy and cognition. In opposition, the peripheral route of persuasion results from associating ideas or making conjectures that are unrelated to the logic and quality of the information presented. This method could be called a heuristic method as it does not ensure to be optimal or even sufficient in achieving its objective of finding correct or truthful information. The peripheral route takes very little energy and cognition. This model fits well with both our results and recent studies on the behavior of sharing online information. Given the similarity of fake content to satire, we hypothesize that fake content targets the peripheral route, helping the reader use simple heuristics to assess veracity of the information. The significant features support this finding: fake news places a high amount substance and claims into their titles and places much less logic, technicality, and sound arguments in the body text of the article. Several studies have argued that the majority of the links shared or commented on in social networks are never clicked, and thus, only the titles of the articles are ever read (Wang, Ramachandran, and Chaintreau 2016) Titles of fake news often present claims about people and entities in complete sentences, associating them with actions. Therefore, titles serve as the main mechanism to quickly make claims which are easy to assess whether they are believable based on the reader’s existing knowl- edge base. The body of fake news articles add relatively little new information, but serves to repeat and enhance the claims made in the title. The fake content is more negative in general, similar to findings in past work (Zollo et al. 2015b). Hence, fake news is assessed through the peripheral route. We hypothesize that if users were to utilize the central route of persuasion, they would have a much lower chance of being convinced by the body content of fake news, as we have shown the fake news content tend to be short, repetitive and lacking arguments. Conclusions and Future work In this paper, we showed that fake and real news articles are notably distinguishable, specifically in the title of the articles. Fake news titles use significantly fewer stop-words and nouns, while using significantly more proper nouns and verb phrases. We also conclude that the complexity and style of content in fake news is more closely related to the content of satire news. We also showed that our features can be used to significantly improve the prediction of fake and satire news, achieving between 71% and 91% accuracy in separating from real news stories. These results lead us to consider the Elaboration Likelihood Model as a theory to explain the spread and persuasion of fake news. We conclude that real news articles persuade users through sound arguments while fake news persuades users through heuristics. This finding is concerning as a person may be convinced of fake news simply out of having low energy, not just contempt, negligence, or low cognition. Unfortunately, misleading claims in the titles of fake news articles can lead to established beliefs which can be hard to change through reasoned arguments. As a starting point, articles that aim to counter fake claims should consider packing the counter-claim into their titles. This work has some limitations and room for improvement. First, we would like to extensively expand the news data set. Typically it is very difficult to obtain a non-noisy ground truth for fake and real news, as the real news is becoming increasingly opinion based and many times more detailed fact checking is needed. We would like to make more objective clusters of fake, real, and satire news through unsupervised machine learning methods on a variety of news sources. With an expanded data set and stronger ground truth comes the ability to do more sophisticated classification and more in-depth natural language feature engineering, all with the hope to stop the spread of malicious fake news quickly. With more data and more in-depth features, our arguments could be made much stronger. Second, we would like to conduct user studies to more directly capture the persuasion mechanisms of fake and real news. For users studies such as this to be valid, careful planning and ethical considerations are needed. Largely, we hope this work helps the academic community continue to build technology and a refined understanding of malicious fake news. Acknowledgments Research was sponsored by the Army Research Laboratory and was accomplished under Cooperative Agreement Number W911NF-09-2-0053 (the ARL Network Science CTA). The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the Army Research Laboratory or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for Government purposes notwithstanding any copyright notation here on. References [Bakshy, Messing, and Adamic 2015] Bakshy, E.; Messing, S.; and Adamic, L. A. 2015. Exposure to ideologically diverse news and opinion on facebook. Science 348(6239):1130–1132. [Bessi et al. 2014] Bessi, A.; Scala, A.; Rossi, L.; Zhang, Q.; and Quattrociocchi, W. 2014. The economy of attention in the age of (mis)information. Journal of Trust Management 1(1):105–13. [Bessi et al. 2015] Bessi, A.; Coletto, M.; Davidescu, G. A.; Scala, A.; Caldarelli, G.; and Quattrociocchi, W. 2015. Science vs Conspiracy: Collective Narratives in the Age of Misinformation. PLoS ONE 10(2):e0118093–17. [Bird 2006] Bird, S. 2006. Nltk: the natural language toolkit. In Proc. of the COLING/ACL, 69–72. [Burfoot and Baldwin 2009] Burfoot, C., and Baldwin, T. 2009. Automatic satire detection: Are you having a laugh? In Proc. ACL-IJCNLP 2009, 161–164. [Chakraborty et al. 2016] Chakraborty, A.; Paranjape, B.; Kakarla, S.; and Ganguly, N. 2016. Stop clickbait: Detecting and preventing clickbaits in online news media. In ASONAM, 9–16. IEEE. [Davies 2008 ] Davies, M. 2008-. The corpus of contemporary american english: 520 million words, 1990present. Available at “http://corpus.byu.edu/coc a/”. Accessed: 7/21/2015. [de Marneffe, MacCartney, and Manning 2006] de Marneffe, M.-C.; MacCartney, B.; and Manning, C. D. 2006. Generating typed dependency parses from phrase structure parses. In LREC 2006. [Dillard and Pfau 2002] Dillard, J. P., and Pfau, M. 2002. The persuasion handbook: Developments in theory and practice. Sage Publications. [Engel 2014] Engel, P. 2014. Here are the most- and least-trusted news outlets in america. Available at “http://www.businessinsider.com/here-are -the-most-and-least-trusted-news-outle ts-in-america-2014-10”. [Horne et al. 2016] Horne, B. D.; Nevo, D.; Freitas, J.; Ji, H.; and Adali, S. 2016. Expertise in social networks: How do experts differ from other users? In ICWSM. [OKeefe 2008] OKeefe, D. J. 2008. Elaboration likelihood model. International encyclopedia of communication 4:1475–1480. [Petty and Cacioppo 1986] Petty, R. E., and Cacioppo, J. T. 1986. The elaboration likelihood model of persuasion. In In Communication and Persuasion. New York: Springer. 1–24. [Randolph 1942] Randolph, M. C. 1942. The structural design of the formal verse satire. Philological Quarterly 21:368. [Rubin et al. 2016] Rubin, V. L.; Conroy, N. J.; Chen, Y.; and Cornwell, S. 2016. Fake news or truth? using satirical cues to detect potentially misleading news. In Proceedings of NAACL-HLT, 7–17. [Rubin, Conroy, and Chen 2015] Rubin, V. L.; Conroy, N. J.; and Chen, Y. 2015. Towards news verification: Deception detection methods for news discourse. In Rapid Screening Technologies, Deception Detection and Credibility Assessment Symposium, HICSS48. [Sebastiani 2002] Sebastiani, F. 2002. Machine learning in automated text categorization. ACM computing surveys (CSUR) 34(1):1–47. [Shao et al. 2016] Shao, C.; Ciampaglia, G. L.; Flammini, A.; and Menczer, F. 2016. Hoaxy: A platform for tracking online misinformation. In WWW 2016, 745–750. Available at “http://dx.doi.org/10.1145/2872518. 289009”. [Silverman 2016] Silverman, C. 2016. This analysis shows how fake election news stories outperformed real news on facebook. Available at“https://www.buzzfeed.com /craigsilverman/viral-fake-election-new s-outperformed-real-news-on-facebook?ut m_term=.gl5qYwrP#.gt1ygzDN”. [Sydell 2016] Sydell, L. 2016. We tracked down a fake-news creator in the suburbs. here’s what we learned. Available at “http://www.npr.org/sections/alltechcons idered/2016/11/23/503146770/npr-finds-t he-head-of-a-covert-fake-news-operation -in-the-suburbs”. [Tausczik and Pennebaker 2010] Tausczik, Y. R., and Pennebaker, J. W. 2010. The psychological meaning of words: Liwc and computerized text analysis methods. Journal of language and social psychology 29(1):24–54. [Thelwall et al. 2010] Thelwall, M.; Buckley, K.; Paltoglou, G.; Cai, D.; and Kappas, A. 2010. Sentiment strength detection in short informal text. Journal of the American Society for Information Science and Technology 61(12):2544–2558. [Wang, Ramachandran, and Chaintreau 2016] Wang, L. X.; Ramachandran, A.; and Chaintreau, A. 2016. Measuring click and share dynamics on social media: a reproducible and validated approach. In ICWSM. [Zimdars 2016] Zimdars, M. 2016. False, misleading, clickbait-y, and satirical news sources’. Available at “https://docs.google.com/document/d/10e A5-mCZLSS4MQY5QGb5ewC3VAL6pLkT53V_81Zyi tM/preview”. [Zollo et al. 2015a] Zollo, F.; Bessi, A.; Del Vicario, M.; Scala, A.; Caldarelli, G.; Shekhtman, L. M.; Havlin, S.; and Quattrociocchi, W. 2015a. Debunking in a World of Tribes. CoRR. [Zollo et al. 2015b] Zollo, F.; Novak, P. K.; Del Vicario, M.; Bessi, A.; Mozetič, I.; Scala, A.; Caldarelli, G.; and Quat- trociocchi, W. 2015b. Emotional Dynamics in the Age of Misinformation. PLoS ONE 10(9):e0138740–22.
© Copyright 2024 Paperzz