Skip to main content

Positive words carry less information than negative words

Abstract

We show that the frequency of word use is not only determined by the word length [1] and the average information content [2], but also by its emotional content. We have analyzed three established lexica of affective word usage in English, German, and Spanish, to verify that these lexica have a neutral, unbiased, emotional content. Taking into account the frequency of word usage, we find that words with a positive emotional content are more frequently used. This lends support to Pollyanna hypothesis [3] that there should be a positive bias in human expression. We also find that negative words contain more information than positive words, as the informativeness of a word increases uniformly with its valence decrease. Our findings support earlier conjectures about (i) the relation between word frequency and information content, and (ii) the impact of positive emotions on communication and social links.

PACS Codes:89.65.-s, 89.70.Cf, 89.90.+n.

1 Introduction

One would argue that human languages, in order to facilitate social relations, should be biased towards positive emotions. This question becomes particularly relevant for sentiment classification, as many tools assume as null hypothesis that human expression has neutral emotional content [4, 5], or reweight positive and negative emotions [6] without a quantification of the positive bias of emotional expression. We have tested and measured this bias in the context of online written communication by analyzing three established lexica of affective word usage. These lexica cover three of the most used languages on the Internet, namely English [7], German [8], and Spanish [9]. The emotional content averaged over all the words in each of them is neutral. Considering, however, the everyday usage frequency of these words we find that the overall emotion of the three languages is strongly biased towards positive values, because words associated with a positive emotion are more frequently used than those associated with a negative emotion.

Historically, the frequency of words was first analyzed by Zipf [1, 10] showing that frequency predicts the length of a word as result of a principle of least effort. Zipf’s law highlighted fundamental principles of organization in human language [11], and called for an interdisciplinary approach to understand its origin [1214] and its relation to word meaning [15]. Recently Piantadosi et al. [2] extended Zipf’s approach by showing that, in order to have efficient communication, word length increases with information content. Further discussions [1517] highlighted the relevance of meaning as part of the communication process as, for example, more abstract ideas are expressed through longer words [18]. Our work focuses on one particular aspect of meaning, namely the emotion expressed in a word, and how this is related to word frequency and information content. This approach requires additional data beyond word length and frequency, which became available thanks to large datasets of human behaviour on the Internet. Millions of individuals write text online, for which a quantitative analysis can provide new insights into the structure of human language and even provide a validation of social theories [19]. Sentiment analysis techniques allow to quantify the emotions expressed through posts and messages [5, 6]. Recent studies have provided statistical analyses [2023] and modelling approaches [24, 25] of individual and collective emotions on the Internet.

An emotional bias in written expressions, however, would have a strong impact, as it shifts the balance between positive and negative expressions. Thus, for all researchers dealing with emotions in written text it would be of particular importance to know about such bias, how it can be quantified, and how it affects the baseline, or reference point, for expressed emotions. Our investigation is devoted to this problem by combining two analyses, (i) quantifying the emotional content of words in terms of valence, and (ii) quantifying the frequency of word usage in the whole indexable web [26]. We provide a study of the baseline of written emotional expression on the Internet in three languages that span more than 67.7% of the websites [27]: English (56.6%), German (6.5%), and Spanish (4.6%). These languages are used everyday by more than 805 million users, who create the majority of the content available on the Internet.

In order to link the emotionality of each word with the information it carries, we build on the recent work of Piantadosi et al. [2]. This way, we reveal the importance of emotional content in human communication which influences the information carried by words. While the rational process that optimizes communication determines word lengths by the information they carry, we find that the emotional content affects the word frequency such that positive words appear more frequently. This points towards an emotional bias in used language and supports Pollyanna hypothesis [3], which asserts that there is a bias towards the usage of positive words. Furthermore, we extend the analysis of information content by taking into account word context rather than just word frequency. This leads to the conclusion that positive words carry less information than negative ones. In other words, the informativeness of words highly depends on their emotional polarity.

We wish to emphasize that our work distinguishes itself both regarding its methodology and its findings from a recent article [28]. There, the authors claim a bias in the amount of positive versus negative words in English, while no relation between emotionality and frequency of use was found. A critical examination of the conditions of that study shows that the quantification of emotions was done in an uncontrolled setup through the Amazon Mechanical Turk. Participants were shown a scale similar to the ones used in previous works [79], as explained in [23]. Thanks to the popular usage of the Mechanical Turk, the authors evaluated more than 10,000 terms from the higher frequency range in four different corpora of English expression. However, the authors did not report any selection criterion for the participant reports, opposed to the methodology presented in [29] where up to 50% of the participants had to be discarded in some experiments.

Because of this lack of control in their experimental setup, the positive bias found in [28] could be easily explained as an acquiescent bias [30, 31], a result of the human tendency to agree in absence of further knowledge or relevance. In particular, this bias has been repeatedly shown to exist in self assessments of emotions [32, 33], requiring careful response formats, scales, and analyses to control for it. Additionally, the wording used to quantify word emotions in [28] (happiness), could imply two further methodological biases: The first one is a possible social desirability bias [34], as participants tend to modify their answers towards more socially acceptable answers. The positive social perception of displaying happiness can influence the answers given by the participants of the study. Second, the choice of the word happiness implies a difference compared with the standard psychological term valence[35]. Valence is interpreted as a static property of the word while happiness is understood as a dynamic property of the surveyed person when exposed to the word. This kind of framing effects have been shown to have a very large influence in survey results. For example, a recent study [36] showed a large change in the answers by simply changing voting for being a voter in a voter turnout survey.

Hence, there is a strong sensitivity to such influences which are not controlled for in [28]. Because of all these limitations, in our analysis we chose to use the current standard lexica of word valence. These lexica, albeit limited to 1,000 to 3,000 words, were produced in three controlled, independent setups, and provide the most reliable estimation of word emotionality for our analysis. Our results on these lexica are consistent with recent works on the relation between emotion and word frequency [37, 38] for English in corpora of limited size.

2 Results

2.1 Frequency of emotional words

In detail, we have analyzed three lexica of affective word usage which contain 1,034 English words, 2,902 German words and 1,034 Spanish words, together with their emotional scores obtained from extensive human ratings. These lexica have effectively established the standard for emotion analyses of human texts [39]. Each word in these lexica is assigned a set of values measuring different aspects of word emotionality. The three independent studies that generated the lexica for English [7], German [8], and Spanish [9] used the Self-Assessment Mannequin (SAM) method to ask participants about the different emotion values associated to each word in the lexicon. One of these values, a scalar variable v called valence, represents the degree of pleasure induced by the emotion associated with the word, and it is known to explain most of the variance in emotional meaning [35]. In this article, we use v to quantify word emotionality.

In each lexicon, words were chosen such that they evenly span the full range of valence.a In order to compare the emotional content of the three different languages, we have rescaled all values of v to the interval [1,1]. As shown in the left panel of Figure 2, indeed, the average valence, as well as the median, of all three lexica is very close to zero, i.e. they do not provide an emotional bias. This analysis, however, neglects the actual frequency of word usage, which is highly skew distributed [1, 10]. For our frequency estimations we have used Google’s N-gram dataset [26] which, with 1012 tokens, is one of the largest datasets available about real human text expressions on the Internet. For our analysis, we have studied the frequency of the words which have an affective classification in the respective lexicon in either English, German, or Spanish. Figure 1 shows emotion word clouds for the three languages, where each word appears with a size proportional to its frequency. The color of a word is chosen according to its valence, ranging from red for v=1 to green for v=+1. It is clear that green dominates over red in the three cases, as positive emotions predominate on the Internet. Some outliers, like “home”, have a special higher frequency of appearance in websites, but as we show later, our results are consistent with frequencies measured from traditional written texts like books.

Figure 1
figure 1

Emotion word clouds with frequencies calculated from Google’s crawl. In each word cloud for English (left), German (middle), and Spanish (right), the size of a word is proportional to its frequency of appearance in the trillion-token Google N-gram dataset [26]. Word colors are chosen from red (negative) to green (positive) in the valence range from psychology studies <abbrgrp>789</abbrgrp>. For the three languages, positive words predominate on the Internet.

In a general setup, the different usage of words with the same valence is quite obvious. For example, both words “party” and “sunrise” have the same positive valence of 0.715, however the frequency of “party” is 144.7 per one million words compared to 6.8 for “sunrise”. Similarly, both “dead” and “distressed” have a negative valence of −0.765, but the former appears 48.4 times per one million words, the latter only 1.6 times. Taking into account all frequencies of word usage, we find for all three languages that the median shifts considerably towards positive values. This is shown in the right panel of Figure 2. Wilcoxon tests show that the means of these distributions are indeed different, with an estimated difference in a 95% confidence interval of 0.257±0.032 for English, 0.167±0.017 for German, and 0.287±0.035 for Spanish. Hence, with respect to usage we find evidence that the language used on the Internet is emotionally charged, i.e. significantly different from being neutral. This affects quantitative analyses of the emotions in written text, because the “emotional reference point” is not at zero, but at considerably higher valence values (about 0.3).

Figure 2
figure 2

Distributions of word emotions weighted by the frequency of word usage. (left panel) Distributions of reported valence values for words in English (top panel, lexicon: [7], 1,034 entries), German (middle panel, lexicon: [8], 2,902 entries), and Spanish (bottom panel, lexicon: [9], 1,034 entries), normalized by the size of the lexica. Average valence (median) 0.048 (0.095) for English, 0.021 (0.067) for German, and −0.065 (−0.006) for Spanish. (right panel) Normalized distributions of reported valence values weighted by the frequency of word usage, obtained from the same lexica. Average valence (median) 0.314 (0.375) for English, 0.200 (0.216) for German, and 0.238 (0.325) for Spanish. The dashed lines indicate the median. Inset numbers: ratio of positive and negative areas in the corresponding distributions.

2.2 Relation between information and valence

Our analysis suggests that there is a definite relation between word valence and frequency of use. Here we study the role of emotions in the communication process building on the relation between information measures and valence. While we are unable to measure information perfectly, we can approximate its value given the frequencies of words and word sequences. First we discuss the relation between word valence and information content estimated from the simple word occurrences, namely self-information. Then we explain how this extends when the information is measured taking into account the different contexts in which a word can appear. The self-information of a word, I(w)[40] is an estimation of the information content from its probability of appearance, P(w), as

I(w)=logP(w)
(1)

Frequency-based information content metrics like self-information are commonly used in computational linguistics to systematically analyze communication processes. Information content is a better predictor for word length than word frequency [2, 41], and the relation between information content and meaning, including emotional content, is claimed to be crucial for the way humans communicate [1517]. We use the self-information of a word as an estimation of information content for a context size of 1, to build up later on larger context sizes. This way, we frame our analysis inside the larger framework of N-gram information measures, aiming at an extensible approach that can be incorporated in the fields of computational linguistics and sentiment analysis.

For the three lexica, we calculated I(w) of each word and linked it to its valence, v(w). As defined in Equation 1, very common words provide less information than very unusual ones, but this nonlinear mapping between frequency and self-information makes the latter more closely related to word valence than the former. The first two lines of Table 1 show the Pearson’s correlation coefficient of word valence and frequency ρ(v,f), followed by the correlation coefficient between word valence and self-information, ρ(v,I). For all three languages, the absolute value of the correlation coefficient with I is larger than with f, showing that self-information provides more knowledge about word valence than plain frequency of use.

Table 1 Correlations between word valence and information measurements.

The right column of Figure 3 shows in detail the relation between v and I. From the clear negative correlation found for all three languages (between −0.3 and −0.4), we deduce that words with less information content carry more positive emotions, as the average valence decreases along the self-information range. As mentioned before the Pearson’s correlation coefficient between word valence and self-information, ρ(v,I), is significant and negative for the three languages (Table 1). Our results outperform a recent finding [37] that, while focusing on individual text production, reported a weaker correlation (below 0.3) between the logarithm of word usage frequency and valence. This previous analysis was based on a much smaller data set from Internet discussions (in the order of 108 tokens) and the same English lexicon of affective word usage [7] we used. Using a much higher accuracy in estimating word frequencies and extending the analysis to three different languages, we were able to verify that there is a significant relation between the emotional content of a word and its self-information, impacting the frequency of usage.

Figure 3
figure 3

Relation between information measures and valence. Each graphic on the left column shows the relation between valence and information content measured up to a context of size four. Each bar represents a bin containing 10% of the words in the lexicon, with a size proportional to the average information content of the words in the bar. The color of each bar ranges from red to green, representing the average valence of the words in the corresponding bin. Each bar has a color gradient according to the standard error of the valence mean. Information content has been rescaled so it can be compared among context sizes. For all three languages and context sizes, negativity increases with information content. The second column shows the relation between word self-information and valence for English, German, and Spanish. Average valence is shown for bins that contain 5% of the data, with error bars showing the standard error. For all the three languages, valence clearly decreases with the self-information of the word, i.e. positive words carry less information than negative words.

Eventually, we also performed a control analysis using alternative frequency datasets, to account for possible anomalies in the Google dataset due to its online origin. We used the word frequencies estimated from traditional written corpuses, i.e. books, as reported in the original datasets for English [42], German [43], and Spanish [44]. Calculating the self-information from these and relating them to the valences given, we obtained similar, but slightly lower Pearson’s correlation coefficients ρ(v, I ) (see Table 1). So, we conclude that our results are robust across different types of written communication, for the three languages analyzed.

It is not surprising to find a larger self-information for negative words, as their probability of appearance is generally lower. The amount of information carried by a word is also highly dependent on its context. Among other factors, the context is defined by the word neighborhood in the sentence. For example, the word “violent” contains less information in the sentence “dangerous murderers are violent” than in “fluffy bunnies are violent”, as the probability of finding this particular word is larger when talking about murderers than about bunnies. For this reason we evaluate how the context of a word impacts its informativeness and valence. The intuition behind measuring information depending on the context is that the information content of a word depends primarily on i) the amount of contexts it can appear and ii) the probability of appearance in each one of these contexts. Not only the most infrequent, but the most specific and unexpectable words are the ones that carry the most information. Given each context c i where a word w appears, the information content is defined as

1 N i = 1 N log(P(W=w|C= c i ))
(2)

where N is the total frequency of the word in the corpus used for the estimation. These values were calculated as approximations of the information content given the words surrounding w up to size 4.

We analyzed how word valence is related to the information content up to context size 4 using the original calculations provided by Piantadosi et al. [2]. This estimation is based on the frequency of sequences of N words, called N-grams, from the Google dataset [26] for N{2,3,4}. This dataset contains frequencies for single words and N-grams, calculated from an online corpus of more than a trillion tokens. The source of this dataset is the whole Google crawl, which aimed at spanning a large subset of the web, providing a wide point of view on how humans write on the Internet. For each size of the context N, we have a different estimation of the information carried by the studied words, with self-information representing the estimation from a context of size 1.

The left column of Figure 3 shows how valence decreases with the estimation of the information content for each context size. Each bar represents the same amount of words within a language and has an area proportional to the rescaled average information content carried by these words. The color of each bar represents the average valence of the binned words. The decrease of average valence with information content is similar for estimations using 2-grams and 3-grams. For the case of 4-grams it also decreases for English and Spanish, but this trend is not so clear for German. These trends are properly quantified by Pearson’s correlation coefficients between valence and information content for each context size (Table 1). Each correlation coefficient becomes smaller for larger sizes of the context, as the information content estimation includes a larger context but becomes less accurate.

2.3 Additional analysis of valence, length and self-information

In order to provide additional support for our results, we tested different hypotheses impacting the relation between word usage and valence. First, we calculated Pearson’s and Spearman’s correlation coefficients between the absolute value of the valence and the self-information of a word, ρ(abs(v),I) (see Table 2). We found both correlation coefficients to be around 0.1 for German and Spanish, while they are not significant for English. The dependence between valence and self-information disappears if we ignore the sign of the valence, which means, indeed, that the usage frequency of a word is not just related to the overall emotional intensity, but to the positive or negative emotion expressed by the word.

Table 2 Additional correlations between valence, self-information and length.

Subsequently, we found that the correlation coefficient between word length and self-information (ρ(l,I)) is positive, showing that word length increases with self-information. These values of ρ(l,I) are consistent with previous results [1, 2]. Pearson’s and Spearman’s correlation coefficients between valence and length ρ(v,l) are very low or not significant. In order to test the combined influence of valence and length to self-information, we calculated the partial correlation coefficients ρ(v,I|l) and ρ(l,I|v). The results are shown in Table 2, and are within the 95% confidence intervals of the original correlation coefficients ρ(v,I) and ρ(l,I). This provides support for the existence of an additional dimension in the communication process closely related to emotional content rather than communication efficiency. This is consistent with the known result that word lengths adapt to information content [2], and we discover the independent semantic feature of valence. Valence is also related to information content but not to the symbolic representation of the word through its length.

Finally, we explore the sole influence of context by controlling for word frequency. In Table 3 we show the partial correlation coefficients of valence with information content for context sizes between 2 and 4, controlling for self-information. We find that most of the correlations keep significant and of negative sign, with the exception of I 2 for English. The weaker correlation for context sizes of 2 is probably related to two word constructions such as negations, articles before nouns, or epithets. These high-frequent, low-information constructions lead to the conclusion that I 2 does not explain more about the valence than self-information in English, as short range word interactions change the valence of the whole particle. This finding supports the assumption of many lexicon-based unsupervised sentiment analysis tools, which consider valence modifiers for two-word constructions [5, 6]. On the other hand, the significant partial correlation coefficients with I 3 and I 4 suggest that word information content combines at distances longer than 2, as longer word constructions convey more contextual information than 2-grams. Knowing the possible contexts of a word up to distance 4 provides further information about word valence than sole self-information.

Table 3 Partial correlation coefficients between valence and information content.

3 Discussion

Our analysis provides strong evidence that words with a positive emotional content are used more often. This lends support to Pollyanna hypothesis [3], i.e. positive words are more often used, for all the three languages studied. Our conclusions are consistent for, and independent of, different corpuses used to obtain the word frequencies, i.e. they are shown to hold for traditional corpuses of formal written text, as well as for the Google dataset and cannot be attributed as artifacts of Internet communication.

Furthermore, we have pointed out the relation between the emotional and the informational content of words. Words with negative emotions are less often used, but because of their rareness they carry more information, measured in terms of self-information, compared to positive words. This relation remains valid even when considering the context composed of sequences of up to four words (N-grams). Controlling for word length, we find that the correlation between information and valence does not depend on the length, i.e. it is indeed the usage frequency that matters.

In our analysis, we did not explore the role of syntactic rules and grammatical classes such as verbs, adjectives, etc. However, previous studies have shown the existence of a similar bias when studying adjectives and their negations [38]. The question of how syntax influences emotional expression is beyond the scope of the present work. Note that the lexica we use are composed mainly of nouns, verbs and adjectives, due to their emotional relevance. Function words such as “a” or “the” are not considered to have any emotional content and therefore were excluded from the original studies. In isolation, these function words do not contain explicit valence content, but their presence in text can modify the meaning of neighboring words and thus modify the emotional content of a sentence as a whole. Our analysis on partial correlations show that there is a correlation between the structure of a sentence and emotional content beyond the simple appearance of individual words. This result suggests the important role of syntax in the process of emotional communication. Future studies can extend our analysis by incorporating valence scores for word sequences, exploring how syntactical rules represent the link between context and emotional content.

The findings reported in this paper suggest that the process of communication between humans, which is known to optimize information transfer [2], also creates a bias towards positive emotional content. A possible explanation is the basic impact of positive emotions on the formation of social links between humans. Human communication should reinforce such links, which it both shapes and depends on. Thus, it makes much sense that human languages on average have strong bias towards positive emotions, as we have shown (see Figure 2). Negative expressions, on the other hand, mostly serve a different purpose, namely that of transmitting highly informative and relevant events. They are less used, but carry more information.

Our findings are consistent with emotion research in social psychology. According to [45], the expression of positive emotions increases the level of communication and strengthens social links. This would lead to stronger pro-social behaviour and cooperation, giving evolutionary advantage to societies whose communication shows a positive bias. As a consequence, positive sentences would become more frequent and even advance to a social norm (cf. “Have a nice day”), but they would provide less information when expressed. Our analysis provides insights on the asymmetry of evaluative processes, as frequent positive expression is consistent with the concept of positivity offset introduced in [46] and recently reviewed in [47]. In addition, Miller’s negativity bias (stronger influence of proximal negative stimuli) found in experiments provides an explanation for the higher information content of negative expression. When writing, people could have a tendency to avoid certain negative topics and bring up positive ones just because it feels better to talk about nice things. That would lower the frequency of negative words and lower the amount of information carried by positive expression, as negative expression would be necessary to transmit information about urgent threats and dangerous events.

Eventually, we emphasize that the positive emotional “charge” of human communication has a further impact on the quantitative analysis of communication on the Internet, for example in chatrooms, forums, blogs, and other online communities. Our analysis provides an estimation of the emotional baseline of human written expression, and automatic tools and further analyses will need to take this into account. In addition, this relation between information content and word valence might be useful to detect anomalies in human emotional expression. Fake texts supposed to be written by humans could be detected, as they might not be able to reproduce this spontaneous balance between information content and positive expression.

Author’s contributions

All authors designed and performed research, analyzed data, and wrote the article.

Endnote

The lexica focus on single words rather than on phrases or longer expressions.

References

  1. Zipf GK: The psycho-biology of language. Houghton Mifflin, Oxford; 1935.

    Google Scholar 

  2. Piantadosi ST, Tily H, Gibson E: Word lengths are optimized for efficient communication. Proc Natl Acad Sci USA 2011,108(9):3526. 10.1073/pnas.1012551108

    Article  Google Scholar 

  3. Boucher J, Osgood C: The Pollyanna hypothesis. J Verbal Learn Verbal Behav 1969,8(1):1–8. 10.1016/S0022-5371(69)80002-2

    Article  Google Scholar 

  4. Pang B, Lee L: Opinion mining and sentiment analysis. Found Trends Inf Retr 2008,2(12):1–135.

    Article  Google Scholar 

  5. Thelwall M, Buckley K, Paltoglou G: Sentiment in Twitter events. J Am Soc Inf Sci Technol 2011,62(2):406–418. 10.1002/asi.21462

    Article  Google Scholar 

  6. Taboada M, Brooke J, Voll K: Lexicon-based methods for sentiment analysis. Comput Linguist 2011,37(2):267–307. 10.1162/COLI_a_00049

    Article  Google Scholar 

  7. Bradley MM, Lang PJ (1999) Affective norms for English words (ANEW): instruction manual and affective ratings. Technical report, The Center for Research in Psychophysiology, University of Florida

    Google Scholar 

  8. Võ ML-H, Conrad M, Kuchinke L, Urton K, Hofmann MJ, Jacobs AM: The Berlin affective word list reloaded (BAWL-R). Behav Res Methods 2009,41(2):534–538. 10.3758/BRM.41.2.534

    Article  Google Scholar 

  9. Redondo J, Fraga I, Padrón I, Comesaña M: The Spanish adaptation of ANEW (affective norms for English words). Behav Res Methods 2007,39(3):600–605. 10.3758/BF03193031

    Article  Google Scholar 

  10. Zipf GK: Human behavior and the principle of least effort. Addison-Wesley, New York; 1949.

    Google Scholar 

  11. Ferrer i Cancho R, Sole RV: Least effort and the origins of scaling in human language. Proc Natl Acad Sci USA 2003,100(3):788–791. 10.1073/pnas.0335980100

    Article  MATH  MathSciNet  Google Scholar 

  12. Hauser MD, Chomsky N, Fitch WT: The faculty of language: what is it, who has it, and how did it evolve? Science 2002,298(5598):1569–1579. 10.1126/science.298.5598.1569

    Article  Google Scholar 

  13. Kosmidis K, Kalampokis A, Argyrakis P: Statistical mechanical approach to human language. Physica A 2006, 366: 495–502.

    Article  Google Scholar 

  14. Havlin S: The distance between Zipf plots. Physica A 1995,216(1–2):148–150. 10.1016/0378-4371(95)00069-J

    Article  Google Scholar 

  15. Piantadosi ST, Tily H, Gibson E: Reply to Reilly and Kean: clarifications on word length and information content. Proc Natl Acad Sci USA 2011,108(20):E109-E109. 10.1073/pnas.1103550108

    Article  Google Scholar 

  16. Griffiths TL: Rethinking language: how probabilities shape the words we use. Proc Natl Acad Sci USA 2011,108(10):3825–3826. 10.1073/pnas.1100760108

    Article  Google Scholar 

  17. Reilly J, Kean J: Information content and word frequency in natural language: word length matters. Proc Natl Acad Sci USA 2011.,108(20): author reply E109

    Article  Google Scholar 

  18. Reilly J, Kean J: Formal distinctiveness of high- and low-imageability nouns: analyses and theoretical implications. Cogn Sci 2007,31(1):157–168. 10.1080/03640210709336988

    Article  Google Scholar 

  19. Lazer D, Pentland A, Adamic L, Aral S, Barabasi A-L, Brewer D, Christakis N, Contractor N, Fowler J, Gutmann M, Jebara T, King G, Macy M, Roy D, Van Alstyne M: Social science. Computational social science. Science 2009,323(5915):721–723. 10.1126/science.1167742

    Article  Google Scholar 

  20. Bollen J, Mao H, Zeng X-j: Twitter mood predicts the stock market. Journal of Computational Science 2011, 2: 1–8. 10.1016/j.jocs.2010.12.007

    Article  Google Scholar 

  21. Golder SA, Macy MW: Diurnal and seasonal mood vary with work, sleep, and daylength across diverse cultures. Science 2011,333(6051):1878–1881. 10.1126/science.1202775

    Article  Google Scholar 

  22. Chmiel A, Sienkiewicz J, Thelwall M, Paltoglou G, Buckley K, Kappas A, a Hoyst J (2011) Collective emotions online and their influence on community life. PLoS ONE 6(7):e22207

    Article  Google Scholar 

  23. Dodds PS, Harris KD, Kloumann IM, a Bliss C, Danforth CM (2011) Temporal patterns of happiness and information in a global social network: Hedonometrics and Twitter. PLoS ONE 6(12):e26752

    Article  Google Scholar 

  24. Schweitzer F, Garcia D: An agent-based model of collective emotions in online communities. Eur Phys J B 2010,77(4):533–545. 10.1140/epjb/e2010-00292-1

    Article  Google Scholar 

  25. Garcia D, Schweitzer F: Emotions in product reviews – empirics and models. In Proceedings of 2011 IEEE international conference on privacy, security, risk, and trust, and IEEE international conference on social computing, PASSAT/SocialCom. Edited by: Bilof R. IEEE Computer Society, Boston; 2011:483–488.

    Google Scholar 

  26. Brants T, Franz A (2009) Web 1T 5-gram, 10 European languages version 1. Linguistic Data Consortium. Philadelphia

    Google Scholar 

  27. Global internet usage. en.wikipedia.org/wiki/Global_Internet_usage

  28. Kloumann IM, Danforth CM, Harris KD, a Bliss C, Dodds PS: Positivity of the English language. PLoS ONE 2012.,7(1):

    Article  Google Scholar 

  29. Bohannon J: Social science for pennies. Science 2011,334(October):2011.

    Google Scholar 

  30. Knowles ES, Nathan KT: Acquiescent responding in self-reports: cognitive style or social concern? J Res Pers 1997,301(31):293–301.

    Article  Google Scholar 

  31. Bentler PM: Semantic space is (approximately) bipolar. J Psychol 1969,71(1):33–40.

    Article  MathSciNet  Google Scholar 

  32. Russell JA: Affective space is bipolar. J Pers Soc Psychol 1979,37(3):345–356.

    Article  Google Scholar 

  33. Yik M, a Russell J, Steiger JH: A 12-point circumplex structure of core affect. Emotion 2011,11(4):705–731.

    Article  Google Scholar 

  34. Robinson JP, Shaver PR, Wrightsman LS: Measures of personality and social psychological attitudes. Elsevier, Amsterdam; 1991.

    Google Scholar 

  35. Russell JA: A circumplex model of affect. J Pers Soc Psychol 1980,39(6):1161–1178.

    Article  Google Scholar 

  36. Bryan CJ, Walton GM, Rogers T, Dweck CS: Motivating voter turnout by invoking the self. Proc Natl Acad Sci USA 2011,108(31):12653–12656. 10.1073/pnas.1103343108

    Article  Google Scholar 

  37. Augustine AA, Mehl MR, Larsen RJ: A positivity bias in written and spoken English and its moderation by personality and gender. Social Psychological and Personality Science 2011,2(5):508–515. 10.1177/1948550611399154

    Article  Google Scholar 

  38. Rozin P, Berman L, Royzman E: Biases in use of positive and negative words across twenty natural languages. Cogn Emot 2010,24(3):536–548. 10.1080/02699930902793462

    Article  Google Scholar 

  39. Dodds PS, Danforth CM: Measuring the happiness of large-scale written expression: songs, blogs, and presidents. Journal of Happiness Studies 2009,11(4):441–456.

    Article  Google Scholar 

  40. Cover TM, Thomas JA: Elements of Information theory. Wiley, New York; 1991.

    Book  MATH  Google Scholar 

  41. Jaeger TF: Redundancy reduction: speakers manage syntactic information density. Cogn Psychol 2010,61(1):23–62. 10.1016/j.cogpsych.2010.02.002

    Article  MathSciNet  Google Scholar 

  42. Kucera H, Francis WN: Computational analysis of presentday American English. Brown University Press, Providence; 1967.

    Google Scholar 

  43. Baayen RH, Piepenbrock R, van Rijn H (1993) The CELEX lexical database [CD-ROM]. Technical report, University of Pennsylvania, Linguistic Data Consortium, Philadelphia

    Google Scholar 

  44. Sebastián N, Martí MA, Carreiras MF, Cuetos F: LEXESP: Léxico informatizado del español. Ediciones de la Universitat de Barcelona, Barcelona; 2000.

    Google Scholar 

  45. Rime B: Emotion elicits the social sharing of emotion: theory and empirical review. Emotion Review 2009,1(1):60–85. 10.1177/1754073908097189

    Article  Google Scholar 

  46. Miller NE: Some recent studies of conflict behavior and drugs. Am Psychol 1961,16(1):12–24.

    Article  Google Scholar 

  47. Norman GJ, Norris CJ, Gollan J, a Ito T, Hawkley LC, Larsen JT, Cacioppo JT, Berntson GG: Current emotion research in psychophysiology: the neurobiology of evaluative bivalence. Emotion Review 2011,3(3):349–359. 10.1177/1754073911402403

    Article  Google Scholar 

Download references

Acknowledgements

This research has received funding from the European Community’s Seventh Framework Programme FP7-ICT-2008-3 under grant agreement no 231323 (CYBEREMOTIONS).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to David Garcia.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2

Authors’ original file for figure 3

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License (https://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Garcia, D., Garas, A. & Schweitzer, F. Positive words carry less information than negative words. EPJ Data Sci. 1, 3 (2012). https://doi.org/10.1140/epjds3

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1140/epjds3

Keywords