Hostname: page-component-745bb68f8f-s22k5 Total loading time: 0 Render date: 2025-02-12T09:49:22.137Z Has data issue: false hasContentIssue false

Measuring Emotional Response: Comparing Alternative Approaches to Measurement*

Published online by Cambridge University Press:  25 November 2015

Rights & Permissions [Opens in a new window]

Abstract

Scholarly interest in the role of emotion in accounting for how people react to political figures, events, and messages has escalated over the past two plus decades in political science and psychology. However, research on the validity of the measurement of subjective self-report of emotional responses is rather limited. We introduce here a new measurement approach, a “slider” format and compare it with the long used “radio button” item format. We assess the reliability and validity of these two approaches to the measurement of affect. The study examines self-report measures of emotion to three generated news stories about terrorist threats. We report that both measurement formats are able to extract the expected threefold affect structure from a ten affect word battery. The slider format is, however, modestly more reliable, and more efficient in time to complete, has the ability to limit missing data, and generates continuous data that is less truncated than data derived from the radio button format. Finally, we report on three tests of construct validity. Both approaches exhibit equivalent results on two of those tests. However, the radio button format does poorly on one test of construct validity, that on the anticipated relationship between anxiety and interest in novel information. We present an assessment of two methods for measuring emotional reactions to stimuli such as political issues, political figures, or events. Both methods are suitable for use in online surveys or computer-driven experiments. The traditional method utilizes labeled “radio buttons” that enable a participant in a study to select by clicking on one of an array of typically five response options, ranging from lower to higher of some identified affect term (e.g., how angry one might feel). Second, the slider method offers a participant the ability to move an “arrow” up or down to indicate how much (up) or little (down) they feel. The goal of both measures is to ascertain the level of a targeted emotion, i.e., how little or how much, say anger. The slider method has been specifically developed to be used with participants using a computer. The slider approach falls within the category of visual analog scales. This method for measuring affective responses to stimuli of whatever sort has not hitherto been examined to determine its reliability and validity. The literature on the reliability and validity of these measurement strategies is thin and we found no studies including an explicit comparison.

Type
Original Articles
Copyright
© The European Political Science Association 2015 

It is widely recognized that the responses obtained in surveys are not themselves direct measures of beliefs, values, or emotional states. Rather they represent accessible assessments of what are, in effect, latent, i.e., unobserved phenomena, such as attitudes, intelligence, partisan orientations, and other theoretically vital concepts (Lord, Ross and Lepper Reference Lord, Ross and Lepper1979; Nunnally and Bernstein Reference Nunnally and Bernstein1994; Borsboom, Mellenbergh and van Heerden Reference Borsboom, Mellenbergh and van Heerden2003). Out of this research tradition come two clear standards for evaluating measures of latent concepts: reliability and validity. Reliability, of course, identifies the measure’s stability and consistency. Validity refers to how well a measure captures what it purports to assess. Validity has various forms, but one variant, construct validity, is assessed by examining whether relationships between theoretically relevant dependent variables and the test measures under evaluation conform to theoretically specified relationships (Campbell and Fiske Reference Campbell and Fiske1959; Alwin Reference Alwin2010). Additionally, other considerations may influence which method one might adopt: time to complete (typically shorter is better); numerical quality of scores (preferring continuous data over ordinal level data, and ordinal level data over dichotomous data); and minimizing missing data so as to obtain valid data from most, if not all, participants (Krosnick and Presser Reference Krosnick and Presser2010). A variety of methodologies have been used to obtain measures of emotion that do not rely on survey items, among them electromyography (EMG), electroencephalography (EEG), and functional magnetic resonance imaging (fMRI). Mauss and Robinson (Reference Mauss and Robinson2009) provide a review. Ruef and Levenson (Reference Ruef and Levenson2007) have proposed the use of physical “affect dials.”Footnote 1 However, the psychometric properties of measuring people’s emotional reactions suitable for experiments and surveys is relatively sparse.

This lapse is of particular importance because the social scientific research literature on emotion and its impact on decision-making and action has expanded quite remarkably over recent years (Brader and Marcus Reference Brader and Marcus2013). While there is research on the reliability and validity of available measures of subjective self-report of emotional response, the extant published research has largely focused on which emotions to measure (Watson Reference Watson1988b; Watson and Clark Reference Watson and Tellegen1994; Watson and Clark Reference Watson, Wiese, Vaidya and Tellegen1997; Marcus et al. Reference Marcus and MacKuen2006) and less so on the particular ways in which respondents can indicate how much, or little, they are experiencing a targeted emotion.

Reliability can be easily evaluated provided, as is the case here, multiple indicators of each latent concept are available. Assessing validity is less straightforward. While face validity is always available for assessment, it represents a rather weak and potentially subjective standard. We turn, instead, to construct validity for a more demanding test. Construct validity requires a theoretical foundation so that empirical relationships between variables can be used to determine whether prospective measures are functioning as theoretically expected. In the section that follows, we outline the theoretical basis for identifying which emotions to examine and the theoretical basis for the substantive relationships used to assess the different measures of emotion we examine below. Thereafter we turn to the research design, data, results, and discussion.

Emotion and its Anticipated Effects

Interest in emotion in politics was evident when the behavior approach to the study of politics was launched supported by the then new interest in survey research. Much of that flowed from discovery that partisan identification was key to understanding the American electorate’s political behavior (Campbell et al. Reference Campbell, Converse, Miller and Stokes1960; Campbell et al. Reference Campbell, Converse, Miller and Stokes1966). Partisan identification was defined as an “affective orientation to an important group object” (Campbell et al. Reference Campbell, Converse, Miller and Stokes1960, 121). This definition is derivative of the notion affect in attitude theory wherein an attitude is understood as having three facets (Allport Reference Allport1954; McGuire Reference McGuire1969). Attitudes are constructed from what we know (the cognitive component); how we feel about what we know (the affective component); and what we do with that knowledge and assessment (the behavior component). As with the then concurrent development of feeling thermometers, the affect component was understood as a single valence dimension (ranging from how liked to how disliked) and as the passive receptacle of evaluations passed on from the cognitive component, thereby being conceived then as an “affect tag” (Fiske and Taylor Reference Fiske and Taylor1991). But little research was done to ascertain whether people actually derived their affective reactions via cognitive assessments and that such evaluations could be validly described as a single valence like–dislike dimension.

During this same period a more complex view of emotion emerged, cognitive appraisal theories. Two features stood out. First, as with the attitude view, affect was understood to derive from conscious assessments (Izard Reference Izard1971; Izard Reference Izard1977; Roseman Reference Roseman1979; Roseman Reference Roseman1984; Ortony, Clore and Collins Reference Ortony, Clore and Collins1989). But, unlike the simple structure of affect as valence, cognitive appraisal theories held that most people would, at the end of the appraisal process, arrive at a single discrete emotional state, there being anywhere from 8–12 or more such “basic” emotions (different theorists argued for somewhat different lists of these basic discrete emotions). In this view, for example, something bad would be felt as anger if the bad act was seen as being done to one by another person, but would be felt as guilt if understood to be one’s own act.

For this reason, an individual would be in but one of a number of discrete affective states at any given moment. It was for that reason that from the outset American National Election Studies (ANES) measures of affective reactions to candidates presented survey participants with the question “have you ever felt” (then followed a list of the various discrete affects) because it was understood that the cognitive evaluations would result in people reporting that they felt this but not that discrete affect. And hence it made good sense to offer as the response option of yes or no. These approaches were largely asserted rather than tested. Hence, the response options presented to study participants, a thermometer scale so that people could indicate how much they liked or disliked the object in question or the question whether thinking about a candidate had ever made them feel, say, angry, largely presumed the validity of the conception and presumed the validity of the measurement. The two conceptions, emotion as valence and emotion as derived from a cognitive appraisal, were challenged on both scores by research published in the early 1980s (Zajonc Reference Zajonc1980; Abelson et al. Reference Abelson, Kinder, Peters and Fiske1982).

What resulted in the 1980s was a two orthogonal dimensional view of emotion, a positive dimension and a negative dimension (i.e., that people commonly had both a positive affective reaction and a negative reaction). Attention to measurement was largely executed by psychologist David Watson who developed the Positive Affect and Negative Affect Scale measure (Watson Reference Watson1988b; Watson, Clark and Tellegen Reference Watson and Clark1988). In that work Watson explored a number of measurement features, among them assessing different affect terms best suited to measure the positive and the negative dimensions, whether different time frames impacted the responses given by study participants, and different response format. But in that work the variations in response options considered focused on comparing a four-option frequency format (how often an affect was experienced) and a five-option extent (how much of that emotion was experience) variants. No consideration was given to response options that offered an expressly continuous response format.

In sum, over the past 60 years affect moved from a debate as to whether affect could be understood as valence or as 8–12 mutually exclusive discrete emotional states to a view of emotion as a twofold dimensional space, most often understood as a circumplex (Plutchik and Conte Reference Plutchik and Conte1997). More recently, two separate research programs have converged in identifying the same three dimensions of emotional response as minimally necessary to identify the “affective space”: a positive dimension, enthusiasm; and two negative dimensions, anxiety and anger.Footnote 2 Roger Masters, with various colleagues, found these three characteristic emotional displays expressed by political leaders both American and non-American (McHugo et al. Reference McHugo, Lanzetta, Sullivan, Masters and Englis1985; Sullivan and Masters Reference Sullivan and Masters1988; Masters and Sullivan Reference Masters and Sullivan1989; Masters and Sullivan Reference Masters and Sullivan1993; Way and Masters Reference Way and Masters1996; Masters Reference Masters2001). Marcus et al., as well as research conducted by others, find the same three dimensions of affective response to a wide array of political stimuli (Redlawsk Reference Redlawsk2002; Brader Reference Brader2006; Marcus et al. Reference Marcus and MacKuen2006; Valentino et al. Reference Valentino, Hutchings, Banks and Davis2008; MacKuen et al. Reference MacKuen, Miller, Marcus and Neuman2010; Valentino et al. Reference Valentino, Brader, Groenendyk, Gregorowicz and Hutchings2011; Gadarian and Albertson Reference Gadarian and Albertson2014). Additionally, the research of psychologist Jennifer Lerner also identifies two negative affective, of aversion and of anxiety(Lerner and Keltner Reference Lerner, Gonzalez, Small and Fischhoff2000; Lerner and Keltner Reference Lerner and Keltner2001; Lerner et al. Reference Lerner and Keltner2003). Finally, Neuman et al. examined the emotional reactions of a national sample of adults to a wide array of news stories, some 50 in all, covering diverse topics such as celebrity news, economic stories, crime, foreign affairs, among others, find that those who read the stories report varying levels of anxiety, aversion, and enthusiasm (Neuman, Marcus and MacKuen Reference Neuman, Marcus and MacKuen2013). For example, they report that a story on the USAir airplane landing in the Hudson river measured very high on enthusiasm but very low on aversion, while a story on the recent Times Square attempted bombing elicited very high levels of both anxiety and aversion. Thus, it is a bit of a simplification to claim that this, or that, media message, elicits “more emotion” than another. In sum, though not all to the same degree, it appears that people respond to political and non-political events with varying levels of enthusiasm, varying levels of anxiety, and varying levels of aversion/anger. But reliance on radio button format collapses the range to four or five response options truncating what is now generally understood to be a continuous distribution. To that end we developed the slider format so that the response option is consistent with the theoretical formulation and is able to generate continuous distribution of three affect dimensions: enthusiasm, aversion, and anxiety.

Construct Validity and Testable Theoretical Claims

The construct validity of a new measure is traditionally assessed by testing relationships between that measure and dependent variables. For this study we made use of three such dependent variables taken from the affective intelligence literature. Of particular importance in this literature is when do people express interest in gathering more information, and which of two approaches to judgment they adopt as they confront a political matter: reliance on established disposition, such as partisan and ideological convictions; or a more open-minded thoughtful consideration of the matter. The distinction between the fast “automatic” reliance on established heuristics and shifting to a slower more thoughtful approach to judgment has come to be identified in psychology as the dual process model of judgment (Chaiken and Trope Reference Chaiken and Trope1999; Haidt Reference Haidt2001; Kahneman Reference Kahneman2011).

The affective intelligence theory, an example of a political science dual process model, advances theoretical claims that we can use to test the validity of measuring affective response. The theory holds that increases in anxiety identify novelty and initiates interest in learning, i.e., acquiring contemporaneously available information pertinent to the issue at hand (Marcus and MacKuen Reference Marcus, MacKuen, Wolak and Keele1993). Hence, anxiety is expected to be related to interest in novel information. The theory also identifies heightened anxiety as leading to thoughtful consideration informed by taking into account contemporaneously acquired information on the relevant options. Hence greater anxiety is expected to be related to deliberative style orientation. Finally, the theory holds that heightened aversion leads to people strengthening their attachment to their extant political habits such as partisan and ideological cues (Marcus, Neuman and MacKuen Reference Marcus, Neuman and MacKuen2000; Marcus Reference Marcus2002), hence heightened aversion is expected to strengthen reliance on the extant convictions judgment style.

In addition, we wish to assess how successful each method is in obtaining relatively pure “state” measures of affect. Various traits have been shown to predict affective reactions. For example, Jost et al. have argued that ideological identification is a stable difference that inclines some to be more, or less, likely to express fear or disgust (Jost, Glaser and Sulloway Reference Jost, Glaser, Kruglanski and Sulloway2003). Similarly, many, if not all, of the primary personality traits are thought to be expressed as affective disposition, for example, to display fear (Zuckerman Reference Zuckerman1991). Another array of research findings suggest that liberals are more engaged by novelty, with conservatives being more engaged by a need for certainty (Amodio et al. Reference Amodio, Jost, Master and Yee2007; Jost and Amodio Reference Jost and Amodio2012). Including a measure of ideological identification enables us to test whether conservatives are more inclined to endorse partisan orientation, while, on the other hand, liberals are expected to be more inclined to endorse a more thoughtful orientation. To enable us to assess whether the affect methods are measuring proximal reactions to the presented stimuli, rather than just expressing stable dispositions, we included an array of trait measures that can be used to assess to what extent either method is able to isolate and measure state affect. To that end we include measures of the Big Five personality traits, Need for Cognition, as well as two political traits, Partisan Identification and Ideological Identification.Footnote 3 We use these same added factors for each of the construct validity tests with three dependent variables, Interest in Novel Information, Self-Confident Resolve (SCR), and Social Open Mindedness (SOM).

In sum, we have three theoretical claims to test. First, we expect that measures of anxiety should be significantly correlated with the interest in novel information. Second, we expect that anxiety should be significantly correlated with a more deliberative style of judgment (as the perception of novelty, marked by higher anxiety, is expected to lead to an openness to consider a wider array of alternatives). And, third, we expect that higher levels of aversion should be significantly correlated with a more partisan style of decision-making (as the perception of familiar foes, marked by higher aversion, is expected to strengthen reliance on extant habituated courses of action). And, by incorporating measures of traits also known to be linked to these three dependent variables, multivariate testing enables us to “control” for these added factors thereby isolating our test of the comparative abilities of two “state” measures of affective response.

We next turn to the development of stories, the stimuli that are intended to elicit affective response, the specific items that make up the measures for each target affect, as well as the other elements in the design of the survey.

Method

Affective Stimulus Materials

As noted above, had we executed this study in the 1950s we might have chosen but a single question to determine how people emotionally reacted to some political stimuli. At that time “feeling thermometers” were developed so that people could signal how much they liked or disliked some designated target (e.g., political party, leader, policy position, social group, etc.). However, while efficient, feeling thermometers constrain people to reporting their affective experience as a single valence, i.e., liking or disliking. Had we executed this study some 30 years later, in the 1980s, we would have likely chosen stories to ensure that we could be sure we elicit each of two anticipated orthogonal affective dimensions, one positive and one negative, reflecting the latest research that found that people experienced concurrent positive and negative affective reactions (Abelson et al. Reference Abelson, Kinder, Peters and Fiske1982; Watson and Tellegen Reference Watson and Clark1985; Mayer and Gaschke Reference Mayer and Gaschke1988; Watson, Clark and Tellegen Reference Watson and Clark1988).Footnote 4 However, as noted above, current research in the study of affective responses find three largely independent dimensions.Footnote 5 To ensure that we can elicit sufficient affective responses to each of these affective dimensions we developed three stories, each intended to ensure heightened, respectively, anxiety, aversion, and enthusiasm (though, of course, any political stimulus is likely to impact self-report on all three affective dimensions).

One story presented the possibility of unknown multiple sleeper cells of terrorists on American soil. It was designed to evoke anxiety. A second story described a successful border protection program marshaled by American security institutions. It was designed to evoke enthusiasm. The third story reviewed the life of Osama bin Laden and his often successful terror attacks against the United States. It was designed to evoke aversion (at the time of this study bin Laden was alive and at large). The three stories are included in Appendix A.

Having three stories and multiple indicators for each affect dimension enable us to assess the efficacy of the stories, to compare the two methods in their ability to elicit the targeted affect, and to assess whether each method identified the anticipated structural model of affect.

Administration

After reading one of the three news stories each participant was randomly assigned to either radio buttons or slider formats to report on their affective state. The former requires selecting one of the five labeled buttons. The latter presents a vertical scale on which participants find an “arrow” that they can move up or down to report how much or how less of an indicated affective reaction they are then experiencing. As numerous studies have demonstrated that participants experience multiple concurrent affective responses thus requiring that researchers use a battery of affect items to accurately differentiate emotional responses. Among many such demonstration we cite here research by Watson, Clark and Tellegen (Reference Watson and Clark1988), Watson (1988Reference Watsona), Marcus et al. (Reference Marcus and MacKuen2006), and Brader, Groenendyk and Valentino (Reference Brader, Groenendyk and Valentino2010).Footnote 6 A fuller description, with figures, is given below. After completing this battery participants continued on to complete batteries for the dependent variables and the other measures (see Appendix B for listing of the items).

Participants Sample

The data collection for this project was executed by Knowledge Networks (KN) under our direction.Footnote 7 The data obtained are intended to be representative of American adults collected online using KN standard methodology. KN fielded this survey to a national sample age 18+ of panel members who had personal computers and internet access at home. The combined sample size for the two methods was 289 participants.Footnote 8 Which of the two methods each participant used was determined by random assignment. The requirement of having a home personal computer with internet access was needed to support distribution and administration of this survey. KN reports a completion rate of 55 percent for the survey used in these analyses. This survey was in the field from September 28 to October 16, 2009. KN provides a weighting variable with each of their samples so that, when applied, the demographic composition, age, gender, income, and so on, of their samples mimic that of the adult population of the United States.

Measures

Choosing word “markers” to measure three dimensions of affective response

We use ten semantic emotional “markers” to assess these three dimensions of affect. For the latent concept Enthusiasm, the words enthusiastic, hopeful, and proud have proven reliable markers; for the latent concept Anxiety, the words used have been scared, worried, and afraid; and, for the latent concept Aversion, hateful, angry, bitter, and resentful have proven reliable (Marcus et al. Reference Marcus and MacKuen2006).

The most common survey method for determining the level of affective response is to provide an array of semantically labeled alternative responses. Typically, five alternatives are presented and the respondent selects a response by clicking on one of the radio buttons (or check boxes in computer or paper-based surveys). Figure 1 shows a screen shot of that method. The screen shot shows five of the ten affect words. A second screen displayed the remaining five items. Random rotation of the ten affect items was used throughout to ensure that presentation order did not introduce an unnecessary complication.

Fig. 1 Radio button/multiple choice format

Radio buttons are a familiar and widely used technique. The semantic label gives clarity to each of the available five response options. While this response format does not produce data that meet the standard of continuous measurement, this sort of ordinal data are generally acceptably treated as continuous for statistical analyses (Bollen and Barb Reference Bollen and Barb1981). Moreover, as is the case here, when multiple measures are available one can easily generate simple summated scales that mitigates concerns about ordinal measurement. Nonetheless, one of our goals is to obtain data that, at the level of individual items, is straightforwardly continuous.

The clarity of semantic labels, while generally an advantage, raises a concern that the time and cognitive effort involved in discriminating among the semantic labels may compromise the capacity of the respondent to capture and identify their emotional state accurately. This led us to explore alternative response formats.

With the slider method format participants are presented with a vertical bar along with an “arrow” that they can move up or down with the computer’s mouse to indicate how much (up) or less (down) of the specific affect they experience. Presenting participants with a physical distance on which participants can place a response builds on earlier work by Lodge (Lodge Reference Lodge and Tursky1979; Lodge and Tursky Reference Lodge and Tursky1981).

One design consideration was where to locate the arrow on the slider as a starting point. We previously experimented with the arrow at the midpoint or alternatively at the bottom of the continuum. Starting with an arrow at the midpoint raised the concern that it might encourage respondents to leave it there as “close enough” to their assessment of the dimension or as an easily selected non-response. Starting with an arrow at the bottom, on the other hand, might bias estimates downward. Using a bottom-point starting position significantly increased the number of instances where respondents left the arrow unmoved (Marcus, Neuman and MacKuen 2009Reference Marcus, Neuman and MacKuena; Marcus, Neuman and MacKuen Reference Marcus, Neuman and MacKuen2009b). In that work we found that the midpoint starting position does not seem to have been used as a convenient “no opinion” option. We found that while respondents might leave one or two of the ten arrows unmoved, they did move most of them. It appears that participants left the slider arrow at the midpoint because the midpoint was their chosen response (Marcus, Neuman and MacKuen 2009Reference Marcus, Neuman and MacKuena; Marcus, Neuman and MacKuen Reference Marcus, Neuman and MacKuen2009b). Had participants left the arrows unmoved so as to indicate no response then their response would add “noise” (error) and, to the extend they did, so would generate a weaker relationship with our dependent variables. We found, in preliminary studies, that using a midpoint position for the slider now generated the highest construct validity. Hence, we concluded that the midpoint starting position works best.

Figure 2 displays the screen image of the slider format variant. And, as with the radio button format five affect terms are presented on each of two pages, with the order of the ten items randomized to prevent any systematic order effects.

Fig. 2 Slider format with midpoint start position and polar labels

Dependent Variables

A scale of interest in novel information was generated by creating a simple summated scale from three items. The statements “My Googling these issues and events to find out more is” and “My looking for websites that present new and different views on these issues and events is” and “My looking for websites that might challenge my points of view on these issues and events is” were offered to which participants could then respond, for each, to the query “this statement is extremely true, very true, moderately true, slightly true, or not at all true.” The resulting scale had a high level of reliability (Cronbach’s α=0.86).

Judgment orientations were measured with batteries of items from which two simple summated scales were also built. From each battery we chose the best items, meaning those items that best defined the latent concept (Hayduk and Littvay Reference Hayduk and Littvay2012). From among those available we chose the two statements for each orientation style. For SCR orientation: “These issues and events provide no room for compromise” and “I am certain that my point of view on these issues and events is the right one. It’s time to move on” were used. For the SCR orientation the two items are correlated at r=0.33. For SOM orientation: “These sorts of issues and events should be resolved so that everybody’s needs are met” and “To solve these sorts of issues and events everyone’s concerns should be heard” were used. These items are correlated at r=0.53. It is important to note that these orientation style items depict two separate styles, the styles are orthogonal to each other (the largest correlation between the four correlations between each array is that between “my point of view …. right one” and “should be heard” at r=−0.10). These orientations are not polar opposites (Marcus Reference Marcus2008; MacKuen et al. Reference MacKuen, Miller, Marcus and Neuman2010). The availability of two styles resonate with the current scholarly and popular debate in the United States concerning polarization, rhetorical styles, and civility (Ansolabehere and Iyengar Reference Ansolabehere and Iyengar1995; Mutz and Reeves Reference Mutz and Reeves2005; Fiorina and Abrams Reference Fiorina and Abrams2008).

In sum, participants in this study begin by reading one of three stories: which story they were given was determined by random assignment. Thereafter they reported on their affective responses to the story just read either, again by random assignment, reliance on the radio button or slider method for self-report. After completing this battery of ten items participants then completed the remaining questions, the interest in novel information battery, the two orientation batteries, the need for cognition battery, the Big Five personality trait battery, partisan and ideological identification, and answered an array of demographic queries. Finally, to facilitate comparison of all measures and all results, each measure, whether an individual response or a scale, was rescaled to a common 0–1 range. This preserves the underlying variance of each measure and scale. Using standardized coefficients, as in ordinary least squares, offers an alternative way of enhancing comparison but at the cost of altering the variance of each measure or scale.

Results

In this section, we begin by examining the administrative costs of using each of the two affect responses methods. Thereafter, we examine the ability of each method to reproduce the expected measurement model of three affect dimensions. Thereafter, we turn to reliability and validity concerns.

Assessing Administrative Costs: Radio Buttons Versus Sliders

Two considerations are relevant to assessing the costs and benefits of alternative measurement strategies: time to complete and missing data. The slider format, as we have described its application in this study, precludes missing data, hence for each of the 154 participants, all recorded responses to the ten affect items are treated as valid.Footnote 9 The radio button does enable participants to skip those items they wish to avoid. For the radio button method, with 135 participants, missing values ranged from a low of four for the affect item “hopeful” to a high of ten for the affect item “hateful” with the other eight items intermediate (see Table 1).

Table 1 Means and Missing Data for Ten Affect Items by Method

Source: Knowledge Networks Survey 2009.

The time to complete does appear to convey an advantage to the slider method. The radio button array of ten items took an average of 103 seconds to complete. The slider array took 85 seconds though the difference is not statistically significant (t-test=0.91, p=0.18). This result is at variance with that found by Couper et al. (Reference Couper, Tourangeau, Conrad and Singer2006). They found that their horizontal slider format took longer to complete than the radio button format, however, that may be due to their visual analog design, which differed from ours both as to its orientation (horizontal rather than vertical) and the density of semantic material used to label their version as compared with the more minimal approach we took.

Table 1 presents the means for the ten affect items as well as showing the missing data for each item. As is evident in comparison of the means, the radio button format generates lower means than does the slider format (a result that analysis of variance, reported below, confirms is significant and not a result of random measurement error). We explore the distribution of responses, by method, in Figure 3. We here examine whether either measurement method is particularly prone to truncation, which is generating responses that “pile up” at either the higher or lower ends of the distributions (i.e., either ceiling or floor effects). Figure 3 shows box plots for the ten affect terms, grouped by the three enthusiasm items (hopeful, proud, enthusiastic), the four aversion items (angry, hateful, bitter, resentful), and anxiety (afraid, scared, worried) with the radio button box plots above and the slider format box plots below. Although we rescale the raw data to a common 0–1 range below, here we show the raw data with the original ranges.Footnote 10 The box plots shown in Figure 3 show that the slider button format generated a more normal distribution and with clear tails both above and below the mean than did the radio button format. We will take up the question of how robust and significant that difference is when we turn to scales built from the ten affect items after we have considered whether these ten affect words successfully serve as measures of the intended latent concepts, Anxiety, Enthusiasm, and Aversion. We turn to that topic next.

Fig. 3 Box plots of ten affect measures by method: radio button and slider format Source: Knowledge Networks Survey 2009.

The Dimensional Structure of Affective Responses

We use both exploratory and confirmatory factor analyses to examine the uni-dimensionality of each of three sets of affect words. The intention is to see if each mode of statistical analysis identifies three factors and that each of the selected affect terms specifically defines (i.e., “loads”) on the target latent concept. The exploratory factor analysis supports three affect dimensions with the expected affect “markers” loading on each of the three dimensions, anxiety, aversion, and enthusiasm, for both methods. For the radio button data, the first four eigenvalues reported are 4.82, 1.70, 0.36, and 0.04. For the slider data, the first four eigenvalues reported are 4.29, 2.05, 0.89, and 0.12.Footnote 11 Thus, using both the scree test and the eigenvalues >1.00 criteria, exploratory factor analyses lead to a three-factor solution.

A more compelling test of the expected measurement model is reliance on structural equation modeling (SEM). That is so because unlike exploratory factor analysis, SEM requires that the researcher identify two features of the model, one that defines the measurement theory (i.e., which items define each latent concept) and one that defines the substantive relationships that are expected to exist between the latent concepts (i.e., the substantive theoretical linkages). This is equivalent to also holding that each affect term measures only the latent concept to which it is linked (i.e., no cross-over linkages to other affect measures or to other affect concepts). The array of goodness of fit measures then enable examination of the relative success of each method in generating acceptable goodness of fit. And, there is one further benefit from adopting confirmatory factor analysis. The three chosen dependent variables are not expected to be orthogonal—indeed the anxiety and anger—aversion dimension are often found to be highly correlated (Lerner and Keltner Reference Lerner and Keltner2001; Lerner et al. Reference Lerner and Keltner2003). Confirmatory factor analysis enables us to differentiate the structural relationships between three affect concepts and the word markers used to identify each of the three affect concepts.Footnote 12 We obtain two SEM models, one for each method of measuring three dimensions.

Figure 4 shows the confirmatory factor results for each method with goodness of fit measures therein as well.

Fig. 4 Structural equation models of (a) radio button and (b) slider methods of affect measurement-standardized solutionSource: Knowledge Networks Survey 2009.

The SEM analysis shows the expected pattern with the affect terms hopeful, proud, and enthusiastic defining the Enthusiasm latent factor; angry, hateful, bitter, and resentful defining the Aversion latent factor; and afraid, scared, and worried defining the Anxiety latent factor. The fit criteria are acceptable for both measurement methods. Both models fit their respective data quite well. The SRMR (i.e., the standardized root mean square residual) value for the radio button method is 0.059, while that for the slider method is 0.60. In sum, the ten affect items, whether measured by the radio button method or by the slider method, generate the expected three factors with each of the ten affect items “loading” on the expected latent factors. These results are consistent with earlier research on these markers and their ability to define the three factors (Marcus et al. Reference Marcus and MacKuen2006). We thus conclude that responses to the ten affect items, whether obtained using the radio button or the slider method, can successfully identify the latent factors. We turn next to the reliability of the two sets of three scales derived from each array of ten affect items.

For assessment we used simple summated scales for each of the three dimensions of affective response, each coded to a common 0–1 range. Table 2 shows the reliabilities of the affect scales. We note also that the reliability values reported here are similar to those shown in Figure 4. Each method secures reliable scales. Though the reliabilities reported show the slider format marginally higher, only in the case of the scales for Enthusiasm are the difference significant (Feldt Reference Feldt1969; Feldt, Woodruff and Salih Reference Feldt, Woodruff and Salih1987). Notwithstanding that one difference, in all instances the reliabilities for each method are quite high.

Table 2 Cronbach’s α and Feldt Test for Radio Button and Slider Measures of Affect

Source: Knowledge Networks Survey 2009.

Before we turn to the dependent variables to test the construct validity of the two affect methods we turn to the question of whether the three stories elicited the targeted affective responses. To that end, we conducted analysis of variance for each of the three affect scales in a 3×2 design (three stories×two methods). In each case, the effect of the story treatment and the effect of the affect method used to present the response options (as we have seen in means of the individual items, Table 1) show the expected pattern. For anxiety, the effects of the story (F (2, 281)=21.44, p<0.0001) and method (F (1, 281)=19.67, p<0.0001) are substantial as well as significant. For enthusiasm, the effects of the story are also robust and significant (F (2, 281)=36.11, p<0.0001) as is method (F (1, 281)=21.57; p<0.0001). Finally, for aversion, we find the same pattern, with the effects of story (F (2, 278)=47.07, p<0.0001) and method (F (1, 278)=24.37, p<0.0001) being substantial as well as significant.Footnote 13

The impact of item format and story type can be seen in Figures 57. Both measurement formats demonstrate a clear ability to mark the differing affective responses to each of the three stories. The slider format, however, produces, on balance, a cleaner and more discriminating array of results than does the radio button format. For example, when measuring anxiety, the slider format reports 0.65 on the Hidden Cells story and 0.36 on the Safe Harbors story, a difference of 0.29, while for the radio button anxiety response to the same stories, 0.46 and 0.27, respectively, a difference of 0.19. More importantly, the two formats record the theoretically expected affective response for each of our three stories: the anxiety story produces anxiety, the enthusiasm story produces enthusiasm, and the aversion story produces aversion. And, though the two negative stories produce elevated levels of both negative affects, both measurement methods show that the story targeting uncertainty stimulates higher levels of anxiety than of aversion, while the story presenting a familiar foe produces higher levels of aversion than of anxiety. In sum, both approaches identify the distinctive affective responses that theory associates with each stimulus story.

Fig. 5 Levels of anxiety by story and by affect methodSource: Knowledge Networks Study, Fall 2009.

Fig. 6 Levels of enthusiasm by story and by affect methodSource: Knowledge Networks Study, Fall 2009.

Fig. 7 Levels of aversion by story and by affect methodSource: Knowledge Networks Study, Fall 2009.

A more decisive test examines the construct validity of each affect method. For that test we turn to the three dependent variables to assess the relative validity of the two measurement approaches.

Construct Validity of Affective Response Measures with Two Method Formats

We utilize three dependent variables: interest in searching for novel information, and the two judgment orientations identified by the dual model of judgment, the default reliance on extant convictions or the shift toward a more deliberative style of political decision-making. To compare formats, we regress each of these three dependent variables on the three affect measures. To exclude confounding factors we include an array of covariates known to be related to the dependent variables. This approach is intended to isolate the ability of each method to measure the state affective responses.

Table 3 displays the regressions of emotion on the first dependent variable, interest in novel information. As expected, anxiety is a robust and significant predictor of interest in obtaining novel information when the data are collected using the slider format. We do note that the equivalent coefficient for impact of anxiety, when anxiety is measured by the radio button format, is both substantively weaker and not statistically significant.Footnote 14 We next turn to the two measures of orientation toward controversy.

Table 3 Affect and Interest in Novel Information by Two Affect Methods

Source: Knowledge Networks Survey 2009.

Note: *p<0.05.

We turn next to the inclination to engage as defense of extant conviction, the SCR orientation. Here the expected relationship is between aversion and the inclination to remain committed to one’s convictions and to avoid the views of others who do not share one’s partisan views. And, as expected, increased aversion does have a significant positive impact on commitment to convictions in both analyses and both methods reporting near identical relationships between aversion and a partisan orientation (Table 4).

Table 4 Affect and Self-Confident Resolve (SCR) Orientation by Two Affect Methods

Source: Knowledge Networks Survey 2009.

Note: *p<0.05.

The anticipated relationship of affect and adopting a more deliberative orientation is that heightened anxiety is expected to be positively associated with activating the inclination to cooperate and that aversion is an inhibiting factor (i.e., a negative relationship). The results are displayed in Table 5. The direct effects of anxiety is consistent with the hypothesized expectation. Also shown in Table 5, the radio button method show a robust and significant negative relationship between aversion and inclination to adopt a deliberative orientation. This specific finding invites additional research for it suggests that aversion might interpose a contrary substantive resistance to engaging in thoughtful consideration.

Table 5 Affect and Social Open Mindedness (SOM) Orientation by Two Affect Methods

Source: Knowledge Networks Survey 2009.

Note: *p<0.05; #p<0.10.

Discussion

Our interest is in developing a methodology that is both capable of producing reliable and valid measurement of affective responses and that can also be used with a wide array of stimuli. We explored the new slider format with a variety of simulated online stories, but the method is applicable to any array of political stimuli. We contrasted the new slider response format with the traditional multiple choice radio buttons. The latter allows participants relying on semantic cues to identify and report their affective circumstances of the moment. While the former allows people to move an arrow up or down along a vertical slide to a location that best identifies their current feeling state for the specified affect. Our interest in developing the slider format grew from a concern that the additional cognitive effort of semantic processing associated with discriminating among the radio buttons may diffuse the capacity of the respondent to quickly capture and express their sense of emotional state. And, indeed, overall, the slider format proved to take less time to administer thus enabling people to generate a swifter response though the difference did not prove to be statistically significant. Of some added benefit, the slider format offers continuous data, and generated robust relationships with each of the three dependent variables.

Both formats generate the same structural pattern among the ten affect words, each fitting and finding the expected three dimensions, anxiety, enthusiasm, and aversion. Both arrays of data, one using the radio button format and the other using the slider format, report generally similar relationships between the expected affect dimensions with the two judgment orientation-dependent variables.

On the basis of those comparisons we conclude that the slider format represents an important and useful addition to the methodological array for securing measures of affective response and one that exhibits modestly superior psychometric properties. Nonetheless, some qualifications and interesting questions arise from our analysis.

We should not put aside some concerns that this study does not address. First, this analysis is based on one data set.Footnote 15 We have found that the slider format proved effective when used for a wide array of stories, having deployed this method in a study with some 50 brief news stories (Neuman, Marcus and MacKuen Reference Neuman, Marcus and MacKuen2013). And our more detailed work using Economics and Food Safety stories shows that the slider format consistently passes the construct validity test over a range of subjects (MacKuen et al. Reference MacKuen, Wolak, Keele and Marcus2013). Nonetheless, research using a wider array of stimuli types, perhaps using just images, or video, or stimuli not given in the convention of newspaper stories, would be helpful. Even as research makes use of more modern technologies, it is likely that research will move away from computers and to tablets. The slider method may be even more useful with tablet use of finger touch to provide participant input than the computer’s use of mouse to record input. Finally, the slider format may well prove a useful way of recording responses generally and not just to affective responses. Whether that proves to be the case of course will have to await additional research.

We remark here on two substantive issues worth further research. Two substantive findings warrant more research. First, the radio button measurement of affective response reported (see Table 3) that anxiety is unrelated to interest in novel information. This result seems to be at variance with a large array of reported research that does report a robust relationship between anxiety and interest in acquiring novel information.Footnote 16 Why this discrepancy between methods here occurs remains unresolved. Second, and of some importance, the results reported in Table 5 find that both methods agree that anxiety encourages people to adopt a deliberative stance to decision-making. However, the methods offer different conclusions about the role of aversion. The radio button method finds that heightened aversion acts to inhibit adopting this stance. On the other hand, the slider format data suggest that anxiety is the sole affective factor driving the inclination to deliberate. This difference merits further research to assess which conclusion is valid. In this time of heightened polarization with public anger being in plentiful supply, knowing whether heightened anxiety is by itself solely capable in turning people from steadfast defense of their extant beliefs or whether that shift requires both heightened anxiety and diminished anger has both theoretical and practical application.

Notwithstanding these suggested trajectories of future research, the two methods for collecting emotional self-report, these two methods, the conventional radio button and the new slider, work well. In the main, the two methods offer reliable and valid methods for obtaining data on how much, or how little, of three affective appraisals people experience.

Supplementary Material

To view supplementary material for this article, please visit http://dx.doi.org/10.1017/psrm.2015.65

Footnotes

*

George E. Marcus, Professor of Political Science, Department of Political Science, Williams College, Williamstown, MA 01267 (gmarcus@williams.edu). W. Russell Neuman, Professor of Communications, New York University, MAGNET, 2 MetroTech Center, Suite 864, Brooklyn NY, 11201. Michael B. MacKuen, Professor of Political Science, Department of Political Science, University of North Carolina Chapel Hill, Chapel Hill NC 27599-3265. To view supplementary material for this article, please visit http://dx.doi.org/10.1017/psrm.2015.65

1 While “affect dials” or as they are also called “people meters,” are widely used especially in efforts to get “real-time” estimates of people’s reactions to such televised events as political speeches and campaign debates, they constrain responses to a single valence, like–dislike, assessment. There is no recent research we know of that demonstrates affective responses to any single stimulus (event, group, individual, or whatever) is other than a confounded measure of multiple concurrent reactions. See Marcus (Reference Marcus2003) for a review of the relevant literature.

2 Some scholars use different terminology to identify these dimensions. For example, variations in positive, enthusiasm, and hope have all been used to identify the same latent factor, though some hold that “hope” and “enthusiasm” differ (Just, Crigler and Belt Reference Just, Crigler and Belt2007). Similarly, anger, disgust, and aversion are terms often used to label what seems to be the same latent factor.

3 Appendix B presents the items used to measure each of these factors We use a ten-item array developed for measuring the “Big Five” personality traits (Rammstedt and John Reference Rammstedt and John2007) and for Need for Cognition, the items developed by Cacioppo and Petty (Reference Cacioppo, Gardner and Berntson1982).

4 A then current dispute pitted two differently located orthogonal axes to depict a two-dimensional space. One view primarily advanced by Russell (Russell and Carroll Reference Russell and Carroll1999) held that the appropriate dimensions were “arousal” and “valence,” while others held that a 45° rotation, locating the dimensions as “positive” and “negative” (Cacioppo, Gardner and Berntson Reference Cacioppo and Petty1999; Watson et al. Reference Watson, Clark and Tellegen1999).

5 Reviews of the evolution of conceptions of affect can be found in Cornelius (Reference Cornelius1996), Marcus (Reference Marcus2003), Brader and Marcus (Reference Brader and Marcus2013).

6 Others have explored the more general use of slider approaches to measurement. Couper et al. have explored a horizontal version of the slider approach to measurement (Couper et al. Reference Couper, Tourangeau, Conrad and Singer2006). The resulting method, called visual analog scales, was compared with text box for input and radio buttons. They found visual analog scales generate comparable response distributions to the other response options, but took longer to complete and generated higher rates of missing data. However, they were not evaluating subjective appraisals of feeling states but rather attitudes. Further, they arrayed the visual analog scales they examined on the horizontal plane, while, in the studies reported below, we arrayed the “slider” scales vertically.

7 This project was funded by a National Science Foundation grant to the principals, grant number: SES-0617098.

8 The survey included another study which consumed half of the study participants. The 553 participants were randomly assigned to one of four conditions, of these four, two were the two response variants for measuring emotion used in this research.

9 As noted above, in so far as some of the sliders left unmoved ought to be treated as missing values, that introduces error variance that should, if present, weaken the relationship of these scales with the dependent variables.

10 The scoring of the slider format data can be set to any level of precision a scholar opts to apply. Here we opted for a range of 0–100 and did not seek to add further differentiation.

11 The results reported are for the unrotated results. For each, when rotated to an oblique solution, there are three factors with eigenvalues >1. The full results are available from the authors.

12 One of the anonymous referees suggested that because the radio button method offers participants but five response options it might be appropriate to treat these data as categorical. As such SEM analysis should be based on polychromic correlations and suitable estimation measures. However, Bollen and Barb (Reference Bollen and Barb1981) have shown with data such as these correlations derived measures built from collapsed variables (i.e., the underlying concept is continuous but the measurement crudely collapses the anticipated continuous distribution into relatively few categories) generate correlations that closely mimic those using continuous variables. In light of that research we favor using the same estimation method for both the radio button data and the slider data.

13 In each case, we also added an interaction term, method by story, to examine any possible combined effects. In none of the three analyses was the term significant (for the anxiety story, F (2, 281)=0.88, p not significant; for the enthusiasm story, F (2, 281)=1.28, p not significant; and, for the aversion story, F (2, 278)=0.22, p not significant).

14 This finding is not a result of including the array of other factors. A simple model, including just the three affect factors each measured using the radio button format, supports the same conclusion.

15 Those interested in how the slider format was developed and tested with other data (both Student and general population samples) and the variety of slider formats examined may find the cited conference papers useful (Marcus, Neuman and MacKuen 2009Reference Marcus, Neuman and MacKuena; Marcus, Neuman and MacKuen Reference Marcus, Neuman and MacKuen2009b).

16 The most recently published review of the literature on affect in politics is Brader and Marcus (Reference Brader and Marcus2013).

References

Abelson, Robert P., Kinder, Donald R., Peters, Mark D., and Fiske, Susan T.. 1982. ‘Affective and Semantic Components in Political Personal Perception’. Journal of Personality and Social Psychology 42:619630.CrossRefGoogle Scholar
Allport, Gordon W. 1954. The Nature of Prejudice. Cambridge, MA: Addison-Wesley.Google Scholar
Alwin, Duane F. 2010. ‘How Good is Survey Measurement? Assessing the Reliability and Validity of Survey Measures’. In Peter V. Wright Marsden and James D. Bingley (eds), Handbook of Survey Research, 405426. UK: Emerald Group.Google Scholar
Amodio, David M., Jost, John T., Master, Sarah L., and Yee, Cindy M. 2007. ‘Neurocognitive Correlates of Liberalism and Conservatism’. Nature Neuroscience 4:12461247.CrossRefGoogle Scholar
Ansolabehere, Stephen D., and Iyengar, Shanto. 1995. Going Negative: How Political Advertisements Shrink and Polarize the Electorate. New York: The Free Press.Google Scholar
Bollen, Kenneth A., and Barb, Kenney. 1981. ‘Pearson’s R and Coarsely Categorized Measures’. American Sociological Review 46:232239.CrossRefGoogle Scholar
Borsboom, Denny, Mellenbergh, Gideon J., and van Heerden, Jaap. 2003. ‘The Theoretical Status of Latent Variables’. Psychological Review 110:203219.CrossRefGoogle ScholarPubMed
Brader, Ted. 2006. Campaigning for Hearts and Minds: How Emotional Appeals in Political Ads Work. Chicago, IL: University of Chicago Press.Google Scholar
Brader, Ted, Groenendyk, Eric W., and Valentino, Nicholas. 2010. ‘Fight or Flight? When Political Threats Arouse Public Anger and Fear’. Annual Meeting of the Midwest Political Science Association, April 22–25, 2010, Chicago, IL.Google Scholar
Brader, Ted, and Marcus, George E.. 2013. ‘Emotion and Political Psychology’. In Leonie Huddy, Jack S. Levy and David O. Sears (eds), Oxford Handbook of Political Psychology, 165204. New York: Oxford University Press.Google Scholar
Cacioppo, John T., and Petty, Richard W. 1982. ‘The Need for Cognition’. Journal of Personality and Social Psychology 42:116131.CrossRefGoogle Scholar
Cacioppo, John T., Gardner, Wendi L., and Berntson, Gary G.. 1999. ‘The Affect System has Parallel and Integrative Processing Components: Form Follows Function’. Journal of Personality and Social Psychology 76:839855.CrossRefGoogle Scholar
Campbell, Angus, Converse, Philip E., Miller, Warren E., and Stokes, Donald E.. 1960. The American Voter. New York: John Wiley and Sons.Google Scholar
Campbell, Angus, Converse, Philip E., Miller, Warren E., and Stokes, Donald E.. 1966. Elections and the Political Order. New York: John Wiley and Sons, Inc. Google Scholar
Campbell, Donald T., and Fiske, Donald W.. 1959. ‘Convergent and Discriminant Validation by the Multitrait-Multimethod Matrix’. Psychological Bulletin 56:81105.CrossRefGoogle ScholarPubMed
Chaiken, Shelly, and Trope, Yaacov (eds). 1999. Dual Process Models in Social Psychology. New York: Guilford Press.Google Scholar
Cornelius, Randolf R. 1996. The Science of Emotion: Research and Tradition in the Psychology of Emotions. Upper Saddle River, NJ: Prentice Hall.Google Scholar
Couper, Mick P., Tourangeau, Roger, Conrad, Frederick G., and Singer, Eleanor. 2006. ‘Evaluating the Effectiveness of Visual Analog Scales: A Web Experiment’. Social Science Computer Review 24:227245.CrossRefGoogle Scholar
Feldt, Leonard S. 1969. ‘A Test of the Hypothesis That Cronbach’s Alpha or Kuder-Richardson Coefficient Twenty is the Same for Two Tests’. Psychometrika 34:363373.CrossRefGoogle Scholar
Feldt, Leonard S., Woodruff, David J., and Salih, Fathi A.. 1987. ‘Statistical Inference for Coefficient Alpha’. Applied Psychological Measurement 11:93101.CrossRefGoogle Scholar
Fiorina, Morris P., and Abrams, Samuel J.. 2008. ‘Political Polarization in the American Public’. Annual Review of Political Science 11:563588.CrossRefGoogle Scholar
Fiske, Susan T., and Taylor, Shelley E.. 1991. Social Cognition. New York: McGraw-Hill.Google Scholar
Gadarian, Shana K., and Albertson, Bethany. 2014. ‘Anxiety, Immigration, and the Search for Information’. Political Psychology 35:133164.CrossRefGoogle Scholar
Haidt, Jonathan. 2001. ‘The Emotional Dog and its Rational Tail: A Social Intuitionist Approach to Moral Judgment’. Psychological Review 108:814834.CrossRefGoogle ScholarPubMed
Hayduk, Leslie A., and Littvay, Levente. 2012. ‘Should Researchers Use Single Indicators, Best Indicators, or Multiple Indicators in Structural Equation Models?’. BMC Medical Research Methodology 12:117.CrossRefGoogle ScholarPubMed
Izard, Carroll E. 1971. Patterns of Emotions: A New Analysis of Anxiety and Depression. New York: Academic Press.Google Scholar
Izard, Carroll. E. 1977. Human Emotions. New York: Plenum Press.CrossRefGoogle Scholar
Jost, John T., and Amodio, David M.. 2012. ‘Political Ideology as Motivated Social Cognition: Behavioral and Neuroscientific Evidence’. Motivation and Emotion 36:5564.CrossRefGoogle Scholar
Jost, John T., Glaser, Jack, Kruglanski, Arie W., and Sulloway, Frank J.. 2003. ‘Political Conservatism as Motivated Social Cognition’. Psychological Bulletin 129:339375.CrossRefGoogle ScholarPubMed
Just, Marion R., Crigler, Ann N., and Belt, Todd L.. 2007. ‘Don’t Give Up Hope: Emotions, Candidate Appraisals, and Votes’. In W. Russell Neuman, George E. Marcus, Ann N. Crigler and Michael B. MacKuen (eds), The Affect Effect, 231259. Chicago, IL: University of Chicago Press.CrossRefGoogle Scholar
Kahneman, Daniel. 2011. Thinking, Fast and Slow. New York: Farrar, Straus and Giroux.Google Scholar
Krosnick, Jon A., and Presser, Stanley. 2010. ‘Question and Questionnaire Design’. In Peter V. Wright Marsden and James D. Bingley (eds), Handbook of Survey Research, 263314. UK: Emerald Group.Google Scholar
Lerner, Jennifer S., and Keltner, Dacher. 2000. ‘Beyond Valence: Toward a Model of Emotion-Specific Influences on Judgement and Choice’. Cognition and Emotion 14:473493.CrossRefGoogle Scholar
Lerner, Jennifer S., and Keltner, Dacher. 2001. ‘Fear, Anger, and Risk’. Journal of Personality and Social Psychology 81:146159.CrossRefGoogle ScholarPubMed
Lerner, Jennifer S., Gonzalez, Roxana M., Small, Deborah A., and Fischhoff, Baruch. 2003. ‘Effects of Fear and Anger on Perceived Risks of Terrorism: A National Field Experiment’. Psychological Science 14:144150.CrossRefGoogle ScholarPubMed
Lodge, G. Milton, and Tursky, Bernard. 1979. ‘Comparisons Between Category and Magnitude Scaling of Political Opinion Employing SRC/CPS Items’. American Political Science Review 73:5066.CrossRefGoogle Scholar
Lodge, G. Milton, and Tursky, Bernard. 1981. ‘On the Magnitude Scaling of Public Opinion in Survey Research’. American Journal of Political Science 25:376419.CrossRefGoogle Scholar
Lord, Charles G., Ross, Lee, and Lepper, Mark R.. 1979. ‘Biased Assimilation and Attitude Polarization: The Effects of Prior Theories on Subsequently Considered Evidence’. Journal of Personality and Social Psychology 37:20982109.CrossRefGoogle Scholar
MacKuen, Michael B., Wolak, Jennifer, Keele, Luke, and Marcus, George E.. 2010. ‘Civic Engagements: Resolute Partisanship or Reflective Deliberation’. American Journal of Political Science 54:440458.CrossRefGoogle Scholar
MacKuen, Michael B., Miller, Patrick R., Marcus, George E., and Neuman, W. Russell. 2013. ‘Affective Signatures and Attention: The Persistent Impact of Emotional Responses to the News’. Midwest Political Science Association Annual Meeting, April 11–13, 2013, Chicago, IL.Google Scholar
Marcus, George E. 2002. The Sentimental Citizen: Emotion in Democratic Politics. University Park, PA: Pennsylvania State University Press.Google Scholar
Marcus, George E. 2003. ‘The Psychology of Emotion and Politics’. In David O. Sears, Leonie Huddy and Robert Jervis (eds), Oxford Handbook of Political Psychology, 182221. Oxford: Oxford University Press.Google Scholar
Marcus, George E. 2008. ‘Different Situations, Different Responses: Threat, Partisanship, Risk, and Deliberation’. Critical Review 20:7589.CrossRefGoogle Scholar
Marcus, George E., and MacKuen, Michael B.. 1993. ‘Anxiety, Enthusiasm and the Vote: The Emotional Underpinnings of Learning and Involvement During Presidential Campaigns’. American Political Science Review 87:688701.CrossRefGoogle Scholar
Marcus, George E., MacKuen, Michael B., Wolak, Jennifer, and Keele, Luke. 2006. ‘The Measure and Mismeasure of Emotion’. In David Redlawsk (ed.), Feeling Politics: Emotion in Political Information Processing, 3145. New York: Palgrave Macmillan.CrossRefGoogle Scholar
Marcus, George E., Neuman, W. Russell, and MacKuen, Michael B. 2000. Affective Intelligence and Political Judgment. Chicago, IL: University of Chicago Press.Google Scholar
Marcus, George E., Neuman, W. Russell, and MacKuen, Michael B. 2009a. ‘Assessing Subjective Emotional Responses: Two Approaches to Measurement’. Midwest Political Science Association Annual Meetings. Chicago, IL.Google Scholar
Marcus, George E., Neuman, W. Russell, and MacKuen, Michael B. 2009b. ‘Measuring Subjective Emotional Response: New Evidence on an Alternative Method’. Annual Scientific Meeting of the International Society of Political Psychology, July 14–17, 2009. Dublin, Ireland.Google Scholar
Masters, Roger D. 2001. ‘Biology and Politics: Linking Nature and Nurture’. Annual Review of Political Science 4:345369.CrossRefGoogle Scholar
Masters, Roger D., and Sullivan, Denis. 1993. ‘Nonverbal Behavior and Leadership: Emotion and Cognition in Political Attitudes’. In Shanto Iyengar and William McGuire (eds), Explorations in Political Psychology, 150182. Durham, NC: Duke University Press.Google Scholar
Masters, Roger D., and Sullivan, Denis G. 1989. ‘Facial Displays and Political Leadership in France’. Behavioral Processes 19:130.CrossRefGoogle ScholarPubMed
Mauss, Iris B., and Robinson, Michael D.. 2009. ‘Measures of Emotion: A Review’. Cognition and Emotion 23:209237.CrossRefGoogle ScholarPubMed
Mayer, John D., and Gaschke, Yvonne N.. 1988. ‘The Experience and Meta-Experience of Mood’. Journal of Personality and Social Psychology 55:102111.CrossRefGoogle ScholarPubMed
McGuire, William J. 1969. ‘The Nature of Attitudes and Attitude Change’. In Gardner Lindzey and Elliot Aronson (eds), The Handbook of Social Psychology, 136314. Reading, MA: Addison-Wesley.Google Scholar
McHugo, Gregory J., Lanzetta, John T., Sullivan, Denis G., Masters, Roger D., and Englis, Basil. 1985. ‘Emotional Reactions to Expressive Displays of a Political Leader’. Journal of Personality and Social Psychology 49:15121529.CrossRefGoogle Scholar
Mutz, Diana C., and Reeves, Byron. 2005. ‘The New Videomalaise: Effects of Televised Incivility on Political Trust’. American Political Science Review 99:115.CrossRefGoogle Scholar
Neuman, W. Russell, Marcus, George E., and MacKuen, Michael B.. 2013. ‘Hardwired for News: Affective Intelligence and Political Attention’. Paper presented at the American Political Science Association Annual Meetings, Chicago, IL.Google Scholar
Nunnally, Jum C., and Bernstein, Ira H.. 1994. Psychometric Theory. New York: McGraw-Hill.Google Scholar
Ortony, Andrew, Clore, Gerald L., and Collins, Allan. 1989. The Cognitive Structure of Emotions. New York: Cambridge University Press.Google Scholar
Plutchik, Robert, and Conte, Hope R.. 1997. Circumplex Models of Personality and Emotions. Washington, DC: American Psychological Association.CrossRefGoogle Scholar
Rammstedt, Beatrice, and John, Oliver P.. 2007. ‘Measuring Personality in One Minute or Less: A 10-Item Short Version of the Big Five Inventory in English and German’. Journal of Research in Personality 41:203212.CrossRefGoogle Scholar
Redlawsk, David P. 2002. ‘Hot Cognition or Cool Consideration? Testing the Effects of Motivated Reasoning on Political Decision Making’. Journal of Politics 64:10211044.CrossRefGoogle Scholar
Roseman, Ira. 1979. ‘Cognitive Aspects of Emotion and Emotional Behavior’. Paper presented at the 87th Annual Convention of the American Psychological Association, New York, September 1–5, 1979.Google Scholar
Roseman, Ira. 1984. ‘Cognitive Determinants of Emotions: A Structural Theory’. In Phillip R. Shaver (ed.), Review of Personality and Social Psychology, 1136. Beverly Hills, CA: Sage.Google Scholar
Ruef, Anna M., and Levenson, Robert W. 2007. ‘Continuous Measurement of Emotion: The Affect Rating Dial’. In James A. Coan and John J. B. Allen (eds), The Handbook of Emotion Elicitation and Assessment, 286297. New York: Oxford University Press.CrossRefGoogle Scholar
Russell, James A., and Carroll, James M. 1999. ‘On the Bipolarity of Positive and Negative Affect’. Psychological Bulletin 125:330.CrossRefGoogle ScholarPubMed
Sullivan, Dennis, and Masters, Roger. 1988. ‘Happy Warriors: Leaders’ Facial Displays, Viewers Emotions, and Political Support’. American Journal of Political Science 32:345368.CrossRefGoogle Scholar
Valentino, Nicholas A., Brader, Ted, Groenendyk, Eric W., Gregorowicz, Krysha, and Hutchings, Vincent L.. 2011. ‘Election Night’s Alright for Fighting: The Role of Emotions in Political Participation’. Journal of Politics 73:156170.CrossRefGoogle Scholar
Valentino, Nicholas A., Hutchings, Vincent L., Banks, Antoine J., and Davis, Anne K.. 2008. ‘Is a Worried Citizen a Good Citizen? Emotions, Political Information Seeking, and Learning Via the Internet’. Political Psychology 29:247273.CrossRefGoogle Scholar
Watson, David. 1988aIntraindividual and Interindividual Analyses of Positive and Negative Affect: Their Relation to Health Complaints, Perceived Stress, and Daily Activities’. Journal of Personality and Social Psychology 54:10201030.CrossRefGoogle ScholarPubMed
Watson, David. 1988bThe Vicissitudes of Mood Measurement: Effects of Varying Descriptors, Time Frames, and Response Formats on Measures of Positive and Negative Affect’. Journal of Personality and Social Psychology 55:128141.CrossRefGoogle ScholarPubMed
Watson, David, and Tellegen, Auke. 1985. ‘Toward a Consensual Structure of Mood’. Psychological Bulletin 98:219235.CrossRefGoogle Scholar
Watson, David, Wiese, David, Vaidya, Jatin, and Tellegen, Auke. 1999. ‘The Two General Activation Systems of Affect: Structural Findings, Evolutionary Considerations, and Psychobiological Evidence’. Journal of Personality and Social Psychology 76:820838.CrossRefGoogle Scholar
Watson, David, and Clark, Lee Anna. 1994. The PANAS-X: Manual for the Positive and Negative Affect Schedule-Expanded Form, 35.Google Scholar
Watson, David, and Clark, Lee Anna. 1997. ‘Measurement and Mismeasurement of Mood: Recurrent and Emergent Issues’. Journal of Personality Assessment 68:267296.CrossRefGoogle ScholarPubMed
Watson, David, Clark, Lee Anna, and Tellegen, Auke. 1988. ‘Development and Validation of Brief Measures of Positive and Negative Affect: The PANAS Scales’. Journal of Personality and Social Psychology 54:10631070.CrossRefGoogle ScholarPubMed
Way, Baldwin M., and Masters, Roger D.. 1996. ‘Political Attitudes: Interactions of Cognition and Affect’. Motivation and Cognition 20:205236.Google Scholar
Zajonc, Robert B. 1980. ‘Feeling and Thinking: Preferences Need No Inferences’. American Psychologist 35:151175.CrossRefGoogle Scholar
Zuckerman, Marvin. 1991. Psychobiology of Personality. Cambridge, England: Cambridge University Press.Google Scholar
Figure 0

Fig. 1 Radio button/multiple choice format

Figure 1

Fig. 2 Slider format with midpoint start position and polar labels

Figure 2

Table 1 Means and Missing Data for Ten Affect Items by Method

Figure 3

Fig. 3 Box plots of ten affect measures by method: radio button and slider format Source: Knowledge Networks Survey 2009.

Figure 4

Fig. 4 Structural equation models of (a) radio button and (b) slider methods of affect measurement-standardized solutionSource: Knowledge Networks Survey 2009.

Figure 5

Table 2 Cronbach’s α and Feldt Test for Radio Button and Slider Measures of Affect

Figure 6

Fig. 5 Levels of anxiety by story and by affect methodSource: Knowledge Networks Study, Fall 2009.

Figure 7

Fig. 6 Levels of enthusiasm by story and by affect methodSource: Knowledge Networks Study, Fall 2009.

Figure 8

Fig. 7 Levels of aversion by story and by affect methodSource: Knowledge Networks Study, Fall 2009.

Figure 9

Table 3 Affect and Interest in Novel Information by Two Affect Methods

Figure 10

Table 4 Affect and Self-Confident Resolve (SCR) Orientation by Two Affect Methods

Figure 11

Table 5 Affect and Social Open Mindedness (SOM) Orientation by Two Affect Methods

Supplementary material: PDF

Marcus supplementary material

Appendix

Download Marcus supplementary material(PDF)
PDF 798.5 KB