1 Introduction
Without information about political actors’ ideological positions, our ability to study politics would be severely limited. We would, for instance, not be able to assess whether the positions of political representatives reflect those of their voters, or examine the degree of polarization within each group. Neither would we be able to study the extent to which ideological proximity influences individual voting decisions, which has been a key question in electoral research. In other words, mapping the positions of different actors within a common ideological space forms a key step toward answering these and other important questions about politics.
In recent years, a number of approaches have been developed to perform such mapping. One approach relies on models for scaling legislators based on their roll-call votes (Poole and Rosenthal Reference Poole and Rosenthal1985; Clinton et al. Reference Clinton, Jackman and Rivers2004). By combining legislative voting data with survey responses from ordinary citizens regarding their positions on specific policy proposals, researchers have been able to estimate the positions of both legislators and citizens on the same scale (e.g. Jessee Reference Jessee2009, Reference Jessee2016; Bafumi and Herron Reference Bafumi and Herron2010). However, a potential obstacle to using this approach is that it requires data from specifically tailored survey questions, which in practice may limit its feasibility. Another promising approach is to use social media data (e.g. Barberá Reference Barberá2015), but this alternative also has a few disadvantages. The sample will tend to be self-selected, and may still need to be matched with surveys to obtain data on other variables, such as vote choice.Footnote 1
In this light, approaches relying only on electoral surveys are still attractive, as such surveys exist in virtually every democracy, and aim to cover a representative sample of voters. Electoral studies typically ask respondents to place themselves as well as political stimuli (such as parties or candidates) on the same ideological scale, which can be very useful. However, there are also well-known challenges to using such data. One of these is that respondents tend to interpret the scale differently, potentially disagreeing on what constitutes a centrist or moderate position, as well as how distances from this position should be mapped onto the scale. In other words, respondents shift and stretch the ideological space in different ways—a form of differential item functioning (DIF). In their classic contribution, Aldrich and McKelvey (Reference Aldrich and McKelvey1977) provide an approach to estimate and correct for DIF, and their approach is likely to see more use in the future, as new implementations have recently become available (Poole et al. Reference Poole, Rosenthal, Lewis, Lo and Carroll2013; Hare et al. Reference Hare, Armstrong, Bakker, Carroll and Poole2015).
While these implementations of the Aldrich–McKelvey (AM) model offer improved estimation procedures, the basic modeling assumption—the sufficiency of an intercept and a stretch parameter—has been retained for more than 40 years. This is potentially problematic because it ignores the pervasive effects of rationalization in the form of projection bias; that is, the tendency for respondents to place stimuli they like too close to themselves, and ones they do not like too far away (see, e.g. Johnston, Fournier, and Jenkins Reference Johnston, Fournier and Jenkins2000). The failure to model these mechanisms is likely to bias the estimated shift and stretch parameters of the AM model, and hence also to bias the estimates of respondents’ ideal points. In fact, the Monte Carlo simulations reported in this study show that the bias can be considerable. As one might expect, the stimuli position estimates are generally more robust to rationalization, but even these can be misleading.
This study addresses these issues by developing an alternative scaling approach, incorporating more information than the AM model to jointly estimate DIF and rationalization bias. As a by-product of this effort, the study also provides an improved version of the Bayesian AM implementation by Hare et al. (Reference Hare, Armstrong, Bakker, Carroll and Poole2015). The models are first tested and compared to existing models in Monte Carlo simulations. The differences between the models are then illustrated with an application to a difficult empirical case. Finally, posterior predictive checks demonstrate the degree of rationalization in the data and show the extent to which the models manage to fit such patterns.
2 Differential Item Functioning
A well-known challenge for inter-personal comparisons of survey data is that respondents tend to interpret questions differently (Brady Reference Brady1985; King et al. Reference King, Murray, Salomon and Tandon2004). Thus, respondents who are in perfect ideological agreement may place themselves at different positions on an ideological scale—or they may disagree considerably and nevertheless place themselves at the same position. In particular, it is likely that respondents disagree on two things: What constitutes the center of the scale (i.e. who the moderate actors are), as well as what a step on the scale means (i.e. how extreme the off-center actors are).
2.1 The Bayesian Aldrich–McKelvey (BAM) Model
Aldrich and McKelvey (Reference Aldrich and McKelvey1977) develop a model to capture such differences, estimating individual shift and stretch parameters that relate latent stimuli positions to reported perceptions of these. This approach has proven useful in a number of studies (e.g. Saiegh Reference Saiegh2009; Lo, Proksch, and Gschwend Reference Lo, Proksch and Gschwend2014),Footnote 2 and it has also been tested and developed further.Footnote 3 Hare et al. (Reference Hare, Armstrong, Bakker, Carroll and Poole2015) develop a Bayesian implementation of the AM model (referred to as BAM), which allows for heteroskedasticity, and provides measures of uncertainty for all estimates of interest. The general question addressed in this literature is how respondent $i\in \{1,\ldots ,N\}$ transforms the latent position
$\unicode[STIX]{x1D703}_{j}$ of political actor
$j\in \{1,\ldots ,J\}$ into the reported position
$Y_{ij}$. If we let
$\unicode[STIX]{x1D719}(\cdot |\cdot ,\cdot )$ be the probability density function of the normal distribution, then the likelihood function for the BAM model as specified by Hare et al. (Reference Hare, Armstrong, Bakker, Carroll and Poole2015) can be written as:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_eqn1.png?pub-status=live)
where $\unicode[STIX]{x1D6FC}_{i}$ is an intercept (or shift) parameter,
$\unicode[STIX]{x1D6FD}_{i}$ is a stretch (or weight) parameter, and
$\unicode[STIX]{x1D70E}_{ij}^{2}$ is a variance term, constructed as
$\unicode[STIX]{x1D70E}_{ij}^{2}=(\unicode[STIX]{x1D70F}_{i}\unicode[STIX]{x1D702}_{j})^{-1}$ to allow for heteroskedasticity.Footnote 4 (Priors and further details on this model are discussed in a later section.) To calculate respondents’ ideal points, the authors scale the reported self-placements (
$V_{i}$) for each posterior draw as follows:Footnote 5
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_eqn2.png?pub-status=live)
Like other ideological scaling models, this approach assumes it is meaningful to map actors onto ideological dimensions that summarize preferences across a number of issue domains. This assumption underpins a large existing literature, using a variety of approaches (e.g. Jessee Reference Jessee2009; Bafumi and Herron Reference Bafumi and Herron2010; Imai, Lo, and Olmsted Reference Imai, Lo and Olmsted2016). However, a number of authors also question whether all voters hold strong ideological convictions (e.g. Barber and Pope Reference Barber and Pope2019), and whether their issue preferences are sufficiently consistent for summary measures to be meaningful (Broockman Reference Broockman2016). This concern is particularly relevant for studies of representation, which may be better off using issue-specific data (Ahler and Broockman Reference Ahler and Broockman2018).
Nevertheless, ideological dimensions do play an important role in both political discourse and academic research, and the goal of this study is therefore to improve the models currently in use. The observation that some voters fit less well within spatial models may in fact also have relevant modeling implications. As discussed further below, the existing literature suggests that such voters still try to behave in accordance with a spatial model, by rationalizing their preferences. Such behavior would invalidate the assumption of the AM model that the shift and stretch parameters are sufficient to capture the differences in how respondents report their perceptions of political positions. In other words, such behavior would call for a more comprehensive model, which is what this study aims to develop.
3 Rationalization Bias
The notion of rationalization bias challenges the assumption that voters first form impressions of where political actors are located ideologically and then form preferences over these actors. It suggests a more complicated picture where respondents also adjust their (reported) perceptions of stimuli positions to fit their already formed political affinities and antipathies (Markus and Converse Reference Markus and Converse1979; Conover and Feldman Reference Conover and Feldman1982; Krosnick Reference Krosnick1990). The existing literature suggests that voters rationalize their preferences via two forms of projection (van der Brug Reference van der Brug2001; Merrill, Grofman, and Adams Reference Merrill, Grofman and Adams2001). First, respondents are likely to place stimuli that they like closer to themselves than they truly are, which is often referred to as assimilation effects. Second, respondents are likely to place stimuli that they do not like further away from themselves than they truly are, which is often referred to as contrast effects. Such behavior appears to be pervasive, posing a major challenge for studies using survey data on actor placements. In the words of Johnston, Fournier, and Jenkins (Reference Johnston, Fournier and Jenkins2000, 1151): “Perceptions are rife with bias”.
Modeling such bias poses a couple of challenges. First, given the low number of stimuli in most settings, a parsimonious model is required: The number of individual-level parameters needs to be strictly limited. Second, the existing literature does not offer formalizations that can readily be included in a new model. In particular, contrast effects are somewhat complicated to model, and the existing literature contains very few attempts to do so.Footnote 6 I argue, however, that assimilation and contrast effects can be seen as facets of the same underlying phenomenon. As Westholm (Reference Westholm1997) notes, what constitutes a good fit to individual preferences will depend on which spatial model respondents have in mind. Recent experiments suggest that most voters behave in accordance with a proximity model—appearing to penalizing distant actors and rewarding close ones (Claassen Reference Claassen2007; Tomz and Van Houweling Reference Tomz and Van Houweling2008; Lacy and Paolino Reference Lacy and Paolino2010). When voters’ preferences diverge from this pattern, because they like actors that are ideological distant, or dislike actors that are close to them, cognitive dissonance theory (Festinger Reference Festinger1957; Aronson Reference Aronson1997) would predict that voters reduce this discrepancy by misperceiving the actors’ stances as necessary.
I thus argue that the main consequence of both assimilation and contrasting processes is to exaggerate the fit of the proximity model of spatial voting. This point is not very clear in the literature on rationalization bias, but it can be very useful for modeling rationalization. While modeling assimilation and contrast effects separately might call for two new individual-level parameters, modeling these effects as part of the same underlying mechanism is more parsimonious. It also ensures that the effects are modeled in a theoretically and formally consistent way. The approach taken here is therefore to use a proximity model to define “ideal” stimuli positions and estimate the extent to which respondents move stimuli toward these positions.
3.1 The Intercept–Stretch–Rationalization (ISR) Model
The aim of this study is to build a model that accounts for rationalization bias as well as DIF. While the AM model is well suited for capturing DIF, accounting for rationalization bias requires a more elaborate model employing additional data. In particular, the model needs to incorporate information on whether the stimuli are liked or disliked by the respondents, as well as where the respondents are located.Footnote 7 Fortunately, surveys asking respondents to place themselves and political actors on ideological scales also tend to ask respondents for their assessment of these actors (e.g. in terms of thermometer scales or probabilities of ever voting for them). In the following, individual $i$’s assessment of stimulus
$j$ will be denoted
$U_{ij}$, and the original data will be linearly transformed so that
$U_{ij}\in [0,1]$, where each respondents’ least preferred stimulus is rated zero and the most preferred is rated one.Footnote 8
As discussed above, I use the proximity model to define “ideal” stimuli positions, conditional on respondent self-placements and preferences, and estimate the extent to which respondents move stimuli toward these positions. As a first step, the standard proximity model can be defined as follows:Footnote 9
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_eqn3.png?pub-status=live)
where $\unicode[STIX]{x1D705}_{i}$ is an intercept,
$\unicode[STIX]{x1D70B}_{i}$ is a coefficient, and
$p_{ij}^{\ast }$ is a hypothetical party or candidate position. In the present case,
$\unicode[STIX]{x1D705}_{i}=\max (U)=1$. The model can also be written somewhat differently, taking into account whether
$p_{ij}^{\ast }$ is on one or the other side of
$V_{i}$, and letting
$\unicode[STIX]{x1D70B}_{i}$ vary accordingly:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_eqn4.png?pub-status=live)
Assuming that preferences drop to zero when a stimulus reaches the upper bound ($B$) or the lower bound (
$-B$), Equation (4) can be solved for
$\unicode[STIX]{x1D70B}_{i1}$ and
$\unicode[STIX]{x1D70B}_{i2}$, which yields:
$\unicode[STIX]{x1D70B}_{i1}=(B+V_{i})^{-1}$ and
$\unicode[STIX]{x1D70B}_{i2}=(B-V_{i})^{-1}$.Footnote 10 Substituting these expressions for
$\unicode[STIX]{x1D70B}_{i1}$ and
$\unicode[STIX]{x1D70B}_{i2}$ in Equation (4) and solving for
$p_{ij}^{\ast }$, we can define the “ideal” position for each stimuli as follows:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_eqn5.png?pub-status=live)
However, Equation (5) introduces a challenge, as it requires information on whether the respondent perceives the stimuli to be on one side or the other. We cannot, for instance, use the observed condition $V_{i}>Y_{ij}$ to define
$p_{ij}^{\ast }$ as this would bias the estimates of rationalization. Instead, we need an estimate of the direction rationalization will take. We could potentially use the condition
$V_{i}>\unicode[STIX]{x1D703}_{ij}^{\ast }$, where
$\unicode[STIX]{x1D703}_{ij}^{\ast }$ represents stimulus
$j$’s latent position translated to the respondent’s own scale:
$\unicode[STIX]{x1D703}_{ij}^{\ast }=\unicode[STIX]{x1D6FC}_{i}+\unicode[STIX]{x1D6FD}_{i}\unicode[STIX]{x1D703}_{j}$. Yet this implies an abrupt shift as
$\unicode[STIX]{x1D703}_{j}^{\ast }$ passes
$V_{i}$, especially for the most disliked stimuli. This would in turn cause trouble for Hamiltonian Monte Carlo approaches, which require continuously differentiable functions. The alternative used here is inspired by the standard approach to handling discrete parameters when direct sampling is infeasible, namely marginalization. In particular, the model is specified as a mixture of two components, representing the two sides on which respondents may perceive the stimuli to be.
The model assumes that the direction of rationalization is at least partly determined by a latent DIF-process.Footnote 11 More specifically, it estimates the extent to which the mixing proportion $l_{ij}\in [0,1]$ depends on a logistic transformation of the distance between
$V_{i}$ and
$\unicode[STIX]{x1D703}_{ij}^{\ast }$:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_eqn6.png?pub-status=live)
where the parameter $\unicode[STIX]{x1D6FF}\in [0,1]$ weights the DIF-based estimate against the alternative of equiprobability, and the constant
$z$ determines how rapidly the indicator shifts from zero to one.Footnote 12 The result is a smooth transition in the mixing proportions as
$\unicode[STIX]{x1D703}_{ij}^{\ast }$ passes
$V_{i}$.
To incorporate rationalization effects in the scaling model, the individual-level intercept and stretch parameters of the BAM model are supplemented with a rationalization component, resulting in a model that will be referred to as ISR. If we again let $\unicode[STIX]{x1D719}(\cdot |\cdot ,\cdot )$ be the probability density function of the normal distribution, then the likelihood function for the ISR model can be written as:
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_eqn7.png?pub-status=live)
where the parameter $\unicode[STIX]{x1D6FE}_{i}\in [0,1]$ captures the extent of rationalization for each respondent.
4 Priors and Identification
It should be noted that the likelihoods discussed above are unidentified. Both the BAM model and the ISR model are characterized by reflection invariance and scaling invariance (see, e.g. Jackman Reference Jackman2001; Bafumi et al. Reference Bafumi, Gelman, Park and Kaplan2005). Hare et al.’s solution is to fix two stimuli positions on the latent scale while placing wide, uniform priors on $\unicode[STIX]{x1D6FC}$ and
$\unicode[STIX]{x1D6FD}$:
$\unicode[STIX]{x1D6FC}_{i},\unicode[STIX]{x1D6FD}_{i}\sim \text{Unif}(-100,100)$. The authors further place standard normal priors on the remaining latent stimuli positions,
$\unicode[STIX]{x1D703}_{j}\sim \text{Normal}(0,1)$, and gamma priors on the precision terms:
$\unicode[STIX]{x1D702}_{j}\sim \text{Gamma}(0.1,0.1)$, and
$\unicode[STIX]{x1D70F}_{i}\sim \text{Gamma}(\unicode[STIX]{x1D708},\unicode[STIX]{x1D714})$. For the individual-specific precision terms, they also place gamma hyperpriors on the shape and rate parameters:
$\unicode[STIX]{x1D708},\unicode[STIX]{x1D714}\sim \text{Gamma}(0.1,0.1)$.
However, Hare et al.’s approach would not be sufficient to ensure identification of the ISR model. The approach taken here is rather to use more informative priors on $\unicode[STIX]{x1D6FC}$ and
$\unicode[STIX]{x1D6FD}$. Within the Bayesian paradigm, including prior information about the parameters is not only straightforward, but may also facilitate estimation and yield more accurate estimates (as the Monte Carlo simulations below also illustrate). The goal here is to include enough prior information to rule out unreasonable results and permit efficient estimation, while still allowing the results to be driven by the data (for similar arguments, see e.g. Gelman et al. Reference Gelman, Jakulin, Pittau and Su2008; Stan Development Team 2019).
In particular, $\unicode[STIX]{x1D6FD}$ is given a normal prior with a mean of one, which implies an approximate one-to-one relationship between the observed and the latent scale. This is harmless because the latent stimuli positions are given a weakly informative prior,
$\unicode[STIX]{x1D703}_{j}\sim \text{Normal}(0,10^{2})$, allowing for adjustment in accordance with the
$\unicode[STIX]{x1D6FD}$’s. A more important issue is the variance of the
$\unicode[STIX]{x1D6FD}$-prior, which for the ISR model is set to one to allow for a considerable degree of stretching, while ruling out unreasonable values:
$\unicode[STIX]{x1D6FD}_{i}\sim \text{Normal}(1,1)$.Footnote 13 The intercept
$\unicode[STIX]{x1D6FC}$ is given a normal prior with a mean of zero, reflecting the assumption that both scales are centered at this point:
$\unicode[STIX]{x1D6FC}_{i}\sim \text{Normal}(0,\unicode[STIX]{x1D706}^{2})$. To permit a reasonable degree of shrinking, the standard deviation of this prior is given a weak half-Cauchy hyperprior:
$\unicode[STIX]{x1D706}\sim \text{Cauchy}^{+}(0,B)$.
Turning to the other parameters of the ISR model, $\unicode[STIX]{x1D6FF}$ is given a beta prior, reflecting the belief that the direction of rationalization most likely depends on a DIF-process:
$\unicode[STIX]{x1D6FF}\sim \text{Beta}(3,1.1)$. The
$\unicode[STIX]{x1D6FE}$’s are also given beta priors, with hyperparameters reflecting the estimated population distribution:
$\unicode[STIX]{x1D6FE}_{i}\sim \text{Beta}(\unicode[STIX]{x1D6FC}_{\unicode[STIX]{x1D6FE}},\unicode[STIX]{x1D6FD}_{\unicode[STIX]{x1D6FE}})$, where
$\unicode[STIX]{x1D6FC}_{\unicode[STIX]{x1D6FE}},\unicode[STIX]{x1D6FD}_{\unicode[STIX]{x1D6FE}}\sim \text{Gamma}(1.5,0.5)$.Footnote 14 Finally,
$\unicode[STIX]{x1D70E}_{ij}$ is decomposed into the product of a unit
$J$-simplex vector,
$\unicode[STIX]{x1D70C}$, and the parameter
$\unicode[STIX]{x1D702}_{i}$, which represents the average variance of respondent
$i$, implicitly increased by a factor of
$J^{2}$.Footnote 15 In other words:
$\unicode[STIX]{x1D70E}_{ij}=\unicode[STIX]{x1D70C}_{j}\sqrt{\unicode[STIX]{x1D702}_{i}}$, where
$\unicode[STIX]{x1D702}_{i}$ is modeled with a scaled inverse chi-squared distribution,
$\unicode[STIX]{x1D702}_{i}\sim \text{Scale-inv-}\unicode[STIX]{x1D712}^{2}(\unicode[STIX]{x1D708}_{\unicode[STIX]{x1D702}},\unicode[STIX]{x1D70F}_{\unicode[STIX]{x1D702}}^{2})$, using weakly informative hyperpriors:
$\unicode[STIX]{x1D708}_{\unicode[STIX]{x1D702}}\sim \text{Cauchy}^{+}(0,50)$, and
$\unicode[STIX]{x1D70F}_{\unicode[STIX]{x1D702}}\sim \text{Cauchy}^{+}(0,JB)$. The simplex vector is given a symmetric Dirichlet prior:
$\unicode[STIX]{x1D70C}\sim \text{Dirichlet}(5)$.
This means that the ISR model differs from the BAM specification by Hare et al. in two ways: (1) the inclusion of preference data to model rationalization, and (2) the choice of priors and the approach to identify the model. Because the model specified by Hare et al. is the most recent contribution to this literature, it is useful to test it as it is, and the Stan implementation of this model will simply be referred to as BAM. However, it is also useful to examine whether the BAM model would perform better with different priors. The analyses below will therefore also include a version of the BAM model using the same priors as the ISR model, and this version will be referred to as BAM2. In addition, the analyses will include the original AM model, as this is useful for replicating existing studies (in particular Lo, Proksch, and Gschwend Reference Lo, Proksch and Gschwend2014). For convenience, the AM, BAM, and BAM2 models will be referred to as AM-type models.
5 Simulation Study
A key question is how well the AM-type models perform in the presence of rationalization. We would expect the estimates of voter positions to deteriorate as the degree of rationalization increases, and the same might apply to the estimates of stimuli positions. Another question is how the performance of the ISR model compares to the other models—whether rationalization is present or not. To answer these questions, this section reports a simulation study.
The data in this study are generated using an ISR-type data generating process. I focus on three scenarios, where the $\unicode[STIX]{x1D6FE}$-parameters are drawn from beta distributions producing increasing degrees of rationalization:
$\text{Beta}(0,1)$,
$\text{Beta}(1,3)$, and
$\text{Beta}(1,1)$. In other words, the first scenario entails no rationalization (only DIF), while the second and third scenarios entail
$\unicode[STIX]{x1D6FE}$-averages around 0.25 and 0.5, respectively. The results reported here entail moderately heteroskedastic errors with an average standard deviation close to 0.5, while a study with larger errors is reported as supplementary material. In both studies, the values of all other parameters are set approximately equal to those estimated in the empirical application below.
For each of the three scenarios, I generate 200 datasets of a typical size ($N=500$,
$J=8$,
$B=5$). I fit the AM model using the maximum likelihood (ML) implementation in the R package basicspace. The BAM, BAM2, and ISR models are fit using a version of the No-U-Turn Sampler (NUTS) by Hoffman and Gelman (Reference Hoffman and Gelman2014), which is an automatically tuned form of Hamiltonian Monte Carlo, implemented in Stan (Carpenter et al. Reference Carpenter, Gelman, Hoffman, Lee, Goodrich, Betancourt, Brubaker, Guo, Li and Riddell2017; Stan Development Team 2018, 2019). Because the models produce differently scaled results, I assess the voter and stimuli position estimates by their correlations with the true values.Footnote 16 Additional test criteria are reported as supplementary material.Footnote 17
Figure 1 reports the main results. The left panel shows correlations between estimated and true stimuli positions. As we would expect, all models perform well in the first scenario, where there is no rationalization. Increasing the degree of rationalization moderately does not have a large impact on the performance of the AM-type models, although the tail gradually increases, reflecting poorer performance in a few cases. When we reach the scenario with the most rationalization, the performance of the AM-type models is notably worse than that of the ISR model, but it is still reasonably good in most cases.
The right panel shows similar correlations for voter positions. These estimates are particularly important to test, as they are hard to validate outside of a simulation context and are likely to be affected by rationalization. As we would expect, the AM-type models’ ability to recover voter positions deteriorates with increasing rationalization, while the ISR model’s performance is more stable. The correlations for the AM-type models are increasingly variable, with some results being very poor—to the extent that there is hardly any positive correlation at all. It is thus worth noting that even when the AM-type models perform well in terms of recovering stimuli positions (for which estimates are generally easier to validate), the voter position estimates may still be poor. Another point to notice is that the BAM2 model performs better than the two other AM-type models on this criteria, illustrating that more informative priors can make the model more robust to rationalization. The results from a similar simulation study with larger errors (reported as supplementary material) show that the BAM2 model is also more robust to random noise.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_fig1.png?pub-status=live)
Figure 1. Correlations between estimates and true positions in simulations.
6 Empirical Application
Having seen how the models perform in a controlled setting, this section applies the models to real data. Among the existing studies employing AM-type models, Lo, Proksch, and Gschwend (Reference Lo, Proksch and Gschwend2014) use a particularly comprehensive and suitable set of survey data, and these will be used here as well. The data come from the European Election Study 2009 (EES 2011), where the respondents were asked to place parties as well as themselves on 11-point left–right scales (see the supplementary materials for exact question wording).Footnote 18 To facilitate analysis and presentation, the scales are given a range from $-5$ to
$+5$, with
$0$ as a center category. The EES respondents were also asked to report their party preferences, expressed as the probability of ever voting for a party.Footnote 19 In addition to the EES data, party position data from the Chapel Hill Expert Survey (CHES; Bakker et al. Reference Bakker, De Vries, Edwards, Hooghe, Jolly, Marks, Polk, Rovny, Steenbergen and Vachudova2015) will be used for validation purposes.
Among the EES countries, I focus on a difficult case where the differences between the models are likely to be notable.Footnote 20 I examine the UK, which is a special case in several ways: The sample mean stimuli placements only have a correlation of 0.76 with the mean CHES placements, while most countries have correlations above 0.9. Furthermore, of all the countries Lo, Proksch, and Gschwend (Reference Lo, Proksch and Gschwend2014) analyze, the UK is where their model is reported to explain the least of the variance—with the exception of Romania. In addition, 44% of the respondents are reported to have negative weights (again, only Romania with 46% has a higher share). This number is extreme, as it would suggest that nearly half the sample cannot tell the political left from right. In short, the AM model does not seem to fit very well, but Lo, Proksch, and Gschwend (Reference Lo, Proksch and Gschwend2014) nevertheless obtain party positions estimates that are much more in line with expert placements than the mean placements are.
To ensure comparability, this case study uses the exact same data selection criteria as Lo, Proksch, and Gschwend, which yield a selection of 536 individuals.Footnote 21 Their results are replicated using the ML implementation of the AM model in the R package basicspace. However, in order to obtain outputs that are comparable to the other models (e.g. a posterior predictive distribution), the AM model is also implemented in a Bayesian fashion, using a model referred to as AM*. A key difference between the AM and BAM models is the AM model’s assumption of homoskedastic errors, and the AM* model shares this assumption. Apart from this feature, the AM* model has the same specification and priors as the BAM model.Footnote 22
6.1 Results
Table 1 summarizes the results. The ML-estimated AM model yields an exact replication of the results reported by Lo, Proksch, and Gschwend (Reference Lo, Proksch and Gschwend2014), including 44% negative weights. The correlation between the party position estimates and CHES estimates is 0.92, which is an improvement over the naive approach of using sample mean placements $(r=\text{0.76})$. Accordingly, the party positions estimates also differ considerably from the mean placements—the correlation between the two is only 0.49. The AM* model largely replicates these results, although with minor deviations.
Table 1. Summary and validation of model fits for the UK.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_tab1.png?pub-status=live)
The BAM model yields very different results. Interestingly, the share of negative weights is reduced from 44% to a more plausible 15%.Footnote 23 However, the party position estimates deviate notably from the CHES estimates, showing a modest correlation of 0.56. The party position estimates are instead very similar to the sample mean placements $(r=\text{0.97})$, which would have been fine if these had greater convergent validity with the expert placements. We also see that the BAM2 model (which has the same priors as the ISR model) shows a similar performance to the BAM model, but has a slightly stronger correlation with the CHES estimates (0.60) and fewer negative weights (10%). In sum, the homoskedastic AM models appear to produce more valid party position estimates, at the cost of an extremely high share of negative weights, while the heteroskedastic BAM models show the opposite characteristics.
While none of the AM-type specifications shows a fully satisfactory performance in this case, the ISR model produces more convincing results. The party position estimates have a very strong correlation with the expert placements, and the share of negative weights is more plausible than for the homoskedastic AM models. The table also reports the Watanabe–Akaike information criterion (WAIC; Watanabe Reference Watanabe2010), which implies that the ISR model fits the data significantly better than the other models even when the effective number of parameters is penalized.
Given their considerable differences, the party position estimates of the BAM and ISR models deserve further attention, and these are shown in Figure 2. Interestingly, the BAM model appears to have particular problems recovering the positions of parties on the right side of the scale. The two parties that experts consider to be far-right—the UK Independence Party (UKIP) and the British National Party (BNP)—are estimated to be close to the center, while the Conservatives—which experts consider more moderate—are estimated to be more extreme. The BAM model also appears to place the Greens too close to the center, relative to the other parties.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_fig2.png?pub-status=live)
Figure 2. Party position estimates over expert placements.
Figure 3 further shows the relationships between the voter position estimates from three of the models, along with their densities. This illustrates the trade-off presented by the AM model, which in this case yields convincing party position estimates, at the cost of a large share of negative weights that are likely to give misleading voter position estimates in a number of cases. In fact, the voter estimates of the AM model only correlate with those of the ISR model and the BAM model at 0.42 and 0.38, while the correlation between the two latter is 0.67. (A similar plot including the BAM2 model is reported as supplementary material.)
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_fig3.png?pub-status=live)
Figure 3. Correlations and densities of estimated voter positions.
A useful feature of the ISR model is that it allows us to assess the extent of rationalization in data, as well as whether there is heterogeneity in the tendency to rationalize. Figure 4 shows the relationship between the estimated $\unicode[STIX]{x1D6FE}$-parameters and voter self-placements, by fitting penalized smoothing splines for each of 100 posterior samples. This reveals a pattern where voters on the left show only a weak to moderate tendency to rationalize, while those on the right do so to a much larger extent. This result is validated and discussed further below.
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_fig4.png?pub-status=live)
Figure 4. Estimated rationalization over voter positions.
6.2 Observed and Predicted Patterns
This section checks how well the models fit the data, focusing on any observed patterns of rationalization. When a model fits well, data generated under the model should look similar to the observed data—this is the basic idea underlying posterior predictive model checking (Rubin et al. Reference Rubin1984; Gelman et al. Reference Gelman, Carlin, Stern, Dunson, Vehtari and Rubin2014). The tests below are based on 1500 draws of simulated outcome data from the posterior predictive distribution for each model. The key question is whether the predictions on average are consistent with observed response patterns across various combinations of self-placements, party positions, and preferences.
In particular, Figure 5 reports the observed mean placement of each party along with densities of mean predictions, conditional on whether the respondents place themselves left ($V_{i}<0$) or right of center (
$V_{i}>0$), as well as whether they like (
$U_{ij}\geqslant 0.5$) or dislike (
$U_{ij}<0.5$) the party in question. With 8 parties, this gives 32 panels, which have been ordered so that the first row belongs to the leftmost party (according to CHES data), and the last row belongs to the rightmost. For reference, the average CHES placements are also shown with dashed vertical lines (these have been rescaled to range from
$-2$ to 2, which is very close to the range of the observed EES survey means).
![](https://static.cambridge.org/binary/version/id/urn:cambridge.org:id:binary:20200531103532134-0451:S1047198719000421:S1047198719000421_fig5.png?pub-status=live)
Figure 5. Posterior predictive densities and observed mean party placements.
Figure 5 shows a remarkable degree of rationalization, but mostly among respondents on the right. As shown in the rightmost column of the plot, when respondents on the right dislike a party, they place it on the opposite side of the scale, no matter its actual position. Even the Conservatives, and parties such as BNP and UKIP—which experts consider to be far-right—are reported to be on the left by the rightist voters who do not like them. In other words, we see a clear example of contrasting. As shown in the third column of the plot, respondents on the right also show a notable degree of assimilation: When they like a party, they tend to place it on the right, even if experts place it firmly on the left. The key examples here are the Labour Party and the Greens (while SNP and Plaid Cymru are placed very close to the middle).
Among respondents who place themselves left of center (shown in the two leftmost columns), such patterns are much less pronounced. In fact, when it comes to placing parties on one or the other side of the center, these respondents agree with the experts in every single case. Comparing the two columns, we see that their placements depend considerably less on whether they like or dislike the parties—in most cases there is little difference. The clearest pattern of rationalization among these respondents is some degree of assimilation for BNP and UKIP, but overall they exhibit a much weaker tendency to rationalize than respondents on the right do. These observed patterns thus serve to validate the results shown in Figure 4, which also imply considerably more rationalization among respondents on the right than on the left.
The question is to what extent the models capture these patterns, and whether the observed data look plausible under each model. Ideally, we would like the observed means to lie within a range of high probability for a given model in every scenario. In practice, however, this would be a demanding requirement, as any uncaptured heterogeneity in how respondents treat specific parties may throw the predictions off. Indeed, what we see for the ISR model is that in those cases where its predictions are somewhat off, the observed mean tends to stand out compared to those of the neighboring parties. However, such cases are relatively rare for the ISR model—its predictions are in general close to the observed means. The two AM-type models show a less consistent performance, with very poor predictions in several cases. The predictions of the BAM2 model (which are reported as supplementary material) turn out slightly worse than those of the BAM model—which is not surprising for a more constrained model.
7 Conclusion
While survey data can provide very useful information on where voters and their representatives are located along ideological dimensions, we know that respondents often display a number of biases when answering surveys. The AM approach was developed to deal with the challenge of DIF, and it serves that purpose well. However, DIF is only one of the key biases survey respondents display, and this study aims to narrow the gap between our knowledge of political psychology and the models we employ.
The results presented above show that estimates of both voter and stimuli positions can be improved by modeling rationalization along with DIF. Monte Carlo simulations show that AM-type models produce increasingly inaccurate estimates of voter positions as the degree of rationalization increases. The effect on estimates of stimuli positions is considerably weaker, but can still be substantial—as the case of the UK also illustrates. In practice, when sample mean placements have high convergent validity (e.g. strong correlations with expert placements), the AM-type models may provide very accurate stimuli position estimates, but the ISR model still tends to fit the data significantly better.Footnote 24 Together with the Monte Carlo simulations, this implies that the voter position estimates from AM-type models are at risk of being biased even in cases where the stimuli positions are accurately estimated.
However, the results also show that the performance of AM-type models can be significantly improved by using more informative priors and an alternative approach to identify the parameters. The BAM2 model uses the same priors as the ISR model, and compared to the AM and BAM models, it is more robust to rationalization, as well as random noise. The BAM2 model should thus be preferred to the existing models, and could provide a relevant alternative to the ISR model—which requires more data. The ISR model is particularly useful when we are substantively interested in rationalization, or when we need estimates that are corrected for such bias. Yet, the ISR model performs as well as the BAM2 model even when there is no rationalization, so there is no clear reason not to consider it the default model.
It is, however, important to consider when these types of models are appropriate. A key assumption of the models examined here is that each stimulus has a single latent position on the scale, and this may seem uncontroversial if we agree that an actor can only have one position at a given time. The strength of these models is to map each actor within a common ideological space that exists beyond the actors’ subjective perceptions. However, it is entirely possible that voters genuinely perceive party positions differently, because they receive different messages, or interpret them in different ways (Lewis and King Reference Lewis and King1999). For some research questions, idiosyncratic perceptions (even biased ones) may provide the most appropriate data. As Bølstad and Dinas (Reference Bølstad and Dinas2017) argue, this may be the case when we are more interested in subjective cognition than the role of external stimuli. In most other situations, however, ISR scaling should be a useful way to map voters and representatives on the same scale, based on a very common form of data.
In their conclusion, Hare et al. (Reference Hare, Armstrong, Bakker, Carroll and Poole2015, 772) note that “while the relationship between the Bayesian AM stimuli estimates and external measures of ideology [...] appears to be strong, it is not perfect”, and ask “[w]hy are some stimuli perceived to be more or less ideologically extreme than is indicated by [such measures]? Are these deviations random noise or the result of ideological maneuvering?” This study may provide a piece of the answer: The deviations do not just represent random noise (or ideological maneuvering), but partly reflect rationalization bias. By taking this bias into account and providing more accurate estimates, the ISR model should facilitate further research on a number of topics in the areas of political representation and political behavior.
Acknowledgements
I have received many useful comments at various stages of this project, and I would like to thank Elin Haugsgjerd Allern, Pablo Barberá, Elias Dinas, Zoltán Fazekas, Bjørn Høyland, Carl Henrik Knutsen, Liam McGrath, Till Weber, Tore Wig, and the anonymous reviewers. I would also like to thank the Stan Development Team for their excellent support and advice. All remaining errors are my own.
Data Availability Statement
Replication materials are available on the Political Analysis Dataverse (Bølstad Reference Bølstad2019). In order to facilitate future use, the replication folder includes a brief guide on how to use the provided functions.
Supplementary material
For supplementary material accompanying this paper, please visit https://doi.org/10.1017/pan.2019.42.