Printer Friendly

Polls and elections: preelection poll accuracy in the 2008 general elections.

Voters in the 2008 presidential election decidedly endorsed Democrat Barack Obama over his Republican opponent John McCain in the first-ever contest between two sitting U.S. senators. Amid growing economic turmoil and frustration with the war in Iraq and with two-term incumbent Republican president George W. Bush, Obama bested McCain at the polls by nearly 10 million votes. The historic election attracted 61.6% of the eligible electorate to the polls on election day--the highest turnout since 1968--and produced the first African American president of the United States (McDonald 2009). Obama ultimately captured 52.9% of the popular vote, while McCain received 45.7%. The Democratic nominee carried 28 states plus the District of Columbia and prevailed in Nebraska's Second Congressional District to garner 365 electoral votes; McCain earned 173 electoral votes from victories in 22 states. (1)

The 2008 election was the first election in 56 years that did not include an incumbent president or vice president. Both parties held spirited primaries that attracted a broad range of candidates, but the Democratic primary race that pitted former first lady and U.S. senator Hillary Clinton, the early front-runner, against Barack Obama was especially vivacious. The 17-month contest remained close throughout, and Obama did not secure enough delegates to clinch the nomination until June 3, 2008.

Despite noteworthy preelection polling setbacks during the primaries (most pollsters predicted an Obama victory in New Hampshire, for example, but Clinton won) (Liss 2008), the final national preelection polls in the general election unanimously projected a Democratic victory. Estimates of Obama's margin over McCain were among the most precise on record. In this essay, I analyze national and subnational preelection polls conducted in the 2008 cycle to present an in-depth evaluation of poll accuracy.

Preelection Poll Volume and Poll Aggregation in 2008

The number of preelection polls conducted during a typical presidential election cycle has grown substantially over the past few decades. Michael Traugott reports that the "explosion" in polls started in the 1980s and that the occurrence of the standard trial heat items in polls increased about 900% between 1984 and 2000 (2005, 644). The overall volume of preelection polling conducted during the 2008 cycle remained high. Based on national samples reported on Pollster.com, 221 non-overlapping presidential trial heats were conducted between Labor Day (September 1) and election day 2008, substantially more than in 2004 (although fewer than in the 2000 cycle). (2) Several polling organizations conducted daily tracking polls in 2008; Gallup and Rasmussen tracked preferences daily for at least six months leading up to the November election. In addition, DailyKos, George Washington University Battleground, Hotline, and Zogby provided daily tracking estimates during the fall campaign, and two additional organizations, IBD/TIPP and ABC News/Washington Post, reported daily tracking poll data toward the end of the campaign. A total of 975 trial heat items in statewide polls fielded between Labor Day and election day probed presidential preferences in 2008, while 394 trial heats reported statewide preferences for U.S. Senate candidates and 110 for gubernatorial candidates over the same duration.

One of the main developments in terms of preelection polling in 2008 was the emergence of online poll aggregators. Web sites such as FiveThirtyEight.com, Pollster.com, the Princeton Election Consortium, and RealClearPolitics.com collected, analyzed, and graphed an unprecedented amount of state and national polling data and helped voters interpret and assess campaign dynamics by providing singular, statistical measures that captured all of the available information daily. Poll aggregators used different mathematical models to contextualize polling information and to refine the predictive capabilities of available polling data (Harmanci 2008). Frequent visits to poll aggregator Web sites became a popular way to monitor preference dynamics in the 2008 cycle; in late-October 2008, Nate Silver, founder of FiveThirtyEight.com, reported that the Web site was receiving more than 600,000 hits per day (Harmanci 2008). RealClearPolitics.com, the most visited site in 2008, reportedly received more than 140 million page views in September alone (Becker 2008).

National Preelection Polls in 2008

National estimates of presidential preferences attracted perhaps the most attention consistently over the course of the 2008 campaign. Here, I assess how well these polls projected the eventual election outcome. Table 1 presents data on 20 final preelection estimates of the 2008 presidential vote based on national samples. Overall, pollsters fared quite well in the 2008 general election cycle. Obama's actual margin over McCain in the popular vote was 7.2 percentage points. All of the final poll estimates were within the range of [+ or -] 4 percentage points of the actual electoral result, and all showed Obama in the lead.

Several measures have been proposed in the literature to assess preelection poll accuracy. Two commonly used measures were developed by Frederick Mosteller and colleagues (1949). Mosteller's Measure 3 is the average absolute difference between the poll estimate for each of the leading candidates in the final estimate, while Mosteller's Measure 5 is the absolute value of the difference between the margin separating the two leading candidates in the poll and the difference in their margins in the actual vote (Traugott 2001, 2005).

An alternative method for assessing poll accuracy was developed by Elizabeth Martin, Michael Traugott, and Courtney Kennedy (2005). The new measure of predictive accuracy (A) is based on the natural logarithm of the odds ratio of the outcome in a poll and the actual election outcome (see Martin, Traugott, and Kennedy 2005 for a complete description). Among several advantages associated with this measure is the ability to compare accuracy across elections and polling firms and to detect the direction of bias because a signed statistic is produced (not an absolute value). A positive sign indicates a pro-Republican bias, while a negative sign indicates a pro-Democratic bias (Traugott 2005). (3)

Table 1 presents the values for the three measures for each of the final national 2008 polls that I evaluate. The average value for the Mosteller Measure 3 is 1.53 for the 20 polls included in the analysis, while the average value for the Mosteller Measure 5 is 1.55. Table 2 helps situate poll accuracy in 2008 in historical context by presenting summaries of Mosteller's Measures 3 and 5 for elections since 1956 (see Traugott 2005). The evidence reveals that the 2008 polls overall performed better than average based on the Mosteller Measure 3 (the average error for the 1956-2004 period was 1.9). Using the Mosteller Measure 5, preelection polls in 2008 were more accurate, on average, than in any presidential election cycle since 1956.

Although there is no comparable time series of values for the measure of predictive accuracy developed by Martin, Traugott, and Kennedy (2005), the authors computed the average values of the statistic for 1948, 1996, 2000, and 2004. They report that the average value of A for the final preelection polls conducted in 1996 was -0.0838, suggesting a slight Democratic bias that overestimated Bill Clinton's margin over Bob Dole. In 2000, polls overestimated support for George W. Bush, and the average value for A was 0.0630. For the 2004 election, the average value of A was -0.024, suggesting that Bush's electoral margin of victory was slightly underestimated (see Traugott 2005). Using the 20 polls that I analyze, the average value of A in 2008 was -0.013, indicating that polls reflected a slight Democratic bias that overestimated Obama's margin over McCain. Assuming a tied election, our estimate of bias implies that polls favored Obama by about one-third of a percentage point on average. (4) However, the standard error associated with the mean value for A that I report for the full sample of national polls is 0.009, indicating that the bias overall is not statistically significant; thus, the final national polls as a whole were not significantly biased in 2008. Moreover, based on this measure of poll accuracy, the average value of A was smaller in 2008 than in any of the preceding cycles for which estimates of the measure are available, suggesting that polls were more precise in 2008 than in 1996, 2000, and 2004. In fact, the gain in precision over the 2004 cycle was considerable; the average value of A in 2008 was about 50% smaller than in 2004.

Statewide Preelection Polls in 2008

As I noted earlier, state-level preelection polling was widespread across the nation during the 2008 cycle, with pollsters assessing statewide preferences for presidential as well as U.S. Senate and gubernatorial candidates. Following the elections, the National Council on Public Polls (NCPP) compiled and analyzed a compendium of 507 final state-level preelection polls conducted after October 15, 2008 (see NCPP 2008 for details and a complete list of polls included). The NCPP reported that most state polls (53.3%) were conducted by telephone using live interviewers, while 28.2% were conducted using Interactive Voice Response (IVR), 18.3% were Internet polls, and one poll was conducted by mail (NCPP 2008).

I use the complete set of polls included in the NCPP report to assess predictive accuracy in statewide polls in 2008. Using each poll as a single (unweighted) observation, I present the frequency distribution of A in Figure 1. I note that the polls include estimates of support for presidential as well as other statewide candidates (U.S. Senate and governor). In the absence of overall bias, I would expect the distribution to be centered on zero. The mean value of A in the complete sample of polls is -0.002, suggesting a negligible bias favoring Democratic candidates in statewide polls overall, but the bias is not statistically significant (standard error = .005). Assuming all races were perfectly tied, this would translate into a difference of merely 0.05 percentage point (see footnote 2).

The measure of predictive accuracy (A) developed by Martin, Traugott, and Kennedy (2005) permits us to compare accuracy across a range of poll characteristics. Table 3 presents mean levels of (A) and the corresponding standard errors for statewide polls grouped by a variety of characteristics, including election type, survey mode, sample type, interviewing period, and sponsor. I begin by comparing the poll performance of 11 individual polling organizations that conducted at least 10 statewide polls and that polled in multiple (more than three) states. (5) The evidence presented in Table 3 for each organization reveals that slight biases in a Democratic direction can be detected for statewide polls conducted by four organizations, while pro-GOP bias appears for the remaining seven polling entities. The standard errors associated with these estimates indicate that the biases are statistically insignificant across polling organizations, with one exception: polls conducted by Research 2000 in collaboration with the Democrati-cleaning blog DailyKos exhibit a significant bias in favor of Democratic candidates.

[FIGURE 1 OMITTED]

Our sample includes polls conducted by partisan as well as nonpartisan organizations. Mean values of A for statewide preelection polls conducted by Democratic and Republican organizations and for all nonpartisan polling firms combined are presented in Table 3. I detect hints of bias among partisan polls in 2008 in favor of the respective party's candidates, but only Republican polls are significantly biased toward GOP candidates in statewide polls. The pro-Democratic bias among the 59 Democratic organizations included in our sample (the mean value for A is -0.030) implies an average bias of 0.75 percentage point in favor of Democratic contenders in tied races, but the bias is not statistically significant (standard error = .019). By contrast, among the 18 polls that I analyze conducted by Republican polling organizations, the mean value for A is 0.042. The pro-Republican bias that I detect in polls conducted by Republican organizations is statistically significant (standard error = .016). In perfectly tied races, our estimate implies that Republican polls overestimated support for GOP candidates by 1.05 percentage points on average, relative to election day results. While this advantage is admittedly quite modest, biases of this magnitude can be important in close races. On the whole, nonpartisan polls reflect virtually no bias in 2008; the mean value for A is -0.0003, suggesting that nonpartisan polls overstated support for Democratic candidates by about 0.01 percentage point on average (given a tie), but this bias is, as expected, statistically insignificant (standard error = .006).

Using A, I can investigate bias in subsamples of polls by the level of electoral contests. In statewide presidential polls (N = 327), the mean value of A is -0.004, suggesting a slight but statistically insignificant (standard error = .006) bias favoring Obama. Similarly, statewide polls in U.S. Senate races (N = 145) reveal a slight but statistically insignificant pro-Democratic bias; the mean value of A is a mere -0.001 (standard error = .012). Statewide gubernatorial polls (N = 35) appear to reflect a modest bias in favor of Republican candidates--the mean value for A is 0.011--but the bias is also statistically insignificant (standard error = .032). I conclude from these results that overall bias was minimal in the final statewide polls conducted in 2008, whether examining statewide presidential, U.S. Senate, or gubernatorial preelection polls.

I turn next to examining overall bias by poll mode. The key results are presented in Table 3. Mean values of A across modes suggest that polls conducted via the Internet reflect the highest degree of bias. Among the 93 Internet polls, the mean value for A is 0.039, suggesting a statistically significant bias in the Republican direction (standard error = .014). IVR polls (N = 143) reveal the lowest degree of overall bias (0.018), but these polls appear to have significantly favored Republican candidates as well (standard error = .008). By contrast, the mean value of A is -0.027 (standard error = .008) for statewide polls conducted by telephone (N = 270), implying a significant bias in a Democratic direction.

There has been considerable debate in recent election cycles about the range of procedures employed by polling organizations in their estimations of likely voters in preelection polls (Erikson, Panagopoulos, and Wlezien 2004; Martin, Traugott, and Kennedy 2005). Table 3 displays mean levels of predictive accuracy (A) for three survey sample types that were included in the overall pool of 2008 statewide polls: samples of likely voters, registered voters, and adult populations. The initial evidence that I present suggests that likely voter samples were indeed more accurate than samples of registered voters or adults in 2008. The directions of the biases suggested by the estimates indicate pro-GOP bias only for registered voter samples, but the biases are statistically insignificant across sample types.

[FIGURE 2 OMITTED]

Longitudinal analysis can also be used to analyze the dynamics of poll accuracy, relative to electoral outcomes, and to gauge whether the predictive capacity of preelection polls improves as election day approaches. Scholarly evidence about the relationship between poll timing and accuracy is mixed; while some studies find that accuracy improves over the course of a campaign (Crespi 1988), others find no significant impact of poll timing on accuracy (Lau 1994; Martin, Traugott, and Kennedy 2005). Figure 2 presents lowess-smoothed levels of overall bias and accuracy in the 2008 statewide polls by the number of days until election day. The solid line presents the smoothed pattern of the absolute value of A over this period and suggests that preelection poll accuracy improved steadily within the final three weeks of the election cycle (the absolute value of A trended toward zero). Accuracy improved most dramatically during the final few days prior to the election. The dashed line in Figure 2 plots lowess-smoothed levels of mean predictive accuracy (A) over the same duration. The pattern suggests that statewide preelection polls initially reflected a pro-Democratic bias (two to three weeks prior to election day), but this bias eroded steadily; within the final few days of the cycle, statewide polls overall increasingly reflected a pro-Republican bias. These initial patterns suggests overall poll accuracy and bias can change over the course of a campaign, particularly during the final campaign period that I examine; substantively, however, these changes may not account for very much. I investigate this matter more rigorously next using multivariate techniques.

To explain overall levels of poll accuracy and bias more rigorously, I conduct a series of multivariate regression analyses. One advantage of A as a measure of predictive accuracy is that it (or variants of A as discussed later) is amenable to explanation using multivariate techniques that utilize various poll attributes as explanatory variables. The results of two such estimations are presented in Table 4. In both regressions, I include controls (fixed effects) for "house" (polling organization) and state effects. In Model 1, the dependent variable is the absolute value of A; as such, higher values represent less accurate poll estimates, relative to election outcomes. The results of the regression analysis reveal that statewide preelection polls for U.S. Senate candidates and gubernatorial candidates were significantly less accurate in 2008, compared to statewide presidential polls (the excluded category); all else being equal, levels of the absolute value of A were higher on average relative to presidential polls. Controlling for other factors, I also find that statewide preelection polls conducted via the Internet in 2008 were significantly less accurate than polls conducted by telephone (the excluded category), while the overall accuracy of mail and IVR polls did not differ significantly from phone polls. The results also suggest that registered voter samples were, in fact, more accurate than likely voters samples (the excluded category), while samples of adults were less accurate than likely voters samples, all else being equal. Once controls for other factors are incorporated, I find no evidence that polls conducted by nonpartisan organization were significantly more accurate overall compared with partisan polls. All else being equal, the analysis reveals poll timing did not significantly impact overall accuracy in statewide polls conducted in 2008.

Model 2 presents the results of a probit regression analysis in which the dependent variable is coded 1 if the preelection poll reflected a pro-Republican bias (A > 0) and 0 if the poll reflected a pro-Democratic bias (A < 0). Using the same poll attributes as in Model 1 to explain whether polls reflected a pro-Republican bias overall, I find that both U.S. Senate and gubernatorial polls were, all else being equal, significantly more likely to be biased in a Republican direction in 2008, compared to presidential polls (the excluded category). Controlling for other poll characteristics, Internet and IVR polls were significantly less likely than polls conducted by phone (the excluded category) to exhibit bias favoring Republican contenders. With controls in place for other factors, samples of registered voters were more likely than likely voter samples to exhibit bias in favor of GOP contenders, all else being equal. Nonpartisan polls overall were not significantly biased, relative to polls conducted by partisan organizations. Similarly, poll timing did not exert a significant impact on overall bias, all else being equal.

The multivariate analyses described here are useful in that they reveal the impact of a range of poll attributes on overall accuracy and the direction of bias while simultaneously controlling for other poll characteristics. In several instances, the results may cause analysts to reconsider, update, or confirm initial conclusions about the impact of various factors on accuracy and bias in preelection polls. For example, the interview period does not appear to influence overall levels of accuracy or bias. Moreover, the multivariate results suggest that preference estimates based on registered voter samples may actually be more accurate than likely voter samples, even as they are prone to pro-Republican bias. Curiously, although the value of multivariate approaches along these lines to explain poll bias and accuracy more rigorously is trumpeted in the literature (Martin, Traugott, and Kennedy 2005), such analyses are infrequently advanced. (6)

Discussion

Unprecedented interest in the 2008 election campaign and in preelection poll estimates of preference dynamics stimulated widespread awareness of persistent and emerging challenges associated with preelection polling and led to a national dialogue about poll quality and polling methodology. Concerns about polling procedures, including the estimation of likely voters, the presence of a so-called Bradley effect, Internet-based or IVR polls, and a variety of other considerations fueled tremendous speculation about eventual poll performance in 2008. The results that I present in this report indicate that, despite these apprehensions, polls across the board performed quite well in 2008. That said, concerns about poll methodologies should not be wholly and readily dismissed. While the evidence suggests that improvements in accuracy and declining bias in preelection polls as a whole, compared to previous election cycles, sources of inaccuracy and bias can also be detected using 2008 polls. Pollsters are wise to devote attention to monitoring and to investigating these sources further in the pursuit of still greater improvements in accuracy and bias.

References

Becker, Bernie. 2008. "Political Polling Sites Are in a Race of Their Own." New York Times, October 27, p. A14.

Crespi, Irving. 1988. Pre-Election Polling: Sources of Accuracy and Error. New York: Russell Sage Foundation.

DeSart, Jay, and Thomas Holbrook. 2003. "Campaigns, Polls and the States: Assessing the Accuracy of Statewide Presidential Trial-Heat Polls." Political Research Quarterly 56 (December): 431-39.

Erikson, Robert, Costas Panagopoulos, and Christopher Wlezien. 2004. "Likely (and Unlikely) Voters and the Assessment of Poll Dynamics." Public Opinion Quarterly 68 (Winter): 588-601.

Harmanci, Reyhan. 2008. "Poll Analysis Sites Put New Spin on Statistics." San Francisco Chronicle, October 18.

Lau, Richard. 1994. "An Analysis of 'Trial-Heat' Polls During the 1992 Presidential Election." Public Opinion Quarterly 58 (Spring): 2-20.

Liss, Sharon. 2008. "Good Night for Clinton But Not So Good for Pollsters, Pundits." Fox News, January 9. http://www.foxnews.com/politics/elections/2008/01/09/ good-night-for-clinton-butnot-so-good-for-pollsters-pundits/[accessed August 17, 2009].

Martin, Elizabeth, Michael Traugott, and Courtney Kennedy. 2005. "A Review and Proposal for a New Measure of Poll Accuracy." Public Opinion Quarterly 69 (Fall): 342-69.

McDonald, Michael P. 2009. "The Return of the Voter: Voter Turnout in the 2008 Presidential Election." The Forum 6 (4). http://www.bepress.com/forum/v016/iss4/art4 [accessed August 17, 2009].

Mosteller, Frederick, Herbert Hyman, Philip McCarthy, Eli Marks, and David Truman. 1949. The Pre-Election Polls of 1948: Report to the Committee on Analysis of Pre-Election Polls and Forecasts. New York: Social Science Research Council.

National Council on Public Polls (NCCP). 2008. "Analysis of Final Presidential Pre-Election Polls, 2008." http://www.ncpp.org/files/NCPP_2008_analysis_of_election_polls_121808 %20pdf_0.pdf [accessed August 17, 2009].

Traugott, Michael. 2005. "The Accuracy of the National Preelection Polls in the 2004 Presidential Election." Public Opinion Quarterly 69 (5): 642-54.

--. 2001. "Assessing Poll Performance in the 2000 Campaign." Public Opinion Quarterly 65 (Fall): 389-419.

COSTAS PANAGOPOULOS

Fordham University

(1.) Nebraska and Maine are the only two states do not adopt the traditional winner-take-all method of electoral vote allocation and permit electoral votes to be split based on candidate vote shares in congressional districts.

(2.) I am grateful to Mark Blumenthal from Pollster.com for providing 2008 data. See Martin, Traugott, and Kennedy (2005) for comparable data for previous cycles. For 2008, we include only non-overlapping poll releases; for tracking polls that reported three-day, rolling averages daily, we include only every third release.

(3.) Polls reflect no bias when A equals zero.

(4.) See Martin, Traugott, and Kennedy (2005, 11 n. 11) for a discussion and formula used to convert the parameter A to a percentage point difference.

(5.) This is consistent with the approach adopted in Martin, Traugott, and Kennedy (2005, 362).

(6.) See Lau (1994) and DeSart and Holbrook (2003) for exceptions analyzing presidential trial heat data.

Costas Panagopoulos is an assistant professor of political science and director of the Center for Electoral Politics and Democracy at Fordham University.
TABLE 1
Final National Preelection Poll Results and Poll
Accuracy Estimates, 2008

 Field
Firm Period McCain Obama Barr

Election Result 45.7 52.9
Democracy Corps (D) 10/30-11/2 44 51 2
FOX News/Opinion Dynamics 11/1-11/2 43 50 --
CNN/Opinion Research 10/30-11/1 4G 53 --
Ipsos/McClatchy 10/30-11/1 4G 53 --
American Research Group 11/1-11/3 45 53
IBD/TIPP 11/1-11/3 44 52 --
Harris Interactive 10/30-11/3 44 52 --
YouGov/Polimetrix 10/18-11/1 45 51 --
Pew 10/29-11/1 46 52 1
Rasmussen 11/1-11/3 46 52 --
NBC News/Wall Street
 Journal 11/1-11/2 43 51 --
GWU (Lake/Tarrance) 11/2-11/3 44 49 --
ABC News/Washington Post 10/30-11/2 44 53 --
Diageo/Hotline 10/31-11/2 45 50 --
DailyKos.com (D)/
 Research 2000 11/1-11/3 46 51 1
Marist College 11/3 43 52 --
CBS News 10/31-11/2 42 51 --
Gallup 10/31-11/2 44 55 --
Reuters/ C-SPAN/Zogby 10/31-11/3 43 54 --
CBS News/New York Times 10/25-10/29 41 52 --
Average

 Predictive
 Obama Accuracy
Firm Nader Other Lead (A)

Election Result 7.2
Democracy Corps (D) 1 1 7 -0.001
FOX News/Opinion Dynamics -- 2 7 -0.005
CNN/Opinion Research -- 7 0.005
Ipsos/McClatchy -- 1 7 0.005
American Research Group 8 -0.017
IBD/TIPP -- 4 8 -0.021
Harris Interactive 1 2 8 -0.021
YouGov/Polimetrix -- 2 6 0.021
Pew 1 6 0.024
Rasmussen -- 6 0.024
NBC News/Wall Street
 Journal -- 2 8 -0.024
GWU (Lake/Tarrance) -- 5 0.039
ABC News/Washington Post -- 2 9 -0.040
Diageo/Hotline -- 5 0.041
DailyKos.com (D)/
 Research 2000 1 0 5 0.043
Marist College -- 3 9 -0.044
CBS News -- 9 -0.048
Gallup -- 11 -0.077
Reuters/ C-SPAN/Zogby -- 11 -0.081
CBS News/New York Times -- 11 -0.091
Average -0.013

 Mosteller Mosteller
Firm Measure 3 Measure 5

Election Result
Democracy Corps (D) 1.80 0.20
FOX News/Opinion Dynamics 2.80 0.20
CNN/Opinion Research 0.20 0.20
Ipsos/McClatchy 0.20 0.20
American Research Group 0.40 0.80
IBD/TIPP 1.30 0.80
Harris Interactive 1.30 0.80
YouGov/Polimetrix 1.30 1.20
Pew 0.60 1.20
Rasmussen 0.60 1.20
NBC News/Wall Street
 Journal 2.30 0.80
GWU (Lake/Tarrance) 2.80 2.20
ABC News/Washington Post 0.90 1.80
Diageo/Hotline 1.80 2.20
DailyKos.com (D)/
 Research 2000 1.10 2.20
Marist College 1.80 1.80
CBS News 2.80 1.80
Gallup 1.90 3.80
Reuters/ C-SPAN/Zogby 1.90 3.80
CBS News/New York Times 2.80 3.80
Average 1.53 1.55

NOTE: To be consistent with previous years' analyses of poll
accuracy, I include poll estimates produced within the final
week of the election. The analysis excludes polls which
completed interviewing prior to October 29, 2008.

TABLE 2
Average Errors in Presidential Polls, 1948-2008

 Morteller Morteller
 Measure 3 Measure 5
 # of # of Average Average
Year Polls Candidate Error (%) Error (%)

2008 20 2 1.5 1.5
2004 19 2 1.7 2.1
2000 19 3 1.7 3.5
1996 9 3 1.7 3.6
1992 6 3 2.2 2.7
1988 5 2 1.5 2.8
1984 6 2 2.4 4.4
1980 4 3 3.0 6.1
1976 3 3 1.5 2.0
1972 3 2 2.0 2.6
1968 2 3 1.3 2.5
1964 2 2 2.7 5.3
1960 1 2 1.0 1.9
1956 1 2 1.8 3.5
Yearly
 Average
1956-2008 1.9 3.2

NOTE: Data for the 1956-2004 period were obtained from
Traugott (2005, 649). The 2008 update was compiled by
author.

TABLE 3
Mean Predictive Accuracy (A) by Poll Characteristics,
2008 Statewide Polls

 Mean
 Predictive
Poll Characteristics Number Accuracy Standard
(Type/Sponsor) of Pools (A) Error

Presidential 327 -.004 .006
U.S. Senate 145 -.001 .120
Governor 35 .011 .032
Democratic 59 -.030 .019
Republican 18 .042 .016
Nonpartisan 430 -.0003 .006
Internet 93 .039 .014
Phone 270 -.027 .008
IVR 143 .018 .008
Likely voters 377 -.007 .006
Registered voters 123 .015 .011
Adults 7 -.041 .065

Sponsor *
AP-GfK 14 -.055 .041
ARG 16 -.030 .027
Allstate/National
 Journal/FD 11 -.011 .019
DailyKos (D)/
 Research 2000 28 -.085 .034
Mason-Dixon 25 .034 .019
Public Policy Polling (D) 30 .018 .015
Rasmussen 44 .016 .014
Research 2000 16 .044 .039
Strategic Vision (R) 15 .031 .016
SurveyUSA 49 .017 .014
YouGov/Polimterix 84 .028 .014

* Consistent with Martin, Traugott, and Kennedy (2005), only
polling organizations that conducted at least 10 statewide
polls and that polled in multiple (three or more) states in
2008 are included in the analysis.

TABLE 4
The Impact of Poll Attributes on Bias and Accuracy in
statewide Preelection Polls, 2008

 Model 2:
 Model 1: Pro-Republican
Independent variables Accuracy Bias
(Poll characteristics)
U.S. Senate .031 *** .448 **
 (.008) .183
Governor .079 *** 1.129 ***
 (.015) (.355)
Internet .114 *** -4.943 ***
 (.035) (1.850)
IVR -.010 -8.266 ***
 (.058) (.962)
Mail -.046 --
 (.075)
Registered voters -.291 *** 6.433 ***
 (.101) (1.435)
Adults .157 ** --
 (.075)
Nonpartisan Polling .012 -1.247
 organization (.041) (1.178)
Days to election .000 -.054
 (.002) (.041)
Constant .127 ** -6.533 ***
 (.056) (2.183)
N 507 384
[R.sup.2]/Pseudo [R.sup.2] .53 .28
Log-likelihood -- -191.787

Notes: Model 1: OLS; dependent variable is the absolute
value of A. Model 2: Probit; dependent variable = 1 if A >
0, and 0 if A < U. Fixed effects for polling organization
("house") and state are included. Observations with
covariate patterns that predict outcome perfectly are
excluded from the model, resulting in the smaller number of
cases.

*** Signifies statistical significance at the p < .01 level,

** p < .05, two-tailed. Standard errors in parentheses.
COPYRIGHT 2009 Center for the Study of the Presidency
No portion of this article can be reproduced without the express written permission from the copyright holder.
Copyright 2009 Gale, Cengage Learning. All rights reserved.

Article Details
Printer friendly Cite/link Email Feedback
Author:Panagopoulos, Costas
Publication:Presidential Studies Quarterly
Article Type:Report
Geographic Code:1USA
Date:Dec 1, 2009
Words:4934
Previous Article:The law: Bush, Cheney, and the separation of powers: a lasting legal legacy?
Next Article:The rhetorical presidency today: how does it stand up?
Topics:


Related Articles
Presidential election polls in 2000: a study in dynamics.
Electronic poll books speed process for workers and voters.
Elderly Voters: Some Improvements in Voting Accessibility from 2000 to 2004 Elections, but Gaps in Policy and Implementation Remain.
Help make democracy work on election day.
131 million: voters put the ballot to the test in a record-setting election.
Race and cell phones challenge pollsters.

Terms of use | Privacy policy | Copyright © 2024 Farlex, Inc. | Feedback | For webmasters |