Chattanooga Times Free Press

Why polls struggle to get it right

- BY DANIEL ALEXANDER

I am a professor of mathematic­s, so my ears perk up when I hear someone say that polls seem inaccurate.

The public understand­ably focuses on polling results and how much those results seem to vary.

Take two presidenti­al approval polls from March 21. Polling firm Rasmussen Reports found that

50 percent of Americans approve of President Donald Trump’s performanc­e, while, that same day,

Gallup stated only 37 percent do.

In late February, the website FiveThirty­Eight listed 18 other presidenti­al approval polls in which Trump’s approval ratings ranged from 39 percent to 55 percent.

Some of those pollsters queried likely voters, some registered voters and other adults, regardless of their voting status. Almost half of the polls relied on phone calls, another half on online polling and a few used a mix of the two.

Further complicati­ng matters, it’s not entirely clear how calling cellphones or landlines affects a poll’s results.

Each of those choices has a consequenc­e, and the range of results attests to the degree that the choices can influence results.

Polling is what mathematic­ians might call a “black art,” a tongue-in-cheek way of saying it does not have the precision of pure mathematic­s.

In my view, the popular sense that polls are inaccurate stems not from poor polling practices, but from assumption­s that both pollsters and the public make. For polls to be more useful to consumers, we need to understand their limitation­s.

The practice of polling and how results are communicat­ed could be improved to build better trust with

Polling is what mathematic­ians might call a “black art,” a tonguein-cheek way of saying it does not have the precision of pure mathematic­s.

Like many of you, I had followed FiveThirty­Eight’s projection­s throughout the election. The site used hundreds of state presidenti­al preference polls to model the election’s outcome. Its poll-based projection­s have a stellar track record: Between the 2008 and 2012 presidenti­al elections, FiveThirty­Eight correctly forecast the victor in every state but one, as well as Washington, D.C.

While FiveThirty­Eight’s final projection­s assigned a 71 to 72 percent probabilit­y to Hillary Clinton’s victory, it wasn’t as bullish on her chances as other poll-based models. The New York Times model gave Clinton an 85 percent chance of winning. The Princeton Election Consortium put Clinton’s probabilit­y of victory at greater than 99 percent.

Trump’s “surprise” victory led many to wonder how the polls got things so wrong.

At the national level, however, the polls did get it right. The final average of national polls at RealClearP­olitics had Hillary Clinton ahead by 3.2 percent nationally. Clinton won the popular vote by roughly 2.1 percent, well within the margin of error.

The presidenti­al election is not decided by national votes, but rather at the state level. If the polls did err, it was in a handful of electorall­y important states. The majority of the poll-based models listed on The New York Times site, including those of FiveThirty­Eight and the Princeton Election Consortium, projected that Clinton would win the pivotal states of Florida, Michigan, North Carolina, Pennsylvan­ia and Wisconsin. Most polls in these states put Clinton ahead as well.

Had we more closely attended to the implicatio­ns of the margin of error, we would perhaps not have reacted with shock as the results came in.

THE MARGIN OF ERROR

Every poll has a margin of error. The margin of error means that the true number is not necessaril­y the reported result, but is within a given range.

Pollsters include a margin of error because they are polling a tiny sample of the voting public. While pollsters do an excellent job of making sure their sample is representa­tive of the voting public, it is rarely a perfect mirror, so there is inevitably error.

In other words, true support for a candidate could fall anywhere within a given range of the poll’s results.

For example, the Democratic polling firm Public Policy Polling sampled 957 likely Michigan voters over two days in November, placing Clinton in the lead over Trump, 46 percent to 41 percent. The poll listed a 3.2 percent margin of error.

Rather than a simple total, the polls provide a range of outcomes. The margin of error implied Clinton’s support level was between 42.8 and 49.2 percent — that is, 46 percent plus or minus the margin of error. Trump’s lay between 37.8 and 44.2 percent.

It is possible that both candidates’ votes could be in the overlap of their respective ranges. It is that scenarios exist where Trump is ahead in Michigan. Most November polls in Florida, Michigan, North Carolina, Pennsylvan­ia and Wisconsin had Clinton ahead. In almost every case, the final results fell within a poll’s margins of error.

HIDDEN ASSUMPTION­S

Polling is limited because pollsters make assumption­s, including assumption­s about likely voters and demographi­cs. Out of necessity, those assumption­s are based on voting patterns from past elections.

Pollsters need to project with a great deal of precision the final voting percentage of each of the subpopulat­ions that compose the electorate. Since polling occurs before an election, it is no easy task to predict, for example, how many white working-class men will vote. Likewise, it is extremely difficult to know the degree to which the prospect of electing a black president drew African-Americans to the polls in 2008 and 2012. Pollsters have to make assumption­s about those kinds of things, and each assumption introduces potential error.

The different assumption­s polls make about their samples helps explain the broad range of results we saw in the Trump approval ratings.

It also may help explain why, during the election, Trump outperform­ed the polls in battlegrou­nd states. His support was high among white working-class voters, who evidently came to the polls in greater numbers than expected. Clinton hoped that black voters would turn out close to how they did in 2008 or 2012, which did not occur. Trump’s margin of victory in the pivotal states of Michigan, Pennsylvan­ia and Wisconsin was roughly 77,000 votes out of 15 million cast.

Very slight changes in demographi­c assumption­s could have accounted for those 77,000 votes and resulted in polls that put Trump ahead.

HEDGING YOUR BETS

There are ways to hedge against error. Baseball teams like the Chicago Cubs and the Boston Red Sox mix sophistica­ted analytics with an “eye test”: that is, the input of old baseball hands who rely on observatio­n and feel rather than pure numbers.

Much in the same way, pollsters and modelers could try mixing in human elements. For example, to find out more about personalit­y traits that might impact electabili­ty, the Cook Political Report incorporat­es personal interviews with candidates into their projection­s of House races. Pollsters could also try to gauge the enthusiasm of a presidenti­al candidate’s supporters by measuring social media activity or public signs.

Another way to improve a poll’s accuracy would be to offer multiple demographi­c models. For example, Public Policy Polling could have used three different models for its Michigan poll, each based on different demographi­c assumption­s. One might assume black turnout as being the same as the previous presidenti­al election; the second could assume a slightly greater turnout; and the third a smaller one. While these kinds of results might resist easy reduction to a headline, they would provide a richer range of possibilit­ies and perhaps fewer surprises.

Some poll-based models hedge against error by considerin­g other factors, such as their own demographi­c analyses, incumbent approval ratings and economic indicators. Stating their results as a probabilit­y also serves to highlight the uncertaint­y involved. But they are still based on polls.

An apt analogy is another way to hedge. On the morning of the election, The New York Times observed that Clinton’s chances of winning were roughly the same that a profession­al field goal kicker will make a 37-yard field goal.

But even the best kickers sometimes miss.

Daniel Alexander is a professor of mathematic­s and associate chairman of the mathematic­s department at Drake University.

Newspapers in English

Newspapers from United States