[acb-hsp] How to Make People Tell the Truth

peter altschul paltschul at centurytel.net
Thu Apr 28 05:20:10 GMT 2011


How To Make People Tell The Truth
  BY Gregory Ferenstein Today
  IY survey platforms make constructing questionnaires easy, but 
the results could be biased, contradictory, or deeply misleading.



Online surveys often have to compete for attention against the 
backdrop of Netflix, Gmail alerts, and 25 open browser tabs.  The 
minimal cognitive effort given to answering questions may 
exacerbate all the problems that lead to biased or outright 
distorted results.
  As Facebook adds polling features and SurveyMonkey acquires 
popular document form builder, Wufoo, the proliferation of 
amateur surveying has a big future.  So, we asked a survey expert 
at the famous University of Michigan Institute for Social 
Research, Professor Michael Traugott, about how to make 
questionnaires that get at precisely the data we're digging for.
  Leading Questions Perhaps the biggest no-no that surveys 
violate is poor wording.  Minor adjustments in questions can 
often produce enormous differences.  For example, one study found 
that for the question "Should divorce in this country be easier 
to obtain, more difficult to obtain, or stay as it is now?" 
placing "more difficult" at the end of question caused a 11% 
difference in responses.
  In another study, for the question "Do you think the United 
States should forbid public speeches against democracy?" 
replacing the world "forbid" with "allow" caused a 26% increase 
in respondents' support for free speech (because individuals, on 
average, have an aversion to forbidding rights).  In other words, 
because respondents don't take the time to think about the 
substance of a question, wording matters.
  One way to get around bias, says Traugott, is to ask balanced 
wording.  "Some people think A, while some people think B, how 
about you?" Be as explicit as possible about the wide range of 
beliefs that exist--or, an individual's sheep-like proclivities 
kick in.
  Second, try not to attach an authority's name to a question, 
such as, "The Supreme court recently decidedX, do you agree?" 
Individuals, especially lazy ones happy to pass off the heavy 
thinking to someone else, will give extra weight to authority 
figures who may know more than they do.
  Pre-test, Pre-test, Pre-test Traugott says its a mistake for 
people to believe "that they can write these questions and get 
them correct the first time." Testing out a survey on a few close 
friends may reveal enormous gaps in understanding.  More 
sophisticated pretesting may require iteratively improving the 
question on different sources until a string of unique testers 
gives the same interpretation to the same question.  If that's 
too cumbersome, asking someone in a nearby cubical or over 
Facebook chat may still lead to big improvements.
  In pre-testing, one of the red flags to look out for is 
response categories that don't allow respondents to answer how 
they truly feel.  For instance, Traugott recommends adding a 
"Don't Know" option if the question relates to an issue for which 
a concrete opinion doesn't exist.  When a respondent agrees to an 
interview they may feel a sort of "social contract" to answer 
questions, "even if they haven't thought very much about it." If 
a pre-test is done correctly, a respondent who does not have a 
solid opinion will tell you so and the questionnaire can be 
adapted accordingly.
  Stand On The Shoulders Of (Survey) Giants Fortunately, you 
don't have to be a PhD in qualitative sociology to make a good 
survey; Traugott recommends looking at survey archives to scout 
out what others have tried.  Specifically, the iPoll and Pew 
Research Center have solid archives.  For more academic 
questions, or more indepth organizational, social, or political 
research, try the Interuniversity Consortium for Political and 
Social Research (ICPSR).
  Hypothesize Through Questions Ultimately, all questions begin 
with a hypothesis about the world.  Pollsters ask about President 
Obama's approval ratings after the State of the Union because 
they suspect an eloquent speech might boost his likability among 
conservatives.  A manager may ask workers if they enjoy their job 
because he or she fears low organizational moraleddTherefore, 
Taugott recommends adding in questions that unearth the cause of 
an answer.  Pollsters should ask which party a respondent is 
affiliated with; a manager might ask a worker how long they've 
been at their job.  Without these additional variables, we're 
left in the dark, unable to prove why the results turned out a 
certain way.
  Copyright Ággc) 2011 Mansueto Ventures LLC.  All rights 
reserved.



More information about the acb-hsp mailing list