|
Whew, I'm going to try this in the hope of helping. If not, then you can always delete it! ----------------------------------------------------- What do "statistical probabilities tell us?"
Usually, the odds that something happened so that we can make an informed decision.
If I had 10 people in a room, and everyone put in a check for all they had in the bank, and we drew one name to receive all the money - would you participate?
The odds are easy to compute IF we ASSUME some things. Each person gets their name in the hat ONCE. We will only draw a name out ONCE. Given the assumptions, your odds are 1 in 10 that you will win the hat of checks in one drawing.
Would you play? That depends on IF you are overdrawn or your account includes money that Uncle B. Gates just left you ($100,000,000). The ODDS, even if they meet the assumptions and are calculated precisely don't change what you think is important...they just inform the decision.
Let's say that we could not compute the exact odds because we didn't meet the assumptions: maybe you had the possibility of putting your name in TWICE or maybe there were THREE drawings and the first two names were left out, but the THIRD name won the money. Even if you weren't exactly sure what the odds of winning were, your decision to play is probably as dependent on the value to you as much as the exact odds! ------------------------------------------------------ What about TIA, Febble, OTOH, and similar debates?
More complicated statistical methods have lots of fancy assumptions similar to the hat example. In actual social science research, very few "quasi-experimental" designs (polls where the data comes from those who chose to participate) and lower quality measures ("Do you agree" questions which are not interval level like a ruler with inch marks) virtually NEVER meet ALL the assumptions. Fortunately, the exact computation of the odds is often the best we can do, but we think it is good enough (called robustness) to make the informed judgment with confidence. TIA depends somewhat on robustness when he claims the "law of large numbers" or "central limit theorem".
IF a conservative analysis (Feeble indicates that the evidence is 12 to 1 that pre-election polls differed from the actual election more than chance) or liberal analysis (TIA says the odds are 65,000,000 to 1) use different assumptions that are actually unknown, THEN it may sway your level of confidence. On the other hand, IF you don't think that there should be ANY difference that could ever occur that indicates stealing an election, then even the lower odds are good enough to raise hell. It depends on your values. If Febble or OTOH want to wait until there is absolute assumption compliance, likely it will never happen from the methodology used in surveys and polls. They may get closer with more sophisticated (powerful) techniques (meta analysis for effect sizes, SEM's, and multivariate correlations), but even then, it's throwing out the baby with the bath to expect social science research to follow all the conventions of laboratory math - and Febble correctly suggests this on many occasions! She is taking a "conservative" approach to the math, but admits that some leeway is granted to logical experience and observation. ------------------------------------------------------- What can be done to meet the "assumptions" and use polls to indicate problems in the actual votes IF they are there?
The best way (in Sancho's opinion) is not to argue about which assumptions of which statistical technique are "met"? The control over the questions asked and the sampling designs are in the hands of the various pollsters. The pollsters COULD help improve the process by increasing data collection for key elections, asking the most valid questions on pre-election polls and exit polls, and sharing profusely. They harm the process by being secretive, using manipulations that appear to meet assumptions that they really didn't, and failing to broadcast the changes in the process that would improve the accuracy. IF there appears to be an issue with poll data, the pollsters should try to fix it, tell us what they intend to do to improve the process, and see if it works. Maybe the pollsters say they are doing that, but it doesn't seem like a sincere effort to me at this point when you lock yourself up and don't over sample in Florida's District 13 in 2006!
OTOH (pun intended), those performing analysis would benefit from keeping it simple and avoiding the debates of "sophisticated" theories that are hard to confirm (reluctant responder or gender-based interviewing or vague questions). Those performing analysis may want to be conservative on the probabilities they claim, but make a serious effort to describe up front what the observed discrepancy indicates and why it logically shows something that could not be happening by chance. If an analysis is supposed to meet some assumptions, they can report it, but most people don't care. Even the pure statisticians realize that often the assumptions are violated and there is little we can do about it. We do know that fancy attempts to guess what would have happened if there was a perfect analysis rarely work well. I think EDA's report is closer to this style than TIA's, but there's always room for improvement.
We can set up our own DU polls, but often the infrastructure and experience isn't there, so it might be easier to convince the pollsters to do MORE than describe why people voted, but also help out with the evidence that the election was consistent with the reported outcome! Then assumptions would not matter, because the evidence would be likely become overwhelmingly obvious.
------------------------------------------------------ What do we know now about the elections because of the debates over polls and assumptions?
Even though the exact odds and mathematical assumptions have not been available, we've had 6 years of one analysis after another that suggest a difference between various polls and the posted election results. If the odds were computed that you would win the hat game 1 out of 10 times, but you weren't sure if those odds were precise; and you played 200 times but you never won, you may wonder IF it mattered what the odds were...it's time to quit playing this game!
It's not that TIA or EDA have met the "assumptions" that demonstrate a particular confidence...it's the fact that many polls, precincts, and questions asked on the polls seem inconsistent with the actual election in many unlikely ways - seeming to favor a particular direction in systematic patterns, or outside of any possible expected error in others. Febble and I also agree that going after the obvious problems would be a good use of time and energy. In many cases, TIA or EDA type of reports help focus on important targets to investigate.
If you are a pure statistician at heart, then jump in there and inform others how to "fix" the assumptions WITHOUT a process change by the pollsters. If it can't be "fixed" to satisfy assumptions from today's available data, then direct the email at the pollsters to do better, or let's find a poll process for 2008 that we can rely on!
|