Daily Mail-o-matic
Daily Mail-o-matic
A new Daily Mail headline every time you click the button.
Election Special For a limited time only, every headline is about Nick Clegg.
COULD NICK CLEGG DESTROY YOUR MORTGAGE?
General notes of interest and moans
Daily Mail-o-matic
A new Daily Mail headline every time you click the button.
Election Special For a limited time only, every headline is about Nick Clegg.
COULD NICK CLEGG DESTROY YOUR MORTGAGE?
"multicd.sh is a shell script designed to build a multiboot CD image containing many different Linux distributions and/or utilities."
Interesting interactive plot of how different constituencies will change depending on the swing in votes (to a maximum of 10%). With a 10% swing in any direction there is no chance that my MP will be from a different party to the current one.
With the general election 2010 campaign well and truly underway, it's easy for the key facts to get lost in a barrage of propaganda, counter-accusations and obsfucation, as Labour, the Conservatives and Liberal Democrats battle for the key marginals.
This election will see hundreds of new constituencies as boundary changes come into effect. These figures, compiled by the Press Association, identify every new constituency, and the votes needed to win it.
How much pride do you have in your area? These numbers measure community involvement on every level.
With cutting public spending a key election issue, these are the most comprehensive figures, department by department.
Straight after the budget, government departments announced sometimes swinging cuts. Find out which ones are the losers
Full list of every MP, complete with constituency IDs and expenses claims. See how they add up.
Have things got better or worse under Labour? See what the data says.
BNP membership, seat by seat.
For most, it's not very much. Get the full data and see for yourself.
Find out which newspapers supported which parties - and what it did to the outcome.
See how it got this big and what could bring it down.
"
In ScienceNews this month, there's controversial article exposing the fact that results claimed to be "statistically significant" in scientific articles aren't always what they're cracked up to be. The article -- titled "Odds Are, It's Wrong" is interesting, but I take a bit of an issue with the sub-headline, "Science fails to face the shortcomings of Statistics". As it happens, the examples in the article are mostly cases of scientists behaving badly and abusing statistical techniques and results:
- Authors abusing P-vales to conflate statistical significance with practical significance. A for example, a drug may uncritically be described as "significantly" reducing the risk of some outcome, but the the actual scale of the statistically significant difference is so small that is has no real clinical implication.
- Not accounting for multiple comparisons biases. By definition, a test "significant at the 95% level" has 5% chance of having occurred by random chance alone. Do enough tests, and you'll find some indeed indicate significant differences -- but there will be some fluke events in that batch. There are so many studies, experiments and tests being done today (oftentimes, all in the same paper)that the "false discovery rate" maybe higher than we think -- especially given that most nonsignificant results go unreported.
Statisticians, in general, are aware of these problems and have offered solutions: there's a vast field of literature on multiple comparisons tests, reporting bias, and alternatives (such as Bayesian methods) to P-value tests. But more often than not, these "arcane" issues (which are actually part of any statistical training) go ignored in scientific journals. You don't need to be a cynic to understand the motives of the authors for doing so -- hey, a publication is a publication, right? -- but the cooperation of the peer reviewers and editorial boards is disturbing.
ScienceNews: Odds Are, It's Wrong