Published: Aug. 27, 2016, noon
P-values are universal, but do we really know what they mean? In this episode, Dan and James discuss a recent paper describing the failure to correctly interpret p-values in a sample of academic psychologists.
\n\n
Some of the topics discussed:
\n\n
\n- Common p-value misconceptions
\n- James tests Dan on his p-value knowledge
\n- p-values vs. effect size
\n- The problem of sample size with p-value interpretation
\n- The Facebook mood manipulation study
\n- Data peeking
\n- Equivalent p-values do not represent equivalent results
\n- Meta-analytical thinking
\n- Using significance as a categorical factor
\n- Statistical vs. clinical significance
\n- Clinical trial registration and 'secondary outcome creep'
\n- Dan and James answer listener questions
\n- Science communicator vs. scientist
\n- Grant titles and the 'Pub test'
\n- NASA and social media
\n
\n\n
Links
\n\n
\n\n
http://journal.frontiersin.org/article/10.3389/fpsyg.2016.01247/full
\n\n
\n- Geoff Cumming's book (we got the name completely wrong - sorry Geoff!)
\n
\n\n
http://www.amazon.com/Understanding-The-New-Statistics-Meta-Analysis-ebook/dp/B007M9D76G/ref=pd_sim_kstore_1?ie=UTF8&refRID=1QWKES82EP85DBAEKNT1
\n\n
The story on research passing the 'pub' test
\n\n
https://theconversation.com/if-youre-going-to-ridicule-research-do-your-homework-64238
\n\n
Real scientists
\n\n
http://realscientists.org
\n\n
Facebook page
\n\n
https://www.facebook.com/everythinghertzpodcast/
\n\n
Twitter account
\n\n
https://www.twitter.com
Support Everything Hertz