tag:blogger.com,1999:blog-64398068758813115762024年10月05日 02:14:18 +0000clinical trialsstatisticsRBayesian statisticslying with statisticsbiostatisticsadaptive trialsbloggingSASstatistical programmingOther blogspoliticssafetyFDAgraphicscritical thinkingdata miningeducationblindinggroup sequentialinterpretationleadershipmachine learningmoocpollsskepticismsocial network analysiswasteNNTcareerscausal inferenceclassesconsultingdata collectionepidemiologyp-valuespharmaceutical industryphase 3practice of statisticsprofessionalismrandomizationstatistical significancevaccinationAvandiaBlogkeepingENHANCEInference for RJSMLittle DebateMicrosoft WordPeter RostPharmalotPresentationsPythonRTFTGN1412Type I errorUltraeditalternative medicinealzheimer's diseaseanalyticsbiasbig databookscomparative effectivenesscourseradrug developmentdrugsjoint statistical meetingslearningloessmathematicsmatrixmodelingnatural language processingnoninferiorityobservationalpharmacogenomicsrelationshipssimulationstatistics forumvisualizationANOVAAsk CatoBrexitCETP inhibitorsEMEAEd SilvermanGoogleHal VarianJUPITERJohn TukeyLan-DeMetsMantel-HaenszelMedDRAMicrosoftMicrosoft ExcelPDPKPhase 2Phase IRStudioSAS programmerTaylor seriesWindowsanalysisantibioticsautismbiomarkerscentral nervous systemchaoscompliancecontestscorrelationcrfcritical path initiativedata cleaningdata sciencedatabasesdelta methodechinacaeaeli lillyemailfreegraduate schoolhumorintention to treatlearning from datalinksmarketingmetaanalysismissing datamodern regressionmultiple comparisonsmultiple sclerosisnetworkingneurologyo'brien-flemingoddsopen datapostmarketingprivacyprobabilityproofpublishingrandom number generationrisksciencesimple statisticssmoothingsocial mediasoftwarestatinsstatistical analysis planstatistical leadershipstatistics without borderssurveyst testtablestorcetrapibtwitterwritingRealizations in BiostatisticsBiostatistics, clinical trial design, critical thinking about drugs and healthcare, skepticism, the scientific process.http://realizationsinbiostatistics.blogspot.com/noreply@blogger.com (Anonymous)Blogger242125tag:blogger.com,1999:blog-6439806875881311576.post-12575326670567325732016年12月12日 02:18:00 +00002018年10月23日T10:26:46.809-04:00bloggingopen dataRI set up a new data analysis blogWell, I tried to write a blog post using the <a href="http://www.rstudio.com/" target="_blank">RStudio</a> Rmarkdown system, and utterly failed. Thus, I set up a system where I could write from RStudio. So I set up a Github pages blog at <a href="http://randomjohn.github.io/">randomjohn.github.io</a>. There I can easily write and publish posts involving data analysis.http://realizationsinbiostatistics.blogspot.com/2016/12/i-set-up-new-data-analysis-blog.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-79956752053341017332016年9月24日 21:34:00 +00002016年09月24日T17:34:08.770-04:00MicrosoftRsoftwareWindowsWindows 10 anniversary updates includes a whole Linux layer - this is good news for data scientistsIf you are on Windows 10, no doubt you have heard that Microsoft included the bash shell in its 2016 Windows 10 anniversary update. What you may not know is that this is much, much more than just the bash shell. This is a whole Linux layer that enables you to use Linux tools, and does away with a further layer like Cygwin (which requires a special dll). However, you will only get the bash shell out of the box. To enable the whole Linux layer, follow instructions <a href="http://www.howtogeek.com/249966/how-to-install-and-use-the-linux-bash-shell-on-windows-10/" target="_blank">here</a>. Basically, this involves enabling developer mode then enabling the Linux layer feature. In the process, you will download some further software from the Windows store.<br />
<br />
Why is this big news? To me, this installs a lot of the Linux tools that have proven useful over the years, such as <span style="font-family: "Courier New", Courier, monospace;">wc</span><span style="font-family: inherit;">, <span style="font-family: "Courier New", Courier, monospace;">sed</span>, <span style="font-family: "Courier New", Courier, monospace;">awk</span>, <span style="font-family: "Courier New", Courier, monospace;">grep</span>, and so forth. In some cases, these tools work much better than software packages such as R or SAS, and their power comes in combining these tools through pipes. You also get <span style="font-family: "Courier New", Courier, monospace;">apt-get</span>, which enables you to install and manage packages such as SQLite, octave, and gnuplot. You can even install R through this method, though I don't know if RStudio works with R installed in this way.</span><br />
<br />
If you're a Linux buff who uses Windows, you can probably think of many more things you can do with this. The only drawback is that I haven't tried using any sort of X Windows or other graphical interfaces.http://realizationsinbiostatistics.blogspot.com/2016/09/windows-10-anniversary-updates-includes.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-40935896030651072832016年6月26日 18:24:00 +00002016年09月24日T17:38:49.927-04:00BrexitpoliticsRWhich countries have Regrexit?This doesn't have a lot to do with bio part of biostatistics, but is an interesting data analysis that I just started. In the wake of the Brexit vote, there is a <a href="https://petition.parliament.uk/petitions/131215" target="_blank">petition for a redo</a>. The data for the petition is <a href="https://petition.parliament.uk/petitions/131215.json" target="_blank">here</a>, in JSON format.<br />
<br />
Fortunately, in R, working with JSON data is pretty easy. You can easily download the data from the link and put it into a data frame. I start on that <a href="https://github.com/randomjohn/regrexit" target="_blank">here</a>, with the RJSONIO package, ggplot2, and a version of the petition I downloaded on 6/26/16.<br />
<br />
One question I had was whether all the signers are British. Fortunately, the petition collects the place of residence of the signer, assuming no fraud. I came up with the following top 9 non-UK countries of origin of signers.<br />
<br />
<div class="separator" style="clear: both; text-align: center;">
<a href="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEingXOfWSxmdCEzuYiI_zKZ52RKe4qzyqYOSWga0TxlyJFin5coWlEND8jGpvSjCulG8KyESY9n4Oaexes26IGu4LAp-Gc-19gUn2ctLLtV4XxOH7kVWv6UL30oAqSRH4GTwgvyDcCG38I/s1600/regrexit.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="195" src="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEingXOfWSxmdCEzuYiI_zKZ52RKe4qzyqYOSWga0TxlyJFin5coWlEND8jGpvSjCulG8KyESY9n4Oaexes26IGu4LAp-Gc-19gUn2ctLLtV4XxOH7kVWv6UL30oAqSRH4GTwgvyDcCG38I/s320/regrexit.png" width="320" /></a></div>
<div class="separator" style="clear: both; text-align: left;">
There are a couple of things to remember when interpreting this graph:</div>
<div class="separator" style="clear: both; text-align: left;">
</div>
<ol>
<li>I left off the UK. The number of signatures is over 3 million, and contains by far the largest percentage of signatories.</li>
<li>5 of the 9 top countries are neighbors, including the top 2. The other 4 are Australia, the US, Canada, and New Zealand, who are all countries that have strong ties to the UK.</li>
<li>This assumes no petition fraud, which I can't guarantee. I saw at least one Twitter posting telling people to use her (if the profile pic is to be believed) residence code. There is a section of the petition data showing constituency, so I'm wondering if it would be possible to analyze the petition for fraud. I'm not as familiar with British census data as I am with US, but I imagine a mashup of the two would be useful.</li>
</ol>
(Update: Greg Jefferis posted a nice analysis <a href="http://rpubs.com/jefferis/192311" target="_blank">here</a>. See comments below.)<br />
<br />http://realizationsinbiostatistics.blogspot.com/2016/06/which-countries-have-regrexit.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-78867713193696678432016年6月05日 19:44:00 +00002016年06月05日T15:44:40.374-04:00biostatisticsclinical trialsLittle Debatep-valuesLittle Debate: defining baselineIn an April 30, 2015 <a href="http://www.dartmouth.edu/~bio125/Leek.&.Peng.Nature.2015.p-values.pdf" target="_blank">note</a> in <i>Nature</i> (vol 520, p. 612), Jeffrey Leek and Roger Peng note that <i>p</i>-values get intense scrutiny, while all the decisions that lead up to the <i>p</i>-values get little debate. I wholeheartedly agree, and so I'm creating a <i>Little Debate</i> series to shine some light on these tiny decisions that may not get a lot of press. Yet these tiny decisions can have a big influence on statistical analysis. Because my focus here is mainly biostatistics, most of these ideas will be placed in the setting of clinical trials.<br />
<br />
Defining baseline seems like an easy thing to do, and conceptually it is. Baseline is where you start before some intervention (e.g. treatment, or randomization to treatment or placebo). However, the details of the definition of baseline in a biostatistics setting can get tricky very quickly.<br />
<h4>
The missing baseline</h4>
<div>
Baseline is often defined as the value at a randomization or baseline visit, i.e. the last measurement before the beginning of some treatment or intervention. However, a lot of times things happen - a needle breaks, a machine stops working, or study staff just forget to do procedures or record times. (These are not just hypothetical cases ... these have all happened!) In these cases, we end up with a missing baseline. A missing baseline will make it impossible to determine the effect of an intervention for a given subject.</div>
<div>
<br /></div>
<div>
In this case, we have accepted that we can use previous values, such as those taken during the screening of a subject, as baseline values. This is probably the best we can do under the circumstances. However, I'm unaware of any research on what effect this has on statistical analysis.</div>
<div>
<br /></div>
<div>
To make matters worse, a lot of times people without statistical training or expertise will make these decisions, such as putting down a post-dose value as baseline. Even with good documentation, these sorts of mistakes are not easy to find, and, when they are, they are often found near the end of the study, right when data management and statisticians are trying to produce results, and sometimes after interim analyses.</div>
<h4>
The average baseline</h4>
<div>
Some protocols specify that baseline consists of the average of three repeated measurements. Again, this decision is often made before any statisticians are consulted. The issue with such a statistical analysis is that averages are not easily comparable to raw values. Let's say that a baseline QTc (a measure of how fast the heart charge recovers from a pump, corrected for heart rate) is defined based on 3 electrocardiogram (ECG) measurements. The standard deviation of a raw QTc measurement (i.e. based on one ECG), let's say, is <i>s</i>. The standard deviation of the average of those three (assuming independence) is <i>s</i>/√3, or just above half the standard deviation of the raw ECG. Thus, a change of 1 unit in the average of 3 ECGs is a lot more noteworthy than a change of 1 unit in a single ECG measurement. And yet we compare that to single measurements for the rest of the study.</div>
<div>
<br /></div>
<div>
To make matters worse, if the ECG machine screws up one measurement, then the baseline becomes the average of two. A lot of times we lose that kind of information, and yet analyze the data as if the mystery average is a raw measurement.</div>
<h4>
The extreme baseline</h4>
<div>
In one observational study, the sponsor wanted to use the maximum value over the last 12 months as a baseline. This was problematic for several reasons. Like the average, the extreme baseline (here the maximum) is on a different scale, and even has a different distribution, than the raw measurement. The <a href="https://en.wikipedia.org/wiki/Fisher%E2%80%93Tippett%E2%80%93Gnedenko_theorem" target="_blank">Fisher-Tippett</a> (extreme value) theorem states that the maximum of <i>n</i> values converges to one of three extreme value distributions (Gumbel, Frechet, or Weibull). These distributions are then being compared to, again, single measurements taken after baseline. What's worse, any number of measurements could have been taken for those subjects within 12 months, leading to a major case of shifting sands regarding the distribution of baseline.</div>
<div>
<br /></div>
<div>
Comparing an extreme value with a later singular measurement will lead to an unavoidable case of <a href="https://en.wikipedia.org/wiki/Regression_toward_the_mean" target="_blank">regression to the mean</a>, thus creating an apparent trend in the data where none may exist. Without proper context, this may lead to overly optimistic interpretations of the effect of an intervention, and overly small <i>p</i>-values. (Note that a Bayesian analysis is not immune to the misleading conclusions that might arise from this terrible definition of baseline.)</div>
<h4>
Conclusion</h4>
<div>
The definition of baseline is a "tiny decision" that can have major consequences in a statistical analysis. Yet, the impact of this decision has not been well studied, especially in the context of a clinical trial where a wide range of definitions may be written into a protocol without the expert advice of a statistician. Even a definition that has been well-accepted -- that baseline is the last singular pre-dose value before intervention -- has not been well-studied in the scenario of missing baseline day measurement. Other decisions are often made without considering the impact on analysis, including some that may lead to wrong interpretations.</div>
http://realizationsinbiostatistics.blogspot.com/2016/06/little-debate-defining-baseline.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-35472516704396201092016年5月20日 12:30:00 +00002016年07月16日T14:29:09.456-04:00clinical trialsRsimulationSimulating a Weibull conditional on time-to-event is greater than a given time<div class="separator" style="clear: both; text-align: center;">
<a href="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEglSD3oLfZNQQW_Wit-ELDsvOv4TCMitbjv8vUBieBqjofqRTIMovN0FmG1wdJ3CznTuIM_6cIqB4IIwKHKNsud-Ev36Zap5ZVcJ7ZEDxbGyUfyryazascZNP-hJdofCI2hshndSio5z9Q/s1600/Rplot.png" imageanchor="1" style="clear: right; float: right; margin-bottom: 1em; margin-left: 1em;"><img border="0" height="121" src="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEglSD3oLfZNQQW_Wit-ELDsvOv4TCMitbjv8vUBieBqjofqRTIMovN0FmG1wdJ3CznTuIM_6cIqB4IIwKHKNsud-Ev36Zap5ZVcJ7ZEDxbGyUfyryazascZNP-hJdofCI2hshndSio5z9Q/s200/Rplot.png" width="200" /></a></div>
Recently, I had to simulate a time-to-event of subjects who have been on a study, are still ongoing at the time of a data cut, but who are still at risk of an event (e.g. progressive disease, cardiac event, death). This requires the simulation of a conditional Weibull. To do this, I created the following function:<br />
<br />
<br />
<br />
<br />
<span style="font-family: "courier new" , "courier" , monospace;"># simulate conditional Weibull conditional on survival > T ---------------</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"><br /></span>
<span style="font-family: "courier new" , "courier" , monospace;"># reliability function is exp{-(T+t/b)^a} / exp{-(T/b)^a} = 1-F(t)</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"># n = number of points to return</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"># shape = shape parm of weibull</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"># scale = scale parm of weibull (default 1)</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"># t is minimum (default is 0, which makes the function act like rweibull)</span><br />
<span style="font-family: "courier new" , "courier" , monospace;">my_rweibull <- function(n,shape,scale=1,t=0) {</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"> if (length(t)!=1 && length(t)!=n) {</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"> stop("length(t) is not 1 or n")</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"> }</span><br />
<span style="font-family: "courier new" , "courier" , monospace;"> return(scale*(-log(runif(n))+(t/scale)^shape)^(1/shape))</span><br />
<span style="font-family: "courier new" , "courier" , monospace;">}</span><br />
<div>
<br /></div>
<span style="font-family: "courier new" , "courier" , monospace;"><br /></span>
<span style="font-family: "courier new" , "courier" , monospace;"><br /></span>
You use this function just like <span style="font-family: "courier new" , "courier" , monospace;">rweibull</span>, with the exception that you pass in another vector <span style="font-family: "courier new" , "courier" , monospace;">t</span> of minimum times or a scalar representing the minimum time of all simulated values. The idea is that the probability that the random variable will be at least T is given by <span style="font-family: "courier new" , "courier" , monospace;">exp{-(T+t/b)^a} / exp{-(T/b)^a}</span><span style="font-family: "times" , "times new roman" , serif;">, so you can simulate this with a uniform random variate. I use the inversion method on the reliability function (just like using the inversion method on the distribution function, with the insight that if </span><span style="font-family: "courier new" , "courier" , monospace;">U</span><span style="font-family: "times" , "times new roman" , serif;"> is uniform(0,1), so is </span><span style="font-family: "courier new" , "courier" , monospace;"><span style="font-family: "times" , "times new roman" , serif;">1-</span><span style="font-family: "times" , "times new roman" , serif;">U</span></span><span style="font-family: "times" , "times new roman" , serif;">).</span><br />
<span style="font-family: "times" , "times new roman" , serif;"><br /></span>
<span style="font-family: "times" , "times new roman" , serif;">Truth be told, I ought to buckle down and learn how to do packages in R, but for now I'll just pass on some code on my blog if I think it will be helpful (or if I need to find it while doing a Google search later).</span><br />
<span style="font-family: "times" , "times new roman" , serif;"><br /></span>
<span style="font-family: "times" , "times new roman" , serif;">(Edit on 7/16: <b>OOPS!</b> A previous version of this had the scale and shape parameters switched. I've corrected it now. If you copied this before 7/16/2016, please check again.)</span>http://realizationsinbiostatistics.blogspot.com/2016/05/simulating-weibull-conditional-on-time.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-86049477500744811362016年1月14日 20:01:00 +00002016年01月14日T16:57:40.821-05:00analyticsdata scienceRstatisticsTalk to Upstate Data Science Group on CaretLast night I gave an <a href="https://github.com/randomjohn/caretdemo2" target="_blank">introduction and demo</a> of the <span style="font-family: Courier New, Courier, monospace;">caret</span><span style="font-family: Times, Times New Roman, serif;"> R package to the Upstate Data Science group, meeting at Furman University. It was fairly well attended (around 20 people), and well received.</span><br />
<span style="font-family: Times, Times New Roman, serif;"><br /></span>
<span style="font-family: Times, Times New Roman, serif;">It was great to get out of my own comfort zone a bit (since graduate school, I've only really given talks on some topic in biostatistics) and meeting statisticians, computer scientists, and other sorts of data scientists from many different fields. This is a relatively new group, and given the interest over the last couple of months or so I think this has been sorely needed in the Upstate South Carolina region.</span><br />
<span style="font-family: Times, Times New Roman, serif;"><br /></span>
<span style="font-family: Times, Times New Roman, serif;">We'll be participating in Open Data day in March of this year, so if you are in the Upstate SC region, or don't mind making the trek from Columbia or Western NC, find us on <a href="http://www.meetup.com/Greenville-Data-Science-Analytics-Meetup/" target="_blank">Meetup</a>. Our next meeting is a data hack night which promises to be interesting.</span>http://realizationsinbiostatistics.blogspot.com/2016/01/talk-to-upstate-data-science-group-on.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-75665814260439665632015年11月25日 13:00:00 +00002016年06月05日T15:43:45.065-04:00Little Debatepractice of statisticsRstatisticsEven the tiniest error messages can indicate an invalid statistical analysisThe other day, I was reading in a data set in R, and the function indicated that there was a warning about a parsing error on one line. I went ahead with the analysis anyway, but that small parsing error kept bothering me. I thought it was just one line of goofed up data, or perhaps a quote in the wrong place. I finally opened up the CSV file in a text editor, and found that the reason for the parsing error was that the data set was duplicated within the CSV file. The parsing error resulted from the reading of the header twice. As a result, anything I did afterward was suspect.<br />
<br />
Word to the wise: track down the reasons for even the most innocuous-seeming warnings. Every stage of a statistical analysis is important, and small errors anywhere along the way and have huge consequences downstream. Perhaps this is obvious, but you still have to slow down and take care of the details.<br />
<br />
(Note that I'm editing this to be a part of my <a href="http://realizationsinbiostatistics.blogspot.com/search/label/Little%20Debate" target="_blank">Little Debate</a> series, which discusses the tiny decisions dealing with data that are rarely discussed or scrutinized, but can have a major impact on conclusions.)http://realizationsinbiostatistics.blogspot.com/2015/11/even-tiniest-error-messages-can.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-54921275906244383302015年10月22日 15:33:00 +00002015年10月22日T11:33:18.505-04:00Statisticians ruin the day again, this time with a retraction<a href="http://retractionwatch.com/2015/10/21/authors-retract-second-study-about-medical-uses-of-honey/">Authors retract second study about medical uses of honey - Retraction Watch at Retraction Watch</a>: <br /><br />
<br /><br />
<div class="separator" style="clear: both; text-align: center;"><a href="https://encrypted-tbn1.gstatic.com/images?q=tbn:ANd9GcSl9epYMW9q_OIFWVRmawkZ8YkyqbZPlHSFC7FU6vb_swzk4bc1" imageanchor="1" style="clear: left; float: left; margin-bottom: 1em; margin-right: 1em;"><img border="0" height="170" src="https://encrypted-tbn1.gstatic.com/images?q=tbn:ANd9GcSl9epYMW9q_OIFWVRmawkZ8YkyqbZPlHSFC7FU6vb_swzk4bc1" width="200" /></a></div>For the second time, authors of manuscripts have had to retract their papers because of serious data analysis errors. While the details of the actual errors are scant, we do know that the article was published, a company tried to replicate the results but failed, the journal editor employed a third-party statistician who found serious errors in the data analysis, and the errors were serious enough that the paper, to stay accepted, would have had to go through a major revision and further peer review.<br /><br />
<br /><br />
Better to get a statistician to ruin your day before publication (and require another revision) than to have to eat crow because a third party did it. Other thoughts:<br /><br />
<br /><br />
<ul><li>Did the authors have a statistician review this before submitting?</li>
<li>Did the journal have this go through a statistical peer review before accepting?</li>
<li>Come to think of it, was a statistician involved in the planning of this study?</li>
</ul>http://realizationsinbiostatistics.blogspot.com/2015/10/statisticians-ruin-day-again-this-time.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-10643682690494637472015年10月15日 05:50:00 +00002015年10月15日T01:50:04.235-04:00The thirty-day trialSteve Pavlina wrote about a self-help technique called the thirty-day trial. To perform the technique, you commit 30 days of some new habit, such as quitting smoking or writing in a journal. The idea is that it’s psychologically easier to commit to something for 30 days than to make a permanent change, but after the 30 days you break addiction to old habits and have the perspective of whether to continue on with the new habit, go back, or go a completely different direction.<br />
<br />
For activities like journaling or quitting smoking, this technique might work. After all, psychologist Jeremy Dean announced that it takes 21 days to form a new habit. (This has been treated with some due skepticism, which it should.) However, if you try a 30 day quit-smoking trial and it doesn’t work, try again. And if that doesn’t work, try it again. Until you succeed.<br />
<br />
However, for activities such as a new diet, treatments, or any sort of healthcare advice, the 30 day trial should not be used.<br />
<br />
For cases where one might try a 30 day trial just to see if it feels better, such as an elimination diet, confirmation bias is likely to be at play. This is especially true in the case of an elimination diet, where one eliminates some aspect of a diet, like dairy or gluten, and see if some kinds of symptoms, such as bloating or fatigue, go away. In such trials, the results may be due to the eliminated item in question, placebo/nocebo effect, or some third confounded eliminated item. For instance, bloating from gluten-containing foods probably comes from medium-length carbohydrates called FODMAPs. Just because you feel better for 30 days after discontinuing gluten-containing foods doesn’t mean that gluten is the culprit. In fact, it probably isn’t, as determined by a well-designed clinical trial. Likewise, eliminating milk from a diet isn’t likely to do too much unless lactose intolerance is the culprit, and there are tests for that.<br />
<br />
Returning to Pavlina’s advice, he recommends a 30 day trial over the results of a clinical trial. This is sloppy and irresponsible advice. First, it is very unlikely that an individual will have access to screening assays, animal models, or the millions of dollars needed to discover and develop a drug. That is, unless said individual is up for trying millions of compounds, many potentially toxic and will probably tragically shorten the trial. Instead, a pharmaceutical should be used under the supervision of a doctor, who should be aware of the totality of literature (ideally consisting of randomized controlled trials if possible and a systematic review) and can navigate the benefits and possible adverse effects. A 30-day trial of a pharmaceutical or medical device may be completely inappropriate to realize benefits or assess risks. Here is the primary case where the plural of anecdote is not data.<br />
<br />
The bottom line is that the 30 day trial is one of several ways (and perhaps not the best one) to change habits that you know from independent confirmation need to be done, like quitting smoking. It’s a terrible way to adjust a diet or determine if a course of treatment is the right one. Treatments should be based on science and under the supervision of a physician who can objectively determine whether a treatment is working correctly.http://realizationsinbiostatistics.blogspot.com/2015/10/the-thirty-day-trial.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-51300518685394617602015年5月08日 16:47:00 +00002015年05月08日T12:47:57.610-04:00Statistics: P values are just the tip of the iceberg : Nature News & Comment<a href="http://www.nature.com/news/statistics-p-values-are-just-the-tip-of-the-iceberg-1.17412">Statistics: P values are just the tip of the iceberg : Nature News & Comment</a>: <br /><br />
<br /><br />
This article is very important. Yes, p-values reported in the literature (or in your own research) need scrutiny, but so does every step in the analysis process, starting with the design of the study.<br /><br />
<br />http://realizationsinbiostatistics.blogspot.com/2015/05/statistics-p-values-are-just-tip-of.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-707362099058929542015年4月17日 12:45:00 +00002015年04月17日T08:45:00.081-04:00biasclinical trialsThe many faces of the placebo response<p>This week, a study was <a href="http://www.sciencedaily.com/releases/2015/04/150413140906.htm" target="_blank">published</a> that claimed that the placebo response is mediated by genetics. Though I need to dig a little deeper and figure out exactly what this article is saying, I do think we need to take a step back and remember what can constitute a placebo response before we start talking about what this means for medical treatment and clinical trials.</p> <p>In clinical trials, the placebo response can refer to a number of apparent responses to sham treatment:</p> <ul> <li>The actual placebo response, i.e. a body’s physiological response to something perceived to be a treatment</li> <li>Natural course of a disease, including fluctuations, adaptations</li> <li>Investigator and/or subject bias on subjective instruments (hopefully mitigated by blinding/masking treatment arms in a study)</li> <li>Regression to the mean (an apparent time-based relationship caused by one measurement that varies markedly from the average measurement)</li> <li>... and many, many other sources</li></ul> <p>This week’s discovery does suggest that there is something physiological to the actual placebo response, and certainly genetics can influence this response. This may be useful in enriching clinical trials where a large placebo response is undesirable, e.g. by removing those subjects who are likely to response well to anything active or inert. After all, you don’t want an estimate of a treatment effect contaminated by a placebo response, nor do you want an impossibly high bar for showing an effect.</p> <p>But we still need to remember the mundane sources of "placebo response" and lower those before we get too excited about genetic tests for clinical trials.</p> http://realizationsinbiostatistics.blogspot.com/2015/04/the-many-faces-of-placebo-response.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-61264538792324690542015年4月14日 19:32:00 +00002015年04月14日T22:22:40.877-04:00critical thinkinglying with statisticsHelicopter parenting because your mind is "terrible at statistics" (or, rather, you are unaware of the denominator)<div class="separator" style="clear: both; text-align: center;">
<a href="http://www.todaysparent.com/wp-content/uploads/2014/04/1iStock_000000434344Small-1.jpg" imageanchor="1" style="clear: left; float: left; margin-bottom: 1em; margin-right: 1em;"><img border="0" src="http://www.todaysparent.com/wp-content/uploads/2014/04/1iStock_000000434344Small-1.jpg" height="200" width="200" /></a></div>
In Megan McArdle's <a href="http://www.bloombergview.com/articles/2015-04-13/seven-reasons-we-hate-free-range-parenting">Seven Reasons We Hate Free-Range Parenting - Bloomberg View</a>:, she states that because of the news cycle, and because our minds are terrible at statistics, we think the world is a much less safe place than 30 years ago. It's true. We have more opportunity (because of the internet and cable news) to hear about tragedies and crime from far-away places. It's worse if such tragedy strikes closer to home. Thus, we tend to think the world is very unsafe. (And therefore are encouraged to become helicopter parents.) We are acutely aware of the numerator<br />
<br />
What we do not hear, because it does not sell on cable news, is is the denominator. For all (very few) children abducted by strangers, for instance, we do not hear of the ones who successfully played at the park, or walked to the library, or went down to the field to play stickball (or I guess nerf softball because we shouldn't be throwing hard objects anymore) without getting abducted. This is because those stories do not sell.<br />
<br />
I guess the second best is reporting on trends in parenting, and how they are driven by how bad we are at statistics (even statisticians).<br />
<br />
<br />
<br />
<br />http://realizationsinbiostatistics.blogspot.com/2015/04/helicopter-parenting-because-your-mind.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-28701934076957403452015年3月19日 12:45:00 +00002015年03月19日T08:45:00.510-04:00alternative medicineBayesian statisticslying with statisticsLying with statistics, CAM version<p>Full disclosure here: at one time I wanted to be a complementary and alternative (CAM) researcher. Or integrative, or whatever the cool kids call it these days. I thought that CAM research would yield positive fruit if they could just tighten up their methodology and leave nothing to question. While this is not intended to be a discussion of my career path, I’m glad I did not go down that road.</p> <p><a href="http://www.ejinme.com/article/S0953-6205(14)00150-2/pdf" target="_blank">This article</a> is a discussion of why. The basic premise of the article is that positive clinical trials do not really provide strong evidence of an implausible therapy, for much the same reason that doctors will give a stronger test to an individual who tests positive for HIV. A positive test will provide some, but not conclusive, evidence for HIV simply because HIV is so rare in the population. The predictive value of even a pretty good test is poor. And the predictive value of a pretty good clinical trial is pretty poor if the treatment has not been established.</p> <p>Put it this way, if we have a treatment that has zero probability of working (the "null hypothesis" in statistical parlance), there will be a 5% probability that it will show a significant result in a conventional clinical trial. But let’s turn that on its head using <a href="http://en.wikipedia.org/wiki/Bayes%27_rule" target="_blank">Bayes Rule</a>:</p> <p>Prob (treat is useless| positive trial) * Prob(positive trial) = Prob (positive trial | treatment is useless) * Prob (treat is useless) (ok, this is just the definition of Prob (treat is useless AND positive trial)</p> <p>Expanding, and using the law of total probability:</p> <p>Prob (treat is useless| positive trial) = Prob (positive trial | treatment is useless)* Prob (treat is useless) / ((Prob positive trial|treat is useless)*Prob(treat is useless) + Prob(positive trial|treat is not useless)*Prob(treat is not useless))</p> <p>Now we can substitute, assuming that our treatment is in fact truly useless:</p> <p>Prob (treat is useless| positive trial) = p-value * 1 / (p-value * 1 + who cares * 0) = 1</p> <p>That is to say, if we know the treatment is useless, the clinical trial is going to offer no new knowledge of the result, even if it was well conducted.</p> <p>Drugs that enter in human trials are required to have some evidence for efficacy and safety, such as that gained from <em>in vitro </em>and animal testing. The drug development paradigm isn’t perfect in this regard, but the principle of the requirement of scientific and empirical evidence for safety and efficacy is sound. When we get better models for predicting safety and efficacy we will all be a lot happier. The point is to reduce the probability of futility to something low and maximize the probability of a positive trial given the treatment is not useless, which would result in something like:</p> <p>Prob (treat is useless | positive trial) = p-value * <something tiny> / (p-value * something tiny + something large * something close to 1) = something tiny</p> <p>Of course, there are healthy debates regarding the utility of the p-value. I question it as well, given that it requires a reference to trials that can never be run. These debates need to be had among regulators, academia, and industry to determine the best indicators of evidence of efficacy and safety.</p> <p>But CAM studies have a long way to go before they can even think about such issues.</p> http://realizationsinbiostatistics.blogspot.com/2015/03/lying-with-statistics-cam-version.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-49057071662192980582015年3月16日 12:30:00 +00002015年03月16日T08:30:00.830-04:00lying with statisticsvaccinationLying with statistics, anti-vax edition 2015<p><img src="http://origin.ih.constantcontact.com/fs061/1101053106216/img/17.jpg?a=1101671877326" width="147" align="left" height="143"></p> <p>Sometimes Facebook’s suggestions of things to read lead to some seriously funny material. After clicking on a link about vaccines, Facebook recommended I read an <a href="http://healthimpactnews.com/2011/new-study-vaccinated-children-have-2-to-5-times-more-diseases-and-disorders-than-unvaccinated-children/" target="_blank">article about health outcomes</a> in unvaccinated children. Reading this rubbish made me as annoyed as a <a href="http://scienceblogs.com/insolence/" target="_blank">certain box of blinking lights</a>, but it again affords me the opportunity to describe how people can confuse, bamboozle, and twist logic using bad statistics.</p> <p>First of all, Health Impact News has all the markings of a crank site. For instance, its banner claims it is a site for "News that impacts your health that other media sources may censor." This in itself ought to be a red flag, just like Kevin Trudeau’s <em>Natural Cures They Don’t Want You to Know About</em>.</p> <p>But enough about that. Let’s see how this article and the referred study abuses statistics.</p> <p>First of all, this is a bit of a greased pig. Their link leads to a malformed PDF file on a site called vaccineinjury.info. The site’s apparent reason for existence is to host a <a href="http://www.vaccineinjury.info/vaccinations-in-general/health-unvaccinated-children.html" target="_blank">questionnaire</a> for parents who did not vaccinate their children. So I’ll have to go on what the article says. There appe<em>No study</em> of health outcomes of vaccinated people versus unvaccinated has ever been conducted in the U.S. by CDC or any other agency in the 50 years or more of an accelerating schedule of vaccinations (now over 50 doses of 14 vaccines given before kindergarten, 26 doses in the first year).ars to be another <a href="http://www.vaccineinjury.info/vaccinations-in-general/health-unvaccinated-children/survey-results-illnesses.html" target="_blank">discussion</a> on the vaccineinjury.info site, which I’ll get to in a moment.</p> <p>The authors claim</p> <blockquote> <p><em>No study</em> of health outcomes of vaccinated people versus unvaccinated has ever been conducted in the U.S. by CDC or any other agency in the 50 years or more of an accelerating schedule of vaccinations (now over 50 doses of 14 vaccines given before kindergarten, 26 doses in the first year).</p></blockquote> <p><a href="http://www.webmd.com/children/vaccines/news/20140519/delaying-measles-related-vaccines-may-raise-seizure-risk-study" target="_blank">Here’s one</a>. A simple Pubmed search will bring up others fairly quickly. These don’t take long to find. What happens after this statement is a long chain of unsupported assertions about what data the CDC has and has not collected, that I really don’t have an interest in debunking right now (and so leave as an exercise).</p> <p>So on to the good stuff. They have a pretty blue and red bar graph that’s just itching to be shredded, so let’s do it. This blue and red bar graph is designed to demonstrate that vaccinated children are more likely to develop certain medical conditions, such as asthma and seizures, than unvaccinated children. Pretty scary stuff, if their evidence were actually true.</p> <p>One of the most important principles in statistics is defining your population. If you fail at that, you might as well quit, get your money back from <a href="http://www.sas.com" target="_blank">SAS</a>, and call it a day, because nothing that comes after that is meaningful. You might as well make up a bunch of random numbers if that’s the case, because that will be just as meaningful.</p> <p>This study fails miserably at defining its population. The best I can tell, the comparison is between a population in an observation study called <a href="http://www.kiggs-studie.de/" target="_blank">KIGGS</a> and respondents to an open invitation survey conducted at vaccineinjury.info.</p> <p>What could go wrong? Rhetorical question.</p> <p>We don’t know who responded to the vaccineinjury.info questionnaire, but it is aimed at parents who did not vaccinate their children. This pretty much tanks the rest of their argument. From what I can tell, these respondents seem to be motivated to give answers favorable to the antivaccine movement. That the data they present are supplemented with testimonials gives this away. They are comparing apples to rotten oranges.</p> <p>The right way to answer a question like this is a matched case-control study of vaccinated and unvaccinated children. An immunologist is probably the best one to determine which factors need to be included in the matching. That way, an analysis conditioned on the matching can clearly point to the effect of the vaccinations rather than leave open the questions of whether the differences in cases were due to differences in inherent risk factors.</p> <p>I’m wondering if there isn’t some ascertainment bias going on as well. Though I really couldn’t tell what the KIGGS population was, it was represented as the vaccinated population. So in addition to imbalances in risk factors, I’m wondering if the "diagnosis" in the unvaccinated population was derived from the parents were asked which medical conditions their children have. In that case, we have no clue what the real rate is like, because we are comparing parents’ judgments (and parents probably more likely to ignore mainstream medicine at that) with, presumably, a GP’s more rigorous diagnosis. That’s not to say that no children in the survey were diagnosed by an MD, but without that documentation (which this web-based survey isn’t going to be able to provide), the red bars in the pretty graph are essentially meaningless. (Which they were even before this discussion.)</p> <p>But let’s move on.</p> <p>The vaccineinjury.info cites some other studies that seem to agree with their little survey. For instance, McKeever, <em>et al.</em> published a <a href="http://ajph.aphapublications.org/doi/abs/10.2105/AJPH.94.6.985" target="_blank">study</a> in the <em>American Journal of Public Health</em> in 2004 from which the vaccineinjury.info site claims an association between vaccines and the development of allergies. However, that apparent association, as stated in the study, is possibly the result of ascertainment bias (the association was only strong in a stratum with the least frequent GP visits). Even <a href="http://ajph.aphapublications.org/doi/pdf/10.2105/AJPH.2004.050583" target="_blank">objections</a> to the discussion of ascertainment bias leave the evidence of association of vaccines and allergic diseases unclear.</p> <p>The vaccineinjury.info site also cites the <a href="http://www.ncbi.nlm.nih.gov/pubmed/11110734" target="_blank">Guinea-Bisseau study</a> reported by Kristensen <em>et al.</em>in <em>BMJ</em> in 2000. They claim, falsely, that the study showed a higher mortality in vaccinated children.</p> <p>They also cite a New Zealand study.</p> <p>What they don’t do is describe how they chose the studies to be displayed on the web site. What were the search terms? Were these studies cherry-picked to demonstrate their point? (Probably, but they didn’t do a good job.)</p> <p>What follows the discussion of other studies is an utter waste of internet space. They report the results of their "survey," I think. Or somebody else’s survey. I really couldn’t figure out what was meant by "Questionnaire for my unvaccinated child ("Salzburger Elternstudie")". The age breakdown for the "children" is interesting, for 2 out of the 1004 "children" were over 60! At any rate, if you are going to be talking about diseases in children, you need to present it by age, because, well, age is a risk factor in disease development. But they did not do this.</p> <p>What is interesting about the survey, though, is the reasons the parents did not vaccinate their children, if only to give a preliminary notion of the range of responses.</p> <p>In short, vaccineinjury.info, and the reporting site Health Impact News, present statistics that are designed to scare rather than inform. Proper epidemiological studies, contrary to the sites’ claims, have been conducted and provide no clear evidence to the notion that vaccinations cause allergies except in rare cases. In trying to compile evidence for their claims, they failed to provide evidence that they did a proper systematic review, and even misquoted the conclusions of the studies they presented.</p> <p>All in all, a day in the life of a crank website.</p> http://realizationsinbiostatistics.blogspot.com/2015/03/lying-with-statistics-anti-vax-edition.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-4910516452835014482014年12月31日 16:07:00 +00002014年12月31日T11:07:36.894-05:00analyticsmachine learningsocial mediaAlgorithmic cruelty<p>By now, most of us know about Facebook’s algorithmic retrospectives, and, of course, how some people thought it could <a href="http://flowingdata.com/2014/12/29/inadvertent-algorithmic-cruelty/" target="_blank">be cruel</a>. Indeed, posts about some sorts of issues, such as divorces or deaths, can get a lot of "likes" (where the "like" button means something other than "like") and comments, and therefore get flagged by whatever algorithm the Facebook data scientists came up with as a post worthy of a retrospective.</p> <p>There are a lot of issues here. When someone "likes" a post, they do not necessarily mean they "like" the event the post is about. It could mean a number of different things, such as "I hear you," or "I’m empathizing with you," or even "Hang in there." However, the algorithms treat all likes equal.</p> <p>Comments, of course, carry much more sophisticated meaning, but are much harder to analyze especially in the presence of sarcasm. And algorithms that do analyze comments (or any free text) for sentiment will require a large training set of hand-coded comments. (Which I suppose Facebook does have the resources to generate.)</p> <p>Which leaves a few ways of handling this problem:</p> <ul> <li>Do nothing different. Which is probably my favorite solution, because I’d like to look back on the good, the bad, and the ugly. It’s my life, and I want to remember it. Besides, the event that really sucked at the time (say, a torn ACL leading to surgery) may lead to good things.</li> <li>Add a "I don’t want to see this" button. Which was already accomplished by including an "X" button, but maybe not so obvious.</li> <li>Eliminate the retrospective, which I don’t think anybody agrees is a good solution.</li></ul> <p>I suppose one day Facebook’s algorithm will be smart enough to withhold posts it knows people don’t want to review, but then that will open up another can of worms.</p> http://realizationsinbiostatistics.blogspot.com/2014/12/algorithmic-cruelty.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-80088245772331454102014年12月09日 23:00:00 +00002014年12月10日T10:23:11.375-05:00epidemiologylying with statisticsskepticismNo, a study did not link genetically engineered crops to 22 diseases<p>In my Facebook feed, a friend posted a very <a href="http://www.organic-systems.org/journal/92/JOS_Volume-9_Number-2_Nov_2014-Swanson-et-al.pdf" target="_blank">scary-looking study</a> that links genetically engineered (GE) crops to the rise in 22 diseases. These are pretty fearsome diseases, too, like bile duct cancer and pelvis cancer. For instance<sup>1</sup>:</p> <p><a href="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEjNknX3GcBA9NlMWTPABN-6sjf1UYhEI7fR1dCfcr8zzkLD5IEqPrho_sfSQjhKQOs3blUoqz0qa2l03JvpKqEhhrbhm6sCVEgNr9TAxTELsCq9wpcDHCmsCBEVaV8hTMJuvRQq4dadDTE/s1600-h/image2.png"><img title="image" style="border-left-width: 0px; border-right-width: 0px; background-image: none; border-bottom-width: 0px; padding-top: 0px; padding-left: 0px; margin: 0px; display: inline; padding-right: 0px; border-top-width: 0px" border="0" alt="image" src="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEhzpqTmyG5CzyiHVMKUIk7S84wH2pN59YYIGDXgn-2IidvJS08ZtrVB9EsI5y8aNoNELfAUgFsCZW64Nf4H4U-JSVDfDGJgZoQL3K6Fat7EkU12cwPJRwCNqwklSEa7VF0fyrRSDraA5pc/?imgmax=800" width="244" height="183"></a><br>There are a few ways to respond to this article:</p> <p>First, it has not passed my attention that the second author has published a book Myths of Safe Pesticides, which has been <a href="http://www.sciencebasedmedicine.org/pesticides-just-how-bad-are-they/" target="_blank">analyzed and debunked by Harriet Hall</a>.</p> <p>Second, I could just say "correlation is not causation." QED. Article debunked, and can be swept to the dustbin.</p> <p>Third, I can point out the <a href="http://kfolta.blogspot.com/2013/02/organic-food-causes-autism.html" target="_blank">correlation between sales of organic produce and autism</a>. (Yikes!) In fact, using the methods of this article, I can probably prove a significant correlation between sales of organic produce and bile cancer, kidney cancer, autism, or lipoprotein disorder deaths. We can all grab our glyphosate-coated pitchforks and demand reform!</p> <p>However, I think there are some statistical lessons here, and it's sometimes good to deconstruct some misused and abused statistics. And trust me, the statistics in this article are seriously misused. In fact, it might be an interesting project for an introductory graduate statistics class to collect articles like this and critique them. I'll do it for fun here. There are others that can speak to the scientific aspects of the article, like how it disagrees with the results of a review of over a trillion meals fed that incorporate GE products. There's also other quibbles with the article, like how it sometimes conflates pesticide discussions with glyphosate (an herbicide), that others can deconstruct.</p> <p>When deciding on how to summarize and analyze data statistically, it is essential to work with the nature of the data. This article fails on several counts. First, it smashes together data from two complete different sources without considering how the data are related. Now, I'm generally excited to see data from disparate sources linked and analyzed together, but it has to be done carefully. This is how they obtained their data on GE use:</p> <blockquote> <p><br>From 1990-2002, glyphosate data were available for all three crops, but beginning in 2003 data were not collected for all three crops in any given year. Data on the application rates were interpolated for the missing years by plotting and calculating a best fit curve. Results for the application rates for soy and corn are shown in Figures 2 and 3. Because the PAT was relatively small prior to about 1995, the sampling errors are much larger for pre-1995 data, more so for corn than for soy. Also, data were not missing until 2003 for soy and 2004 for corn. For these reasons, the interpolated curves begin in 1996 for soy and 1997 for corn in Figures 2 and 3. <br></p></blockquote> <p>This is how they obtained epidemiological data:</p> <blockquote> <p>Databases were searched for epidemiological data on diseases that might have a correlation to glyphosate use and/or GE crop growth based on information given in the introduction. The primary source for these data was the Centers for Disease Control and Prevention (CDC). These data were plotted against the amount of glyphosate applied to corn and soy from Figure 6 and the total %GE corn and soy crops planted from Figure 1. The percentage of GE corn and soy planted is given by: (total estimated number of acres of GE soy + total estimated number of acres of GE corn)/(total Estimated acres of soy + total estimated acres of corn)x100, where the estimated numbers were obtained from the USDA as outlined above.</p></blockquote> <p>This seems innocent enough, but there's already a lot of wrong happening here. It's good that they explained some of their data cleaning, though we can always stand for more transparency behind this step. It's not scientifically glorious to describe how you handle missing or sparse data, but mishandling such can certainly sink your Nobel prize work. It's also good to explain derived variables, though I haven't gone back and checked their math.</p> <p>The first fatal error is how they link the data. They simply merge it by year. It's the obvious-seeming step that already tanks their analysis. This is the same kind of merging that links, say, sales of organic crops to autism. Mashing up data needs to be done in a scientifically valid way, and simply merging disparate data by year isn't going to cut it here. All these data they gathered are crude summaries, and they just strung them together by year without giving any thought to whether the subjects in the epidemiological database have any connection to the subjects in the GE database. Sloppy, and that right there can be enough to tank any analysis, even if the analysis were well done. Which this one wasn't.</p> <p>The second fatal error is how they present the data. Take the Figure 16 above. This graph breaks so many rules of data presentation that Edward Tufte's head would probably explode just from looking at it. But let's dig a little deeper. The authors say they plotted incidence of disease (in Figure 16 it's age-adjusted deaths due to lipoprotein disorder) against GE and glyphosate use. However, if you want to get technical about it, they plot all three of these versus time. This is a very important distinction. If they plotted incidence versus GE use, then they would put GE use on the x-axis. However, they show incidence in bar graphs by time, GE use in a line graph by time, and glyphosate use by time. I'll explain why this is important in the discussion of the third fatal flaw. But let's move ahead with the graph. From what I've been able to figure out, the left y-axis goes with the bar graph and is in deaths per hundred thousand. The axis on the right does double duty and covers both % of GE planted and 1000 tons of glyphosate used. It took me a while to figure that out, and it's very sloppy design anyway (the two scales have nothing to do with each other). If you ever see a line plot with a left and right y-axis, get skeptical. Here, the left axis starts at 0 and ends at 2.75 or so, and the right axis starts at -20 (!) and ends at 85 or so. I can see why they chose the y-axis, but the right axis is very curious. The -20 is a terrible choice for the start of the right axis. It's an invalid value of % of GE crops planted and 1000s of tons of glyphosate used. "Yes, Monsanto, I used -20,000 tons of glyphosate. You owe me 50,000ドル." It seems that the origin and scale of the right y-axis was chosen specifically to make GE and glyphosate use appear to track closely with deaths. I usually choose incompetence over malice to explain motivations, but it's very challenging to support incompetence in this case. It takes talent and/or effort to choose axes like this. I'll leave a deconstruction of the other graphs as an exercise, perhaps for your graduate-level stats class.</p> <p>The third and final fatal error is how they analyze the data. Their analysis is the statistical equivalent to bringing a knife to a gunfight. They basically take all the GE and epidemiological data, ignore the time component, and send it through your Stat 101 Pearson correlation estimator formula. They construct some p-values, unsurprisingly find a massively small p-value, declare victory, and hit the publish button. Problem is, they compute the wrong statistical summary using the wrong formula and use it to make the wrong inference. The Pearson correlation estimator they use is designed for independent data, not time series data (and they know it's time series data because they say so on p. 11). Time series data has a complex correlation structure, and thus estimating second-order parameters like correlations is a bit of a challenge. For instance, GE use this year is going to be heavily correlated to GE use last year, as are deaths from lipoprotein disorders. Does the correlation reflect a relationship between death and GE use, or death this year and death last year? The naïve estimate assumes the correlation is between death and GE use, and accounts nothing of the relationship between deaths this year and last year (in the stat world we call this autocorrelation). Though I haven't done the math, my guess is that the correlation between death and GE use will be greatly reduced if not disappear altogether if time is taken into account. And even if there is a nonzero, significant correlation, the fact of the matter is that there needs to be a stronger link than time between the GE data and epidemiological data.</p> <p>As a bonus, the paper claims to find a link between GE crop use, glyphosate use, and a whole bunch of nasty stuff, but they never try to tease out whether the nasty stuff is attributable to glyphosate or GE crops.</p> <p>In conclusion, the paper claims to find a strong link between GE crop use and glyphosate use, and a host of diseases. Given that their paper was so deeply methodologically flawed, they are unable to support their conclusions. This paper should not be considered as evidence of the dangers of GE crop use or glyphosphate use, but should rather be used as a showcase of "How Not to Do It."</p><p>Edit: I need to learn how to spell glyphosate.</p> <p>Footnotes:<br><sup>1</sup>Swanson, Leu, Abrahamson, and Wallet. "Genetically engineered crops, glyphosphate and the deterioration of health in the United States." Journal of Organic Systems. 9(2), 2014. Figure 16.</p> http://realizationsinbiostatistics.blogspot.com/2014/12/no-study-did-not-link-genetically.htmlnoreply@blogger.com (Anonymous)Americas, null54.525963 -105.25512tag:blogger.com,1999:blog-6439806875881311576.post-51453611903989782272013年8月07日 16:02:00 +00002013年08月07日T12:02:29.214-04:00joint statistical meetingspractice of statisticsstatisticsJoint statistical meetings 2013<p dir="ltr">Every year, the first week of August, we statisticians meet to get our statistics, networking, dancing, and beer on. With thousands in attendance, it's exhausting. I wonder about the quality of statistical work the second week of August.</p>
<p dir="ltr">Each conference seems to have a life of its own, so I tend to reflect on each one. Here's my reflection on this year's:</p>
<p dir="ltr">First, being in Montreal, most of us couldn't use smartphones. Thankfully, <a href="http://www.revolutionanalytics.com">Revolution Analytics</a> sponsored free WiFi. They also do great work with <a href="http://www.r-project.org">R</a>. So we were all for the most part able to tweet.</p>
<p dir="ltr">The quality of talks was pretty good this year, and I've learned a lot. We even had one person describe simulations with a flowchart rather than indecipherable equations, and I strongly encourage that practice.</p>
<p dir="ltr">As a member of the biopharmaceutical section, I was struck by how few people take advantage of our awards. Of course, everybody giving a contributed or topic contributed talks is automatically entered into the best contributed paper competition. But we have a poster competition and student paper competition that have to be explicitly entered, and participation is low. This is a great opportunity.</p>
<p dir="ltr">The highlight of the conference, of course, was Nate Silver's talk, and he delivered admirably. The perhaps thousand statisticians in attendance needed the message: learn to communicate with journalists and teach them numbers need context. I also like his response to the question "statistician or data scientist?" Which was, of course, "I don't care what you call yourself, just do good work."<br><br></p>
http://realizationsinbiostatistics.blogspot.com/2013/08/joint-statistical-meetings-2013.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-36718790019002616712013年7月15日 12:45:00 +00002013年07月15日T08:45:00.606-04:00Bayesian statisticsWasserman on noninformative priors<p>Larry Wasserman calls the use of <a href="http://normaldeviate.wordpress.com/2013/07/13/lost-causes-in-statistics-ii-noninformative-priors/">noninformative priors</a> a "lost cause." I agree for the reasons he stated, and the fact that there are always better alternatives anyway. At the very least, there are the heavy-tailed "weakly informative priors" that put nearly all weight on something reasonable, such as small to moderate values of a variance, and little weight on stupid prior values, such as mean values on the order of 10<sup>100</sup>.</p> <p>However, they’ll be around for years to come. Noninformative priors are nice security blankets, and we get to think that we are approaching a problem with an open mind. I guess open minds can have stupid properties as well.</p> <p>I hope, though, that we will start thinking more deeply about the consequences of our assumptions especially about noninformative priors rather than feeling nice about them.</p> http://realizationsinbiostatistics.blogspot.com/2013/07/wasserman-on-noninformative-priors.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-66323823367479064412013年4月29日 03:21:00 +00002013年04月28日T23:21:10.452-04:00moocMOOCs–a low-risk way to explore outside your field<p>One of the things I'm realizing from Massively Open Online Courses (MOOCs) -- those online free classes from universities that have seem to sprung up from almost nowhere in the last year and a half -- is that they offer a perfect opportunity to explore outside my field. At first (and this was even before the term MOOC was coined), I took classes there were just outside my field. For instance, I've been in clinical and postmarketing pharmaceutical statistics for over 10 years, and my first two classes were in databases and machine learning. I did this because I was aching to learn something new, but I figured that with a class in databases I could make our database guys in IT sweat a bit just by dropping some terms and showing some understanding of the basics. It worked. In addition, I wanted to understand what this machine learning field was all about, and how it was different from statistics. I accomplished that goal, too.</p> <p>Since then, I have taken courses in the area of artificial intelligence/machine learning, sociology and networks, scientific computing (separately from statistical computing), and even entrepreneurship. I have also encouraged others to take part in MOOCs, though I don't know the result of that. Finally, I have come back to some classes I've already taken as a community TA, or former student who actively takes part in discussions to help new students take the class.</p> <p>This is all valuable experience, and I could write several blog entries on the benefits. The main one I'm feeling right now is the feeling that I'm coming up for air, and taking a sampling of other points of view in a low-risk way. For example, though I don't actively use Fourier analysis in my own work, one recent class and one current class both use it to do different things (solve differential equations and process signals). Because these classes involve programming assignments, I've now deepened my understanding of the spectral theorem, which I only studied from a theoretical point of view in graduate school. I'm also thinking about this work from the point of view of time series analysis, which is helping me think about some problems involving longitudinal data at work.</p> <p>From a completely different standpoint, another class helped me think about salary negotiations in terms of expected payoff (i.e. combination of probability of an offer being accepted vs. salary). This sort of analysis invited further analysis of the value of that job vs. what I would be paid and the insecurity of moving to a different job. In the end, I turned down what would have been a pretty good offer, because I decided it did not compensate for the risks I was incurring. The cool thing is that these were all applying concepts I already understood (expected value, expected payoff), but applied in a different way from what I was already doing.</p> <p>The best thing about MOOCs is that the risk is low. All that is required is an internet connection and a decent computer. Some math courses may require a better computer to do high-powered math, but I've seen few that require expensive textbooks or expensive software. Even Mathworks is now offering Matlab at student pricing to people who are taking some classes, and Octave remains a free option for people unable to take advantage of it. And, if you are unable to keep up the work, there is now downside. You can simply unenroll.</p> http://realizationsinbiostatistics.blogspot.com/2013/04/moocsa-low-risk-way-to-explore-outside.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-29751345765021856682013年4月16日 02:33:00 +00002013年04月15日T22:33:14.288-04:00RRStudioRStudio is reminding me of the older Macs<div class="separator" style="clear: both; text-align: center;">
<a href="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEg6nSS_aWY0lp2UrJAJ4aHGxe4jjD4REzWXi80IK5pXh4AzsOG1JQfN8TZZU9xicUhlRgfsxVD5B4pxuHefOEtZJzHwU-mAVWSkWP3HJkiKjxYJTxb-_zi3wsmwbCcyghI0WKu1H2PPjc8/s1600/rbomb.png" imageanchor="1" style="margin-left: 1em; margin-right: 1em;"><img border="0" height="212" src="https://blogger.googleusercontent.com/img/b/R29vZ2xl/AVvXsEg6nSS_aWY0lp2UrJAJ4aHGxe4jjD4REzWXi80IK5pXh4AzsOG1JQfN8TZZU9xicUhlRgfsxVD5B4pxuHefOEtZJzHwU-mAVWSkWP3HJkiKjxYJTxb-_zi3wsmwbCcyghI0WKu1H2PPjc8/s320/rbomb.png" width="320" /></a></div>
The only thing missing is the cryptic ID number.<br />
<br />
Well, the only bad thing is that I am trying to run a probabilistic graphical model on some real data, and having a crash like this will definitely slow things down.http://realizationsinbiostatistics.blogspot.com/2013/04/rstudio-is-reminding-me-of-older-macs.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-82857686383623112592013年3月30日 18:21:00 +00002013年03月30日T14:21:11.312-04:00PresentationsPresenting without slides<p>Tired of slides, I’ve been experimenting with different ways of presenting. At the recent Conference on Statistical Practice, I decided only to use slides for an outline and references. As it turns out, the most critical feedback I got had to do with the fact that the audience couldn’t follow the organization because I had no slides.</p> <p>I tried presenting without slides because, well, I started to use them as a crutch. I also saw a lot of people presenting essentially by putting together slides and reading from them. So I figured I would expand my horizons.</p> <p>Next time I present, I’ll do slides, I guess, but I may try something a bit different.</p> http://realizationsinbiostatistics.blogspot.com/2013/03/presenting-without-slides.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-37089678607007896652013年3月27日 20:49:00 +00002013年03月27日T16:49:52.131-04:00learning from datamachine learningmoocLast session of Caltech's Learning from Data course starts April 2<div class="tr_bq">
I just received this email:</div>
<br />
<blockquote>
<span style="background-color: white; color: #222222; font-family: arial, sans-serif; font-size: 13px;">Caltech's Machine Learning MOOC is coming to an end this spring, with the final session starting on April 2. There will be no future sessions. The course has attracted more than 200,000 participants since its launch last year, and has gained wide acclaim. This is the last chance for anyone who wishes to take the course (</span><a href="http://work.caltech.edu/telecourse" style="background-color: white; color: #1155cc; font-family: arial, sans-serif; font-size: 13px;" target="_blank">http://work.caltech.edu/<wbr></wbr>telecourse</a><span style="background-color: white; color: #222222; font-family: arial, sans-serif; font-size: 13px;">).</span><br style="background-color: white; color: #222222; font-family: arial, sans-serif; font-size: 13px;" /><span style="background-color: white; color: #222222; font-family: arial, sans-serif; font-size: 13px;">Best.</span><br style="background-color: white; color: #222222; font-family: arial, sans-serif; font-size: 13px;" /><span style="background-color: white; color: #222222; font-family: arial, sans-serif; font-size: 13px;">The Caltech Team</span></blockquote>
I strongly <a href="http://realizationsinbiostatistics.blogspot.com/2013/03/review-of-caltech-learning-from-data-e.html" target="_blank">recommend this course</a> if you can take it, even if you have taken other machine learning classes. It lays a great theoretical foundation for machine learning, sets it off nicely from classical statistics, and gives you some experience working with data as well.<br />
<br />
If you were for some reason waiting for the right time, it looks to be now or never.http://realizationsinbiostatistics.blogspot.com/2013/03/last-session-of-caltechs-learning-from.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-67948861130248872622013年3月21日 02:47:00 +00002013年03月20日T22:47:32.637-04:00machine learningmoocReview of Caltech's Learning from Data e-course<p dir="ltr">Caltech has an online course <a href="http://work.caltech.edu/telecourse.html">Learning from Data</a>, taught by Professor Yaser Abu-Mostafa, that seeks to make the course material accessible to everybody. Unlike most of the online courses I've taken, this one is independently offered through a platform created just for the class. I took the course for its second offering in Jan-March 2013.</p>
<p dir="ltr">The platform on which the course is offered isn't as slick as Coursera. The lectures are offered through a Youtube playlist, and the homeworks are graded through multiple choice. That's perhaps a weakness of the class, but somehow the course faculty made it work.</p>
<p dir="ltr">The class's content was its strong point. Abu-Mostafa weaved theory and pragmatic concerns throughout the class, and invited students to write code in just about any platform (I, of course, chose <a href="http://www.r-project.org">R</a>) to explore the theoretical ideas in a practical setting. Between this class and Andrew Ng's Machine Learning class on the Coursera platform, a student will have a very strong foundation to apply these techniques to a real-world setting.</p>
<p dir="ltr">I have only one objection to the content, which came in the last lecture. In his description of Bayesian techniques, he claimed that in most circumstances you could only model a parameter with a delta function. This, of course, falls in line with the frequentist notion that you have a constant, but unknowable "state of nature." I felt this way for a long time, but don't really believe it any more in a variety of contexts. I think he played up the Bayesian v. frequentist squabble a bit much, which may have been appropriate 20 years ago but is not so much an issue now.</p>
<p dir="ltr">Otherwise, I found the perspective from the course extremely valuable, especially in the context of supervised learning.</p>
<p dir="ltr">If you plan on taking the course, I recommend leaving a lot of time for it or having a very strong statistical background.</p>
http://realizationsinbiostatistics.blogspot.com/2013/03/review-of-caltech-learning-from-data-e.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-46012305995877324722013年3月12日 23:12:00 +00002013年03月12日T19:12:06.753-04:00practice of statisticsRDistrust of R<p dir="ltr">I guess I've been living in a bubble for a bit, but apparently there are a lot of people who still mistrust R. I got asked this week why I used R (and, specifically, the package rpart) to generate classification and regression trees instead of SAS Enterprise Miner. Never mind the fact that rpart code has been around a very long time, and probably has been subject to more scrutiny than any other decision tree code. (And never mind the fact that I really don't like classification and regression trees in general because of their limitations.)</p>
<p dir="ltr">At any rate, if someone wants to pay the big bucks for me to use SAS Enterprise Miner just on their project, they can go right ahead. Otherwise, I have got a bit of convincing to do.</p>
http://realizationsinbiostatistics.blogspot.com/2013/03/distrust-of-r.htmlnoreply@blogger.com (Anonymous)tag:blogger.com,1999:blog-6439806875881311576.post-61114594030425732512013年3月01日 01:49:00 +00002013年02月28日T20:49:55.430-05:00publishingsciencestatistical leadershipBad statistics in high impact journals<p dir="ltr"><a href="http://blogs.discovermagazine.com/neuroskeptic/2013/02/19/better-journals-worse-statistics/#.UTAHJ1Mo6aw">Better Journals… Worse Statistics? : </a><a href="http://blogs.discovermagazine.com/neuroskeptic/2013/02/19/better-journals-worse-statistics/#.UTAHJ1Mo6aw">Neuroskeptic</a></p>
<p dir="ltr">In the linked blog entry, Neuroskeptic notes that high impact journals often have fewer statistical details than other journals. The research reported in these journals is often heavily amended, if not outright contradicted, by later research. I don't think this is nefarious, though, nor is it worthless. The kind of work reported in Science and Nature, for instance, generates interest and, therefore, more scrutiny (funding, studies, theses, etc.).</p>
<p dir="ltr">But as with all other research, if statistical details are included it might direct subsequent research in these topics a bit better.</p>
http://realizationsinbiostatistics.blogspot.com/2013/02/bad-statistics-in-high-impact-journals.htmlnoreply@blogger.com (Anonymous)