## Archive for David Spiegelhalter

## RSS tribute

Posted in Statistics, University life with tags David Spiegelhalter, Joyce Fienberg, Pittsburgh, Royal Statistical Society, shooting, StatsLife, Steve Fienberg on November 4, 2018 by xi'an## absurd graph [if relevant warning]

Posted in pictures, Statistics, Wines with tags alcohol, bad graph, David Spiegelhalter, graphics, health care, risk, The Guardian, The Lancet on August 28, 2018 by xi'an**A** pretty silly graph opposing countries with an overwhelming majority of non-Muslims and countries with an overwhelming majority of Muslims in terms of alcohol consumption. Surprise, surprise! And not incorporating the average amount or anything useful… In a Guardian article reporting on a Lancet paper about the lack of health benefit from drinking even moderate amounts of alcohol. Although, as pointed out by David Spiegelhalter at the bottom of the article, an increased risk of 0.5% associated with one unit of alcohol a day [half a pint] , as opposed to 7% for two units [a pint!], should not get occasional drinkers too worried: “Come to think of it, there is no safe level of living, but nobody would recommend abstention.”

## extra glass of wine? 30mn, please…

Posted in pictures, Statistics, Wines with tags David Spiegelhalter, life expectancy, RSS, The Guardian, The Lancet, wine on April 20, 2018 by xi'an**A**s I was reading The Guardian early today, I came across this entry on how an extra glass (17.5cl) glass of wine was equivalent to 30mn less of life (expectancy), above the recommended maximum of five glass a week. As explained by Prof of Risk David Spiegelhalter himself! The Lancet study behind this analysis stated that “early deaths rose when more than 100g per week, which is five to six glasses of wine or pints of beer, was consumed.” So be careful!!!

## double yolk priors [a reply from the authors]

Posted in Books, Statistics, University life with tags Brad Efron, David Spiegelhalter, egg double yolks, empirical Bayes methods, frequentist inference, reply on March 14, 2018 by xi'an*[Here is an email I received from Subhadeep Mukhopadhyay, one of the authors of the paper I discussed yesterday.}*

*Leg*and

_{j}(u)_j*T*

_{j}=*Leg*. One is orthonormal polyn of

_{j}(G(θ))*L*and the other one is

_{2}[0,1]*L*. The second one is poly of rank-transform

_{2}[G]*G(θ)*.

*g⁻¹(θ)*, which is the d(u;G,F) over unit interval. Now, this new transformed function is a proper density.

*d(G(θ))*which can be expended into (NOT in Leg-basis) in , in eq (2.2), as it lives in the Hilbert space

*L*

_{2}(G)*g*.

## double yolk priors

Posted in Statistics with tags Brad Efron, David Spiegelhalter, egg double yolks, empirical Bayes methods, frequentist inference on March 13, 2018 by xi'an

“To develop a “defendable and defensible” Bayesian learning model, we have to go beyond blindly ‘turning the crank’ based on a “go-as-you-like” [approximate guess] prior. A lackluster attitude towards prior modeling could lead to disastrous inference, impacting various fields from clinical drug development to presidential election forecasts. The real questions are: How can we uncover the blind spots of the conventional wisdom-based prior? How can we develop the science of prior model-building that combines both data and science [DS-prior] in a testable manner – a double-yolk Bayesian egg?”

**I** came through R bloggers on this presentation of a paper by Subhadeep Mukhopadhyay and Douglas Fletcher, Bayesian modelling via goodness of fit, that aims at solving all existing problems with classical Bayesian solutions, apparently! (With also apparently no awareness of David Spiegelhalter’s take on the matter.) As illustrated by both quotes, above and below:

“The two key issues of modern Bayesian statistics are: (i) establishing principled approach for distilling statistical prior that is consistent with the given data from an initial believable scientific prior; and (ii) development of a Bayes-frequentist consolidated data analysis work ow that is more effective than either of the two separately.”

(I wonder who else in this Universe would characterise “modern Bayesian statistics” in such a non-Bayesian way! And love the notion of distillation applied to priors!) The setup is actually one of empirical Bayes inference where repeated values of the parameter θ drawn from the prior are behind independent observations. Which is not the usual framework for a statistical analysis, where a single value of the parameter is supposed to hide behind the data, but most convenient for frequency based arguments behind empirical Bayes methods (which is the case here). The paper adopts a far-from-modern discourse on the “truth” of “the” prior… (Which is always conjugate in that Universe!) Instead of recognising the relativity of a statistical analysis based on a given prior.

When I tried to read the paper any further, I hit a wall as I could not understand the principle described therein. And how it “consolidates Bayes and frequentist, parametric and nonparametric, subjective and objective, quantile and information-theoretic philosophies.”. Presumably the lack of oxygen at the altitude of Chamonix…. Given an “initial guess” at the prior, g, a conjugate prior (in dimension one with an invertible cdf), a family of priors is created in what first looks like a form of non-parametric exponential tilting of g. But a closer look [at (2.1)] exposes the “family” as the tautological π(θ)=g(θ)x π(θ)/g(θ). The ratio is expanded into a Legendre polynomial series. Which use in Bayesian statistics dates a wee bit further back than indicated in the paper (see, e.g., Friedman, 1985; Diaconis, 1986). With the side issue that the resulting approximation does not integrate to one. Another side issue is that the coefficients of the Legendre truncated series are approximated by simulations from the prior [Step 3 of the Type II algorithm], rarely an efficient approach to the posterior.