Re: RE: Standard errors of estimates for strictly positive parameters

From: Nick Holford Date: February 12, 2015 technical Source: mail-archive.com
Hi, The original quote about electrons comes from a remark I made in 1999 on nmusers. http://www.cognigencorp.com/nonmem/nm/99nov121999.html Lewis Sheiner agreed in the same thread. Thanks to the wonders of living on a sphere Lewis appears to agree with me the day before I made the comment :-) My comments then about the relative merits of asymptotic SEs ("some kind of rough diagnostic"), log-likelihood profiling ("may require reparameterization") and non-parametric bootstrap ("reasonable values for a confidence interval") remain unchanged. The only change I would make is that I currently find 100 bootstrap replicates is good enough for model selection and evaluation. With the availability of grid computing bootstrap runs can be more readily included in the model selection strategy by identifying parameters which are poorly estimable. Most published papers I have read recently and most presentations at PAGE use bootstrap estimates to describe uncertainty as part of model evaluation. Unlike Leonid my experience is that bootstrap confidence intervals are often asymmetrical and do not agree with asymptotic SE predictions (e.g. see Matthews 2004). This may be because I deal with more non-linear problems. There is also the problem that when models get sufficiently complex to be interesting NONMEM is often unable to calculate an asymptotic SE (Holford's Rule -- http://www.cognigencorp.com/nonmem/current/2008-July/0017.html ). Matthews I, Kirkpatrick C, Holford N. Quantitative justification for target concentration intervention--parameter variability and predictive performance using population pharmacokinetic models for aminoglycosides. Br J Clin Pharmacol. 2004;58(1):8-19. Best wishes, Nick
Quoted reply history
On 13/02/2015 4:51 a.m., Bob Leary wrote: > Hi Leonid - > a) I recall the general occasion of the Lewis Sheiner quote - it was at the > PAGE meeting in Paris in 2002 at the end of a talk I gave there, > but not the specific question that inspired it (but I did show a profiling > graph there, so maybe it was related to that). > b) Profiling is NOT a local method - that's the whole point of why it is > useful, at least for single parameters of particular interest. > It explores the objective function surface or actually the optimal objective > function and parameters conditioned on a selected parameter moving > along a line that extends to points that are potentially far away from the > original optimum. If you are willing to focus on just a few parameters > of interest, it actually avoids some of the potential pitfalls of bootstrapping > . Moreover, it is theoretically quite sound (no asymptotic arguments > required) as long as the objective function is really the log likelihood or a > decent approximation to it. > > -----Original Message----- > From: Leonid Gibiansky [mailto:[email protected]] > Sent: Thursday, February 12, 2015 10:09 AM > To: Bob Leary; Chaouch Aziz; Eleveld, DJ; [email protected] > Cc: [email protected] > Subject: Re: [NMusers] RE: Standard errors of estimates for strictly positive > parameters > > I think we are back to the discussion of usefulness of SE/RSE provided by > Nonmem (search in archives will recover many e-mails on the subject). > I am reluctant to disagree with Lewis Sheiner (if this indeed was his citation > that was not taken out of context) but SEs are very useful (not perfect but > useful) as an indicator of identifiability of the problem and some measure of > precision. One cannot do bootstrap on each and every step of the modeling > (moreover, bootstrap CI are not that different from asymptotic CI, especially > for well-estimated parameters). Profiling is also local, and unlikely to give > something significantly different from the Nonmem SE results unless RSEs are > very large (and profiling is as time-consuming as bootstrap). So SEs serve as a > very useful indicator whether data support the model. > > Note that Nonmem SEs are local; they rely on approximation of the OF surface, > essentially, give the curvature of this surface, so they do not take into > account any constrains. In this sense, form of the model > (CL=THETA(1) or CL=log(THETA(2)) just define the rule of extrapolation beyond > the locality of the estimate. For practical purposes, SEs of these two > representations are related (approximately) as > > SE(THETA2)=SE(THETA(1))/THETA(1) > > (propagation of error rule). Given SE(THETA(1)) and THETA(1) you can estimate > SE(THETA(2)) and simulate accordingly. > > Thanks > Leonid > > -------------------------------------- > Leonid Gibiansky, Ph.D. > President, QuantPharm LLC > web: www.quantpharm.com > e-mail: LGibiansky at quantpharm.com > tel: (301) 767 5566 > > On 2/12/2015 9:04 AM, Bob Leary wrote: > > > Dear Aziz - > > The approximate likelihood methods in NONMEM such as FO, FOCE,and > > LAPLACE optimize an objective function than is parameterized > > internally by the Cholesky factor L of Omega, regardless of whether > > the matrix is diagonal (the EM -based methods do something > > considerably different and work directly with Omega rather than the > > Cholesky factor.) > > > > Thus for the approximate likelihood methods, the SE's computed > > internally by $COV from the Hessian or Sandwich or Fisher score > > methods are first computed with respect to these Cholesky parameters , and then > > the corresponding SE's of the full Omega=LL' are computed by a 'propagation of > > errors' approach which skews the results, particularly if the SE's are large. > > Thus in a sense regarding your dilemma about whether Model 1 or Model 2 is > > better with respect to applicability of $COV results, one answer is that both > > are fundamentally distorted by the propagation of errors method with respect > > to the Omega elements. > > > > But regarding your fundamental question 'can we trust the output of $COV '- > > all of this makes very little difference. Standard errors computed by $COV are > > inherently dubious - the applicability of the usual asymptotic arguments is > > very questionable for the types/sizes of data sets we often deal with. > > As Lewis Sheiner used to say of these results, 'they are not worth the > > electrons used to compute them'. They are the best we can do for the level > > of computational investment put into them - > > If you want something better, try a bootstrap or profiling method. > > > > ________________________________________ > > > >