SAEM and IMP

9 messages 6 people Latest: May 19, 2014

SAEM and IMP

From: Pavel Belo Date: May 15, 2014 technical
Hello NONMEM Users, As SAEM does not provide a useful objective function, the manuals recommend using IMP after SAEM. It works well in many cases when IMP works well. When IMP works well, SAEM is not always needed. SAEM is really needed when the other methods do not work well. The issue is that there are hard cases when SAEM works very well and IMP does not work at all. SAEM provides meaningful and consistent PK/PD parameters across very different runs, while IMP provides objective function, which varies so greatly that it looks meaningless. Another potential issue with IMP is that even when it works well with a problem, it occasionally provides low values of objective functions after SAEM or as the first estimate (less frequently in the middle of IMP run) and then becomes unstable or jumps to much higher objective function and then converges to something between the low and the high values for a long time. It almost looks like IMP does not show some kind of integration/computation errors and keeps running providing a funny objective function. It seems like we cannot estimate objective function when SAEM runs well and IMP does not. It reduces the value of SAEM. Is there a way around it? Thanks, Pavel

RE: SAEM and IMP

From: Robert Bauer Date: May 15, 2014 technical
Pavel: Try using the latest nonmem 7.3 version if you can, and see if this still occurs. Improvements are continually being made. Try also the following after an SAEM step: $EST METHOD=IMP EONLY=1 MAPITER=0 (...plus other options as neded) Or $EST METHOD=IMPMAP EONLY=1 (...plus other options as needed) Robert J. Bauer, Ph.D. Vice President, Pharmacometrics, R&D ICON Development Solutions 7740 Milestone Parkway Suite 150 Hanover, MD 21076 Tel: (215) 616-6428 Mob: (925) 286-0769 Email: [email protected] Web: www.iconplc.com
Quoted reply history
-----Original Message----- From: [email protected] [mailto:[email protected]] On Behalf Of Pavel Belo Sent: Thursday, May 15, 2014 11:02 AM To: [email protected] Subject: [NMusers] SAEM and IMP Hello NONMEM Users, As SAEM does not provide a useful objective function, the manuals recommend using IMP after SAEM. It works well in many cases when IMP works well. When IMP works well, SAEM is not always needed. SAEM is really needed when the other methods do not work well. The issue is that there are hard cases when SAEM works very well and IMP does not work at all. SAEM provides meaningful and consistent PK/PD parameters across very different runs, while IMP provides objective function, which varies so greatly that it looks meaningless. Another potential issue with IMP is that even when it works well with a problem, it occasionally provides low values of objective functions after SAEM or as the first estimate (less frequently in the middle of IMP run) and then becomes unstable or jumps to much higher objective function and then converges to something between the low and the high values for a long time. It almost looks like IMP does not show some kind of integration/computation errors and keeps running providing a funny objective function. It seems like we cannot estimate objective function when SAEM runs well and IMP does not. It reduces the value of SAEM. Is there a way around it? Thanks, Pavel

RE: SAEM and IMP

From: Brian Sadler Date: May 15, 2014 technical
Pavel, You might try IMP after SAEM with EONLY=1 (and a limited number of iterations) so that the parameters remain stationary and only the expectation step is executed. Cheers... Brian
Quoted reply history
-----Original Message----- From: [email protected] [mailto:[email protected]] On Behalf Of Pavel Belo Sent: Thursday, May 15, 2014 11:02 AM To: [email protected] Subject: [NMusers] SAEM and IMP Hello NONMEM Users, As SAEM does not provide a useful objective function, the manuals recommend using IMP after SAEM. It works well in many cases when IMP works well. When IMP works well, SAEM is not always needed. SAEM is really needed when the other methods do not work well. The issue is that there are hard cases when SAEM works very well and IMP does not work at all. SAEM provides meaningful and consistent PK/PD parameters across very different runs, while IMP provides objective function, which varies so greatly that it looks meaningless. Another potential issue with IMP is that even when it works well with a problem, it occasionally provides low values of objective functions after SAEM or as the first estimate (less frequently in the middle of IMP run) and then becomes unstable or jumps to much higher objective function and then converges to something between the low and the high values for a long time. It almost looks like IMP does not show some kind of integration/computation errors and keeps running providing a funny objective function. It seems like we cannot estimate objective function when SAEM runs well and IMP does not. It reduces the value of SAEM. Is there a way around it? Thanks, Pavel

Re: SAEM and IMP

From: Emmanuel Chigutsa Date: May 15, 2014 technical
Hi Pavel I have experienced a similar problem. In my case, the following code for IMP after SAEM (using NM7.3) greatly reduced the Monte Carlo OFV noise from variations of about +/- 60 points to variations of +/- 6 points (though still not good enough for covariate testing): $EST METHOD=IMP LAPLACE INTER NITER=15 ISAMPLE=3000 EONLY=1 DF=7 IACCEPT=0.3 ISAMPEND=10000 STDOBJ=2 MAPITER=0 PRINT=1 SEED=123456 RANMETHOD=3S2 The settings are explained in the NM7.3 guide. If you are using NM7.3, you can also try IACCEPT=0.0 whereupon "NONMEM will determine the most appropriate IACCEPT level for each subject". Of course the settings for DF and IACCEPT in the above code will depend on the type of data you have. Which brings me to my own question. If I have both continous and categorical DVs in the dataset (which would mean different optimal settings) and I am using F_FLAG accordingly, what would the 'right' values of DF and IACCEPT be? I have noticed that the DF automatically chosen by NONMEM for individuals in the dataset can vary from 0-8 and this appears to be random. Emmanuel
Quoted reply history
From: Pavel Belo <[email protected]> >To: [email protected] >Sent: Thursday, May 15, 2014 11:01 AM >Subject: [NMusers] SAEM and IMP > > >Hello NONMEM Users, > >As SAEM does not provide a useful objective function, the manuals >recommend using IMP after SAEM. It works well in many cases when IMP >works well. When IMP works well, SAEM is not always needed. SAEM is >really needed when the other methods do not work well. > >The issue is that there are hard cases when SAEM works very well and IMP >does not work at all. SAEM provides meaningful and consistent PK/PD >parameters across very different runs, while IMP provides objective >function, which varies so greatly that it looks meaningless. Another >potential issue with IMP is that even when it works well with a problem, >it occasionally provides low values of objective functions after SAEM or >as the first estimate (less frequently in the middle of IMP run) and >then becomes unstable or jumps to much higher objective function and >then converges to something between the low and the high values for a >long time. It almost looks like IMP does not show some kind of >integration/computation errors and keeps running providing a funny >objective function. > >It seems like we cannot estimate objective function when SAEM runs well >and IMP does not. It reduces the value of SAEM. Is there a way around >it? > >Thanks, >Pavel > > >

RE: SAEM and IMP

From: Bob Leary Date: May 15, 2014 technical
Hi Emmanuel, While I am a strong advocate of using quasi-random rather than pseudo- random sequences for importance sampling in EM methods like IMP, there is a theoretical (and very real) problem with their use in the context you suggested in your message, namely with a multivariate t distribution as the importance sampling distribution. The 3S2 option implies you are using a Sobol quasi-random sequence, while the DF=7 implies the use of a multivariate T-distribution with 7 degrees of freedom. The standard way of generating a p-dimensional multivariate t -random variable with DF degrees of freedom is to generate a p-dimensional multivariate normal and then divide by an additional independent random variable which is basically the square root of a 1-d chi square random variable with DF degrees of freedom. Thus to generate a p-dimensional importance sample, you actually need to use p+1 independent random variables. If you simply use a p+1 dimensional Sobol vector as the base quasi-random draw, the nonlinear mapping from p+1 dimensions to the final p dimensional result destroys the low discrepancy property of the final sequence in the p-dimensional space and in fact introduces a significant amount of bias in the final result. The problem arises directly from the p+1 vs p dimensional mismatch. There is no problem if the final p-dimensional result can be generated from a p-dimensional quasi-random sequence, which is the case for multivariate normal Importance samples. So quasi random sequences should really only be used for the DF=0 multivariate normal importance sampling distribution case, not the multivariate DF>0 multivariate t case. I ran across this effect in testing the Sobol-based importance sampling EM algorithm QRPEM in Phoenix NLME. It is very real and the net effect is to introduce a significant bias. There is a partial fix that works but gives up some of the benefit of using low-discrepancy sequences - namely use a p-dimensional quasi-random vector to generate the p-dimensional multivariate normal, but then use a 1-d pseudo-random sequence to generate the chi-square random variable.
Quoted reply history
From: [email protected] [mailto:[email protected]] On Behalf Of Emmanuel Chigutsa Sent: Thursday, May 15, 2014 1:03 PM To: Pavel Belo; [email protected] Subject: Re: [NMusers] SAEM and IMP Hi Pavel I have experienced a similar problem. In my case, the following code for IMP after SAEM (using NM7.3) greatly reduced the Monte Carlo OFV noise from variations of about +/- 60 points to variations of +/- 6 points (though still not good enough for covariate testing): $EST METHOD=IMP LAPLACE INTER NITER=15 ISAMPLE=3000 EONLY=1 DF=7 IACCEPT=0.3 ISAMPEND=10000 STDOBJ=2 MAPITER=0 PRINT=1 SEED=123456 RANMETHOD=3S2 The settings are explained in the NM7.3 guide. If you are using NM7.3, you can also try IACCEPT=0.0 whereupon "NONMEM will determine the most appropriate IACCEPT level for each subject". Of course the settings for DF and IACCEPT in the above code will depend on the type of data you have. Which brings me to my own question. If I have both continous and categorical DVs in the dataset (which would mean different optimal settings) and I am using F_FLAG accordingly, what would the 'right' values of DF and IACCEPT be? I have noticed that the DF automatically chosen by NONMEM for individuals in the dataset can vary from 0-8 and this appears to be random.

RE: SAEM and IMP

From: Joseph Standing Date: May 15, 2014 technical
Dear Emmanuel and Pavel, Further to Bob's answer, recall also that delta OFV in the likelihood ratio test is only asymptotocaily chi squared distributed,and this is not the only reason why you should not get too hung up on OFV to help choose your models. For example, Lavielle 2010 in Biometrics showed nicely how the SAEM algorithm can estimate parameters of complex differential equation models for joint HIV viral load and CD4 counts. Using OFV-based metrics a latent model was chosen whereby the majority of circulating T-cells were infected with virus - this model also gave nice fits to the data. When immunologists look at CD4 cells of HIV infected patients however, they find that much less than 1% (closer to 0.01%) of circulating T-cells contain virus (most of the virus making up the latent reservoir is stuck to folicular cells in the periphery), so one would have to question the meaning of the parameters identified as the best fit by SAEM. By all means use SAEM to fit ODE models that don't run/converge with other algorithms (I do), but choose models with parameters that make mechanistic sense rather than relying too heavily on OFV-based metrics. A nice VPC always goes down well too. Joe Joseph F Standing MRC Fellow, UCL Institute of Child Health Antimicrobial Pharmacist, Great Ormond Street Hospital Tel: +44(0)207 905 2370 Mobile: +44(0)7970 572435
Quoted reply history
________________________________________ From: [email protected] [[email protected]] On Behalf Of Bob Leary [[email protected]] Sent: 15 May 2014 19:22 To: Emmanuel Chigutsa; Pavel Belo; [email protected] Subject: RE: [NMusers] SAEM and IMP Hi Emmanuel, While I am a strong advocate of using quasi-random rather than pseudo- random sequences for importance sampling in EM methods like IMP, there is a theoretical (and very real) problem with their use in the context you suggested in your message, namely with a multivariate t distribution as the importance sampling distribution. The 3S2 option implies you are using a Sobol quasi-random sequence, while the DF=7 implies the use of a multivariate T-distribution with 7 degrees of freedom. The standard way of generating a p-dimensional multivariate t -random variable with DF degrees of freedom is to generate a p-dimensional multivariate normal and then divide by an additional independent random variable which is basically the square root of a 1-d chi square random variable with DF degrees of freedom. Thus to generate a p-dimensional importance sample, you actually need to use p+1 independent random variables. If you simply use a p+1 dimensional Sobol vector as the base quasi-random draw, the nonlinear mapping from p+1 dimensions to the final p dimensional result destroys the low discrepancy property of the final sequence in the p-dimensional space and in fact introduces a significant amount of bias in the final result. The problem arises directly from the p+1 vs p dimensional mismatch. There is no problem if the final p-dimensional result can be generated from a p-dimensional quasi-random sequence, which is the case for multivariate normal Importance samples. So quasi random sequences should really only be used for the DF=0 multivariate normal importance sampling distribution case, not the multivariate DF>0 multivariate t case. I ran across this effect in testing the Sobol-based importance sampling EM algorithm QRPEM in Phoenix NLME. It is very real and the net effect is to introduce a significant bias. There is a partial fix that works but gives up some of the benefit of using low-discrepancy sequences – namely use a p-dimensional quasi-random vector to generate the p-dimensional multivariate normal, but then use a 1-d pseudo-random sequence to generate the chi-square random variable. From: [email protected] [mailto:[email protected]] On Behalf Of Emmanuel Chigutsa Sent: Thursday, May 15, 2014 1:03 PM To: Pavel Belo; [email protected] Subject: Re: [NMusers] SAEM and IMP Hi Pavel I have experienced a similar problem. In my case, the following code for IMP after SAEM (using NM7.3) greatly reduced the Monte Carlo OFV noise from variations of about +/- 60 points to variations of +/- 6 points (though still not good enough for covariate testing): $EST METHOD=IMP LAPLACE INTER NITER=15 ISAMPLE=3000 EONLY=1 DF=7 IACCEPT=0.3 ISAMPEND=10000 STDOBJ=2 MAPITER=0 PRINT=1 SEED=123456 RANMETHOD=3S2 The settings are explained in the NM7.3 guide. If you are using NM7.3, you can also try IACCEPT=0.0 whereupon "NONMEM will determine the most appropriate IACCEPT level for each subject". Of course the settings for DF and IACCEPT in the above code will depend on the type of data you have. Which brings me to my own question. If I have both continous and categorical DVs in the dataset (which would mean different optimal settings) and I am using F_FLAG accordingly, what would the 'right' values of DF and IACCEPT be? I have noticed that the DF automatically chosen by NONMEM for individuals in the dataset can vary from 0-8 and this appears to be random.

RE: SAEM and IMP

From: Robert Bauer Date: May 16, 2014 technical
Bob: Yes, when I was developing the code for using the Sobol method in NONMEM, I too found that using the t-distribution with Sobol/Quasi normal method resulted in biased assessments of the objective function, when I used the unmodified standard technique of generating t-samples. So, I experimented with various alternative methods like you did, and the final algorithms used in NONMEM 7.2 and 7.3 appear to avoid the bias, at least when I tried some simple models, such as the example1 problem in the NONMEM ..\examples directory. For such a problem, I tried the following DF settings, with and without using the SOBOL method (RANMETHOD=3 (default), or RANMETHOD=3S2), DF=0, 1, 2, 3, 4, 5, 6, 7, 8, 10, 11, and 12. I think beyond 12 DF is sufficiently normal like, and not much use. In all cases, the OBJF was within 1 unit of each other, and always within a STD of the Monte Carlo noise to the OBJF (I used EONLY=1, 20 iterations to obtain OBJFV replicates for each setting). By the way, the .cnv file publishes the mean and standard deviation of the last CITER objective function values (see nm730.pdf manual on root.cnv). Certainly it is worth assuring this lack of bias under other kinds of, and more complex, models, just to make sure, so it is worth while for modelers to use caution when mixing Sobol with DF>0. I should point out that odd-number DF’s such as 7 makes the corrective algorithm for Sobol with t-distributions work harder, whereas even-numbered DF’s are a little cleaner. So I recommend even-numbered DF’s, or 1, up to 10 to work with Sobol, at least as far as my algorithm is concerned (so DF=1,2,4,6,8,10 seem most efficient to use with Sobol). The algorithm I use for t-distributions is in the subroutine TDEV2, and is in ..source\general.f90. I publish it because I borrowed freely from the literature, and I consider it therefore “open-source”. Just a note for orienting you to the code, IRANM=5 refers to the Sobol method. I am afraid my theoretical expertise in the Sobol/Quasi random methods is not as good as yours, so please feel free to review the code, and perhaps let me and the modeling community know if there may be any potential pitfalls in using this t-distribution algorithm with Sobol. I am also happy to consider improvements you may come up with on this matter. Robert J. Bauer, Ph.D. Vice President, Pharmacometrics, R&D ICON Development Solutions 7740 Milestone Parkway Suite 150 Hanover, MD 21076 Tel: (215) 616-6428 Mob: (925) 286-0769 Email: [email protected]<mailto:[email protected]> Web: http://www.iconplc.com/

RE: SAEM and IMP

From: Bob Leary Date: May 16, 2014 technical
Bob, thanks for your reply and pointer to the source code for tdev2 . I have looked at it but perhaps you can clarify what the source of tdev2 is and how it is used (I cannot see any calls made to it in your source code, so I am not really sure how it is being used). I can see for DF =1, tdev2 returns a Cauchy distributed random variable, which is the 1D t-distribution for DF=1, and for t=2 it looks like it returns the 1D t-distribution for DF=2. For the other DF values it is not so clear, but I am guessing it returns a 1D t-distribution with the specified DF. If you simply assemble a p-dimensional sample vector with components which are independent 1D t random variables with a specified DF , indeed it will avoid the dimensional mismatch and work OK without bias in the QR case, but it will not be a true multivariate t-distribution (the components of a multivariate t-distribution are not independent). The usual argument for using a multivariate t as an importance sampler is that a) It has fatter tails than a multivariate normal b) It preserves the radial symmetry of the multivariate normal in the case where the covariance matrix is the unit matrix (i.e. the density is only a function of the distance from the origin) Assembling a p-dimensional sample vector whose individual components are 1D t distributions with a specified DF satisfies a), but not b). It’s a perfectly valid importance sampling distribution, and indeed has been suggested in the literature as much simpler alternative to the multivariate t, particularly in low dimensions, but it is not a multivariate t (which is the more usual meaning of ‘t-distribution’ in the multivariate importance sampling context) . It is not radially symmetric (which may or may not be important, depending on the posterior multidimensional shape and the dimensionality. There is a directionality bias which increases with dimensionality) The documentation is somewhat ambiguous – DF=4 The proposal density is to be t distribution with 4 degrees of freedom. Based on the code I have seen, I now guess this means that each component of the proposal is an independent t random variable with DF=4. But I am only guessing – perhaps you could clarify what distribution is really being referred to by ‘t-distribution’ in the p-dimensional case – multivariate p-dimensional t with specified DF or Cartesian product of p independent 1D t random variables with specified DF?
Quoted reply history
From: [email protected] [mailto:[email protected]] On Behalf Of Bauer, Robert Sent: Thursday, May 15, 2014 9:57 PM To: [email protected] Subject: RE: [NMusers] SAEM and IMP Bob: Yes, when I was developing the code for using the Sobol method in NONMEM, I too found that using the t-distribution with Sobol/Quasi normal method resulted in biased assessments of the objective function, when I used the unmodified standard technique of generating t-samples. So, I experimented with various alternative methods like you did, and the final algorithms used in NONMEM 7.2 and 7.3 appear to avoid the bias, at least when I tried some simple models, such as the example1 problem in the NONMEM ..\examples directory. For such a problem, I tried the following DF settings, with and without using the SOBOL method (RANMETHOD=3 (default), or RANMETHOD=3S2), DF=0, 1, 2, 3, 4, 5, 6, 7, 8, 10, 11, and 12. I think beyond 12 DF is sufficiently normal like, and not much use. In all cases, the OBJF was within 1 unit of each other, and always within a STD of the Monte Carlo noise to the OBJF (I used EONLY=1, 20 iterations to obtain OBJFV replicates for each setting). By the way, the .cnv file publishes the mean and standard deviation of the last CITER objective function values (see nm730.pdf manual on root.cnv). Certainly it is worth assuring this lack of bias under other kinds of, and more complex, models, just to make sure, so it is worth while for modelers to use caution when mixing Sobol with DF>0. I should point out that odd-number DF’s such as 7 makes the corrective algorithm for Sobol with t-distributions work harder, whereas even-numbered DF’s are a little cleaner. So I recommend even-numbered DF’s, or 1, up to 10 to work with Sobol, at least as far as my algorithm is concerned (so DF=1,2,4,6,8,10 seem most efficient to use with Sobol). The algorithm I use for t-distributions is in the subroutine TDEV2, and is in ..source\general.f90. I publish it because I borrowed freely from the literature, and I consider it therefore “open-source”. Just a note for orienting you to the code, IRANM=5 refers to the Sobol method. I am afraid my theoretical expertise in the Sobol/Quasi random methods is not as good as yours, so please feel free to review the code, and perhaps let me and the modeling community know if there may be any potential pitfalls in using this t-distribution algorithm with Sobol. I am also happy to consider improvements you may come up with on this matter. Robert J. Bauer, Ph.D. Vice President, Pharmacometrics, R&D ICON Development Solutions 7740 Milestone Parkway Suite 150 Hanover, MD 21076 Tel: (215) 616-6428 Mob: (925) 286-0769 Email: [email protected]<mailto:[email protected]> Web: http://www.iconplc.com/

RE: SAEM and IMP

From: Robert Bauer Date: May 19, 2014 technical
Third attempt to send to nmusers. Ignore if you have already received this message. Bob: First. My apologies in case you are getting repeated e-mails from me. I am finding I need to attempt several times for nmusers list-serv system to publish my e-mails. Something to do with >40000 characters. So I need to cut off some of the trailing e-mail portions that have already been published. In my original investigations, the normal and t-distribution were using Box-Mueller type random sample generations, which of course did not work well, so I switched to inverse CDF methods. For normal distribution (DF=0), the IMP uses the modified GASDEV, where you see that when ISELECT=5 (Sobol), DINVNORM is used, and the the Box-Mueller method is used when Sobol method is not used. Robert J. Bauer, Ph.D. Vice President, Pharmacometrics, R&D ICON Development Solutions 7740 Milestone Parkway Suite 150 Hanover, MD 21076 Tel: (215) 616-6428 Mob: (925) 286-0769 Email: [email protected]<mailto:[email protected]> Web: http://www.iconplc.com/
Quoted reply history
From: Bob Leary [mailto:[email protected]] Sent: Monday, May 19, 2014 2:36 PM To: Bauer, Robert; [email protected]<mailto:[email protected]> Subject: RE: [NMusers] SAEM and IMP Thanks, Bob. This indeed has been an interesting and thought provoking discussion. I too took another look at this. I thought, without really doing any analysis, that the directionality bias using independent univariate t-components would get really bad as the dimensionality increased, because that’s the way it works with using fat-tailed independent double exponentials (samples there tend to fail disproportionately near the coordinate axes, and this gets arbitrarily bad as the dimensionality increases without bound). But when I actually did the analysis in the student t-case, it’s not bad at all and simply goes to a limit as the dimensionality increases. So in fact, there may be some real advantages, particularly in the Sobol quasi-random case, to doing it the way you do with independent univariate t’s rather than using the multivariate t. So my original message to Emmanuel to discourage him from using Sobol in combination with the t-distribution was based on the erroneous assumption that IMP was using the multivariate-t rather than Independent univariate t’s. So his suggestion is probably OK (except as you noted, that an even value of DF might be cleaner than DF=7 given the way tdev2 works). But there is still the question of why you were seeing a bias with your original method before going to tdev2 –was it simply a different method of generating the same distribution? Along somewhat similar lines, I did notice that your routine GASDEV for generating normal N(0,1) components (presumably used in the DF=0 case) actually offers two different ways of doing it – Box-Muller and inverse cdf of a uniform 0-1. It is not clear which one you get when you simply specify DF=0. The inverse cdf method of course is by definition OK for the quasi-random case – the cdf of the random vectors generated with the Sobol distribution using the inverse cdf method will simply be that Sobol distribution. But for Box Muller, this certainly will not be the case, and it is not clear what the discrepancy of the Box Muller generated CDF will be like. Certainly Box Muller is obviously OK in the pseudorandom case, and will have the same discrepancy as the inverse cdf method using pseudorandom uniform 0-1 inputs, but whether the CDF of a Box Muller – generated sequence of Normal random vectors using Sobol vector inputs is actually a low discrepancy sequence (or even a relatively low discrepancy sequence) is not at all clear. I know this has been looked at in 2-dimensions and there Box Muller empirically looks OK, but I don’t know if anything has actually been proved. The low discrepancy property tends to be fragile with respect to nonlinear transformations. So it is possible that tdev2 actually does preserve this property , or at least does relatively little damage to it compared to whatever you were using before. From: Bauer, Robert [mailto:[email protected]] Sent: Monday, May 19, 2014 11:08 AM To: Bob Leary; [email protected]<mailto:[email protected]> Subject: RE: [NMusers] SAEM and IMP Bob: Yes, these are 1 Dimensional t-distribution random deviates generated by TDEV2, followed by scaling p sets of them with an offset vector and cholesky matrix to provide covariance and mean. It provides the tails as you say (property a), and is conveniently used in the Sobol process. Your discussion on the properties of various t-distribution sample creation techniques and their possible impact in importance sampling is interesting. With that in mind, I just completed testing example6, which has an 8 dimensional eta space, and found no bias in the objective function evaluation, or in parameter assessment when setting DF=2 or 4 (I have not tried others), with or without using Sobol (RANMETHOD=3S2). Also, using the 1 dimensional t-distribution random deviates retained the Sobol method’s stochastic noise reduction ability in this example. When I used the multivariate t-distribution algorithm, Sobol’s stochastic noise reduction was not as good, confirming what you related earlier. As you point out as well, property b (radial symmetry) may or may not be needed depending on the posterior density. I am inclined to think that since the posterior density is not going to be perfectly t-distributed or normal distributed anyway, then the sampling density matching the posterior density regarding the radial symmetry property may be of less of relevance. The more important aspect may be to choose a sampling density that has long tails in cases where the posterior density also has long tails, to promote the general efficiency of the sampling density for that posterior density. It is possible that because the sampling density is also dynamically scaled to best fit the posterior density, this reduces any inefficiency in fitting the posterior density that might occur from the sampling density not having radial symmetry. Robert J. Bauer, Ph.D. Vice President, Pharmacometrics, R&D ICON Development Solutions 7740 Milestone Parkway Suite 150 Hanover, MD 21076 Tel: (215) 616-6428 Mob: (925) 286-0769 Email: [email protected]<mailto:[email protected]> Web: http://www.iconplc.com/