Choice of models

6 messages 6 people Latest: Jan 24, 2012

Choice of models

From: Toufigh Gordi Date: January 24, 2012 technical
Dear all, I have a general question on the choice of model in a population analysis. I have a set of data set that includes a large number of studies with about ¾ of the data from extensive sampling schemes (phase 1, 2, and 3 studies) and the rest from sparse samples (phase 3 clinical studies). When developing the PK model, a model on the extensive samples only fits the data well and I can get quite reasonable parameter estimates, including covariate effects, and a successful $COV (NONMEM). When all data is used, the model becomes somewhat instable: the same covariates are identified but the model becomes quite sensitive to the initial estimates and the $COV step won’t go through. I could, of course, perform a bootstrap to go around this issue. In general, the fit of the model based on the full data set is not as good as the extensive data set model, although the two models are rather similar with regard to the parameter estimates. However, the range of estimated parameters is wider when using all data and noticeably KA and V2 are skewed to very larger values. Moving forward, I could either use the full data model and simulate steady state profiles for the phase 3 study (sparse samples) data. Or, I could use the model based on the extensive samples only, use the sparse data and generate post-hoc estimates for the sparsely sampled individuals and move forward that way. The advantage with the first option is that all the available data have been used in the modeling process. The disadvantage would be that the model is not as good as the other model, with sparse data distorting the parameter estimates. The advantage of the second option is that the model performs better and there is really no reason why the underlying PK model for the sparsely sampled subjects should be different, which means one should be able to use that model to generate post-hoc estimates. The disadvantage is that not all the available data have been used in the model building process. It would be interesting to hear other people’s thoughts and ideas on this. Toufigh

Re: Choice of models

From: Bill Denney Date: January 24, 2012 technical
Hi Toufigh, I typically think that data quality decreases with phase and with sampling frequency. Given what you described below, I'd think that you're fighting data quality in the sparse, phase 3 studies, and with the parameters you're describing as having trouble, it seems to support that thought. Were I to guess, you could probably pick out the most influential 3% of sparse samples (arbitrary percentage), and look at them in more detail and find that they look more like Cmax than Ctrough or something such the time since last dose appears to be off. Beyond that, philosophically, I think that trough concentrations should not be allowed to affect Ka because the effect is usually so small as not to be measurable (assuming that we're discussing a drug with a reasonable separation between the alpha elimination phase and measurement time). Thanks, Bill
Quoted reply history
On Jan 23, 2012, at 11:45 PM, "Toufigh Gordi" <[email protected]> wrote: > Dear all, > > I have a general question on the choice of model in a population analysis. I > have a set of data set that includes a large number of studies with about ¾ > of the data from extensive sampling schemes (phase 1, 2, and 3 studies) and > the rest from sparse samples (phase 3 clinical studies). When developing the > PK model, a model on the extensive samples only fits the data well and I can > get quite reasonable parameter estimates, including covariate effects, and a > successful $COV (NONMEM). When all data is used, the model becomes somewhat > instable: the same covariates are identified but the model becomes quite > sensitive to the initial estimates and the $COV step won’t go through. I > could, of course, perform a bootstrap to go around this issue. In general, > the fit of the model based on the full data set is not as good as the > extensive data set model, although the two models are rather similar with > regard to the parameter estimates. However, the range of estimated parameters > is wider when using all data and noticeably KA and V2 are skewed to very > larger values. > > Moving forward, I could either use the full data model and simulate steady > state profiles for the phase 3 study (sparse samples) data. Or, I could use > the model based on the extensive samples only, use the sparse data and > generate post-hoc estimates for the sparsely sampled individuals and move > forward that way. The advantage with the first option is that all the > available data have been used in the modeling process. The disadvantage would > be that the model is not as good as the other model, with sparse data > distorting the parameter estimates. The advantage of the second option is > that the model performs better and there is really no reason why the > underlying PK model for the sparsely sampled subjects should be different, > which means one should be able to use that model to generate post-hoc > estimates. The disadvantage is that not all the available data have been used > in the model building process. > > It would be interesting to hear other people’s thoughts and ideas on this. > > Toufigh

Re: Choice of models

From: Jan-Stefan van der Walt Date: January 24, 2012 technical
Hi Toufigh, Recently I used the 90% prediction interval (generated by an appropriately binned VPC) of the rich data (three studies with observed doses) to evaluate the sparse data (one sample on 4 occasions). The sparse data contained more information about the covariates of interest, but the dosing was unobserved. I analysed the rich and sparse data simultaneously first including and then excluding the sparse data outside the 90% PI and compared the results. The eta-shrinkage values decreased considerably when the observations outside the 90% PI were excluded and I had more confidence in the covariate relationships. As a side issue, I estimated a time-after-dose for the observations outside the 90% PI. It was interesting that the difference between the reported and estimated dosing times seemed to increase as the trial progressed (0.92h [month 6], 1.05h [month 12]), 1.11h [month 18] and 3.6h [month 24]. Hope this helps. Regards, Jan-Stefan
Quoted reply history
On 24 January 2012 05:05, Denney, William S. <[email protected]>wrote: > Hi Toufigh, > > I typically think that data quality decreases with phase and with sampling > frequency. Given what you described below, I'd think that you're fighting > data quality in the sparse, phase 3 studies, and with the parameters you're > describing as having trouble, it seems to support that thought. Were I to > guess, you could probably pick out the most influential 3% of sparse > samples (arbitrary percentage), and look at them in more detail and find > that they look more like Cmax than Ctrough or something such the time since > last dose appears to be off. > > Beyond that, philosophically, I think that trough concentrations should > not be allowed to affect Ka because the effect is usually so small as not > to be measurable (assuming that we're discussing a drug with a reasonable > separation between the alpha elimination phase and measurement time). > > Thanks, > > Bill > > On Jan 23, 2012, at 11:45 PM, "Toufigh Gordi" <[email protected]> > wrote: > > Dear all,**** > > ** ** > > I have a general question on the choice of model in a population analysis. > I have a set of data set that includes a large number of studies with about > ¾ of the data from extensive sampling schemes (phase 1, 2, and 3 studies) > and the rest from sparse samples (phase 3 clinical studies). When > developing the PK model, a model on the extensive samples only fits the > data well and I can get quite reasonable parameter estimates, including > covariate effects, and a successful $COV (NONMEM). When all data is used, > the model becomes somewhat instable: the same covariates are identified but > the model becomes quite sensitive to the initial estimates and the $COV > step won’t go through. I could, of course, perform a bootstrap to go around > this issue. In general, the fit of the model based on the full data set is > not as good as the extensive data set model, although the two models are > rather similar with regard to the parameter estimates. However, the range > of estimated parameters is wider when using all data and noticeably KA and > V2 are skewed to very larger values.**** > > ** ** > > Moving forward, I could either use the full data model and simulate steady > state profiles for the phase 3 study (sparse samples) data. Or, I could use > the model based on the extensive samples only, use the sparse data and > generate post-hoc estimates for the sparsely sampled individuals and move > forward that way. The advantage with the first option is that all the > available data have been used in the modeling process. The disadvantage > would be that the model is not as good as the other model, with sparse data > distorting the parameter estimates. The advantage of the second option is > that the model performs better and there is really no reason why the > underlying PK model for the sparsely sampled subjects should be different, > which means one should be able to use that model to generate post-hoc > estimates. The disadvantage is that not all the available data have been > used in the model building process.**** > > ** ** > > It would be interesting to hear other people’s thoughts and ideas on this. > **** > > ** ** > > Toufigh **** > > -- *United Kingdom* Flat 5, 41 Devons Rd, E3 3BF, London +44 20 7987 6688 *(h) * +44 77 9618 4662 *(m)* *South Africa* Ballet & Lodge, 34 Kerk St, George, 6529 Postnet Suite 39, Private Bag, X6590, George, 6530 +27 44 884 1560 *(h)* *Sweden* Pharmacometrics, Department of Pharmaceutical Biosciences PO Box 591, SE-75124 Uppsala +46 73 066 7338 *(m)*

RE: Choice of models

From: Jean Lavigne Date: January 24, 2012 technical
Dear Toufigh, Here are some of my thoughts: 1 - You may have an issue of quality of your data with the sparse samples (patients with un-accurate recording of evens like dosing and sampling) 2 - You have not specified which method of estimate that was used with NONMEM (FO, FOCE, Bayesian, etc...), I would try with a Bayesian approach from your best model with rich samples. 3 - If you use post-hoc estimates for your sparse data, you may want to double check if they are different from the rest of the population Best regards, Jean
Quoted reply history
From: [email protected] [mailto:[email protected]] On Behalf Of Jan-Stefan Van der Walt Sent: Tuesday, January 24, 2012 5:24 AM To: Toufigh Gordi; [email protected] Subject: Re: [NMusers] Choice of models Hi Toufigh, Recently I used the 90% prediction interval (generated by an appropriately binned VPC) of the rich data (three studies with observed doses) to evaluate the sparse data (one sample on 4 occasions). The sparse data contained more information about the covariates of interest, but the dosing was unobserved. I analysed the rich and sparse data simultaneously first including and then excluding the sparse data outside the 90% PI and compared the results. The eta-shrinkage values decreased considerably when the observations outside the 90% PI were excluded and I had more confidence in the covariate relationships. As a side issue, I estimated a time-after-dose for the observations outside the 90% PI. It was interesting that the difference between the reported and estimated dosing times seemed to increase as the trial progressed (0.92h [month 6], 1.05h [month 12]), 1.11h [month 18] and 3.6h [month 24]. Hope this helps. Regards, Jan-Stefan On 24 January 2012 05:05, Denney, William S. <[email protected]<mailto:[email protected]>> wrote: Hi Toufigh, I typically think that data quality decreases with phase and with sampling frequency. Given what you described below, I'd think that you're fighting data quality in the sparse, phase 3 studies, and with the parameters you're describing as having trouble, it seems to support that thought. Were I to guess, you could probably pick out the most influential 3% of sparse samples (arbitrary percentage), and look at them in more detail and find that they look more like Cmax than Ctrough or something such the time since last dose appears to be off. Beyond that, philosophically, I think that trough concentrations should not be allowed to affect Ka because the effect is usually so small as not to be measurable (assuming that we're discussing a drug with a reasonable separation between the alpha elimination phase and measurement time). Thanks, Bill On Jan 23, 2012, at 11:45 PM, "Toufigh Gordi" <[email protected]<mailto:[email protected]>> wrote: Dear all, I have a general question on the choice of model in a population analysis. I have a set of data set that includes a large number of studies with about ¾ of the data from extensive sampling schemes (phase 1, 2, and 3 studies) and the rest from sparse samples (phase 3 clinical studies). When developing the PK model, a model on the extensive samples only fits the data well and I can get quite reasonable parameter estimates, including covariate effects, and a successful $COV (NONMEM). When all data is used, the model becomes somewhat instable: the same covariates are identified but the model becomes quite sensitive to the initial estimates and the $COV step won't go through. I could, of course, perform a bootstrap to go around this issue. In general, the fit of the model based on the full data set is not as good as the extensive data set model, although the two models are rather similar with regard to the parameter estimates. However, the range of estimated parameters is wider when using all data and noticeably KA and V2 are skewed to very larger values. Moving forward, I could either use the full data model and simulate steady state profiles for the phase 3 study (sparse samples) data. Or, I could use the model based on the extensive samples only, use the sparse data and generate post-hoc estimates for the sparsely sampled individuals and move forward that way. The advantage with the first option is that all the available data have been used in the modeling process. The disadvantage would be that the model is not as good as the other model, with sparse data distorting the parameter estimates. The advantage of the second option is that the model performs better and there is really no reason why the underlying PK model for the sparsely sampled subjects should be different, which means one should be able to use that model to generate post-hoc estimates. The disadvantage is that not all the available data have been used in the model building process. It would be interesting to hear other people's thoughts and ideas on this. Toufigh -- United Kingdom Flat 5, 41 Devons Rd, E3 3BF, London +44 20 7987 6688 (h) +44 77 9618 4662 (m) South Africa Ballet & Lodge, 34 Kerk St, George, 6529 Postnet Suite 39, Private Bag, X6590, George, 6530 +27 44 884 1560 (h) Sweden Pharmacometrics, Department of Pharmaceutical Biosciences PO Box 591, SE-75124 Uppsala +46 73 066 7338 (m) This electronic transmission may contain confidential and/or proprietary information and is intended to be for the use of the individual or entity named above. If you are not the intended recipient, be aware that any disclosure, copying, distribution or use of the contents of this electronic transmission is prohibited. If you have received this electronic transmission in error, please destroy it and immediately notify us of the error. Thank you.

RE: Choice of models

From: Juan Jose Perez Ruixo Date: January 24, 2012 technical
Dears, Sometimes assigning different residual variability for rich vs sparse data helps to make the model more stable. Regards, Juan.
Quoted reply history
From: [email protected] [mailto:[email protected]] On Behalf Of Lavigne, Jean Sent: martes, 24 de enero de 2012 8:21 To: Toufigh Gordi; [email protected] Subject: RE: [NMusers] Choice of models Dear Toufigh, Here are some of my thoughts: 1 - You may have an issue of quality of your data with the sparse samples (patients with un-accurate recording of evens like dosing and sampling) 2 - You have not specified which method of estimate that was used with NONMEM (FO, FOCE, Bayesian, etc...), I would try with a Bayesian approach from your best model with rich samples. 3 - If you use post-hoc estimates for your sparse data, you may want to double check if they are different from the rest of the population Best regards, Jean From: [email protected] [mailto:[email protected]] On Behalf Of Jan-Stefan Van der Walt Sent: Tuesday, January 24, 2012 5:24 AM To: Toufigh Gordi; [email protected] Subject: Re: [NMusers] Choice of models Hi Toufigh, Recently I used the 90% prediction interval (generated by an appropriately binned VPC) of the rich data (three studies with observed doses) to evaluate the sparse data (one sample on 4 occasions). The sparse data contained more information about the covariates of interest, but the dosing was unobserved. I analysed the rich and sparse data simultaneously first including and then excluding the sparse data outside the 90% PI and compared the results. The eta-shrinkage values decreased considerably when the observations outside the 90% PI were excluded and I had more confidence in the covariate relationships. As a side issue, I estimated a time-after-dose for the observations outside the 90% PI. It was interesting that the difference between the reported and estimated dosing times seemed to increase as the trial progressed (0.92h [month 6], 1.05h [month 12]), 1.11h [month 18] and 3.6h [month 24]. Hope this helps. Regards, Jan-Stefan On 24 January 2012 05:05, Denney, William S. <[email protected]<mailto:[email protected]>> wrote: Hi Toufigh, I typically think that data quality decreases with phase and with sampling frequency. Given what you described below, I'd think that you're fighting data quality in the sparse, phase 3 studies, and with the parameters you're describing as having trouble, it seems to support that thought. Were I to guess, you could probably pick out the most influential 3% of sparse samples (arbitrary percentage), and look at them in more detail and find that they look more like Cmax than Ctrough or something such the time since last dose appears to be off. Beyond that, philosophically, I think that trough concentrations should not be allowed to affect Ka because the effect is usually so small as not to be measurable (assuming that we're discussing a drug with a reasonable separation between the alpha elimination phase and measurement time). Thanks, Bill On Jan 23, 2012, at 11:45 PM, "Toufigh Gordi" <[email protected]<mailto:[email protected]>> wrote: Dear all, I have a general question on the choice of model in a population analysis. I have a set of data set that includes a large number of studies with about ¾ of the data from extensive sampling schemes (phase 1, 2, and 3 studies) and the rest from sparse samples (phase 3 clinical studies). When developing the PK model, a model on the extensive samples only fits the data well and I can get quite reasonable parameter estimates, including covariate effects, and a successful $COV (NONMEM). When all data is used, the model becomes somewhat instable: the same covariates are identified but the model becomes quite sensitive to the initial estimates and the $COV step won't go through. I could, of course, perform a bootstrap to go around this issue. In general, the fit of the model based on the full data set is not as good as the extensive data set model, although the two models are rather similar with regard to the parameter estimates. However, the range of estimated parameters is wider when using all data and noticeably KA and V2 are skewed to very larger values. Moving forward, I could either use the full data model and simulate steady state profiles for the phase 3 study (sparse samples) data. Or, I could use the model based on the extensive samples only, use the sparse data and generate post-hoc estimates for the sparsely sampled individuals and move forward that way. The advantage with the first option is that all the available data have been used in the modeling process. The disadvantage would be that the model is not as good as the other model, with sparse data distorting the parameter estimates. The advantage of the second option is that the model performs better and there is really no reason why the underlying PK model for the sparsely sampled subjects should be different, which means one should be able to use that model to generate post-hoc estimates. The disadvantage is that not all the available data have been used in the model building process. It would be interesting to hear other people's thoughts and ideas on this. Toufigh -- United Kingdom Flat 5, 41 Devons Rd, E3 3BF, London +44 20 7987 6688 (h) +44 77 9618 4662 (m) South Africa Ballet & Lodge, 34 Kerk St, George, 6529 Postnet Suite 39, Private Bag, X6590, George, 6530 +27 44 884 1560 (h) Sweden Pharmacometrics, Department of Pharmaceutical Biosciences PO Box 591, SE-75124 Uppsala +46 73 066 7338 (m) This electronic transmission may contain confidential and/or proprietary information and is intended to be for the use of the individual or entity named above. If you are not the intended recipient, be aware that any disclosure, copying, distribution or use of the contents of this electronic transmission is prohibited. If you have received this electronic transmission in error, please destroy it and immediately notify us of the error. Thank you.

RE: Choice of models

From: Doug J. Eleveld Date: January 24, 2012 technical
Hi Toufigh, Just a suggestion that you may already be using, do you use the SORT option for estimation. This is i think helpful when the informativeness of individuals vary considerably. I might help stabilise the full data set. Douglas Eleveld
Quoted reply history
________________________________________ From: [email protected] [[email protected]] on behalf of Toufigh Gordi [[email protected]] Sent: Tuesday, January 24, 2012 5:23 AM To: [email protected] Subject: [NMusers] Choice of models Dear all, I have a general question on the choice of model in a population analysis. I have a set of data set that includes a large number of studies with about ¾ of the data from extensive sampling schemes (phase 1, 2, and 3 studies) and the rest from sparse samples (phase 3 clinical studies). When developing the PK model, a model on the extensive samples only fits the data well and I can get quite reasonable parameter estimates, including covariate effects, and a successful $COV (NONMEM). When all data is used, the model becomes somewhat instable: the same covariates are identified but the model becomes quite sensitive to the initial estimates and the $COV step won’t go through. I could, of course, perform a bootstrap to go around this issue. In general, the fit of the model based on the full data set is not as good as the extensive data set model, although the two models are rather similar with regard to the parameter estimates. However, the range of estimated parameters is wider when using all data and noticeably KA and V2 are skewed to very larger values. Moving forward, I could either use the full data model and simulate steady state profiles for the phase 3 study (sparse samples) data. Or, I could use the model based on the extensive samples only, use the sparse data and generate post-hoc estimates for the sparsely sampled individuals and move forward that way. The advantage with the first option is that all the available data have been used in the modeling process. The disadvantage would be that the model is not as good as the other model, with sparse data distorting the parameter estimates. The advantage of the second option is that the model performs better and there is really no reason why the underlying PK model for the sparsely sampled subjects should be different, which means one should be able to use that model to generate post-hoc estimates. The disadvantage is that not all the available data have been used in the model building process. It would be interesting to hear other people’s thoughts and ideas on this. Toufigh ________________________________