1 / 29

Estimation + optimization in PK modeling

Estimation + optimization in PK modeling. Introduction to modeling Estimation criteria Numerical optimization, examples. http://pharmapk.pharmacie.univ-mrs.fr/. Real process and mathematical model. Real process. Fitted model. Math. model. Functional scheme. Measurement noise.

tucker
Download Presentation

Estimation + optimization in PK modeling

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Estimation + optimization in PK modeling • Introduction to modeling • Estimation criteria • Numerical optimization, examples • http://pharmapk.pharmacie.univ-mrs.fr/

  2. Real process and mathematical model Real process Fitted model Math. model

  3. Functional scheme Measurement noise Observation PK process + Administration protocol A priori information Equivalence criterion Prediction PK model Nonlinear programming

  4. Mathematical modeling • Models are defined by : • - their structure ( number and connectivity of compartments, etc ) expressed by mathematical operations involving adjustable parameters : • Ex : 1-cpt, • exponential structure, parameters : • - the numerical value of parameters used : CHARACTERIZATION Structure ESTIMATION Parameters + MODELING System Identification

  5. Checking identifiability • Structural identifiability : given a hypothetical structure with unknown parameters, and a set of proposed experiments (not measurements !), would the unknown parameters be uniquely determinable from these experiments ? • Parametric identifiability : estimating the parameters from measurements with errors and optimize sampling designs. : structural non-identifiableNon-consistent estimate 1 2

  6. Structural identifiability • It depends on the observation site ! • Solutions : • Grouping : ¨ But ONLY identifiable parameters : • Setting : 1 2 1 2

  7. Functional scheme (dynamic) Linear / p model : no loop, one stage estimation. Nonlinear / p model : many loops until convergence. Measurement noise PK process + Administration protocol A priori information Equivalence criterion PK model Arbitrary initial values Nonlinear programming

  8. Arbitrary initial values Iterations, parameter convergence • Ex : Fotemustine neutrophil toxicity : Nonlinear modeling : Optimized final values

  9. Initial parametric error (canceled at the convergence) Residual error Errors in the functional scheme • The existing errors : • experimental, • structural, • parametric. • Residual error : • experimental, • structural (model • misspecification).

  10. Parametric and output spaces Parametric space Output space Observation PK process Real process Comparison Comparison Prediction PK model Artificial mechanism Random component Precision of estimates Measurement error

  11. Optimal estimation • Estimation is the operation of assigning a numerical values to unknown parameters, based on noise-corrupted observations. • Organization of the variables : • The observed drug concentrations over time, (dimensional vector). • The random parameters to be estimated, (dimensional vector). • Consider the joint pdf and then : • the marginal is called prior pdf [ the marginal is not of interest ]. • the conditional is called posterior pdf : • the conditional leads to the likelihood function : MAP MLE

  12. Maximum a posteriori (MAP) • Design : A reasonable estimate of would be the mode of the posterior density for the given observation : • Ex : • if • if • The role of the dispersion.

  13. Maximum likelihood (MLE) • Design : After the observation has been obtained, a reasonable estimate of would be , the value which gives to the particular observation the highest probability of occurrence : • Ex : • if • if • The role of the precision.

  14. The influence of x on the conditional

  15. The influence of y on the likelihood

  16. MLE criterion for single - output • Initial form : • Hypotheses : H0 : The model is an exact description of the real process. • ErrorOutput H1 : Additive error : H2 : Normal error : H3 : Independence :

  17. Estimate ! Variance heterogeneity • The regression model : • assumes that • Need to relax this assumption (particularly when the model is highly nonlinear). • Transformed models • Find a transformation function under which the error assumptions hold, i.e. : • where • Box – Cox transformations : • Other transformations : John – Drapper, Carroll, Huber, etc.

  18. General form of the MLE criterion • For available observed data and under the H3 hypothesis the estimator becomes : • Where : if • is the criterion function to be minimized. • The 1st term is known as the extended SE term. • The 2nd term is called the weighted SE term. It is the only one involving observed data and it is weighted by the uncertainty of experiment.

  19. Criterion and error variance model • After introducing the error variance model : • is minimum along the direction when : • or with • Then : • Nonlinearly unconstrained optimization : • Find : • Assumptions : is computable for all and analytic solution does not exist.

  20. Iterative solutions • Solution for the nonlinear optimization problem • Sequentially approximate starting from an initial value and converging towards • a stationary point. • Design a routine algorithm generating the converging sequence : • Terminology : is the initialization and obtaining from is an iteration. • Assign :

  21. Approximation of functions • Taylor's expansion for smooth multivariate function: • Construct simple approximations of a general function in a neighborhood of the reference point . With • a vector of unit length supplying the direction of search, and • a scalar specifying the step length: • Associate successive approximations to iterations :

  22. Direction of search • Linear approx. of : • The scalar gives the rate of change of at the point along the direction . • To reduce , move along the direction opposite to : the descent direction. • Quadratic approx. of : • The scalar involves the second derivative of . • characterizes an ellipse. To reduce , move along the direction targeting • the center of ellipse : : the Newton direction

  23. Line search • Minimization directions : moving along the Newton direction for quadratic surfaces, near . Elsewhere, move along the descent direction. • Line search : to complete the iteration search for in the direction of search : • or 4.5 Minimization direction 4.5 3.8 3.4

  24. Families of algorithms • Practical : Approximate derivatives by finite-differences instead analytical calculation. • Classify : • Twice-continuously differentiable : • Second derivative : quadratic model of , compute and invert (not numerically stable, time consuming processing). • First derivative : quadratic model of , approximate : • without inverting but directly from by finite-differences of . • quasi-Newton methods (appropriate in many circumstances) : BFGS, DFP,... • Non-derivative : linear model of . Approximate by finite-differences of (for smooth functions) : Powell, Brent,... • No assumptions on differentiability : heuristicalgorithms : NMS, Hooke-Jeeves,...

  25. The information matrix • The Fisher information matrix : For MLE estimation : • Cramér-Rao inequality : • is a lower bound of the covariance matrix , evaluating the precision of . • In practice : With the vector of the sampling times, • Obtain the sensitivity matrix with elements , • Set and the order diagonal matrices having as elements • and respectively.

  26. Covariance (precision) matrix • The order precision matrix is : • Dependence on the sampling protocol: • Graphic interpretation of the precision matrix : • is symmetric, and, if , it is also definite positive (by construction). • If , then is an a dimensional ellipsoid. • The volume of the ellipsoid is : Sums of weighted products of sensitivity functions over the available sampling times The lowest , the most precise

  27. Check the structural identifiability • The sensitivity matrix depends : • On the experiment (not measurements !) and • On the model parametrization (structural and parametric) . • Ensure definite-positivity of the sensitivity matrix : • It must be of full rank for any numeric value of the parameters, e.g., for the arbitrary initial values (several). • Ex : • Observation in the central cpt free • # central cpt fixed • # peripheral cpt ?  ?

  28. Simulation in optimization • 2-cpt model : • Administration : • IV bolus • Observation : • Horizon • nbr • Heteroscedastic

  29. Performances of algorithms

More Related