What is a model? A joint probability distribution!
Contents
Introduction
A model built for real-world applications can involve various types of variable, such as measurements, individual and population parameters, covariates, design, etc. The model allows us to represent relationships between these variables.
If we consider things from a probabilistic point of view, some of the variables will be random, so the model becomes a probabilistic one, representing the joint distribution of these random variables.
Defining a model therefore means defining a joint distribution. The hierarchical structure of the model will then allow it to be decomposed into submodels, i.e., the joint distribution decomposed into a product of conditional distributions.
Tasks such as estimation, model selection, simulation and optimization can then be expressed as specific ways of using this probability distribution.
- A submodel is a conditional distribution derived from this joint distribution.
- A task is a specific use of this distribution.
We will illustrate this approach starting with a very simple example that we will gradually make more sophisticated. Then we will see in various situations what can be defined as the model and what its inputs are.
An illustrative example
A model for the observations of a single individual
Let $y=(y_j, 1\leq j \leq n)$ be a vector of observations obtained at times $\vt=(t_j, 1\leq j \leq n)$. We consider that the $y_j$ are random variables and we denote $\qy$ the distribution (or pdf) of $y$. If we assume a parametric model, then there exists a vector of parameters $\psi$ that completely define $y$.
We can then explicitly represent this dependency with respect to $\bpsi$ by writing $\qy( \, \cdot \, ; \psi)$ for the pdf of $y$.
If we wish to be even more precise, we can even make it clear that this distribution is defined for a given design, i.e., a given vector of times $\vt$, and write $ \qy(\, \cdot \, ; \psi,\vt)$ instead.
By convention, the variables which are before the symbol ";" are random variables. Those that are after the ";" are non-random parameters or variables. When there is no risk of confusion, the non-random terms can be left out of the notation.
-The inputs of the model are the parameters $\psi$ and the design $\vt$.
\( f(t;V,k) = \displaystyle{ \frac{500}{V} }e^{-k \, t} , \)
where $V$ is the volume of distribution and $k$ the elimination rate constant. The concentration is measured at times $(t_j, 1\leq j \leq n)$ with additive residual errors:
\( y_j = f(t_j;V,k) + e_j , \quad 1 \leq j \leq n . \)
Assuming that the residual errors $(e_j)$ are independent and normally distributed with constant variance $a^2$, the observed values $(y_j)$ are also independent random variables and
\(
y_j \sim {\cal N} \left( f(t_j ; V,k) , a^2 \right), \quad 1 \leq j \leq n. \)
|
(1) |
Here, the vector of parameters $\psi$ is $(V,k,a)$. $V$ and $k$ are the PK parameters for the structural PK model and $a$ the residual error parameter. As the $y_j$ are independent, the joint distribution of $y$ is the product of their marginal distributions:
\( \py(y ; \psi,\vt) = \prod_{j=1}^n \pyj(y_j ; \psi,t_j) , \)
A model for several individuals
Now let us move to $N$ individuals. It is natural to suppose that each is represented by the same basic parametric model, but not necessarily the exact same parameter values. Thus, individual $i$ has parameters $\psi_i$. If we consider that individuals are randomly selected from the population, then we can treat the $\psi_i$ as if they were random vectors. As both $\by=(y_i , 1\leq i \leq N)$ and $\bpsi=(\psi_i , 1\leq i \leq N)$ are random, the model is now a joint distribution: $\qypsi$. Using basic probability, this can be written as:
\( \pypsi(\by,\bpsi) = \pcypsi(\by | \bpsi) \, \ppsi(\bpsi) .\)
If $\qpsi$ is a parametric distribution that depends on a vector $\theta$ of population parameters and a set of individual covariates $\bc=(c_i , 1\leq i \leq N)$, this dependence can be made explicit by writing $\qpsi(\, \cdot \,;\theta,\bc)$ for the pdf of $\bpsi$. Each $i$ has a potentially unique set of times $t_i=(t_{i1},\ldots,t_{i \ \!\!n_i})$ in the design, and $n_i$ can be different for each individual.
\( \pypsi(\by , \bpsi; \theta, \bc,\bt)=\pcypsi(\by | \bpsi;\bt) \, \ppsi(\bpsi;\theta,\bc) . \)
- The inputs of the model are the population parameters $\theta$, the individual covariates $\bc=(c_i , 1\leq i \leq N)$ and the measurement times
- $\bt=(t_{ij} ,\ 1\leq i \leq N ,\ 1\leq j \leq n_i)$.</div> </div> <div class="noprint" style="margin-left: 2em; border-left: 6px solid #C04000; marging-left:4%; margin-right:4%"> <div style="padding-left: 1%; padding-top:0.9em;">[[Image:man02.jpg|52px|left|top|link=]] </div> <div style="text-align: left; padding-left: 3%em; font-size:13pt; font-weight:bold"><u>Example</u></div> <br> <div style="text-align: left; padding-left: 3%; padding-bottom:2em">Let us suppose $ N$ patients received the same treatment as the single patient did. We now have the same PK model [[#ex_proba1|(1)]] for each patient, except that each has its own individual PK parameters $ V_i$ and $ k_i$ and potentially its own residual error parameter $ a_i$: {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba2a">\( y_{ij} \sim {\cal N} \left( \displaystyle{\frac{500}{V_i}e^{-k_i \, t_{ij} } } , a_i^2 \right). \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(2) |} Here, $\psi_i = (V_i,k_i,a_i)$. One possible model is then to assume the same residual error model for all patients, and log-normal distributions for $V$ and $k$: <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\begin{eqnarray} a_i &=& a \end{eqnarray}\) </div> {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba2b">\(\begin{eqnarray} \log(V_i) &\sim_{i.i.d.}& {\cal N}\left(\log(V_{\rm pop})+\beta\,\log(w_i/70),\ \omega_V^2\right) \end{eqnarray}\)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(3) |} <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\begin{eqnarray} \log(k_i) &\sim_{i.i.d.}& {\cal N}\left(\log(k_{\rm pop}),\ \omega_k^2\right), \end{eqnarray}\) </div> where the only covariate we choose to consider, $w_i$, is the weight (in kg) of patient $i$. The model therefore consists of the conditional distribution of the concentrations defined in [[#ex_proba2a|(2)]] and the distribution of the individual PK parameters defined in [[#ex_proba2b|(3)]]. The inputs of the model are the population parameters $\theta = (V_{\rm pop},k_{\rm pop},\omega_V,\omega_k,\beta,a)$, the covariates (here, the weight) $(w_i, 1\leq i \leq N)$, and the design $\bt$.</div> </div> <br> ==='"`UNIQ--h-4--QINU`"'A model for the population parameters=== In some cases, it may turn out that it is useful or important to consider that the population parameter $\theta$ is itself random, rather than fixed. There are various reasons for this, such as if we want to model uncertainty in its value, introduce a priori information in an estimation context, or model an inter-population variability if the model is not looking at only one given population. If so, let us denote $\qth$ the distribution of $\theta$. As the status of $\theta$ has therefore changed, the model now becomes the joint distribution of its random variables, i.e., of $\by$, $\bpsi$ and $\theta$, and can be decomposed as follows: {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="proba3a">\( \pypsith(\by,\bpsi,\theta;\bt,\bc) = \pcypsi(\by |\bpsi;\bt) \, \pcpsith(\bpsi|\theta;\bc) \, \pth(\theta) . \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(4) |} <div class="noprint" style="float:none; border-left: 14px solid #DABDAB; border-radius: 0.5em 0.5em 0.5em 0.5em; background-color:beige; padding:12px; margin-left: 2%; margin-right: 8%;"> <div style="text-align: left; padding-left: 0.5em; padding-top:0.5em; font-size:13pt;font-family:Segoe UI;font-weight:bold">Remarks</div><br> <div style="text-align: left; padding-left: 1.2em; padding-bottom:0.7em"><ol> <li> The formula is identical for $\ppsi(\bpsi; \theta)$ and $\pcpsith(\bpsi|\theta)$. What has changed is the status of $\theta$. It is not random in $\ppsi(\bpsi; \theta)$, the distribution of $\bpsi$ for any given value of $\theta$, whereas it is random in $\pcpsith(\bpsi | \theta)$, the conditional distribution of $\bpsi$, i.e., the distribution of $\bpsi$ obtained after observing randomly generated $\theta$. </li><br> <li>If $\qth$ is a parametric distribution with parameter $\varphi$, this dependence can be made explicit by writing $\qth(\, \cdot \,;\varphi)$ for the distribution of $\theta$.</li><br> <li>Not necessarily all of the components of $\theta$ need be random. If it is possible to decompose $\theta$ into $(\theta_F,\theta_R)$, where $\theta_F$ is fixed and $\theta_R$ random, then the decomposition [[#proba3a|(4)]] becomes </li> {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="proba3b">\( \pypsith(\by,\bpsi,\theta_R;\bt,\theta_F,\bc) = \pcypsi(\by |\bpsi;\bt) \, \pcpsith(\bpsi|\theta_R;\theta_F,\bc) \, \pth(\theta_R). \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(5) |} </ol></div> </div> <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em"><li> In this context, the model is the joint distribution of the observations, the individual parameters and the population parameters: <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\pypsith(\by,\bpsi,\theta;\bc,\bt) = \pcypsi(\by |\bpsi;\bt) \, \pcpsith(\bpsi|\theta;\bc) \, \pth(\theta). \) </div> <li> The inputs of the model are the individual covariates $\bc=(c_i , 1\leq i \leq N)$ and the measurement times $\bt=(t_{ij} , 1\leq i \leq N , 1\leq j \leq n_i)$.</div> </div> <div class="noprint" style="margin-left: 2em; border-left: 6px solid #C04000; marging-left:4%; margin-right:4%"> <div style="padding-left: 1%; padding-top:0.9em;">[[Image:man02.jpg|52px|left|top|link=]] </div> <div style="text-align: left; padding-left: 3%em; font-size:13pt; font-weight:bold"><u>Example:</u></div> <br> <div style="text-align: left; padding-left: 3%; padding-bottom:2em">We can introduce [http://en.wikipedia.org/wiki/Prior_probability prior distributions] in order to model the inter-population variability of the population parameters $ V_{\rm pop}$ and $k_{\rm pop}$: {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba3">\(\begin{eqnarray} V_{\rm pop} &\sim& {\cal N}\left(30,3^2\right) \end{eqnarray}\)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(6) |} <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\begin{eqnarray} k_{\rm pop} &\sim& {\cal N}\left(0.1,0.01^2\right). \end{eqnarray}\) </div> As before, the conditional distribution of the concentration is given by [[#ex_proba2a|(2)]]. Now, [[#ex_proba2b|(3)]] is the ''conditional distribution'' of the individual PK parameters, given $\theta_R=(V_{\rm pop},k_{\rm pop})$. The distribution of $\theta_R$ is defined in [[#ex_proba3|(6)]]. Here, the inputs of the model are the fixed population parameters $\theta_F = (\omega_V,\omega_k,\beta,a)$, the weights $(w_i)$ and the design $\bt$.</div> </div> <br> ==='"`UNIQ--h-5--QINU`"'A model for the covariates=== Another scenario is to suppose that in fact it is the covariates $\bc$ that are random, not the population parameters. This may either be in the context of wanting to simulate individuals, or when modeling and wanting to take into account uncertainty in the covariate values. If we note $\qc$ the distribution of the covariates, the joint distribution $\qpsic$ of the individual parameters and the covariates decomposes naturally as: {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="proba4">\( \ppsic(\bpsi,\bc;\theta) = \pcpsic(\bpsi | \bc;\theta) \, \pc(\bc) \, , \) |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(7) |} where $\qcpsic$ is the conditional distribution of $\bpsi$ given $\bc$. <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em"><li>In this context, the model is the joint distribution of the observations, the individual parameters and the covariates: <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \( \pypsic(\by,\bpsi,\bc;\theta,\bt) = \pcypsi(\by | \bpsi;\bt) \, \pcpsic(\bpsi | \bc;\theta) \, \pc(\bc) . \) </div> <li>The inputs of the model are the population parameters $\theta$ and the measurement times $\bt$.</div> </div> <div class="noprint" style="margin-left: 2em; border-left: 6px solid #C04000; marging-left:4%; margin-right:4%"> <div style="padding-left: 1%; padding-top:0.9em;">[[Image:man02.jpg|52px|left|top|link=]] </div> <div style="text-align: left; padding-left: 3%em; font-size:13pt; font-weight:bold"><u>Example:</u></div> <br> <div style="text-align: left; padding-left: 3%; padding-bottom:2em">We could assume a normal distribution as a prior for the weights: {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba4">\( w_i \sim_{i.i.d.} {\cal N}\left(70,10^2\right). \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(8) |} Once more, [[#ex_proba2a|(2)]] defines the conditional distribution of the concentrations. Now, [[#ex_proba2b|(3)]] is the ''conditional distribution'' of the individual PK parameters, given the weight $\bw$, which is now a random variable whose distribution is defined in [[#ex_proba4|(8)]]. Now, the inputs of the model are the population parameters $\theta= (V_{\rm pop},k_{\rm pop},\omega_V,\omega_k,\beta,a)$ and the design $\bt$.</div> </div> <br> ==='"`UNIQ--h-6--QINU`"'A model for the measurement times=== Another scenario is to suppose that there is uncertainty in the measurement times $\bt=(t_{ij})$ and not the population parameters or covariates. If we note $\nominal{\bt}=(\nominal{t}_{ij}, 1\leq i \leq N, 1\leq j \leq n_i)$ the nominal measurement times (i.e., those presented in a data set), then the "true" measurement times $\bt$ at which the measurement were made can be considered random fluctuations around $\nominal{\bt}$ following some distribution $\qt(\, \cdot \, ; \nominal{\bt})$. Randomness with respect to time can also appear in the presence of dropout, i.e., individuals that prematurely leave a clinical trial. For such an individual $i$ who leaves at the random time $T_i$, measurement times are the nominal times before $T_i$: $t_{i} = (\nominal{t}_{ij} \ \ {\rm s.t. }\ \ \nominal{t}_{ij}\leq T_i)$. In such situations, measurement times are therefore random and can be thought to come from a distribution $\qt(\, \cdot \, ; \nominal{\bt})$. <div class="noprint" style="float:none; border-left: 14px solid #DABDAB; border-radius: 0.5em 0.5em 0.5em 0.5em; background-color:beige; padding:12px; margin-left: 2%; margin-right: 8%;"> <div style="text-align: left; padding-left: 0.5em; padding-top:0.5em; font-size:13pt;font-family:Segoe UI;font-weight:bold">Remark</div><br> <div style="text-align: left; padding-left: 1.2em; padding-bottom:0.7em">If there are also other regression variables $\bx=(x_{ij})$, it is of course possible to use the same approach and consider $\bx$ as a random variable fluctuating around $\nominal{\bx}$.</div> </div> <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em"><li> In this context, the model is the joint distribution of the observations, the individual parameters and the measurement times: <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\pypsit(\by , \bpsi,\bt; \theta,\bc,\nominal{\bt})=\pcypsit(\by |\bpsi,\bt) \, \ppsi(\bpsi;\theta,\bc) \, \pt(\bt ; \nominal{\bt}) . \) </div> <li> The inputs of the model are the population parameters $\theta$, the individual covariates $\bc$ and the nominal design $\nominal{\bt}$.</div> </div> <div class="noprint" style="margin-left: 2em; border-left: 6px solid #C04000; marging-left:4%; margin-right:4%"> <div style="padding-left: 1%; padding-top:0.9em;">[[Image:man02.jpg|52px|left|top|link=]] </div> <div style="text-align: left; padding-left: 3%em; font-size:13pt; font-weight:bold"><u>Example:</u></div> <br> <div style="text-align: left; padding-left: 3%; padding-bottom:2em">Let us assume as prior a normal distribution around the nominal times: {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba5">\( t_{ij} \sim_{i.i.d.} {\cal N}\left(\nominal{t}_{ij},0.03^2\right). \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(9) |} Here, [[#ex_proba5|(9)]] defines the distribution of the now random variable $ \bt$. The other components of the model defined in [[#ex_proba2a|(2)]] and [[#ex_proba2b|(3)]] remain unchanged. The inputs of the model are the population parameters $ \theta$, the weights $ (w_i)$ and the nominal measurement times $ \nominal{\bt}$.</div> </div> <br> ==='"`UNIQ--h-7--QINU`"'A model for the dose regimen=== If the structural model is a dynamical system (e.g., defined by a system of [http://en.wikipedia.org/wiki/Ordinary_differential_equation ordinary differential equations]), the ''source terms'' $\bu = (u_i, 1\leq i \leq N)$, i.e., the inputs of the dynamical system, are usually considered fixed and known. This is the case for example for doses administered to patients for a given treatment. Here, the source term $u_i$ is made up of the dose(s) given to patient $i$, the time(s) of administration, and their type (intravenous bolus, infusion, oral, etc.). Here again, there may be differences between the nominal dose regimen stated in the protocol and given in the data set, and the dose regimen that was in reality administered. For example, it might be that the times of administration and/or the dosage were not exactly respected or recorded. Also, there may have been non-compliance, i.e., certain doses that were not taken by the patient. If we denote $\nominal{\bu}=(\nominal{u}_{i}, 1\leq i \leq N)$ the nominal dose regimens (reported in the dataset), then in this context the "real" dose regimens $\bu$ can be considered to randomly fluctuate around $\nominal{\bu}$ with some distribution $\qu(\, \cdot \, ; \nominal{\bu})$. <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em"><li> In this context, the model is the joint distribution of the observations, the individual parameters and the dose regimens: <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\pypsiu(\by , \bpsi,\bu; \theta,\bc,\bt,\nominal{\bu})=\pcypsiu(\by | \bpsi,\bu;\bt) \, \pu(\bu ; \nominal{\bu}) \, \ppsi(\bpsi;\theta,\bc) . \) </div> <li> The inputs of the model are the population parameters $\theta$, the individual covariates $\bc$, the nominal design $\bt$ and the nominal dose regimens $\nominal{\bu}$.</div> </div> <div class="noprint" style="margin-left: 2em; border-left: 6px solid #C04000; marging-left:4%; margin-right:4%"> <div style="padding-left: 1%; padding-top:0.9em;">[[Image:man02.jpg|52px|left|top|link=]] </div> <div style="text-align: left; padding-left: 3%em; font-size:13pt; font-weight:bold"><u>Example:</u></div> <br> <div style="text-align: left; padding-left: 3%; padding-bottom:2em">Suppose that instead of the one dose given in the example up to now, there are now repeated doses $(d_{ik}, k \geq 1)$ administered to patient $i$ at times $(\tau_{ik} , k \geq 1)$. Then, it is easy to see that {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba6b">\( y_{ij} \sim {\cal N}\left(f(t_{ij};V_i,k_i) , a_i^2 \right), \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(10) |} where {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba6a">\( f(t;V_i,k_i) = \sum_{k, \tau_{ik}<t}\displaystyle {\frac{d_{ik} }{V_i} }\, e^{-k_i \, (t- \tau_{ik})} . \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(11) |} The "real" dose regimen administered to patient $i$ can be written $u_i=(d_{ik},\tau_{ik}, k\geq 1)$, and the prescribed dose regimen $\nominal{u}_i=(\nominal{d}_{ik},\nominal{\tau}_{ik}, k\geq 1)$. We can model the random fluctuations of the administration times $\tau_{ik}$ around the nominal times $(\nominal{\tau}_{ij})$: {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba6c">\(\begin{eqnarray} \tau_{ik} &\sim_{i.i.d.}& {\cal N}\left(\nominal{\tau}_{ik},0.02^2\right)\ , \end{eqnarray}\)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(12) |} and non-compliance (here meaning that a dose is not taken): {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ex_proba6d">\(\begin{eqnarray} \pi &=& \prob{d_{ik} = 0} \nonumber \\ &=& 1 - \prob{d_{ik}= \nominal{d}_{ik} }. \end{eqnarray}\)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(13) |} Here, [[#ex_proba6b|(10)]] and [[#ex_proba6a|(11)]] define the conditional distributions of the concentrations $(y_{ij})$, [[#ex_proba2b|(3)]] defines the distribution of $\bpsi$ and [[#ex_proba6c|(12)]] and [[#ex_proba6d|(13)]] define the distribution of $\bu$. The inputs are the population parameters $\theta$, the weights $(w_i)$, the measurement times $\bt$ and the nominal dose regimens $\nominal{\bu}$.</div> </div> <br> ==='"`UNIQ--h-8--QINU`"'A complete model=== We have now seen the variety of ways in which the variables in a model either play the role of random variables whose distribution is defined by the model, or nonrandom variables or parameters. Any combination is possible, depending on the context. For instance, the population parameters $\theta$ and covariates $\bc$ could be random with parametric probability distributions $\qth(\, \cdot \,;\varphi)$ and $\qc(\, \cdot \,;\gamma)$, and the dose regimen $\bu$ and measurement times $\bt$ reported with uncertainty and therefore modeled as random variables with distributions $\qu$ and $\qt$. <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em"><li> In this context, the model is the joint distribution of the observations, the individual parameters, the population parameters, the dose regimens, the covariates and the measurement times: <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \( \pypsithcut(\by , \bpsi, \theta, \bu, \bc,\bt; \nominal{\bu},\nominal{\bt},\varphi,\gamma)=\pcypsiut(\by |\bpsi,\bu,\bt) \, \pcpsithc(\bpsi|\theta,\bc) \, \pth(\theta;\varphi) \, \pc(\bc;\gamma) \, \pu(\bu ; \nominal{\bu}) \, \pt(\bt ; \nominal{\bt}). \) </div> <li> The inputs of the model are the nominal dose regimens $\nominal{\bu}$, the nominal measurement times $\nominal{\bt}$ and the "hyper-parameters" $\varphi$ and $\gamma$.</div> </div> <br> =='"`UNIQ--h-9--QINU`"'Using the model for executing tasks== In the modeling and simulation context, the tasks to execute make specific use of the various probability distributions associated with a model. <br> ==='"`UNIQ--h-10--QINU`"'Simulation=== By definition, simulation makes direct use of the probability distribution that defines the model. Simulation of the global model is straightforward as soon as the joint distribution can be decomposed into a product of easily simulated conditional distributions. Consider for example that the variables involved in the model are those introduced in [[#An illustrative example|the previous section]]: # The population parameters $\theta$ can either be given, or simulated from the distribution $\qth$. # The individual covariates $\bc$ can either be given, or simulated from the distribution $\qc$. # The individual parameters $\bpsi$ can be simulated from the distribution $\qcpsithc$ using the values of $\theta$ and $\bc$ obtained in steps 1 and 2. # The dose regimen $\bu$ can either be given, or simulated from the distribution $\qu$. # The measurement times $\bt$ (resp. regression variables $\bx$) can either be given, or simulated from the distribution $\qt$ (resp. $\qx$). # Lastly, observations $\by$ can be simulated from the distribution $\qcypsiut$ using the values of $\bpsi$, $\bu$ and $\bt$ obtained at steps 3, 4 and 5. <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em">Simulation of a set of variables $w$ using another given set of variables $z$ requires: <ul> * a model, i.e., a distribution $\qw$ if $z$ is treated as a nonrandom variable, or a conditional distribution $\qcwz$ if $z$ is treated as a random variable. * the input $z$, i.e., a value of $z$ which allows the distribution $\qw(\, \cdot \, ; z)$ or the conditional distribution $\qcwz(\, \cdot \, | z)$ to be defined. * an algorithm which allows us to generate $w$ from $\qw$ or $\qcwz$. </ul></div> </div> <div class="noprint" style="margin-left: 2em; border-left: 6px solid #C04000; marging-left:4%; margin-right:4%"> <div style="padding-left: 1%; padding-top:0.9em;">[[Image:man02.jpg|52px|left|top|link=]] </div> <div style="text-align: left; padding-left: 3%em; font-size:13pt; font-weight:bold"><u>Example:</u></div> <br> <div style="text-align: left; padding-left: 3%; padding-bottom:2em">- Imagine instead that the population parameter $\theta$ and the design $(\bu,\bt)$ are given, and we want to simulate the individual covariates $\bc$, the individual parameters $\bpsi$ and the observations $\by$. Here, the variables to simulate are $w=(\bc,\bpsi,\by)$ and the variables which are given are $z=(\theta,\bu,\bt)$. If the components of $z$ are taken to be nonrandom variables, then: <ul> * The model is the joint distribution $\qypsic( \, \cdot \, ;\theta,\bu,\bt)$ of $(\by,\bpsi,\bc)$. * The inputs required for the simulation are the values of $(\theta,\bu,\bt)$. * The algorithm should be able to generate $(\by,\bpsi,\bc)$ from the joint distribution $\qypsic(\, \cdot \, ;\theta,\bu,\bt)$. Decomposing the model into three submodels leads to decomposing the joint distribution as </ul> <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \( \pypsic(\by,\bpsi,\bc ;\theta,\bu,\bt) = \pc(\bc) \, \pcpsic(\bpsi | \bc;\theta) \, \pcypsi(\by | \bpsi;\bu,\bt) . \) </div> The algorithm therefore reduces to successively drawing $\bc$, $\bpsi$ and $\by$ from $\qc$, $\qcpsic(\, \cdot \, | \bc;\theta)$ and $\qcypsi(\, \cdot \, | \bpsi;\bu,\bt)$. - Imagine instead that the individual covariates $\bc$, the observations $\by$, the design $(\bu,\bt)$ and the population parameter $\theta$ are given (in a modeling context for instance, $\theta$ may have been estimated), and we want to simulate the individual parameters $\bpsi$. The only variable to simulate is $w=\bpsi$ and the variables which are given are $z=(\by,\bc,\theta,\bu,\bt)$. Here, we will treat $\by$ as if it is a random variable. The other components of $z$ can be treated as non random variables. Here, <ul> * The model is the conditional distribution $\qcpsiy(\, \cdot \, | \by ;\bc,\theta,\bu,\bt)$ of $\psi$. * The inputs required for the simulation are the values of $(\by,\bc,\theta,\bu,\bt)$. * The algorithm should be able to sample $\bpsi$ from the conditional distribution $\qcpsiy(\, \cdot \, | \by ;\bc,\theta,\bu,\bt)$. [http://en.wikipedia.org/wiki/Markov_chain_Monte_Carlo Markov Chain Monte Carlo] (MCMC) algorithms can be used for sampling from such complex conditional distributions. </ul></div> </div> <br> ==='"`UNIQ--h-11--QINU`"'Estimation of the population parameters=== In a modeling context, we usually assume that we have data that includes the observations $\by$ and the measurement times $\bt$. There may also be individual covariates $\bc$, and in pharmacological applications the dose regimen $\bu$. For clarity, let us consider the most general case where all are present. Any statistical method for estimating the population parameters $\theta$ will be based on some specific probability distribution. Let us illustrate this with two common statistical methods: maximum likelihood and Bayesian estimation. ''Maximum likelihood estimation'' consists in maximizing with respect to $\theta$ the ''observed likelihood'', defined by: <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\begin{eqnarray} {\like}(\theta ; \by,\bc,\bu,\bt) &\eqdef& \py(\by ; \bc,\bu,\bt,\theta) \\ &=& \int \pypsi(\by,\bpsi ; \bc,\bu,\bt,\theta) \, d \bpsi . \end{eqnarray}\) </div> The variance of the estimator $\thmle$ and therefore confidence intervals can be derived from the observed Fisher information matrix, which itself is calculated using the observed likelihood (i.e., the pdf of the observations $\by$): {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="ofim_intro3">\( \ofim(\thmle ; \by,\bc,\bu,\bt) \ \ \eqdef \ \ - \displaystyle{ \frac{\partial^2}{\partial \theta^2} } \log({\like}(\thmle ; \by,\bc,\bu,\bt)) . \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(14) |} <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em">Maximum likelihood estimation of the population parameter $\theta$ requires: * a model, i.e., a joint distribution $\qypsi$. * inputs $\by$, $\bc$, $\bu$ and $\bt$. * an algorithm which allows us to maximize $\int \pypsi(\by,\bpsi ; \bc,\bu,\bt,\theta) \, d \bpsi$ with respect to $\theta$ and to compute $\displaystyle{ \frac{\partial^2}{\partial \theta^2} }\left\{\log\left(\int \pypsi(\by,\bpsi ; \bc,\bu,\bt,\thmle) \, d \bpsi \right)\right\}$.</div> </div> ''Bayesian estimation'' consists in estimating and/or maximizing the conditional distribution <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\begin{eqnarray} \pcthy(\theta | \by ;\bc,\bu,\bt) &=& \displaystyle{ \frac{\pyth(\by , \theta ; \bc,\bu,\bt)}{\py(\by ; \bc,\bu,\bt)} } \\ &=& \frac{\displaystyle{ \int \pypsith(\by,\bpsi,\theta ; \bc,\bu,\bt) \, d \bpsi} }{\py(\by ; \bc,\bu,\bt)} . \end{eqnarray}\) </div> <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em">Bayesian estimation of the population parameter $\theta$ requires: * a model, i.e., a joint distribution $\qypsith(\, \cdot \, ; \bc, \bu, \bt)$ for $(\by,\bpsi,\theta)$. * inputs $\by$, $\bc$, $\bu$ and $\bt$. * algorithms able to estimate and maximize $\pcthy(\theta | \by ;\bc,\bu,\bt)$. MCMC methods can be used for estimating this conditional distribution. For nonlinear models, optimization tools are required for computing its mode, i.e., finding its maximum.</div> </div> <br> ==='"`UNIQ--h-12--QINU`"'Estimation of the individual parameters=== When $\theta$ is given (or estimated), various estimators of the individual parameters $\bpsi$ are available. They are all based on a probability distribution: ''Maximum likelihood estimation'' consists of maximizing with respect to $\bpsi$ the ''conditional likelihood'' <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\begin{eqnarray} {\like}(\bpsi ; \by,\bu,\bt) &\eqdef& \pcypsi(\by | \bpsi ;\bu,\bt) . \end{eqnarray}\) </div> The ''maximum a posteriori'' (MAP) estimator is obtained by maximizing with respect to $\bpsi$ the ''conditional distribution'' <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\begin{eqnarray} \pcpsiy(\bpsi | \by ; \theta,\bc,\bu,\bt) &=& \displaystyle{ \frac{\pypsi(\by , \bpsi;\theta,\bc,\bu,\bt)}{\py(\by ; \theta,\bc,\bu,\bt)} } . \end{eqnarray}\) </div> The ''conditional mean'' of $\bpsi$ is defined as the mean of the conditional distribution $\qcpsiy(\, \cdot \, | \by ; \theta,\bc,\bu,\bt)$ of $\psi$. <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em">Estimation of the individual parameters $\bpsi$ requires: * a model, i.e., a joint distribution $\qypsi(\, \cdot \, ; \theta, \bc, \bu, \bt)$ for $(\by,\bpsi)$. * inputs $\by$, $\theta$, $\bc$, $\bu$ and $\bt$. * algorithms able to estimate and maximize $\pcpsiy(\bpsi | \by ; \theta,\bc,\bu,\bt)$. MCMC methods can be used for estimating this conditional distribution. For nonlinear models, optimization tools are required for computing its mode (i.e., its MAP).</div> </div> <br> ==='"`UNIQ--h-13--QINU`"'Model selection=== Likelihood ratio tests and statistical information criteria ([http://en.wikipedia.org/wiki/Bayesian_information_criterion BIC], [http://en.wikipedia.org/wiki/Akaike_information_criterion AIC]) compare the ''observed likelihoods'' computed under different models, i.e., the probability distribution functions $\py^{(1)}(\by ; \bc,\bu,\bt,\thmle_1)$, $\py^{(2)}(\by ; \bc,\bu,\bt,\thmle_2)$, ..., $\py^{(K)}(\by ; \bc,\bu,\bt,\thmle_K)$ computed under models ${\cal M}_1, {\cal M}_2$, ..., ${\cal M}_K$, where $\thmle_k$ maximizes the observed likelihood of model ${\cal M}_k$, i.e., maximizes $\py^{(k)}(\by ; \bc,\bu,\bt,\theta)$ . <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em">Computing the observed likelihood and information criteria require: * a model, i.e., a joint distribution $\qypsi(\, \cdot \, ; \theta, \bc, \bu, \bt)$ for $(\by,\bpsi)$. * inputs $\by$, $\theta$, $\bc$, $\bu$ and $\bt$. * an algorithm able to compute $\int \pypsi( \by ,\bpsi ;\theta,\bc,\bu,\bt) \, d\bpsi$. For nonlinear models, linearization methods or Monte Carlo methods can be used.</div> </div> <br> ==='"`UNIQ--h-14--QINU`"'Optimal design=== In the design of experiments for estimating statistical models, optimal designs allow parameters to be estimated with minimum variance by optimizing some statistical criteria. Common optimality criteria are [http://en.wikipedia.org/wiki/Functional_%28mathematics%29 functionals] of the [http://en.wikipedia.org/wiki/Eigenvalues_and_eigenvectors eigenvalues] of the expected Fisher information matrix {| align=left; cellpadding="2" |style="width: 550px; text-align: left; font-size:13pt; padding-left: 4em"| <div id="efim_intro3">\( \efim(\theta ; \bu,\bt) \ \ \eqdef \ \ \esps{y}{\ofim(\theta ; \by,\bu,\bt)} , \)</div> |style="width: 100px; text-align: right; font-size:13pt; padding-right: 8em" |(15) |} where $\ofim$ is the observed Fisher information matrix defined in [[#ofim_intro3|(15)]]. For the sake of simplicity, we consider models without covariates $\bc$. <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em">Optimal design for minimum variance estimation requires: * a model, i.e., a joint distribution $\qypsi(\, \cdot \, ; \theta, \bc, \bu, \bt)$ for $(\by,\bpsi)$. * a vector of population parameters $\theta$. * a criteria ${\cal D}(\bu,\bt)$ derived from the expected Fisher information matrix $\efim(\theta ; \bu,\bt)$. * an algorithm able to estimate $\efim(\theta ; \bu,\bt)$ for any design $(\bu,\bt)$ and to maximize ${\cal D}(\bu,\bt)$ with respect to $\bu$ and $\bt$.</div> </div> In a [http://en.wikipedia.org/wiki/Clinical_trial clinical trial] context, studies are designed to optimize the probability of reaching some predefined target ${\cal A}$, i.e., $\prob{(\by, \bpsi,\bc) \in {\cal A} ; \bu,\bt,\theta}$. This may include optimizing safety and efficacy, and things like the probability of reaching [http://en.wikipedia.org/wiki/Sustained_viral_response sustained virologic response], etc. <div class="noprint" style="float:none; border-left: 14px solid #ABCDEF; border-radius: 0.5em 6em 0.5em 6em; background-color:#F0F8FF; padding:12px; margin-left: 2%; margin-right:8%"> <div style="margin-left: 3em; margin-right: 2.5em; margin-top: 0.8em; margin-bottom: 0.8em">Optimal design for clinical trials requires: * a model, i.e., a joint distribution $\qypsic(\, \cdot \, ; \theta, \bu, \bt)$ for $(\by,\bpsi,\bc)$. * a vector of population parameters $\theta$. * a target ${\cal A}$. * an algorithm able to estimate $\prob{(\by, \bpsi,\bc) \in {\cal A} ; \bu,\bt,\theta}$ and to maximize it with respect to $\bu$ and $\bt$.</div> </div> <br> =='"`UNIQ--h-15--QINU`"'Implementing models and running tasks== ==='"`UNIQ--h-16--QINU`"'Example 1 === Consider first the model defined by the joint distribution <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \(\pypsi(\by,\bpsi ; \theta ,\bt) = \pcypsi(\by |\bpsi;\bt) \pcpsic(\bpsi ; \theta),\) </div> where as in our running example, <ul> * $\by = (y_{ij}, 1\leq i \leq N , 1 \leq j \leq n_i)$ are concentrations * $ \bpsi= (\psi_i, 1\leq i \leq N)$ are individual parameters; here $ \psi_i=(V_i,k_i,a_i)$ * $ \theta=(V_{\rm pop},k_{\rm pop},\omega_V,\omega_k,a)$ are population parameters * $ \bt = (t_{ij}, 1\leq i \leq N , 1 \leq j \leq n_i)$ are the measurement times. </ul> We aim to define a joint model for $\by$ and $\bpsi$. To do this, we will characterize each component of the model and show how this can be implemented with $\mlxtran$. {| cellspacing="10" cellpadding="10" |style="width:50%"| <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\( \pypsi(\by,\bpsi ; \theta, \bt) \)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> </div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\( \pcpsic(\bpsi |\theta)\)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> \(\begin{array}{c} \log(V_i) &\sim& {\cal N}\left(\log(V_{\rm pop}), \, \omega_V^2\right) \\ \log(k_i) &\sim& {\cal N}\left(\log(k_{\rm pop}),\, \omega_k^2\right) \end{array}\)</div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\(\pcypsi(y|\bpsi; \bt) \)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> \(\begin{eqnarray} f(t;V_i,k_i) &=& \frac{500}{V_i}e^{-k_i \, t} \\[0.2cm] y_{ij} &\sim& {\cal N} \left(f(t_{ij};V_i,k_i) , a^2 \right) \end{eqnarray}\)</div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> |style = "width:50%" | <div class="noprint" style="background-color:#EFEFEF; border: 1px solid darkgray; border-radius:1em"> <div style="margin-top:1em;margin-left:1em">[[Image:monolix_icon2.png|left|top|link=]]</div> <div style="text-align: left; padding-left: 4em; margin-top:0.5em; font-family:'courier new'; font-size:14pt; font-weight:bold; color: #0095B6; margin-top:1.2em">MLXTran <span style="font-weight:normal; font-size:13pt">Example 1</span></div> <div style="text-align: left; padding-left: 1.2em; font-family:'courier new';margin-bottom:1em">'"`UNIQ--pre-00000000-QINU`"'</div> </div> |} We can then use this model with different tools for executing different tasks: it can be used for example with $\mlxplore$ for model exploration, with $\monolix$ for modeling, with R or Matlab for simulation, etc. It is important to remember that $\mlxtran$ is not a "function" that calculates an output. It is not an imperative but rather a declarative language, one that allows us to describe a model. It is then the tasks we choose to do which use $\mlxtran$ like a function, "requesting" it to give predictions, simulate random variables, compute a pdf, maximizes a likelihood, etc. <br> ==='"`UNIQ--h-17--QINU`"'Example 2=== Consider now a model defined by the joint distribution <div class="noprint" style="text-align: left;font-size: 13pt; padding-left: 6%"> \( \pypsithc(\by,\bpsi, \theta, \bc ; \bt) = \pcypsi(\by|\bpsi;\bt) \pcpsic(\bpsi|\bc ; \theta) \, \pth(\theta) \pc(\bc) , \) </div> where the covariates $\bc$ are the weights of the individuals: $\bc = (w_i, 1\leq i \leq N)$. The other variables and parameters are those already defined in the previous example. We now aim to define a joint model for $\by$, $\bpsi$, $\bc$ and $\theta_R=(V_{\rm pop},k_{\rm pop})$. {| cellspacing="10" cellpadding="10" |style="width:50%" | <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\(\pypsithc(\by,\bpsi, \theta, \bc ; \bt)\)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> </div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\(\pth(\theta)\)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> \(\begin{eqnarray} V_{\rm pop} &\sim& {\cal N}\left(30,3^2\right) \\ k_{\rm pop} &\sim& {\cal N}\left(0.1,0.01^2\right) \end{eqnarray}\)</div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\(\pc(\bc)\)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> \(\begin{eqnarray} w_i &\sim& {\cal N}\left(70,10^2\right) \end{eqnarray}\)</div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\(\pcpsic(\bpsi |\bc;\theta)\)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> \( \begin{eqnarray} \hat{V}_i &=& V_{\rm pop}\left(\frac{w_i}{70}\right)^\beta \\[0.4cm] \log(V_i) &\sim& {\cal N}\left(\log(\hat{V}_i), \, \omega_V^2\right) \\ \log(k_i) &\sim& {\cal N}\left(\log(k_{\rm pop}),\, \omega_k^2\right) \end{eqnarray}\)</div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> <div class="noprint" style="margin-left: 2em; border:none"> <div style="color:#990066;font-size:13pt;margin-top: 14px; margin-left:4%">\(\pcypsi(y|\bpsi; \bt) \)</div><br> <div style="margin-left:5%; margin-bottom:1.2em"> \(\begin{eqnarray} f(t;V_i,k_i) &=& \frac{500}{V_i}e^{-k_i \, t} \\[0.2cm] y_{ij} &\sim& {\cal N} \left(f(t_{ij};V_i,k_i) , a^2 \right) \end{eqnarray}\)</div> <div style="text-align:left;border: 1.5px solid darkgray; margin-left:none; width:100%"></div> </div> |style="width:50%"| <div class="noprint" style="background-color:#EFEFEF; border: 1px solid darkgray; border-radius:1em"> <div style="margin-top:1em;margin-left:1em">[[Image:monolix_icon2.png|left|top|link=]]</div> <div style="text-align: left; padding-left: 4em; margin-top:0.5em; font-family:'courier new'; font-size:14pt; font-weight:bold; color: #0095B6; margin-top:1.2em">MLXTran <span style="font-weight:normal; font-size:13pt">jointModel2.txt</span></div> <div style="text-align: left; padding-left: 1.2em; font-family:'courier new';margin-bottom:1em">'"`UNIQ--pre-00000001-QINU`"'</div> </div> |} We can use the approach described above for various tasks, e.g., simulating $(\by,\bpsi, \bc, \theta_R)$ for a given input $(\theta_F, \bt)$, simulating the population parameters $(V_{\rm pop},k_{\rm pop})$ with the conditional distribution $p_{\theta_R|\by, \bc}( \, \cdot \, | \by, \bc ; \theta_F,\bt)$, estimating the log-likelihood, maximizing the observed likelihood and computing the MAP.