- Survival data analysis at target time (TT)
- Survival analysis using a toxicokinetic-toxicodynamic (TKTD) model: GUTS
- Reproduction data analysis at target-time

The package `morse`

is devoted to the analysis of data
from standard toxicity tests. It provides a simple workflow to
explore/visualize a data set, and compute estimations of risk assessment
indicators. This document illustrates a typical use of
`morse`

on survival and reproduction data, which can be
followed step-by-step to analyze new data sets.

The following example shows all the steps to perform survival
analysis on standard toxicity test data and to produce estimated values
of the \(LC_x\). We will use a data set
of the library named `cadmium2`

, which contains both survival
and reproduction data from a chronic laboratory toxicity test. In this
experiment, snails were exposed to six concentrations of a metal
contaminant (cadmium) during 56 days.

The data from a survival toxicity test should be gathered in a
`data.frame`

with a specific layout. This is documented in
the paragraph on `survData`

in the reference manual, and you
can also inspect one of the data sets provided in the package (e.g.,
`cadmium2`

). First, we load the data set and use the function
`survDataCheck()`

to check that it has the expected
layout:

`## Correct format`

The output `## No message`

just informs that the data set
is well-formed.

`survData`

objectThe class `survData`

corresponds to survival data and is
the basic layout used for the subsequent operations. Note that if the
call to `survDataCheck()`

reports no error (i.e.,
`## No message`

), it is guaranteed that `survData`

will not fail.

```
## # A tibble: 6 × 6
## conc time Nsurv Nrepro replicate Ninit
## <dbl> <int> <int> <int> <fct> <int>
## 1 0 0 5 0 1 5
## 2 0 3 5 262 1 5
## 3 0 7 5 343 1 5
## 4 0 10 5 459 1 5
## 5 0 14 5 328 1 5
## 6 0 17 5 742 1 5
```

The function `plot()`

can be used to plot the number of
surviving individuals as a function of time for all concentrations and
replicates.

Two graphical styles are available, `"generic"`

for
standard `R`

plots or `"ggplot"`

to call package
`ggplot2`

(default). If argument `pool.replicate`

is `TRUE`

, datapoints at a given time-point and a given
concentration are pooled and only the mean number of survivors is
plotted. To observe the full data set, we set this option to
`FALSE`

.

By fixing the concentration at a (tested) value, we can visualize one subplot in particular:

We can also plot the survival rate, at a given time-point, as a
function of concentration, with binomial confidence intervals around the
data. This is achieved by using function `plotDoseResponse()`

and by fixing the option `target.time`

(default is the end of
the experiment).

Function `summary()`

provides some descriptive statistics
on the experimental design.

```
##
## Number of replicates per time and concentration:
## time
## conc 0 3 7 10 14 17 21 24 28 31 35 38 42 45 49 52 56
## 0 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 53 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 78 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 124 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 232 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 284 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
##
## Number of survivors (sum of replicates) per time and concentration:
## 0 3 7 10 14 17 21 24 28 31 35 38 42 45 49 52 56
## 0 30 30 30 30 29 29 29 29 29 28 28 28 28 28 28 28 28
## 53 30 30 29 29 29 29 29 29 29 29 28 28 28 28 28 28 28
## 78 30 30 30 30 30 30 29 29 29 29 29 29 29 29 29 27 27
## 124 30 30 30 30 30 29 28 28 27 26 25 23 21 18 11 11 9
## 232 30 30 30 22 18 18 17 14 13 12 8 4 3 1 0 0 0
## 284 30 30 15 7 4 4 4 2 2 1 1 1 1 1 1 0 0
```

Now we are ready to fit a probabilistic model to the survival data,
in order to describe the relationship between the concentration in
chemical compound and survival rate at the target time. Our model
assumes this latter is a log-logistic function of the former, from which
the package delivers estimates of the parameters. Once we have estimated
the parameters, we can then calculate the \(LC_x\) values for any \(x\). All this work is performed by the
`survFitTT()`

function, which requires a
`survData`

object as input and the levels of \(LC_x\) we want:

The returned value is an object of class `survFitTT`

providing the estimated parameters as a posterior^{1} distribution, which
quantifies the uncertainty on their true value. For the parameters of
the models, as well as for the \(LC_x\)
values, we report the median (as the point estimated value) and the 2.5
% and 97.5 % quantiles of the posterior (as a measure of uncertainty,
a.k.a. credible intervals). They can be obtained by using the
`summary()`

method:

```
## Summary:
##
## The loglogisticbinom_3 model with a binomial stochastic part was used !
##
## Priors on parameters (quantiles):
##
## 50% 2.5% 97.5%
## b 1.000e+00 1.259e-02 7.943e+01
## d 5.000e-01 2.500e-02 9.750e-01
## e 1.227e+02 5.390e+01 2.793e+02
##
## Posteriors of the parameters (quantiles):
##
## 50% 2.5% 97.5%
## b 8.540e+00 3.956e+00 1.613e+01
## d 9.571e-01 9.104e-01 9.861e-01
## e 2.362e+02 2.104e+02 2.535e+02
##
## Posteriors of the LCx (quantiles):
##
## 50% 2.5% 97.5%
## LC10 1.830e+02 1.261e+02 2.139e+02
## LC20 2.013e+02 1.541e+02 2.259e+02
## LC30 2.142e+02 1.746e+02 2.348e+02
## LC40 2.255e+02 1.931e+02 2.435e+02
## LC50 2.362e+02 2.104e+02 2.535e+02
```

If the inference went well, it is expected that the difference between quantiles in the posterior will be reduced compared to the prior, meaning that the data were helpful to reduce the uncertainty on the true value of the parameters. This simple check can be performed using the summary function.

The fit can also be plotted:

This representation shows the estimated relationship between concentration of chemical compound and survival rate (orange curve). It is computed by choosing for each parameter the median value of its posterior. To assess the uncertainty on this estimation, we compute many such curves by sampling the parameters in the posterior distribution. This gives rise to the grey band, showing for any given concentration an interval (called credible interval) containing the survival rate 95% of the time in the posterior distribution. The experimental data points are represented in black and correspond to the observed survival rate when pooling all replicates. The black error bars correspond to a 95% confidence interval, which is another, more straightforward way to bound the most probable value of the survival rate for a tested concentration. In favorable situations, we expect that the credible interval around the estimated curve and the confidence interval around the experimental data largely overlap.

A similar plot is obtained with the style `"generic"`

:

Note that `survFitTT()`

will warn you if the estimated
\(LC_{x}\) lie outside the range of
tested concentrations, as in the following example:

```
data("cadmium1")
doubtful_fit <- survFitTT(survData(cadmium1),
target.time = 21,
lcx = c(10, 20, 30, 40, 50))
```

```
## Warning: The LC50 estimation (model parameter e) lies outside the range of
## tested concentrations and may be unreliable as the prior distribution on this
## parameter is defined from this range !
```

In this example, the experimental design does not include sufficiently high concentrations, and we are missing measurements that would have a major influence on the final estimation. For this reason this result should be considered unreliable.

The fit can be further validated using so-called posterior predictive checks: the idea is to plot the observed values against the corresponding estimated predictions, along with their 95% credible interval. If the fit is correct, we expect to see 95% of the data inside the intervals.

In this plot, each black dot represents an observation made at a
given concentration, and the corresponding number of survivors at target
time is given by the value on the *x-axis*. Using the
concentration and the fitted model, we can produce the corresponding
prediction of the expected number of survivors at that concentration.
This prediction is given by the *y-axis*. Ideally observations
and predictions should coincide, so we’d expect to see the black dots on
the points of coordinate \(Y = X\). Our
model provides a tolerable variation around the predited mean value as
an interval where we expect 95% of the dots to be in average. The
intervals are represented in green if they overlap with the line \(Y=X\), and in red otherwise.

The steps for a TKTD data analysis are absolutely analogous to what
we described for the analysis at target time. Here the goal is to
estimate the relationship between chemical compound concentration, time
and survival rate using the GUTS models. GUTS, for General Unified
Threshold models of Survival, is a TKTD models generalising most of
existing mechanistic models for survival description. For details about
GUTS models, see the vignette *Models in ‘morse’ package*, and
the included references.

Here is a typical session to analyse concentration-dependent time-course data using the so-called “Stochastic Death” (SD) model:

```
# (1) load data set
data(propiconazole)
# (2) check structure and integrity of the data set
survDataCheck(propiconazole)
# (3) create a `survData` object
dat <- survData(propiconazole)
# (4) represent the number of survivors as a function of time
plot(dat, pool.replicate = FALSE)
# (5) check information on the experimental design
summary(dat)
```

To fit the *Stochastic Death* model, we have to specify the
`model_type`

as `"SD"`

:

Then, the `summary()`

function provides parameters
estimates as medians and 95% credible intervals.

Once fitting is done, we can compute posteriors vs. priors
distribution with the function `plot_prior_post()`

as
follow:

The `plot()`

function provides a representation of the
fitting for each replicates

Original data can be removed by using the option
`adddata = FALSE`

A posterior predictive check is also possible using function
`ppc()`

:

You can fix the background mortality, parameter `hb`

.

```
# fit the TKTD model SD with fixed hb value
fit_cstSDFIXhb <- survFit(dat, quiet = TRUE, model_type = "SD", hb_value=FALSE, hb_valueFIXED = 0.2)
```

In the summary the `hb`

is no more return

To have access to this value, you can simply write:

The *Individual Tolerance* (IT) model is a variant of TKTD
survival analysis. It can also be used with `morse`

as
demonstrated hereafter. For the *IT* model, we have to specify
the `model_type`

as `"IT"`

:

We can first get a summary of the estimated parameters:

And the plot of posteriors vs. priors distributions:

Here is a typical session fitting an SD or an IT model for a data set under time-variable exposure scenario.

```
# (1) load data set
data("propiconazole_pulse_exposure")
# (2) check structure and integrity of the data set
survDataCheck(propiconazole_pulse_exposure)
# (3) create a `survData` object
dat_ <- survData(propiconazole_pulse_exposure)
# (4) represent the number of survivor as a function of time
plot(dat_)
# (5) check information on the experimental design
summary(dat_)
```

GUTS models can be used to simulate the survival of the organisms
under any exposure pattern, using the calibration done with function
`survFit()`

from observed data. The function for prediction
is called `predict()`

and returns an object of class
`survFitPredict`

.

```
# (1) upload or build a data frame with the exposure profile
# argument `replicate` is used to provide several profiles of exposure
data_4prediction <- data.frame(time = c(1:10, 1:10),
conc = c(c(0,0,40,0,0,0,40,0,0,0),
c(21,19,18,23,20,14,25,8,13,5)),
replicate = c(rep("pulse", 10), rep("random", 10)))
# (2) Use the fit on constant exposure propiconazole with model SD (see previously)
predict_PRZ_cstSD_4pred <- predict(object = fit_cstSD, data_predict = data_4prediction)
```

If `NA`

are produce an `error`

message is
returned.

From an object `survFitPredict`

, results can ben plotted
with function `plot()`

:

```
# (3) Plot the predicted survival rate under the new exposure profiles.
plot(predict_PRZ_cstSD_4pred)
```

`deSolve`

It appears that with some extreme data set, the fast way used to
compute predictions return `NA`

data, due to numerical error
(e.g. number greater or lower than \(10^{300}\) or \(10^{-300}\)).

When this issue happens, the function `predict()`

returns
an error, with the message providing the way to use the robust
implementation with ODE solver provided by `deSolve`

.

This way is implemented through the use of the function
`predict_ode()`

. Robustness goes often with longer time to
compute. Time to compute can be long, so we use by default MCMC chain
size of 1000 independent iterations.

This new object `predict_PRZ_cstSD_4pred_ode`

is a
`survFitPredict`

object and so it has exactly the same
properties as an object returned by a `predict()`

function.

Note that since `predict_ode()`

can be very long to
compute, the `mcmc_size`

is reduced to 1000 MCMC chains by
default.

See for instance, with the plot:

While the model has been estimated using the background mortality
parameter `hb`

, it can be interesting to see the prediction
without it. This is possible with the argument `hb_value`

. If
`TRUE`

, the background mortality is taken into account, and
if `FALSE`

, the background mortality is set to \(0\) in the prediction.

Following EFSA recommendations, the next functions compute qualitative and quantitative model performance criteria suitable for GUTS, and TKTD modelling in general: the percentage of observations within the 95% credible interval of the Posterior Prediction Check (PPC), the Normalised Root Mean Square Error (NRMSE) and the Survival-Probability Prediction Error (SPPE).

**PPC**

The PPC compares the predicted median numbers of survivors associated to their uncertainty limits with the observed numbers of survivors. This can be visualised by plotting the predicted versus the observed values and counting how frequently the confidence/credible limits intersect with the 1:1 prediction line [see previous plot]. Based on experience, PPC resulting in less than 50% of the observations within the uncertainty limits indicate poor model performance.

**Normalised Root Mean Square Error NRMSE**

NRMSE criterion is also based on the expectation that predicted and observed survival numbers matches the 1:1 line in a scatter plot. The criterion is based on the classical root-mean-square error (RMSE), used to aggregate the magnitudes of the errors in predictions for various time-points into a single measure of predictive power. In order to provide a criterion expressed as a percentage, it is suggested using a normalised RMSE by the mean of the observations.

\[ NRMSE = \frac{RMSE}{\overline{Y}} = \frac{1}{\overline{Y}} \sqrt{\frac{1}{n} \sum_{i=1}^{n} (Y_{obs,i} - Y_{pred,i})^2} \times 100 \]

**Survival Probability Prediction Error (SPPE)**

The SPPE indicator is negative (between 0 and -100%) for an underestimation of effects, and positive (between 0 and 100%) for an overestimation of effects. An SPPE value of 0% means an exact prediction of the observed survival probability at the end of the experiment.

\[ SPPE = \left( \frac{Y_{obs, t_{end}}}{Y_{init}} - \frac{Y_{pred, t_{end}}}{Y_{init}} \right) \times 100 = \frac{Y_{obs, t_{end}} - Y_{pred, t_{end}}}{Y_{init}} \times 100 \]

For *NRMSE* and *SPPE*, we need to compute the number
of survivors. To do so, we use the function `predict_Nsurv()`

where two arguments are required: the first argument is a
`survFit`

object, and the other is a data set with four
columns (`time`

, `conc`

, `replicate`

and `Nsurv`

). Contrary to the function
`predict()`

, here the column `Nsurv`

is
necessary.

`predict_Nsurv_ode`

For the same reason that a `predict_ode`

function as been
implemented to compute `predict`

function using the ODE
solver of *deSolve*, a `predict_Nsurv_ode`

function as
been implemented as equivalent to `predict_Nsurv`

. The time
to compute is subtentially longer than the original function.

When both function work well, their results are identical (or highly similar):

Then, using object produce with the function
`predict_Nsurv()`

we can compute *PPC*, *NRMSE*
and *SPPE* for all models.

When ploting a PPC for a `survFitPredict_Nsurv`

object, 3
types of lines are represented (following EFSA recommendations). - A
plain line corresponding to the 1:1 line (\(y=x\)): prediction match perfectly with
observation when dots are on this line. - A band of dashed lines
corresponding to the range of 25% deviation. - A band of dotted lines
corresponding to the range of 50% deviation.

Following the naming of parameters in the EFSA Scientific Opinion (2018), which differs from our naming of parameters, we add an option to be in agreement with EFSA.

Several names of parameters are used in the TKTD GUTS models. The
‘R-package’ `morse`

, and more specifically since the GUTS
implementation, several name of parameters have been used.

For stability reason of algorithms and package, we do not change
parameters name in implemented algorithms. However, we added argument
`EFSA_name`

to use EFSA naming in the `summary()`

functions, and in the functions `priors_distribution()`

providing the distributions of priors (note: distributions of posteriors
are obtained with `$mcmc`

element of a
`survFit`

object) and `plot_prior_post()`

plotting
priors distributions versus posteriors distributions.

For instance:

Compared to the target time analysis, TKTD modelling allows to
compute and plot the lethal concentration for any *x* percentage
and at any time-point. The chosen time-point can be specified with
`time_LCx`

, by default the maximal time-point in the data set
is used.

```
# LC50 at the maximum time-point:
LCx_cstSD <- LCx(fit_cstSD, X = 50)
plot(LCx_cstSD)
# LC50 at time = 2
LCx(fit_cstSD, X = 50, time_LCx = 2) %>% plot()
## Note the use of the pipe operator, `%>%`, which is a powerful tool for clearly expressing a sequence of multiple operations.
## For more information on pipes, see: http://r4ds.had.co.nz/pipes.html
```

Warning messages are returned when the range of concentrations is not appropriated for one or more LCx calculation(s).

```
# LC50 at the maximum time-point:
LCx_cstIT <- LCx(fit_cstIT, X = 50)
plot(LCx_cstIT)
# LC50 at time = 2
LCx(fit_cstIT, X = 50, time_LCx = 2) %>% plot()
# LC30 at time = 15
LCx(fit_cstIT, X = 30, time_LCx = 15) %>% plot()
```

Using prediction functions, GUTS models can be used to simulate the survival rate of organisms exposed to a given exposure pattern. In general, this realistic exposure profile does not result in any related mortality, but a critical question is to know how far the exposure profile is from adverse effect, that is a “margin of safety”.

This idea is then to multiply the concentration in the realistic exposure profile by a “multiplication factor”, denoted \(MF_x\), resulting in \(x\%\) (classically \(10\%\) or \(50\%\)) of additional death at a specified time (by default, at the end of the exposure period).

The multiplication factor \(MF_x\) then informs the “margin of safety” that could be used to assess if the risk should be considered as acceptable or not.

Computing an \(MF_x\) is easy with
function `MFx()`

. It only requires object
`survFit`

and the exposure profile, argument
`data_predict`

in the function. The chosen percentage of
survival reduction is specified with argument `X`

, the
default is \(50\), and the chosen
time-point can be specified with `time_MFx`

, by default the
maximal time-point in the data set is used.

There is no explicit formulation of \(MF_x\) (at least for the GUTS-SD model), so
the `accuracy`

argument can be used to change the accuracy of
the convergence level.

```
# (1) upload or build a data frame with the exposure profile
data_4MFx <- data.frame(time = 1:10,
conc = c(0,0.5,8,3,0,0,0.5,8,3.5,0))
# (2) Use the fit on constant exposure propiconazole with model SD (see previously)
MFx_PRZ_cstSD_4MFx <- MFx(object = fit_cstSD, data_predict = data_4MFx, ode = TRUE)
```

As the computing time can be long, the function prints the
`accuracy`

for each step of the tested multiplication factor,
for the median and the 95% credible interval.

Then, we can plot the survival rate as a function of the tested multiplication factors. Note that it is a linear interpolation between tested multiplication factor (cross dots on the graph).

In this specific case, the x-axis needs to be log-scaled, what is
possible by setting option `log_scale = TRUE`

:

```
# (3 bis) Plot the survival rate as function of the multiplication factors in log-scale.
plot(MFx_PRZ_cstSD_4MFx, log_scale = TRUE)
```

As indicated, the warning message just remind you how multiplication factors and linear interpolations between them have been computed to obtain the graph.

To compare the initial survival rate (corresponding to a
multiplication factor set to 1) with the survival rate at the asked
multiplication factor leading to a reduction of \(x\%\) of survival (provided with argument
`X`

), we can use option `x_variable = "Time"`

.
The option `x_variable = "Time"`

allows to vizualize
differences in survival rate with and without the multiplication
factor.

```
# (4) Plot the survival rate versus time. Control (MFx = 1) and estimated MFx.
plot(MFx_PRZ_cstSD_4MFx, x_variable = "Time")
```

What is provided with the function `plot()`

is direclty
accessible within the object of class `MFx`

. For instance, to
have access to the median and \(95\%\)
of returned `MFx`

, we simply extract the element
`df_MFx`

which is the following `data.frame`

:

Here is an other example with a 10 percent Multiplication Factor:

```
# (2 bis) fit on constant exposure propiconazole with model SD (see previously)
MFx_PRZ_cstSD_4MFx_x10 <- MFx(object = fit_cstSD, data_predict = data_4MFx, X = 10)
```

The `warning messages`

are just saying that the quantile
at \(2.5\%\) was not possible to
compute. You can see this in the object `df_MFx`

included in
`MFx_PRZ_cstSD_4MFx_x10`

. The reason of this impossibility is
obvious when you plot the multiplication factor-response curve:

Then, you can reduce the threshold of iterations as:

```
# (2 ter) fit on constant exposure propiconazole with model SD (see previously)
MFx_PRZ_cstSD_4MFx_x10_thresh20 <- MFx(object = fit_cstSD, data_predict = data_4MFx, X = 10, threshold_iter = 20)
plot(MFx_PRZ_cstSD_4MFx_x10_thresh20, log_scale = TRUE)
```

After the `plot()`

function, you have the following
message:
`Warning message: Removed 1 rows containing missing values (geom_point).`

This message comes from the use of `ggplot()`

function (see
the `ggplot2`

package) as an echo of the warning message
about the missing point at \(2.5\%\)
that has not been computed.

Multiplication factor is also available for the GUTS IT model (option
`quiet = TRUE`

remove the output):

```
# (2) Use the fit on constant exposure propiconazole with model IT. No print of run messages.
MFx_PRZ_cstIT_4pred <- MFx(object = fit_cstIT, data_predict = data_4MFx, time_MFx = 4, quiet = TRUE)
# (3) Plot the survival rate versus multiplication factors.
plot(MFx_PRZ_cstIT_4pred, log_scale = TRUE)
```

This last example set a reduction of \(10\%\) of the survival rate, remove the
background mortality by setting `hb_value = FALSE`

and is
computed at time `time_MFx = 4`

.

```
# (2) Use the fit on constant exposure propiconazole with model IT. No print of run messages.
MFx_PRZ_cstIT_4pred <- MFx(object = fit_cstIT, X=10, hb_value = FALSE, data_predict = data_4MFx, time_MFx = 4, quiet = TRUE)
plot(MFx_PRZ_cstIT_4pred, log_scale = TRUE)
plot(MFx_PRZ_cstIT_4pred, x_variable = "Time")
```

Once we have obtained the desired multiplication factor inducing the
\(x\%\) reduction of the survival rate,
it can be relevant to explore the sentivity of this parameter by
exploring survival rate over a range of multiplication factors. This is
possible by setting argument `X = NULL`

and providing a range
of wanted multiplication factors, for instance
`MFx_range = c()`

in our first example.

```
# Use the fit on constant exposure propiconazole with model SD.
MFx_PRZ_cstSD_4pred_range <- MFx(object = fit_cstSD, data_predict = data_4MFx, X = NULL, MFx_range = 1:6)
```

The associated plot if given by:

And the argument `x_variable = "Time"`

returns all
computed time series:

To select a specific time series, we can use the element
`ls_predict`

wich is a list of object of class
`survFitPredict`

to wich a plot is defined.

The steps for reproduction data analysis are absolutely analogous to what we described for survival data. Here, the aim is to estimate the relationship between the chemical compound concentration and the reproduction rate per individual-day.

Here is a typical session:

```
# (1) load data set
data(cadmium2)
# (2) check structure and integrity of the data set
reproDataCheck(cadmium2)
```

`## Correct format`

```
# (3) create a `reproData` object
dat <- reproData(cadmium2)
# (4) represent the cumulated number of offspring as a function of time
plot(dat, concentration = 124, addlegend = TRUE, pool.replicate = FALSE)
```

```
# (5) represent the reproduction rate as a function of concentration
plotDoseResponse(dat, target.time = 28)
```

```
##
## Number of replicates per time and concentration:
## time
## conc 0 3 7 10 14 17 21 24 28 31 35 38 42 45 49 52 56
## 0 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 53 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 78 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 124 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 232 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
## 284 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6
##
## Number of survivors (sum of replicates) per time and concentration:
## 0 3 7 10 14 17 21 24 28 31 35 38 42 45 49 52 56
## 0 30 30 30 30 29 29 29 29 29 28 28 28 28 28 28 28 28
## 53 30 30 29 29 29 29 29 29 29 29 28 28 28 28 28 28 28
## 78 30 30 30 30 30 30 29 29 29 29 29 29 29 29 29 27 27
## 124 30 30 30 30 30 29 28 28 27 26 25 23 21 18 11 11 9
## 232 30 30 30 22 18 18 17 14 13 12 8 4 3 1 0 0 0
## 284 30 30 15 7 4 4 4 2 2 1 1 1 1 1 1 0 0
##
## Number of offspring (sum of replicates) per time and concentration:
## 0 3 7 10 14 17 21 24 28 31 35 38 42 45 49
## 0 0 1659 1587 2082 1580 2400 2069 2316 1822 2860 2154 3200 1603 2490 1609
## 53 0 1221 1567 1710 1773 1859 1602 1995 1800 2101 1494 2126 935 1629 2108
## 78 0 1066 2023 1752 1629 1715 1719 1278 1717 1451 1826 1610 1097 1727 2309
## 124 0 807 1917 1423 567 383 568 493 605 631 573 585 546 280 594
## 232 0 270 1153 252 30 0 37 28 46 119 19 9 0 0 0
## 284 0 146 275 18 1 0 0 0 0 0 0 0 0 0 0
## 52 56
## 0 2149 2881
## 53 1686 1628
## 78 1954 1760
## 124 328 380
## 232 0 0
## 284 0 0
```

```
# (7) fit a concentration-effect model at target-time
fit <- reproFitTT(dat, stoc.part = "bestfit",
target.time = 21,
ecx = c(10, 20, 30, 40, 50),
quiet = TRUE)
summary(fit)
```

```
## Summary:
##
## The loglogistic model with a Gamma Poisson stochastic part was used !
##
## Priors on parameters (quantiles):
##
## 50% 2.5% 97.5%
## b 1.000e+00 1.259e-02 7.943e+01
## d 1.830e+01 1.554e+01 2.107e+01
## e 1.488e+02 7.902e+01 2.804e+02
## omega 1.000e+00 1.585e-04 6.310e+03
##
## Posteriors of the parameters (quantiles):
##
## 50% 2.5% 97.5%
## b 3.856e+00 2.811e+00 6.051e+00
## d 1.776e+01 1.538e+01 2.019e+01
## e 1.370e+02 1.134e+02 1.773e+02
## omega 1.441e+00 8.322e-01 2.861e+00
##
## Posteriors of the ECx (quantiles):
##
## 50% 2.5% 97.5%
## EC10 7.733e+01 5.353e+01 1.208e+02
## EC20 9.540e+01 7.105e+01 1.389e+02
## EC30 1.098e+02 8.537e+01 1.527e+02
## EC40 1.232e+02 9.939e+01 1.653e+02
## EC50 1.370e+02 1.134e+02 1.773e+02
```

As in the survival analysis, we assume that the reproduction rate per
individual-day is a log-logistic function of the concentration. More
details and parameter signification can be found in the vignette
*Models in ‘morse’ package*.

For reproduction analyses, we compare one model which neglects the
inter-individual variability (named “Poisson”) and another one which
takes it into account (named “gamma Poisson”). You can choose either one
or the other with the option `stoc.part`

. Setting this option
to `"bestfit"`

, you let `reproFitTT()`

decides
which models fits the data best. The corresponding choice can be seen by
calling the `summary`

function:

```
## Summary:
##
## The loglogistic model with a Gamma Poisson stochastic part was used !
##
## Priors on parameters (quantiles):
##
## 50% 2.5% 97.5%
## b 1.000e+00 1.259e-02 7.943e+01
## d 1.830e+01 1.554e+01 2.107e+01
## e 1.488e+02 7.902e+01 2.804e+02
## omega 1.000e+00 1.585e-04 6.310e+03
##
## Posteriors of the parameters (quantiles):
##
## 50% 2.5% 97.5%
## b 3.856e+00 2.811e+00 6.051e+00
## d 1.776e+01 1.538e+01 2.019e+01
## e 1.370e+02 1.134e+02 1.773e+02
## omega 1.441e+00 8.322e-01 2.861e+00
##
## Posteriors of the ECx (quantiles):
##
## 50% 2.5% 97.5%
## EC10 7.733e+01 5.353e+01 1.208e+02
## EC20 9.540e+01 7.105e+01 1.389e+02
## EC30 1.098e+02 8.537e+01 1.527e+02
## EC40 1.232e+02 9.939e+01 1.653e+02
## EC50 1.370e+02 1.134e+02 1.773e+02
```

When the gamma Poisson model is selected, the summary shows an
additional parameter called `omega`

, which quantifies the
inter-individual variability (the higher `omega`

the higher
the variability).

In `morse`

, reproduction data sets are a special case of
survival data sets: a reproduction data set includes the same
information as in a survival data set plus the information on
reproduction outputs. For that reason, the S3 class
`reproData`

inherits from the class `survData`

,
which means that any operation on a `survData`

object is
legal on a `reproData`

object. In particular, in order to use
the plot function related to the survival analysis on a
`reproData`

object, we can use `survData`

as a
conversion function first:

In Bayesian inference, the parameters of a model are estimated from the data starting from a so-called

*prior*, which is a probability distribution representing an initial guess on the true parameters, before seing the data. The*posterior*distribution represents the uncertainty on the parameters after seeing the data and combining them with the prior. To obtain a point estimate of the parameters, it is typical to compute the mean or median of the posterior. We can quantify the uncertainty by reporting the standard deviation or an inter-quantile distance from this posterior distribution.↩︎