Preliminaries: first, we need to load required libraries.

```
library(ggplot2)
library(nlraa)
library(car)
```

`## Loading required package: carData`

`library(nlme)`

The bootstrap is a technique in statistics which consists of resampling the observed data in order to create an empirical distribution of some statistic (Fox and Weisberg, 2012, 2017). This is often done to evaluate the assumptions of the model or to derive empirical distributions which are difficult to approximate. For nonlinear models, bootstrapping can be useful because often questions arise that a typical analysis does not answer. In many instances the distributions of parameters from a nonlinear model are in fact normal and standard approximations work well, in other instances, however, this is not the case.

In the following sections I will illustrate the use of the bootstrap for nonlinear model (nls), generalized nonlinear models (gnls) and nonlinear mixed models (nlme). In another section I will illustrate the ability to simulate from nonlinear (mixed) models

The first type of models illustrated here are nonlinear models for which we make standard assumptions for error distributions and there are no random effects.

As a simple example, we can use the dataset ‘barley’ in the ‘nlraa’ package. This represents the response of barley yield to different doses of fertilizer over several seasons. We will ignore the effect of ‘year’ in this section.

```
data(barley, package = "nlraa")
## Quick visualization
ggplot(data = barley, aes(x = NF, y = yield)) + geom_point()
```

We can fit a very simple model known as the linear-plateau.

```
## Linear-plateau model
## The function SSlinp is in the 'nlraa' package
<- nls(yield ~ SSlinp(NF, a, b, xs), data = barley)
fit.lp
## Visualize data and fit
ggplot(data = barley, aes(x = NF, y = yield)) +
geom_point() +
geom_line(aes(y = fitted(fit.lp)))
```

At this point several questions might arise:

- What are the confidence intervals for the model parameters?
- How do we describe the uncertainty around the fitted values of the model?
- What is the estimate and confidence interval for the asymptote?

The confidence intervals for the model parameters can be derived using a method called profiling. The generic function ‘confint’ can be used which invokes ‘confint.nls’ from the ‘MASS’ package.

`confint(fit.lp)`

`## Waiting for profiling to be done...`

```
## 2.5% 97.5%
## a 104.821725 167.73660
## b 19.937411 38.00267
## xs 5.849474 12.23099
```

For nonlinear models this method is preferred to simple confidence intervals that would directly use the standard error of the model parameters. In many cases, there is good agreement between the two, but this is not always the case and it can be informative to compare both methods. Those standard errors can be obtained using the ‘summary’ function, along with hypothesis testing.

`summary(fit.lp)`

```
##
## Formula: yield ~ SSlinp(NF, a, b, xs)
##
## Parameters:
## Estimate Std. Error t value Pr(>|t|)
## a 137.067 16.179 8.472 1.83e-12 ***
## b 27.501 5.269 5.219 1.62e-06 ***
## xs 7.682 1.211 6.342 1.68e-08 ***
## ---
## Signif. codes: 0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
##
## Residual standard error: 70.79 on 73 degrees of freedom
##
## Number of iterations to convergence: 2
## Achieved convergence tolerance: 6.548e-16
```

Some interesting plots can be produced which illustrate the symmetry (or lack thereof) of the distribution for the model parameters. Profile plots which show large departures from normality would indicate that the normal approximation for the confidence intervals might not be the best choice.

```
## For the intercept
plot(profile(fit.lp, "a"))
```

```
## This one is fairly symetric and the normal approximation is reasonable
plot(profile(fit.lp, "b"))
```

`plot(profile(fit.lp, "xs"))`

`## These last parameters are less symetrical`

Being able to see the whole profile for a parameter can be interesting in terms of understanding its behavior. For example, in this model, which has a ‘break-point’, we would not expect all the parameters to have identical shaped profiles and that is illustrated above for parameter ‘xs’.

In this case, bootstraping is an alternative to computing the confidence intervals and it can be used as a way to double-check the previous results, but it can also be used when profiling fails. A couple of functions which can perform bootstraping for nonlinear models are ‘boot_nls’ in the ‘nlraa’ pacakge or the ‘Boot’ in the ‘car’ package.

```
## psim = 3 just adds residuals and does not resample parameters
<- boot_nls(fit.lp, psim = 3) fit.lp.Bt
```

`## Number of times model fit did not converge 169 out of 999`

`## Or you can use the Boot function in the car package`

In this case, this function uses the base ‘boot’ package under the hood and it returns an object of that class. This also allows for other options such as defining a function as a combination of model parameters, and the use of more than one core to speed up computation. In this case, since we are also interested in the asymptote, which is ‘a + b * xs’, we can obtain confidence intervals for this parameter by doing the following.

```
## I'm using Boot in the 'car' package here, but it is similar to 'boot_nls'
<- function(x) coef(x)[1] + coef(x)[2] * coef(x)[3]
fn <- Boot(fit.lp, f = fn, labels = "asymptote") fit.lp.Bt.asymp
```

```
##
## Number of bootstraps was 811 out of 999 attempted
```

`confint(fit.lp.Bt.asymp)`

```
## Bootstrap bca confidence intervals
##
## 2.5 % 97.5 %
## asymptote 323.9545 382.7751
```

`hist(fit.lp.Bt.asymp)`

The bootstrap method takes a few seconds here, but it can be computationally much more demanding in larger problems since it re-fits the model many, many times. An alternative is to use the delta method, for which there is a function in the ‘car’ pacakge. The delta method has the advantage of being fast and the disadvantage that it makes the assumption of a normal distribution. In this case the lower bound for the confidence interval is similar to the one obtained with the bootstrap, but the upper bound for the confidence interval is somewhat higher using the bootstrap (381 vs. 370). Since the bootstrap method makes fewer assumptions it is probably the better one to report.

```
<- deltaMethod(fit.lp, "a + b * xs")
fit.lp.Dlt.asymp fit.lp.Dlt.asymp
```

```
## Estimate SE 2.5 % 97.5 %
## a + b * xs 348.326 11.484 325.817 370.84
```

In order to answer our second question above related to the uncertainty around the fitted values we could plug-in the values from the bootstrap sampling and obtain different regression lines.

```
## The object 't' in the bootstrap run has
## the parameter estimate values
## First remove missing values
<- na.omit(fit.lp.Bt$t)
fit.lp.Bt.prms
<- length(unique(barley$NF))
nrb <- nrow(fit.lp.Bt.prms)
nrp
## Set up an empty data.frame
<- data.frame(i = as.factor(rep(1:nrp, each = nrb)), NF = rep(unique(barley$NF), nrp), prd = NA)
prd.dat
## A simple loop can be used to run the model multiple times
for(i in 1:nrp){
<- fit.lp.Bt.prms[i,1]
a.i <- fit.lp.Bt.prms[i,2]
b.i <- fit.lp.Bt.prms[i,3]
xs.i
c(1 + (nrb*(i - 1))):c(i * nrb),3] <- linp(unique(barley$NF), a.i, b.i, xs.i)
prd.dat[
}
## Plot the data with the original fit and the uncertainty
ggplot() +
geom_line(data = prd.dat, aes(x = NF, y = prd, group = i),
color = "gray", alpha = 0.2) +
geom_line(data = barley, aes(x = NF, y = fitted(fit.lp))) +
geom_point(data = barley, aes(x = NF, y = yield)) +
ylab("Yield") + xlab("Nitrogen rate") +
ggtitle("Using results from Boot \n and plug-in into linp")
```

The previous graph shows the black line with the original fit and the variability in the fitted values due to the resampling generated during the bootstrap process. The function ‘predict.nls’ at the moment ignores the arguments ‘se.fit’ and ‘interval’ which means that this functionality is not available in the base ‘stats’ package. (There is an alternative approach in the ‘propagate’ package - see references.)

So we could use the quantiles of ‘prd.dat’ object above to derive confidence intervals for the regression line. The previous example is an attempt to make it clear how the uncertainty could be displayed. Equivalently, we can use ‘Boot’ for this purpose, but with some effort manipulating the data.

```
<- function(x) predict(x, newdata = data.frame(NF = 0:14))
fn2 <- Boot(fit.lp, fn2) fit.lp.Bt2
```

```
##
## Number of bootstraps was 807 out of 999 attempted
```

```
<- na.omit(fit.lp.Bt2$t)
fttd <- c(t(fttd))
prds <- data.frame(i = as.factor(rep(1:nrow(fttd), each = ncol(fttd))),
ndat NF = rep(0:14, nrow(fttd)))
$prd <- prds
ndat
## Essentially the same graph as the one above
ggplot() +
geom_line(data = ndat, aes(x = NF, y = prd, group = i),
color = "gray", alpha = 0.2) +
geom_line(data = barley, aes(x = NF, y = fitted(fit.lp))) +
geom_point(data = barley, aes(x = NF, y = yield)) +
ylab("Yield") + xlab("Nitrogen rate")
```

Implementing bootstrap for more complex models takes extra work. For this, I’m taking the approach of sampling from the vector of fixed parameters and also bootstrapping the standardized residuals for ‘gnls’ and ‘nlme’ objects. (This methodology can be improved in the future.) This takes advantage that we assume that the residuals are normally distributed.

As a first example, we can compare the bootstrapped confidence intervals with the ones obtained by ‘intervals’. Note: I’m running this only a few hundred times in the examples below for efficiency, but it is a good practice to run these a few thousand times.

```
set.seed(101)
## Simplify the dataset to make the set up simpler
<- subset(barley, year < 1974)
barley2
<- gnls(yield ~ SSlinp(NF, a, b, xs), data = barley2)
fit.lp.gnls2
intervals(fit.lp.gnls2)
```

```
## Approximate 95% confidence intervals
##
## Coefficients:
## lower est. upper
## a 138.79981 176.800000 214.800186
## b 13.75248 27.603093 41.453706
## xs 3.64251 6.174127 8.705744
## attr(,"label")
## [1] "Coefficients:"
##
## Residual standard error:
## lower est. upper
## 25.50341 35.17935 56.67543
```

```
## Compare this to the bootstrapping approach
## R = 200 is too low for bootstrap, this is for illustration only
<- boot_nlme(fit.lp.gnls2, R = 200) fit.lp.gnls2.bt
```

`## Number of times model fit did not converge 22 out of 200`

`summary(fit.lp.gnls2.bt)`

```
##
## Number of bootstrap replications R = 178
## original bootBias bootSE bootMed
## 1 176.8000 -1.881489 23.1895 174.7183
## 2 27.6031 1.095069 7.4412 28.4382
## 3 6.1741 -0.072778 1.3834 5.9537
```

`confint(fit.lp.gnls2.bt, type = "perc")`

```
## Bootstrap percent confidence intervals
##
## 2.5 % 97.5 %
## 1 125.320034 224.479385
## 2 14.440280 42.522212
## 3 4.040724 9.591933
```

The confidence intervals obtained by bootstrap are wider (as expected) than the ones obtained using intervals because they consider the uncertainty in the parameters of the nonlinear model. The next example shows a slightly more complex model in which we introduce the (fixed) effect of year for a subset of the data.

```
set.seed(101)
$year.f <- as.factor(barley2$year)
barley2
<- coef(fit.lp.gnls2)
cfs
<- update(fit.lp.gnls2,
fit.lp.gnls3 params = list(a + b + xs ~ year.f),
start = c(cfs[1], 0, 0, 0,
2], 0, 0, 0,
cfs[3], 0, 0, 0))
cfs[
## This bootstraps the vector of parameters
<- boot_nlme(fit.lp.gnls3, R = 300) fit.lp.gnls3.bt
```

`## Number of times model fit did not converge 152 out of 300`

`confint(fit.lp.gnls3.bt, type = "perc")`

```
## Bootstrap percent confidence intervals
##
## 2.5 % 97.5 %
## 1 111.1375192 196.119143
## 2 -45.3383112 86.644389
## 3 -52.9212287 65.101962
## 4 -0.5059473 114.993794
## 5 14.3750448 41.574802
## 6 -20.9463834 15.697041
## 7 -17.7796470 16.955886
## 8 -18.1083851 16.965825
## 9 4.4735481 9.074686
## 10 -2.6628309 4.960814
## 11 -4.1799869 1.682101
## 12 -2.8113050 3.862596
```

`hist(fit.lp.gnls3.bt, 1, ci = "perc")`

This is simply to illustrate the use of bootstrapping for a ‘gnls’ object, which is something that function ‘car::Boot’ does not seem to be able to handle (the deltaMethod function also fails for this type of model, because it cannot handle the names in the vector of coefficients which uses periods).

For illustration, I will continue to use the barley example, but this time the model is fitted to each year individually and then a nonlinear mixed model which assumes a diagonal matrix for the random effects (for simplicity). In this case we want an esimtate of the confidence interval for the asymptote which is not an explicit parameter but rather a combination ‘a + b * xs’ of the three parameters.

```
set.seed(101)
$year.f <- as.factor(barley$year)
barley
<- groupedData(yield ~ NF | year.f, data = barley)
barleyG
<- nlsList(yield ~ SSlinp(NF, a, b, xs), data = barleyG)
fitL.bar
<- nlme(fitL.bar, random = pdDiag(a + b + xs ~ 1))
fit.bar.nlme
## Confidence intervals of the model fixed parameters
intervals(fit.bar.nlme, which = "fixed")
```

```
## Approximate 95% confidence intervals
##
## Fixed effects:
## lower est. upper
## a 109.212647 134.663569 160.11449
## b 23.970656 27.939486 31.90832
## xs 6.823868 7.671139 8.51841
## attr(,"label")
## [1] "Fixed effects:"
```

```
## Function which computes the asymptote
<- function(x) fixef(x)[1] + fixef(x)[2] * fixef(x)[3]
fna
## Bootstrap the model for the asymptote
<- boot_nlme(fit.bar.nlme, f = fna, R = 200) fit.bar.nlme.bt
```

`## Number of times model fit did not converge 90 out of 200`

`confint(fit.bar.nlme.bt, type = "perc")`

```
## Bootstrap percent confidence intervals
##
## 2.5 % 97.5 %
## 1 306.707 382.6985
```

`hist(fit.bar.nlme.bt, ci = "perc")`

For this model the estimate for the asymptote is 349, but from the model fit we do not get a direct confidence interval. Bootstrapping makes this easy (but it takes a bit of time) and it results in 307, 383.

For this example I will use the Orange dataset. The goal here is to place confidence bands around the mean response function.

```
data(Orange)
## This fits a model which considers the fact that
## the variance typically increases as the fitted
## values increase
<- gnls(circumference ~ SSlogis(age, Asym, xmid, scal),
fitg data = Orange, weights = varPower())
## Here we use bootstrapping to investigate
## the uncertainty around the fitted values
<- boot_nlme(fitg, fitted, psim = 1, R = 300) fitg.bt1
```

`## Number of times model fit did not converge 3 out of 300`

```
## Compute 90% quantile confidence bands
<- apply(t(fitg.bt1$t), 1, quantile, probs = 0.05, na.rm = TRUE)
lwr1.q <- apply(t(fitg.bt1$t), 1, quantile, probs = 0.95, na.rm = TRUE)
upr1.q
ggplot() +
geom_point(data = Orange, aes(x = age, y = circumference)) +
geom_line(data = Orange, aes(x = age, y = fitted(fitg))) +
geom_ribbon(aes(x = Orange$age, ymin = lwr1.q, ymax = upr1.q),
fill = "purple", alpha = 0.2) +
ggtitle("Orange fit using the logistic: \n 90% confidence band for the mean function")
```

Here I will continue with the Orange dataset and illustrate different types of simulation which are relevant depending on what the inference scope is. This simulation is, in fact, used inside of the bootstrap function above; it is a necessary step for bootstrap for these type of models.

The first level of predictions we can generate for the Orange dataset are at the ‘population’ level. These are similar to the fitted values for the ‘gnls’ case above. Notice that we do not have the option of requesting standard errors for the predicitons for this model either.

```
<- nlsList(circumference ~ SSlogis(age, Asym, xmid, scal), data = Orange)
fmoL
<- nlme(fmoL, random = pdDiag(Asym + xmid + scal ~ 1))
fmo
## Just one simulation, because with psim = 0 and level = 0, we are
## computing the predicted values at level = 0 (?predict.nlme)
<- simulate_nlme(fmo, nsim = 1, psim = 0, level = 0)
sim00
<- cbind(Orange, prd = as.vector(sim00))
dat00
ggplot(data = dat00) +
geom_point(aes(x = age, y = circumference)) +
geom_line(aes(x = age, y = prd)) +
ggtitle("psim = 0, level = 0")
```

One approach we can use to assess the uncertainty in the response would be to sample from the vector of parameters and the associated variance-covariance matrix. In this case, we could do many realizations, but I’ll keep it to just 100. A frequentist interpretation of the bands below would be that in repeated sampling from this same population we would expect that the true relationship will be within this band with a specified probability. One way to compute these bands (which I’m not doing below), would be to calculate the quantiles of the empirical distribution. These could be called confidence bands, which would be equivalent to the ‘interval = confidence’ in a linear model (?predict.lm).

```
<- simulate_nlme(fmo, nsim = 100, psim = 1, level = 0, value = "data.frame")
sdat10
ggplot(data = sdat10) +
geom_line(aes(x = age, y = sim.y, group = ii), color = "gray", alpha = 0.5) +
geom_point(aes(x = age, y = circumference)) +
ggtitle("psim = 1, level = 0") +
ylab("circumference")
```

In this case, each tree is an ‘individual’ and the previous one is a population level confidence. Next, we would like to produce individual level ‘prediction’ bands. This is, if we were able to resample from a population with a similar structure and if we wanted to have bands that include the true relationship between circumference and age for these trees, we could produce the confidence bands from the quantiles of the empirical distribution of the lines in the figure below (we would need a few thousand samples in order to do this).

```
<- simulate_nlme(fmo, nsim = 100, psim = 1, level = 1, value = "data.frame")
sdat11
## We need a tree simulation ID
## for plotting
$Tree_ID <- with(sdat11, paste0(Tree,"_",ii))
sdat11
ggplot(data = sdat11) +
facet_wrap(~ Tree) +
geom_line(aes(x = age, y = sim.y, color = Tree, group = Tree_ID),
alpha = 0.5) +
geom_point(aes(x = age, y = circumference)) +
ggtitle("psim = 1, level = 1") +
ylab("circumference") +
theme(legend.position = "none")
```

Finally, the next level would be to produce intervals that would likely contain a future observation instead of just the relationship, which was captured in the figure above. The ‘bands’ below are wider than in the previous figure. The reason why the difference is small, is, because in this case the residual variance is comparatively small.

```
<- simulate_nlme(fmo, nsim = 100, psim = 2, level = 1, value = "data.frame")
sdat21
## Here I'm plotting points to emphasize that we are making
## predictions at the level of a single observation
ggplot(data = sdat21) +
facet_wrap(~ Tree) +
geom_point(aes(x = age, y = sim.y, color = Tree),
alpha = 0.5) +
geom_point(aes(x = age, y = circumference)) +
ggtitle("psim = 2, level = 1") +
ylab("circumference") +
theme(legend.position = "none")
```

The previous examples are simulations, but in order to generate confidence intervals or bands we would need to perform bootstrap, but it takes much longer, so I’m not including it in this vignette.

Using the previous simulation functions as a backbone it is possible to perform prediction. In this case, I’m also illustrating how to perform multimodel averaging.

```
## All models should be fitted using Maximum Likelihood
<- nlme(circumference ~ SSlogis(age, Asym, xmid, scal),
fm.L random = pdDiag(Asym + xmid + scal ~ 1),
method = "ML", data = Orange)
<- nlme(circumference ~ SSgompertz(age, Asym, b2, b3),
fm.G random = pdDiag(Asym + b2 + b3 ~ 1),
method = "ML", data = Orange)
<- nlme(circumference ~ SSfpl(age, A, B, xmid, scal),
fm.F random = pdDiag(A + B + xmid + scal ~ 1),
method = "ML", data = Orange)
<- nlme(circumference ~ SSbg4rp(age, w.max, lt.e, ldtm, ldtb),
fm.B random = pdDiag(w.max + lt.e + ldtm + ldtb ~ 1),
method = "ML", data = Orange)
## Let's compare the models
print(IC_tab(fm.L, fm.G, fm.F, fm.B), digits = 2)
```

```
## model df AIC AICc BIC dAIC weight
## 1 fm.L 7 277 281 288 0.0 0.656
## 3 fm.F 9 280 287 294 2.8 0.159
## 2 fm.G 7 280 284 291 3.1 0.137
## 4 fm.B 9 282 290 296 5.3 0.047
```

The function *predict_nlme* can take one or more models and
perform model averaged prediciton.

```
## Each model prediction is weighted according to their AIC values
<- predict_nlme(fm.L, fm.G, fm.F, fm.B)
prd
ggplot(data = Orange, aes(x = age, y = circumference)) +
geom_point() +
geom_line(aes(y = predict(fm.L, level = 0), color = "Logistic")) +
geom_line(aes(y = predict(fm.G, level = 0), color = "Gompertz")) +
geom_line(aes(y = predict(fm.F, level = 0), color = "4P-Logistic")) +
geom_line(aes(y = predict(fm.B, level = 0), color = "Beta")) +
geom_line(aes(y = prd, color = "Avg. Model"), size = 1.2)
```

And we can also compute and visualize confidence intervals.

```
<- predict_nlme(fm.L, fm.G, fm.F, fm.B, interval = "confidence", level = 0.90)
prdc <- cbind(Orange, prdc)
OrangeA
ggplot(data = OrangeA, aes(x = age, y = circumference)) +
geom_point() +
geom_line(aes(y = predict(fm.L, level = 0), color = "Logistic")) +
geom_line(aes(y = predict(fm.G, level = 0), color = "Gompertz")) +
geom_line(aes(y = predict(fm.F, level = 0), color = "4P-Logistic")) +
geom_line(aes(y = predict(fm.B, level = 0), color = "Beta")) +
geom_line(aes(y = prd, color = "Avg. Model"), size = 1.2) +
geom_ribbon(aes(ymin = Q5, ymax = Q95), fill = "purple", alpha = 0.3)
```

In the case of simulating data from a nonlinear model and how this
relates to bootstrapping. I’ve been thinking about a number of things.
In a nonlinear model, the mean function should account for a
*substantial* part of the variability. If we think about models
in terms of

\[ data = signal + noise \] And for mixed models: \[ data = fixed + random + residual \]

I can see how in the linear case often the random part can be of
substantial interest and sometimes more important than the fixed part,
but for nonlinear models, the *fixed* part should dominate,
otherwise we would need a better function or the data are not amenable
to nonlinear modeling. In simulate_nlme, psim = 0, gives you the fitted
values, so should be equivalent to ‘predict.nlme’, if theres is ever a
discrepancy, then that is a bug. For psim = 1, I sample from the vector
of parameters assuming that they are normally distributed. This seems to
be very reasonable in practice and, in fact, much more important is the
variance-covariance matrix of the vector of parameters and in particular
the correlations. The problem is that in nonlinear mixed models, the
relationship among parameters is not linear and can look a bit like a
‘banana’. This can be investigated through the use of bootstrap. For
psim = 2, I add the residual variability which considers the modeling of
the variance (heterogeneous variances) and also the correlation
structure. With the psim = 3 option it is possible to sample new
subjects or individuals and this option is available for both lme and
nlme objects. By default, ‘predict.nlme’ predicts at the deepest level
in the hierarchy (i.e. Q) and changing this argument allows for
simulations at the different levels. A good reason for using psim = 2
(instead of psim = 3) is that, often, there is meaning or interest in
these specific random terms. For example,
sites/locations/subjects/experimenta.units are higher or lower because,
in part, some characteristics that are know and some unknown, but which
are out of our control. For this reason, assigning resampling or
assigning a deviation at random at this level might be undesirable. If
we have 10 sites labeled “A” through “J” and we know that site “A” is
higher due to some characteristics (which we do not control), it would
not make sense to assign the deviation from site “A” to “J” (which might
be the lowest). In our analysis it might still make sense to treat
‘sites’ as random. In this case, when we use simulate_nlme, it makes
more sense to use ‘psim = 2, level = 1’ (‘site’ would be our deepest
level in the hierarchy - like ‘Tree’ in the ‘Orange’ example above) to
generate data which appears similar to the data which was actually
collected. In the current approach, I do not resample new values for the
estimated variance-covariance matrix of the random effects, in part,
because these are often poorly estimated (i.e. very low precision). If
you need this, a fully Bayesian approach would be better.

J. Fox and S. Weisberg. An R Companion to Applied Regression. Sage, Thousand Oaks CA, 2nd edition, 2011. URL http://z.umn.edu/carbook.

J. Fox and S. Weisberg. Bootstrapping regression models in R. Technical report, 2017. URL: https://socialsciences.mcmaster.ca/jfox/Books/Companion-2E/appendix/Appendix-Bootstrapping.pdf

J. Fox and S. Weisberg. Nonlinear Regression, Nonlinear Least Squares, and Nonlinear Mixed Models in R. Appendix, 2018. URL: https://socialsciences.mcmaster.ca/jfox/Books/Companion/appendices/Appendix-Nonlinear-Regression.pdf

Morris, J. S. (2002). The BLUPs Are Not ‘best’ When It Comes to Bootstrapping. Statistics & Probability Letters 56(4): 425–430. https://doi.org/10.1016/S0167-7152(02)00041-X.

http://sia.webpopix.org/nonlinearRegression.html#confidence-intervals-and-prediction-intervals

propagate: https://rmazing.wordpress.com/2013/08/31/introducing-propagate/

For linear mixed models: package ‘lmeresampler’: https://CRAN.R-project.org/package=lmeresampler