7 | Parameter estimation 1: nonlinear least squares Flashcards
Name a method discussed in the lecture to estimate the parameters of a systems biology model from experimental data. When is a set of parameters considered to be optimal in this method?
(2023_1, 2021_1 - question on NFkB Hoffman paper)
Nonlinear least squares.
Optimal when parameters minimize sum of sq. R between observed data, model predictions.
OR
Likelihood based methods.
Optimal when parameters maximize likelihood function
Which of the following regression functions are nonlinear?
f(z;θ) = θ1 + zθ2 + z2θ3
f(z;θ) = θ1 + (z2θ2) / (1 + z2)
f(z;θ)= θ1e−zθ2
The regression function corresponding to the reaction system
∗ ⟶(θ1) S ⟶(θ2) ∗
with observation function h(x)=x
(ungraded quiz)
.
f(z;θ)= θ1e−zθ2
The regression function corresponding to the reaction system
∗ ⟶(θ1) S ⟶(θ2) ∗
with observation function h(x)=x
(nonlinear regression function = nonlinear in the parameters, not x)
What is the difference between the errors and the residuals?
(ungraded quiz)
Errors describe randomness in the data generating process, while residuals describe misfit of a model assuming a certain parameter (e.g., least squares estimate).
True or false
Residuals describe randomness in the data generating process, while errors describe misfit of a model assuming a certain parameter (e.g., least squares estimate).
(ungraded quiz)
F
True or false
Errors describe randomness in the data generating process, while residuals describe misfit of a model assuming a certain parameter (e.g., least squares estimate).
(ungraded quiz)
T
True or false
Errors and residuals can be used synonymously
(ungraded quiz)
F
Consider the nonlinear regression model:
yi = f(zi;θ) + εi.
What can be done if the scatter in the residuals ri increases with values of the dependent variable yi?
☐ Use weighted least squares estimation
☐ This is no problem for the nonlinear regression model.
☐ Use the transform-both-sides-approach
☐ Use a more complex regression function
(ungraded quiz)
Use the transform-both-sides-approach:
involves transforming both the dependent variable and the independent variables to stabilize the variance. For example, taking the logarithm of both sides can help if the variance increases with the mean.
Use weighted least squares estimation
method that assigns different weights to different observations based on the variance of their errors. When the scatter of residuals increases with the dependent variable, WLS can be used to give less weight to observations with higher variance, thereby stabilizing the variance across all observations.
True or false
The least squares estimate can always be found analytically.
(ungraded quiz)
F
True or false
For ODE-based models, the regression function cannot always be stated analytically.
(ungraded quiz)
T
True or false
Residual plots allow to detect misspecifications of the error model, but not the regression function.
(ungraded quiz)
F
True or false
The least squares estimate solves the equation R(θ)=0, where R is the sum of squared residuals.
(ungraded quiz)
F
True or false
For nonlinear regression functions, the least squares estimate cannot be found analytically.
(ungraded quiz)
T
True or false
Residual plots allow to detect misspecifications of the regression function or the error model
(ungraded quiz)
T
Assume we want to find a polynomial of a suitable degree that fits a dataset consisting of x-y pairs.
For a fixed degree K of the polynomial, specify the regression function; then explain how least squares can be used to estimate the coefficients of the polynomial.
(2022_2)
Model:
f(xi; θ) = θ1 + θ2x2 + θ3x3 + θKxK
Observed values vs model:
yi = f(xi;θ) + εi.
How to do nonlinear least squares:
find the coefficients θ that minimize the sum of the squared differences between the observed values and the values predicted by the polynomial model
Assume we want to find a polynomial of a suitable degree that fits a dataset consisting of x-y pairs.
What is the problem with using the sum of squared residuals R to find the best polynomial degree? What could be done instead?
(2022_2)
main problem is overfitting:
- As degree of polynomial increases, R always decrease / stay same, even if higher degree polynomial not actually better model
- (higher degree polynomial has more flexibility to fit the specific data points, including noise and random fluctuations, rather than capturing the true underlying relationship)
As a result, using R alone will always favor the highest degree polynomial possible, which may not generalize well to new data.
Instead of solely relying on the sum of squared residuals, model selection criteria that penalize model complexity can be used, for example
- likelihood ratio test
- Akaike Information Criterion (AIC):