7 | Parameter estimation 1: nonlinear least squares Flashcards

1
Q

Name a method discussed in the lecture to estimate the parameters of a systems biology model from experimental data. When is a set of parameters considered to be optimal in this method?

(2023_1, 2021_1 - question on NFkB Hoffman paper)

A

Nonlinear least squares.
Optimal when parameters minimize sum of sq. R between observed data, model predictions.

OR

Likelihood based methods.
Optimal when parameters maximize likelihood function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Which of the following regression functions are nonlinear?

f(z;θ) = θ1 + zθ2 + z2θ3

f(z;θ) = θ1 + (z2θ2) / (1 + z2)

f(z;θ)= θ1e−zθ2

The regression function corresponding to the reaction system
∗ ⟶(θ1) S ⟶(θ2) ∗
with observation function h(x)=x

(ungraded quiz)
.

A

f(z;θ)= θ1e−zθ2

The regression function corresponding to the reaction system
∗ ⟶(θ1) S ⟶(θ2) ∗
with observation function h(x)=x

(nonlinear regression function = nonlinear in the parameters, not x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the difference between the errors and the residuals?

(ungraded quiz)

A

Errors describe randomness in the data generating process, while residuals describe misfit of a model assuming a certain parameter (e.g., least squares estimate).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

True or false

Residuals describe randomness in the data generating process, while errors describe misfit of a model assuming a certain parameter (e.g., least squares estimate).

(ungraded quiz)

A

F

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

True or false

Errors describe randomness in the data generating process, while residuals describe misfit of a model assuming a certain parameter (e.g., least squares estimate).

(ungraded quiz)

A

T

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

True or false

Errors and residuals can be used synonymously

(ungraded quiz)

A

F

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Consider the nonlinear regression model:

yi = f(zi;θ) + εi.

What can be done if the scatter in the residuals ri increases with values of the dependent variable yi?

☐ Use weighted least squares estimation
☐ This is no problem for the nonlinear regression model.
☐ Use the transform-both-sides-approach
☐ Use a more complex regression function

(ungraded quiz)

A

Use the transform-both-sides-approach:
involves transforming both the dependent variable and the independent variables to stabilize the variance. For example, taking the logarithm of both sides can help if the variance increases with the mean.

Use weighted least squares estimation
method that assigns different weights to different observations based on the variance of their errors. When the scatter of residuals increases with the dependent variable, WLS can be used to give less weight to observations with higher variance, thereby stabilizing the variance across all observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

True or false

The least squares estimate can always be found analytically.

(ungraded quiz)

A

F

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

True or false

For ODE-based models, the regression function cannot always be stated analytically.

(ungraded quiz)

A

T

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

True or false

Residual plots allow to detect misspecifications of the error model, but not the regression function.

(ungraded quiz)

A

F

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

True or false

The least squares estimate solves the equation R(θ)=0, where R is the sum of squared residuals.

(ungraded quiz)

A

F

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

True or false

For nonlinear regression functions, the least squares estimate cannot be found analytically.

(ungraded quiz)

A

T

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

True or false

Residual plots allow to detect misspecifications of the regression function or the error model

(ungraded quiz)

A

T

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Assume we want to find a polynomial of a suitable degree that fits a dataset consisting of x-y pairs.

For a fixed degree K of the polynomial, specify the regression function; then explain how least squares can be used to estimate the coefficients of the polynomial.
(2022_2)

A

Model:
f(xi; θ) = θ1 + θ2x2 + θ3x3 + θKxK
Observed values vs model:
yi = f(xi;θ) + εi.

How to do nonlinear least squares:
find the coefficients θ that minimize the sum of the squared differences between the observed values and the values predicted by the polynomial model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Assume we want to find a polynomial of a suitable degree that fits a dataset consisting of x-y pairs.

What is the problem with using the sum of squared residuals R to find the best polynomial degree? What could be done instead?

(2022_2)

A

main problem is overfitting:
- As degree of polynomial increases, R always decrease / stay same, even if higher degree polynomial not actually better model
- (higher degree polynomial has more flexibility to fit the specific data points, including noise and random fluctuations, rather than capturing the true underlying relationship)

As a result, using R alone will always favor the highest degree polynomial possible, which may not generalize well to new data.

Instead of solely relying on the sum of squared residuals, model selection criteria that penalize model complexity can be used, for example
- likelihood ratio test
- Akaike Information Criterion (AIC):

How well did you know this?
1
Not at all
2
3
4
5
Perfectly