L7 - Gradient Descent, Convex Optimisation, Bracketing Flashcards
What is the purpose of Gradient Descent? What type of function is it?
An iterative optimisation function to determine the minimum point of a convex set. It is a Convex Function.
How does Gradient Descent work to solve Convex Optimisation?
Algorithm iteratively tunes the parameters in order to iterate towards the minimum point. As algorithm nears the minimum point, the jumps towards minimum decrease so as not to overshoot.
In the context of ML, what does the gradient represent?
The relationship between independent and dependent variables.
What is tuned in order to establish the minimum of a convenient function?
Parameter of the GD algorithm.
In ML, what is GD used to establish?
Minimum of the models cost function.
Upon each iteration of GD, what happens to the size of the steps towards the min. value? Why is this needed?
Size of step decreases to safeguard against overshooting the minimum point.
What is a Convex Optimisation problem?
A problem in which we want to find the minimum or maximum of a convex function.
What determines if a function is convex?
A function is convex if for any 2 points in the domain, a line drawn between the points is always above the function graph.
How many minimum points does a convex function have?
1
How do we use Jensens Inequality or the Second Derivative Test to determine whether a function is convex?
- Jensens Inequality - If for any 2 X and Y points in the domain, and for all 0 < t < 1, the line connecting the points is entirely above the function graph.
- Second Derivative Test - If the second derivative of the function is non-zero for all X’s in the domain, the function is known to be convex.
What is the epigraph of a function? How can we used it to determine whether a function is convex?
The set of points in the domain the are located above the function graph.
Give the definition of a Convex Set…
A set in which for any point in the set, the point if above the function graph.
At the minimum of a convex function, what is the value of the derivative?
0
What techniques do we use to find the minimum of a non-convex function?
Downhill Simplex or Bracketing
Explain Bracketing…
Method used to find the minimum of a non-convex set in 1-d space. Start by establishing initial points either side of the minimal area (we must know the rough area of minimum). Iteratively decrease the bracket size, converging towards the minimum point. Repeat until convergence or timeout.