12 - Continual/Lifelong Machine Learning Flashcards
Traditional machine learning dev cycle
- First trained on suitable dataset
- System performance is assessed on a test benchmark
- System is frozen and deployed in an application scenario
- If conditions change the system is decommissioned
Problem with traditional AI systems after deployment
They do not improve after deployment.
- Test data is prepared: balanced classes, interleaved learning, static distributions
^ This may not happen in the real world
Catastrophic Forgetting
When you learn further teasks, you forget previous ones.
Replay based methods (forgetting)
Relay old samples from previous tasks while learning new.
- Learn from online data and keep examples of new clases as they show.
- Interleave current data with data from old tasks
Replay based methods challenges
- How do we set the balance between old and new tasks
- Memory requirements as the data can grow a lot
Regularisation methods (forgetting)
Start from the idea that forgetting comes from important parameters being changed while learning new tasks
- Estimate which are important for task 1
- Devise a loss function that penalises change to those
- Learn new tasks using such a loss fn to limit interference with old tasks
Regularisation methods challenges
How to estimate the importance of parameters
Most parameter could become important so it’s difficult to learn
Parameter Isolation (forgetting)
A big net might require a small subset of parameters to solve a task.
Find and isolate those for a task.
- Find subset and isolate
- Grow net with additional params for new tasks
- Create masks (light binary networks for each task that select sub regions of a network that are useful)