Lecture 1 Flashcards
model
simplified description of a real system, process, phenomenon
based on original system
reflects only the relevant selection of the original system properties
usable in place of the original system for some purposes
we need models for…
simulation, control design and testing, performance / behaviour, evaluation, fundamental insight, online controllers, design, assessment / optimization, safety assessment, certification
properties of a system
- inner structure
- system boundary (interfaces)
- interaction with the environment
subsystems
subsystems consist of elements that are systems themselves
systems engineering
development of systems capable of fulfilling a given task accounting for certain requirements
aircraft systems
- flight control systems
- energy supply and distribution systems
- cabin and payload systems
- flight gear and doors
- propulsion and APU (auxiliary power unit)
- avionics
- safety and emergency systems
types of models
visual, conceptual, functional, systems
mathematical, computational, empirical, statistical, etc.
deterministic model vs probabilistic (stochastic) model
deterministic (ideal):
1. outcome is certain given initial values and boundary conditions
2. given input always gives the same output
3. represents ideal system
probabilistic (real):
1. include a random (probabilistic) component
2. outputs are random variables that can take on values in a particular uncertainty range
3. the same input might produce different outputs
4. more realistic representation of the system
data-driven model vs first-principles model
data-driven model - set of data for input, set of data for output, but we don’t know what is in between
first-principles model - we know exactly what is going on in between input and output (e.g. physics-wise)
statistical model
non-deterministic model
instead or specific values, the variables are random variables with probability distribution
combination of mathematical modelling and statistical assumptions to represent a set of data
based on data from data-generating process
statistical model’s data sets
training set (estimation set):
1. 70% of data
2. used to create the model
3. known input and output
test set (validation set):
1. testing the model’s learnt ability to make correct predictions (from the learning process on training dataset)
accuracy and precision
accuracy: how close are the model’s results to the actual value of the real system
precision: how close are the model’s results to each other
model fidelity levels
the degree of exactness with which something is copied or reproduced
L0: based on empirical knowledge (rough estimation)
L2: simplified physics principles - limited nr of effects considered
L3: physical behaviour represented by more accurate relationships and equations
L4: very high computational costs - only local, distinct effects, high precision, high nr of features captured
big-O notation
measure of computational complexity
model complexity
- underfitting - too simple, too little features
- overfitting - too complex, too many features, model is too fitted to training dataset and the error from test set is big