11 - Validating simulation models Flashcards
Motivation: Why validate simulations?
- when not rigorously calibrated and validated, simulations are neither a reliable research method nor a reliable tool for practical decision support
Definition: Verification, Validation, Calibration
(Steps, that subsume validation)
Verification
Does the code do what the specification asks?
- code could include faults
Definition: Verification, Validation, Calibration
(Steps, that subsume validation)
Structural Validation
Does the model correctly represent the problem space?
- are relevant relationships included in the model etc?
Definition: Verification, Validation, Calibration
(Steps, that subsume validation)
Input/Output validation
Does the simulation correctly represent the problem space?
- some kind of description of the status quo based on data
- parametrise the shape of the distribution
Calibration: adjusting input values to get valid output values
-> danger: overfitting the model when calibrating it
-> agent-based models are subject to over parametrization -> customer learning -> we can make assumptions but we don’t know the parameter for sure (no input parameters)
Simulation Data
Different layers of data
Input data
Process data
Output (result) data
Simulation Data
Different layers of data
Input data
- collected of inferred from empirical status quo
- e.g. customer arrival rate, range of products a company offers
Simulation Data
Different layers of data
Process data
- generated during the simulation process, provides insights into the simulation model that are not empirically available or relevant for the simulation purpose
- process or event logs from the real world can be compared
Simulation Data
Different layers of data
Output (Results) data
- indicators calculated for validation or what-if analysis, matching empirical indicators
- in the status quo these are the indicators we are interested in -> what happens when we change them? Outcome?
Simulation data
Input data
Assumption:
- Structure of relevant simulation components and parameters has been determined
Direct observation:
- for transparent systems
- example: machine run times
- or: what is the actual outline of the shop
Indirect inference:
- based on empirical process and result indicators
- example: customer choice
- > we know what happened but not why, so we apply data analysis to find a model how things work in the real world
Simulation data
Input data: scenarios
Stochastic scenarios
Worst- and best case scenarios
Qualitatively discrete scenarios
Simulation data
Input data: scenarios
Stochastic scenarios
- follow empirical distributions
- stochastic input scenarios lead to stochastic process and result data
- > when we are uncertain about data -> consider stochastic scenarios
Simulation data
Input data: scenarios
Worst- and best-case scenarios
- model extreme cases
- results indicate the modeled systems’ robustness
- > robustness = does it behave the same in each of these cases?
Simulation data
Input data: scenarios
Qualitatively discrete scenarios
- model discrete alternative cases
- test robustness and the necessity of individualized strategies
Simulation data
Example: railway ticket as simulation input
Directly observable:
- supply: products, capacity, price categories, availabilities
- demand: historical sales
Indirect inference (also: estimation):
- customer loyalty (e.g. we have the name on the ticket -> we can analyse how often they bought in the past)
- reference prices
- willingness to pay
- > can use this to calibrate customers in a model
Simulation data
Input data on the future
- Insights:
- analyzing empirical data to parametrize the simulation input data is based on insights, the pre-condition for predictive analytics - Forecast:
- to simulate future scenarios, the future values of input data have to be forecasted