Medical Image Computing Flashcards

1
Q

What are the main steps in medical image formation?

A

Generate Energy - Energy transfer (focus, collimate,…) - Tissue Interaction - Energy collection - Image formation, reconstruction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Explain the main steps for medical image formation based on a modality of your choice.

A

US: Ultrasonic waves generated, penetrate the body, interact with the tissue: reflection, device collects the thrown back waves, generates US image
X-Ray: Electromagnetic waves, guiding waves with magnetic field and shutters through patient, interaction with tissue: Different tissue types have different absorption coefficients (Beer-Law), collect pass through radiation with detector (e.g. flat panel detection), format image by processing sensor dates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the characteristics of the electromagnetic energy spectrum? Which bands are used for medical imaging?

A

Two perpendicular waves ( oscillating electric field and magnetic field)
Bands: 1. Radion (MRI 50 MHz), 2. Visible (Optical Imaging 600 THz), 3. X-ray (300 PHz), 4. Gamma (30EHz, PET SPECT)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What other energy sources are used for medical imaging? Give an example.

A

Acoustic: aeroacoustics, microphones measure sound pressure at different locations
Ultrasonic: US waves
Electronic: sensor placed behind a camera lens to translate an image into an electronic signal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are usual imaging resolution ranges in medical imaging?

A

Imaging resolution: MRI > CT > US > OCT
Imaging penetration depth: MRI = CT > US > OCT
For lungs of Covid-19 patients: CT, due to depth and resolution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Which modalities are useful for full-body imaging?

A

Radiography/X-Ray
Flat panel detectors are used to convert X-rays to light photons which are then measured by CCD/CMOS sensors (indirect measurement)
+ RT: Relatively low cost, Deep penetration with good resolution, Wide availability and standards
- CT: higher radiation exposure than X-Ray – from different angles: carcinogenic
- MR: expensive, takes very long + noisy -> discomfort
- US: cumbersome, since body contact would be necessary everywhere, time-consuming

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is functional imaging? Give an example.

A

= Imaging method to visualize function of organs/tissue
* Information on changes e.g. Metabolism, blood flow (e.g. detection of cancer)
* Opposed to structural imaging, specific areas within a certain tissue or organ imaged
* Often use tracers or probes to reflect spatial distribution within the body based on bounding
Examples: Functional MRI, PET/SPECT, EEG, Optical imaging, Angiography

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

. Which imaging modality would you use for imaging the lungs, e.g. of Covid-19 patients? Give a rationale.

A

CT
+ Good penetration depth to resolution ratio
+ Series of X-ray images from different angles
+ Imaging time in CT is less than 5 minutes
+ CT to visualize moving objects, e.g. heart, lung
+ Increase in detector elements, results in higher imaging speeds / higher resolution
+ In modern CT scanners images consist of 512 × 512 pixels representing the CT number in Hounsfield units (HU): lung -600…-900 HU: here differenciation air or water in the lungs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is multi-modal imaging and how is it used for nuclear imaging?

A

= Combination of different modalities = multi-modal
* Each modality is fitted for certain tasks / body parts
* Get complimentary information
Example ▪ CT/PET ▪ MR/PET ▪ OCT/PAM
Detailed CT image of area and in addition PET image (tracer to bind to tumor cells for representation) to be lied above

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the main characteristics of optical imaging methods? When would you use them?

A

Light and special properties of photons used to probe tissue, similar to ultrasound but using electromagnetic waves
Measuring of reflection, using interferometry
+ Non-ionizing radiation (visible, ultraviolet, and infrared light) => no exposure to harmful radiation => used for repeated procedures to monitor progression of disease
+ Structural and functional imaging early (detect metabolism changes => markers of abnormal functioning)
+ Usually, non-invasive diagnostic tool
+ Combination with other imaging techniques
Examples: OCT (Optical Coherence Tomography), Endoscopy, Photoacoustic imaging…

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the main properties of a digitized image or volume?

A
  1. Image elements: pixel (2D), voxel (3D)
  2. Image depth = range of values for image function: 8, 12 or 16 bit (2^8 …)
  3. Dimension
    Representation as: Surface plot, numerical array, visual intensity array
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What makes medical image formats special compared to conventional consumer image formats?

A

Compressed formats (jpg) not very popular as possible loss of diagnostic information
* Big images (>8bit): need to be handled by special image processing programs
* 3D imaging format with details: Slice thickness, Distance between slices
* Information on relevant patient data/image aquisition is not provided in conventional image header formats
* Structured link from patient to image
Example: .tif

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Explain the DICOM format: Which information can be stored? How is the information organized? What are application entities and how do they communicate?

A

“Digital Imaging and Communications in Medicine” = container for medical imaging data (image exchange, visualization and acquisition)
* 3D images of digital radiography, CT, MRI, US, …
* Link image data to medical procedure, image-acquisition parameters (Patient 3D position, sizes, orientations, slice thickness, radiation doses and exposures, image processing filters)
* Encoding of medical data (e.g. patient name, current patient diagnosis)
* Real-world data are viewed by DICOM as objects with attributes
▪ Image height – „Rows“ attribute
▪ Image width – „Columns“ attribute
▪ Image pixel data – „Pixel data“ attribute
▪ Unique identifier
Communication: Captured data can be transmitted and processed between DICOM devices and software – Application Entities (AE)
▪ AEs provide services to each other
▪ Entity: Clearly identifiable object through which information is stored or processed
▪ Relationship between AEs: Service Class Provider (SCP)  → Service Class User (SCU)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Which information systems are usually involved in a hospital radiological workflow?

A
  • PACS: Pictures archiving and communication system: consists of medical image and data acquisition, storage, and display subsystems
  • HIS: Hospital information system
  • RIS: Radiology information system
  • MDR: Medical data record software
    Tasks: Patient scheduling, billing, financials, reports, worklists, labs, …
  • HL7 standard: Messaging and document implementation in XML
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How can telemedicine be used in the area of medical imaging? Give a scenario.

A

People living on rural areas need CT scan of lung
Establishment available but no expert in Image evaluation/Diagnosis/Therapy
* Tele-diagnosis: „Point-of-care“ diagnosis and treatment: Ability to test and treat patients rapidly at sites close to where they live
* Tele-consultations: Physician in rural area sends medical information such as CT, MRI scans to specialist in a distant location and receives advice
* Tele-management: Immediate management care to the patient: receive information, monitoring of advised therapy (over two instances) …

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What types of digital images are usually used in medical image computing? Give an example for each type.

A
  • Not in medical imaging: Binary image (1 bit)
  • Greyscale image (N bit): usually 8 bits: Breast X-ray
  • Colour image (n x N bit): types of digital images are usually used in medical image: 24 bits
    (8 bits for each colour): PET scan brain
  • Multichannel image (n x N bit): information outside normal human perceptron range (infrared, ultraviolet, x-ray) => Processed, used in medical research: Divide intensity range of brain CT images
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What are point and neighbourhood image transformations and what are they used for?

A

In order to transform an image for extension of intensity levels or compression of intensity levels
Point image transformations: only intensity and location of single pixel is processed, surrounding pixels are irrelevant
* i.e. histogram, thresholding, grayscale transformation, contrast stretching
Neighbourhood image transformations: area is processed, usually centre pixel and surrounding pixels within specific distance
* Correlation, convolution, morphological operators, spartial filtering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What are image histograms?

A

Provide frequency of grey values
Horizontal axis → grey values
Vertical axis → number of pixels with grey value
Give information about contrast and colour (dark or bright): Aim: Create an image with equally distributed brightness levels over the whole brightness scale. ➔ Enhance the contrast!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is spatial (räuml.) filtering and how is a spatial filter defined? Given an example of a spatial filter and what is it used for?

A

= Neighbourhood image transformation
* Compute a function of the local neighbourhood of each pixel in the image
* Function = filter kernel (called mask, template, window) saying how to combine values from neighbours
Used for:
▪ Image enhancement (denoise, resize, etc.)
▪ Extract information (texture, edges, etc.)
▪ Pattern recognition (template matching)
Example: Moving average 2D = smoothing (compressing data)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

How can the Fourier transform be applied to medical images and what is used for?

A

Fourier transform = weighting function
To be multiplied with integral of sines/cosines of function is not periodic
Convolution and correlation can be done efficiently in the Fourier domain
* process takes a complex-valued function (images are complex-valued functions with zero imaginary component) and returns a complex-valued function
* representing the image as a sum of sinusoids
* Fourier transform considering the phase and magnitude of the complex function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What are image artifacts?

A

= false images, or parts of images, that do not represent true anatomic structures
* Noise
* Intensity inhomogeneity
* Motion
But: “Higher noise, higher contrast – lower noise, lower contrast”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

How can you find edges in an image?

A

Edges = change in pixel intensity
* Rate of change found by derivative (Ableitung)
* Usually: neighbourhood processing
* Example algorithms: Sobel, canny edge

Find edges in 2D images:
* Smooth first to remove noise
* Calculate image gradient
* Gradient points in direction of most rapid change in intensity
* First derivatives show where the edges are, Zero crossings of second derivatives can be used to extract them

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Explain the main steps in the Canny Edge detection algorithm.

A

(1) Smooth image with Gaussian filter to remove noise
(2) Compute gradient magnitude and angle images
(3) Apply non-maxima suppression to the gradient images
(4) Use double thresholding (hysteresis) and connectivity analysis to detect and link edges

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What is the Hough transform and what is used for?

A

= feature extraction technique (edge detection)
* Locate shapes by linking edge segments in images
* Extract lines, circles and ellipses (or conic sections)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
You have to binarize an image based on intensity only. How do you proceed to perform the task by setting a manual threshold? Which approach would you choose for an automated segmentation for this problem?
Manual approach for threshold: Histogram -> Minima = Threshold Eventually smoothing first to remove noise Automated approach: Otsu method: Uses grey-value histogram of the given image as input and aims at providing the best threshold in a sense that the overlap between two classes (set of object and background pixels) is minimized
26
How do active contours work? What are they used for?
= Allow model-based segmentation based on regional parameters To segment an image where the data are distorted by noise or artefacts * Represent boundary as a parametric curve * The curve is associated to an energy function E * Snake smoothly follows high intensity gradients if they reliably reflect the object boundary * Internal forces: Designed to keep the model smooth during deformation * External forces: Move the model towards an object boundary or other desired feature within an image * Energy minimization based on internal (contour properties) and external energy (image)
27
What is k-means clustering?
= Method for seeking clusters in data based on observations, e.g. intensity
28
How can images be represented as graphs? Which graph-based segmentation methods do you know?
Representation can help to find features (i.e. edges) Image graph: * Segmentation of the image as weighted, (un)directed graph * Pixels are nodes * Edge is connection between pairs of nearby pixels * Every node is connected to its 4 (or 8) x-y neighbours Goal of segmentation using graphs: Order the nodes in diverse subsets, where similarity within subset is big and out of subset is low Examples: Graph cut image segmentation, Watershed segmentation, Cell segmentation, Dijkstra‘s shortest path algorithm (Compute minimum cost path from one seed to all other pixels)
29
How are segmentation algorithms evaluated? What are useful metrics?
Based on execution time and correctness of the result: * Accuracy: ability of the method to mirror standard of reference measurements or diagnosis. * Precision: ability of the method to provide repeatable measurements i.e. low variability due to noise, different acquisition procedure * Specificity: Ratio of negative cases correctly classified as negative * Sensitivity: Ratio of true cases correctly classified as true
30
What are typical applications for image registration?
Stitching: MR/MR, ... of separate times, Photo stitching Multi-modal imaging: PET/CT, Augmented reality / image guided surgery
31
What are geometric transformations and what kinds of transformations are usually defined for image registration?
= Modify the spatial arrangement of pixels in an image (Ändern räuml. Anordnung), e,g. map features in one image to another Two basic operations: ▪ Spatial transformation of coordinates ▪ Intensity interpolation Kinds of transformations: ▪ Rigid: Translation, rotation ▪ Affine (combination in single matrix): translation (along transformation vector), rotation (around origin), scaling (isotropic or anisotropic: diff. for x & y), shear (= Kippen) ▪ Deformable/non-rigid: all transformation possible (deformation)
32
. Which two general approaches do exist for finding a geometric transformation? What are their pros and cons?
Feature-based registration: * Find correspondence between image features: points, lines, contours * Algorithms based on: ▪ Distance between corresponding points ▪ Similarity metric between feature values, e.g. curvature-based registration Intensity-based registration: * Align the entire images to match up the colours/grey value of as many pixels as possible * No landmark or feature selection is necessary * Algorithms: ▪ Registration by minimizing ▪ Intensity difference ▪ Correlation techniques ▪ Ratio image uniformity ▪ Partitioned Intensity ▪ Uniformity * Needed: Pixel-by-pixel error measure, Mapping technique (transformation), Minimization technique
33
What are 2D (joint) histograms?
Images can be viewed as probability distributions; calculated via joint histograms Frequency of corresponding intensity pairs can be interpreted in terms of probabilities
34
Which similarity measures do you know for image registration?
Similarity measure = quantifies degree of similarity between intensity patterns in two images Multimodal * Mutual information * Normalized mutual information Same modality * Joint entropy * Cross-correlation * Sum of squared intensity differences * Ratio image uniformity
35
Given a defined similarity measure, how can the optimal parameters be found to minimize the criterion?
Using the Gradient descent Repeat: ▪ Compute gradient ▪ Make step in gradient direction (not too small, not to big) ▪ Update mapping equation ▪ Remap image Until convergence ▪ Ideally when gradient = 0
36
Give an application for when to use free-form deformation image registration.
= Transformation of moving (non-rigid) image based on deformation Contrast-enhanced breast MR Transformation in 3D: * Motion of breast is non-rigid (moving) * Approach: Develop a combined transformation which consists of global and local transformation * Overall motion of the breast. Affine transformation in 3D: Rigid (6 degrees of freedom) + Scaling and shearing (6 degrees of freedom)
37
What is feature extraction and which steps are involved? Give an example.
= process of defining a set of features, or image characteristics, which will most efficiently or meaningfully represent the information that is important for analysis Consists of ▪ Feature detection, i.e. finding features such as corners ▪ Feature description, i.e. quantification of attributes such as corner orientation Example: Extract malignant lesion of CT scan lung
38
What is a feature vector and a feature space? What are invariance and co-variance in the context of features?
* Feature vectors ▪ Package/container for numerical descriptors ▪ Column vector (d x 1) or row vector (1 x d) E.g. each pixel is a 3D (RGB) colour feature vector (R value, G value, B value) Or (length; mean; variance; intensity) = 4x1 = f_B1 * Feature space ▪ Contains a point cloud of feature vectors ▪ d-dimensional Example: (f_B1, f_B2, f_B3, f_B4, f_B5) = 4x5 Matrix Invariance: = Features should be invariant (insensitive) to variations (“Invariance to intensity transformations”) ▪ in: Scale, translation, rotation, illumination and viewpoint Value of the feature does not change after the application of the given transformation family => features(transform(image)) = features(image) Co-variance: = Feature changes by the same amount (“Covariance with geometric transformations”) ▪ E.g. Feature Area is covariant with scaling ▪ E.g. Feature Direction is covariant with rotation Covariant detection: If we have two transformed versions of the same image, features should be detected in corresponding locations => features(transform(image)) = transform(features(image))
39
What are chain codes and how can they be extracted?
= Representation of boundary by a set of connected straight lines with defined direction and length * Offers a unified way to analyse the shape of a boundary * Separately encodes each connected component * Application: compression (lossless), recognition
40
. Your task is to analyse a network of blood vessels. How would you extract relevant information?
Method: Medial representations/Skeletonization: Thinning (Medial axis transform): * Original shape can be fully reconstructed * Set of points that are equidistant from the boundary * Like boundaries, skeletons/blood vessels are related to the shape of a region * Reduce a region to a tree or a graph
41
What are geometrical features and how are they extracted? Give an example.
* Length: number of pixels along the boundary * Diameter = max distance where p1 and p2 are points on the boundary * Major axis = line segment of length equal to the diameter and connecting two points on the boundary * Minor axis = line perpendicular to the major axis and of such length until intersection of the boundary * Basic rectangle (bounding box) by minor and major axis intersections with circle * Eccentricity is the ratio of major to minor axis
42
Which basic regional descriptors do you know?
Basic descriptors: area, perimeter, compactness, mean value, circularity, effective diameter, eccentricity, normalized area
43
Which approaches do you know to describe image/region texture?
Statistical approaches ▪ Smooth, coarse, grainy, regularity … Spectral approaches ▪ Global periodicity ▪ Fourier spectrum: High energy, narrow peaks, …
44
What is the biggest limitation of textural features based on histograms only? How can this limitation be overcome?
The histogram does not carry information about the spatial relationship of pixels (räuml. Beziehung) Need to take into account relative positions of pixels => Haralick features  Grey level co-occurrence matrix (GLCM) * Matrix G were measured the number of times intensities occur in a specified position * Position specified with a position operator Q ▪ E.g. Q = “one pixel immediately to the right” ▪ E.g. Q = "one pixel to the right and one pixel above”
45
What are Haralick features?
= Set of 14 textural features which can be extracted from the co-occurrence matrix * Contain information about image texture characteristics such as homogeneity, linearity, contrast * Four directions of adjacency, calculate GLCM for each direction
46
How do local binary patterns work?
* Operator returns a discrete value at each pixel that characterizes the local texture partially invariant to luminance changes * Compares eight neighbourhood pixel intensities to the centre pixel intensity: ▪ 0 if intensity is less then centre pixel ▪ 1 if intensity is greater or equal than centre pixel * Orientation invariant ▪ Shift to minimum * Reduce classes by aggregation ▪ E.g. uniform vs. Non-uniform: 00000000 vs 01010101
47
What is a design matrix?
A design matrix is a matrix containing data about multiple characteristics of several individuals or objects. Each row corresponds to an individual and each column to a characteristic. Gather the feature descriptors for a single image which then can be used for further processing (e.g. kmeans) => here just one feature vector for each object (4x5x1) Features on one axis, classes/samples on other axis
48
What is the principal component analysis? What is it used for?
Given example vectors (design matrix) 2. Calculate covariance matrix, and its eigenvector decomposition 3. Represent data in the new space with reduced dimensionality (also Autoencoder)  Reduce the dimensions of data in space => large feature vectors (Get rid of useless/redundant information) ▪ Feature selection (Visualization) ▪ Shape normalization ▪ Model shape ▪ Classification * To visualize high dimensional datasets * Each example is represented by a vector containing the values for all features (E.g. 3 for colour vectors)
49
What are whole-image features? What are their typical applications?
Features to be extracted from whole images * Corner detection Corners = Junctions of contours / Large local gradient in multiple directions Change of intensity for the displacement of a patch (Intensitätsänderung für die Verschiebung) Harris corner detection: ▪ Look for local patch that produces noticeable difference when moved around ▪ Change of intensity for the displacement [u,v] * SIFT features Applications ▪ Image matching purposes (registration, mosaicking): patches containing a corner has distinct local features ▪ Seed point selection
50
What are SIFT features? What are their properties? (Funktionen + Eigenschaften)
= Scale-Invariant Feature Transform = Local histogram-based descriptor Outline: Input = Image, Output = feature vector * Detection of keypoints (= SIFT features) * Refine selection of keypoints * Describe each keypoint using a feature vector Properties of SIFT feature descriptor: * Invariant to uniform scaling, orientation, illumination changes, and partially invariant to affine distortion * Various feature often combined * Useful in the application of image registration/matching, mosaicking Preferred method for: * Scale changes, rotations, changes in illumination and/or viewpoint * Applications: object recognition, image stitching, 3D modeling, gesture recognition, video tracking, individual identification, …  
51
What are the limitations of hand-designed programs? How can machine learning overcome these limitations?
Limitation: hard to maintain for complex problems Machine learning:  Allow to learn from data and improve the program in fluctuating environments  Can help humans learn, by inspecting to see what they have learned (although this can be tricky for some algorithms)  Getting insights about complex problems and large amounts of data
52
What is the difference between supervised and unsupervised learning? Give a machine learning algorithm for each of the types.
Unsupervised learning: No need for human supervision, Training data is unlabelled, model identifies structures like clusters E.g. k-Means, principal component analysis PCA, association rule learning, GAN Supervised learning: Human supervision, Training data fed to algorithm includes known labels, model learns decision boundarys and replicates labelling E.g. k-Nearest neighbours, linear and logistic regression, support vector machines (SVM), decision trees and random forests, neural networks
53
What is the definition of learning algorithms? Give an example for experience, task and performance.
… are able to learn from data. „A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P improves with experience E.“ Task: how the system should process an example E.g. Classification, Regression, Denoising, Anomaly detection Performance: evaluate the abilities of the ML algorithm E.g. ROC curve, Confusion matrix (with FP, FN, TP, TN, accuracy, specificity, sensitivity) Experience: kind of experience during learning E.g. supervised, unsupervised, design matrix, …
54
Which datasets are usually involved to find the parameters of a machine learning model? What is a useful split for these datasets?
Distinct datasets: * Training set: Data used to train the model * Tuning set: Used to evaluate the performance of different models and used to find the hyperparameter values * (Validation) test set: Used to evaluate the performance with unseen data * (External validation test set)
55
Assuming only a rather small dataset is available for your classification task. How can you evaluate the model throughly? How can you increase the dataset?
Cross-validation: If small dataset => splitting the dataset into training and test set randomly multiple times Most common: k-fold cross validation * Dataset is split into k non-overlapping subsets * k trials are performed * Average test error of k trials is taken
56
What is the curse of dimensionality?
= Performance can be reduced when the amount of features is increased (small number of examples)  Rule of thumb: number of examples/number of features > 10
57
What is a classifier? What is over- and under fitting?
A classifier in machine learning is an algorithm that automatically orders or categorizes data into one or more of a set of “classes.”[Link] Different classifiers with different approaches to get boundary from training data * Capture distribution of training data faithfully * Be fast during training * Be fast during classification of new data * Deal with noise in the training data * Deal with imperfect feature extractors Over-fitting: the classifier is too faithful to the training data, but doesn’t reflect the underlying distribution anymore (Low bias - high variance)  Regularization Under-fitting: the classifier doesn’t reflect the training data or the underlying distribution (High bias - low variance)  Possibly we need a classifier with more parameters
58
How do nearest neighbour classifiers work?
Assign class label based on training example with lowest distance: NN classifier (also k-nearest neighbour, minimum distance classifiers) Principle: * Training set of labelled prototypes and classes * An example to be classified * Find prototype which is closest to this * Example gets label of prototype
59
. What are the advantages of decision trees? What are the design considerations when constructing decision trees?
Feature vectors mainly numerical vectors (E.g. length, width, intensity) Advantages: * Easy to understand and interpret * Fast * Versatile and powerful: Minor changes in training set can cause major changes in the classification Design considerations CART: 1. Branching factor: How many splits should be attached to a node, two (binary) or more? 2. Query selection: Which feature should be tested at a node? 3. Stopping criteria: When does a node should be transferred into a leave? 4. Pruning: If tree gets too big, how can it be made smaller (pruning)? 5. Classification: If a leaf is not pure, how is a decision made?
60
What is ensemble learning? Give an example of a machine learning algorithm based on ensemble learning.
* Aggregate predictions of group of predictors (regressors, classifiers) are often superior than best individual predictor * Group of predictors is called ensemble → ensemble learning Example: Group of decision trees – random forest
61
Which dimensionality reduction methods do you know?
Autoencoders, PCA Principal component analysis
62
How is Bayes' rule applied in pattern recognition?
X is a feature, Y is a class membership * We want to know P(X|Y) => Probability: feature X belongs to class Y * Conditional probability: If we know P(X) we can compute P(X|Y) using Bayes’ rule
63
What is a perceptron and how does it work? What are the limitations?
= Special case of a linear, binary classifier * Assuming a n-dimensional feature vector and two classes * Find a discriminant function which codes the class labels * Goal: Find weights and bias * Is guaranteed to converge if data are linearly separable Limitations: Good for linearly separable classes  Fails for typical XOR problem (not linear separable), Solution – multiple perceptrons
64
What is the definition of a neural network? Explain the basic architecture of feedforward neural networks.
= Interconnected perceptron-like computing elements called artificial neurons * Interconnected neurons are organized in layers where the output of one layer provides the input of the following layer (Hence, a sensitive activation function would affect all subsequent layers) * Activation function make a neuron to “fire” and must be differentiable “Feedforward to next layer”: neuron in hidden layer l goes to all neurons in layer l+1
65
Which parameters define a neural network? How are neural networks trained?
* Weights W * Biases B * Activation functions h Training: use sets of training patterns to estimate these parameters Tool “backpropagation”: 1. Inputting the pattern vectors 2. Forward pass through the network and determine the classification error 3. A backward (backpropagation) pass that feeds output error back to estimate required changes 4. Updating the weights and biases in the network
66
What are convolutional neural networks and what are the differences to feedforward neural networks?
Convolutional neural networks: * Have additional convolutional layers and pooling layers before the NN => Purpose: Automatically detecting features * Accept (training) images directly as inputs * No pre-defined features by human experts needed  CNNs learn directly from raw image data FFN Networks: Convert images to vectors = patterns organized in feature vectors, loss of spatial relationship
67
What is a convolutional layer? What is pooling and what are common pooling methods?
CNN layer: Collection of feature maps (One feature map: Out of receptive field = Neighbourhoods of image; with same weights and single bias) Pooling = Downsampling to reduce spatial resolution (for memory + comp. power) * Typically 2x2 (no overlap) * Pooling methods: Average, Max-pooling, L2 pooling
68
What are the main differences between different deep CNN networks? Name an architecture mentioned in the lecture which you would use for the task of biomedical image segmentation?
Deep CNN networks: Use of pre-trained models U-Net: Convolutional network for biomedical image segmentation  Works with only a few annotated images
69
What is generative modelling?
Goal: Given training data set, generate new samples from same distribution * We train a model from samples drawn from a distribution * It learns an estimate of this distribution
70
What is the goal of unsupervised learning? What are the applications of unsupervised methods?
Learn some underlying structure of data Examples: Clustering dimensionality reduction, density estimation,...
71
Compare autoencoders against general adversarial networks. What are the main differences?
Autoencoders = Neural networks capable of learning efficient representations of the input data * Encoder (recognition network) > Decoder (generative network) * Used for: Dimensionality reduction, Feature, detection, generative modelling GAN = Make neural networks to compete each other and pushing them to excel (=übertreffen) * Generator and Discriminator * Used for generative modelling: Anomaly detection, data augmentation, super-resolution
72
What are the principal steps in image-guided therapy? Which imaging modalities and which image processing steps are mainly involved?
= Aims to use imaging to improve the localization and targeting of diseased tissue, Monitor and control treatments In image-guided surgery and interventional radiology Steps: * Prior to procedure diagnostic imaging is performed * Images are converted and modelled, e.g. in 3D, to represent the patient’s anatomy * Information is used for Pre-operative surgical planning Intraoperative surgical decision making
73
How can medical imaging improve personalized medicine?
Individual treatment check and endpoints – image guided therapy (E.g. Macular Edema Recurrence) * Dose optimization, optimal efficacy * Timely treatment * Prediction of treatment response Main challenge: find predictive markers for future disease progression and treatment response
74
What is radiomics? What are the main steps involved? What is currently the bottleneck in radiomics?
= High-throughput extraction and analysis (mining) of radiological image data * Extraction of quantitative features from imaging data such as MR or CT * Bridge between medical imaging and personalized medicine * Associate features with predictive goals: Diagnosis, Prognosis Motivation: Image guidance to support clinical decisions (radiations, surgery, treatment)