Medical Image Computing Flashcards

You may prefer our related Brainscape-certified flashcards:
1
Q

What are the main steps in medical image formation?

A

Generate Energy - Energy transfer (focus, collimate,…) - Tissue Interaction - Energy collection - Image formation, reconstruction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Explain the main steps for medical image formation based on a modality of your choice.

A

US: Ultrasonic waves generated, penetrate the body, interact with the tissue: reflection, device collects the thrown back waves, generates US image
X-Ray: Electromagnetic waves, guiding waves with magnetic field and shutters through patient, interaction with tissue: Different tissue types have different absorption coefficients (Beer-Law), collect pass through radiation with detector (e.g. flat panel detection), format image by processing sensor dates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the characteristics of the electromagnetic energy spectrum? Which bands are used for medical imaging?

A

Two perpendicular waves ( oscillating electric field and magnetic field)
Bands: 1. Radion (MRI 50 MHz), 2. Visible (Optical Imaging 600 THz), 3. X-ray (300 PHz), 4. Gamma (30EHz, PET SPECT)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What other energy sources are used for medical imaging? Give an example.

A

Acoustic: aeroacoustics, microphones measure sound pressure at different locations
Ultrasonic: US waves
Electronic: sensor placed behind a camera lens to translate an image into an electronic signal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are usual imaging resolution ranges in medical imaging?

A

Imaging resolution: MRI > CT > US > OCT
Imaging penetration depth: MRI = CT > US > OCT
For lungs of Covid-19 patients: CT, due to depth and resolution

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Which modalities are useful for full-body imaging?

A

Radiography/X-Ray
Flat panel detectors are used to convert X-rays to light photons which are then measured by CCD/CMOS sensors (indirect measurement)
+ RT: Relatively low cost, Deep penetration with good resolution, Wide availability and standards
- CT: higher radiation exposure than X-Ray – from different angles: carcinogenic
- MR: expensive, takes very long + noisy -> discomfort
- US: cumbersome, since body contact would be necessary everywhere, time-consuming

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is functional imaging? Give an example.

A

= Imaging method to visualize function of organs/tissue
* Information on changes e.g. Metabolism, blood flow (e.g. detection of cancer)
* Opposed to structural imaging, specific areas within a certain tissue or organ imaged
* Often use tracers or probes to reflect spatial distribution within the body based on bounding
Examples: Functional MRI, PET/SPECT, EEG, Optical imaging, Angiography

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

. Which imaging modality would you use for imaging the lungs, e.g. of Covid-19 patients? Give a rationale.

A

CT
+ Good penetration depth to resolution ratio
+ Series of X-ray images from different angles
+ Imaging time in CT is less than 5 minutes
+ CT to visualize moving objects, e.g. heart, lung
+ Increase in detector elements, results in higher imaging speeds / higher resolution
+ In modern CT scanners images consist of 512 × 512 pixels representing the CT number in Hounsfield units (HU): lung -600…-900 HU: here differenciation air or water in the lungs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is multi-modal imaging and how is it used for nuclear imaging?

A

= Combination of different modalities = multi-modal
* Each modality is fitted for certain tasks / body parts
* Get complimentary information
Example ▪ CT/PET ▪ MR/PET ▪ OCT/PAM
Detailed CT image of area and in addition PET image (tracer to bind to tumor cells for representation) to be lied above

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the main characteristics of optical imaging methods? When would you use them?

A

Light and special properties of photons used to probe tissue, similar to ultrasound but using electromagnetic waves
Measuring of reflection, using interferometry
+ Non-ionizing radiation (visible, ultraviolet, and infrared light) => no exposure to harmful radiation => used for repeated procedures to monitor progression of disease
+ Structural and functional imaging early (detect metabolism changes => markers of abnormal functioning)
+ Usually, non-invasive diagnostic tool
+ Combination with other imaging techniques
Examples: OCT (Optical Coherence Tomography), Endoscopy, Photoacoustic imaging…

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the main properties of a digitized image or volume?

A
  1. Image elements: pixel (2D), voxel (3D)
  2. Image depth = range of values for image function: 8, 12 or 16 bit (2^8 …)
  3. Dimension
    Representation as: Surface plot, numerical array, visual intensity array
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What makes medical image formats special compared to conventional consumer image formats?

A

Compressed formats (jpg) not very popular as possible loss of diagnostic information
* Big images (>8bit): need to be handled by special image processing programs
* 3D imaging format with details: Slice thickness, Distance between slices
* Information on relevant patient data/image aquisition is not provided in conventional image header formats
* Structured link from patient to image
Example: .tif

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Explain the DICOM format: Which information can be stored? How is the information organized? What are application entities and how do they communicate?

A

“Digital Imaging and Communications in Medicine” = container for medical imaging data (image exchange, visualization and acquisition)
* 3D images of digital radiography, CT, MRI, US, …
* Link image data to medical procedure, image-acquisition parameters (Patient 3D position, sizes, orientations, slice thickness, radiation doses and exposures, image processing filters)
* Encoding of medical data (e.g. patient name, current patient diagnosis)
* Real-world data are viewed by DICOM as objects with attributes
▪ Image height – „Rows“ attribute
▪ Image width – „Columns“ attribute
▪ Image pixel data – „Pixel data“ attribute
▪ Unique identifier
Communication: Captured data can be transmitted and processed between DICOM devices and software – Application Entities (AE)
▪ AEs provide services to each other
▪ Entity: Clearly identifiable object through which information is stored or processed
▪ Relationship between AEs: Service Class Provider (SCP)  → Service Class User (SCU)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Which information systems are usually involved in a hospital radiological workflow?

A
  • PACS: Pictures archiving and communication system: consists of medical image and data acquisition, storage, and display subsystems
  • HIS: Hospital information system
  • RIS: Radiology information system
  • MDR: Medical data record software
    Tasks: Patient scheduling, billing, financials, reports, worklists, labs, …
  • HL7 standard: Messaging and document implementation in XML
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How can telemedicine be used in the area of medical imaging? Give a scenario.

A

People living on rural areas need CT scan of lung
Establishment available but no expert in Image evaluation/Diagnosis/Therapy
* Tele-diagnosis: „Point-of-care“ diagnosis and treatment: Ability to test and treat patients rapidly at sites close to where they live
* Tele-consultations: Physician in rural area sends medical information such as CT, MRI scans to specialist in a distant location and receives advice
* Tele-management: Immediate management care to the patient: receive information, monitoring of advised therapy (over two instances) …

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What types of digital images are usually used in medical image computing? Give an example for each type.

A
  • Not in medical imaging: Binary image (1 bit)
  • Greyscale image (N bit): usually 8 bits: Breast X-ray
  • Colour image (n x N bit): types of digital images are usually used in medical image: 24 bits
    (8 bits for each colour): PET scan brain
  • Multichannel image (n x N bit): information outside normal human perceptron range (infrared, ultraviolet, x-ray) => Processed, used in medical research: Divide intensity range of brain CT images
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What are point and neighbourhood image transformations and what are they used for?

A

In order to transform an image for extension of intensity levels or compression of intensity levels
Point image transformations: only intensity and location of single pixel is processed, surrounding pixels are irrelevant
* i.e. histogram, thresholding, grayscale transformation, contrast stretching
Neighbourhood image transformations: area is processed, usually centre pixel and surrounding pixels within specific distance
* Correlation, convolution, morphological operators, spartial filtering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What are image histograms?

A

Provide frequency of grey values
Horizontal axis → grey values
Vertical axis → number of pixels with grey value
Give information about contrast and colour (dark or bright): Aim: Create an image with equally distributed brightness levels over the whole brightness scale. ➔ Enhance the contrast!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is spatial (räuml.) filtering and how is a spatial filter defined? Given an example of a spatial filter and what is it used for?

A

= Neighbourhood image transformation
* Compute a function of the local neighbourhood of each pixel in the image
* Function = filter kernel (called mask, template, window) saying how to combine values from neighbours
Used for:
▪ Image enhancement (denoise, resize, etc.)
▪ Extract information (texture, edges, etc.)
▪ Pattern recognition (template matching)
Example: Moving average 2D = smoothing (compressing data)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

How can the Fourier transform be applied to medical images and what is used for?

A

Fourier transform = weighting function
To be multiplied with integral of sines/cosines of function is not periodic
Convolution and correlation can be done efficiently in the Fourier domain
* process takes a complex-valued function (images are complex-valued functions with zero imaginary component) and returns a complex-valued function
* representing the image as a sum of sinusoids
* Fourier transform considering the phase and magnitude of the complex function

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

What are image artifacts?

A

= false images, or parts of images, that do not represent true anatomic structures
* Noise
* Intensity inhomogeneity
* Motion
But: “Higher noise, higher contrast – lower noise, lower contrast”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

How can you find edges in an image?

A

Edges = change in pixel intensity
* Rate of change found by derivative (Ableitung)
* Usually: neighbourhood processing
* Example algorithms: Sobel, canny edge

Find edges in 2D images:
* Smooth first to remove noise
* Calculate image gradient
* Gradient points in direction of most rapid change in intensity
* First derivatives show where the edges are, Zero crossings of second derivatives can be used to extract them

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Explain the main steps in the Canny Edge detection algorithm.

A

(1) Smooth image with Gaussian filter to remove noise
(2) Compute gradient magnitude and angle images
(3) Apply non-maxima suppression to the gradient images
(4) Use double thresholding (hysteresis) and connectivity analysis to detect and link edges

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What is the Hough transform and what is used for?

A

= feature extraction technique (edge detection)
* Locate shapes by linking edge segments in images
* Extract lines, circles and ellipses (or conic sections)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

You have to binarize an image based on intensity only. How do you proceed to perform the task by setting a manual threshold? Which approach would you choose for an automated segmentation for this problem?

A

Manual approach for threshold: Histogram -> Minima = Threshold
Eventually smoothing first to remove noise

Automated approach: Otsu method: Uses grey-value histogram of the given image as input and aims at providing the best threshold in a sense that the overlap between two classes (set of object and background pixels) is minimized

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

How do active contours work? What are they used for?

A

= Allow model-based segmentation based on regional parameters
To segment an image where the data are distorted by noise or artefacts

  • Represent boundary as a parametric curve
  • The curve is associated to an energy function E
  • Snake smoothly follows high intensity gradients if they reliably reflect the object boundary
  • Internal forces: Designed to keep the model smooth during deformation
  • External forces: Move the model towards an object boundary or other desired feature within an image
  • Energy minimization based on internal (contour properties) and external energy (image)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

What is k-means clustering?

A

= Method for seeking clusters in data based on observations, e.g. intensity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

How can images be represented as graphs? Which graph-based segmentation methods do you know?

A

Representation can help to find features (i.e. edges)
Image graph:
* Segmentation of the image as weighted, (un)directed graph
* Pixels are nodes
* Edge is connection between pairs of nearby pixels
* Every node is connected to its 4 (or 8) x-y neighbours
Goal of segmentation using graphs: Order the nodes in diverse subsets, where similarity within subset is big and out of subset is low

Examples: Graph cut image segmentation, Watershed segmentation, Cell segmentation, Dijkstra‘s shortest path algorithm (Compute minimum cost path from one seed to all other pixels)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

How are segmentation algorithms evaluated? What are useful metrics?

A

Based on execution time and correctness of the result:
* Accuracy: ability of the method to mirror standard of reference measurements or diagnosis.
* Precision: ability of the method to provide repeatable measurements i.e. low variability due to noise, different acquisition procedure
* Specificity: Ratio of negative cases correctly classified as negative
* Sensitivity: Ratio of true cases correctly classified as true

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

What are typical applications for image registration?

A

Stitching: MR/MR, … of separate times, Photo stitching
Multi-modal imaging: PET/CT, Augmented reality / image guided surgery

31
Q

What are geometric transformations and what kinds of transformations are usually defined for image registration?

A

= Modify the spatial arrangement of pixels in an image (Ändern räuml. Anordnung), e,g. map features in one image to another
Two basic operations:
▪ Spatial transformation of coordinates
▪ Intensity interpolation
Kinds of transformations:
▪ Rigid: Translation, rotation
▪ Affine (combination in single matrix): translation (along transformation vector), rotation (around origin), scaling (isotropic or anisotropic: diff. for x & y), shear (= Kippen)
▪ Deformable/non-rigid: all transformation possible (deformation)

32
Q

. Which two general approaches do exist for finding a geometric transformation? What are their pros and cons?

A

Feature-based registration:
* Find correspondence between image features: points, lines, contours
* Algorithms based on:
▪ Distance between corresponding points
▪ Similarity metric between feature values, e.g. curvature-based registration

Intensity-based registration:
* Align the entire images to match up the colours/grey value of as many pixels as possible
* No landmark or feature selection is necessary
* Algorithms:
▪ Registration by minimizing
▪ Intensity difference
▪ Correlation techniques
▪ Ratio image uniformity
▪ Partitioned Intensity
▪ Uniformity
* Needed: Pixel-by-pixel error measure, Mapping technique (transformation), Minimization technique

33
Q

What are 2D (joint) histograms?

A

Images can be viewed as probability distributions; calculated via joint histograms
Frequency of corresponding intensity pairs can be interpreted in terms of probabilities

34
Q

Which similarity measures do you know for image registration?

A

Similarity measure = quantifies degree of similarity between intensity patterns in two images
Multimodal
* Mutual information
* Normalized mutual information
Same modality
* Joint entropy
* Cross-correlation
* Sum of squared intensity differences
* Ratio image uniformity

35
Q

Given a defined similarity measure, how can the optimal parameters be found to minimize the criterion?

A

Using the Gradient descent
Repeat:
▪ Compute gradient
▪ Make step in gradient direction (not too small, not to big)
▪ Update mapping equation
▪ Remap image
Until convergence
▪ Ideally when gradient = 0

36
Q

Give an application for when to use free-form deformation image registration.

A

= Transformation of moving (non-rigid) image based on deformation
Contrast-enhanced breast MR
Transformation in 3D:
* Motion of breast is non-rigid (moving)
* Approach: Develop a combined transformation which consists of global and local transformation
* Overall motion of the breast. Affine transformation in 3D: Rigid (6 degrees of freedom) + Scaling and shearing (6 degrees of freedom)

37
Q

What is feature extraction and which steps are involved? Give an example.

A

= process of defining a set of features, or image characteristics, which will most efficiently or meaningfully represent the information that is important for analysis
Consists of
▪ Feature detection, i.e. finding features such as corners
▪ Feature description, i.e. quantification of attributes such as corner orientation
Example: Extract malignant lesion of CT scan lung

38
Q

What is a feature vector and a feature space? What are invariance and co-variance in the context of features?

A
  • Feature vectors
    ▪ Package/container for numerical descriptors
    ▪ Column vector (d x 1) or row vector (1 x d)
    E.g. each pixel is a 3D (RGB) colour feature vector (R value, G value, B value)
    Or (length; mean; variance; intensity) = 4x1 = f_B1
  • Feature space
    ▪ Contains a point cloud of feature vectors
    ▪ d-dimensional
    Example: (f_B1, f_B2, f_B3, f_B4, f_B5) = 4x5 Matrix
    Invariance:
    = Features should be invariant (insensitive) to variations (“Invariance to intensity transformations”)
    ▪ in: Scale, translation, rotation, illumination and viewpoint
    Value of the feature does not change after the application of the given transformation family => features(transform(image)) = features(image)

Co-variance:
= Feature changes by the same amount (“Covariance with geometric transformations”)
▪ E.g. Feature Area is covariant with scaling
▪ E.g. Feature Direction is covariant with rotation
Covariant detection: If we have two transformed versions of the same image, features should be detected in corresponding locations => features(transform(image)) = transform(features(image))

39
Q

What are chain codes and how can they be extracted?

A

= Representation of boundary by a set of connected straight lines with defined direction and length
* Offers a unified way to analyse the shape of a boundary
* Separately encodes each connected component
* Application: compression (lossless), recognition

40
Q

. Your task is to analyse a network of blood vessels. How would you extract relevant information?

A

Method: Medial representations/Skeletonization:
Thinning (Medial axis transform):
* Original shape can be fully reconstructed
* Set of points that are equidistant from the boundary
* Like boundaries, skeletons/blood vessels are related to the shape of a region
* Reduce a region to a tree or a graph

41
Q

What are geometrical features and how are they extracted? Give an example.

A
  • Length: number of pixels along the boundary
  • Diameter = max distance where p1 and p2 are points on the boundary
  • Major axis = line segment of length equal to the diameter and connecting two points on the boundary
  • Minor axis = line perpendicular to the major axis and of such length until intersection of the boundary
  • Basic rectangle (bounding box) by minor and major axis intersections with circle
  • Eccentricity is the ratio of major to minor axis
42
Q

Which basic regional descriptors do you know?

A

Basic descriptors: area, perimeter, compactness, mean value, circularity, effective diameter, eccentricity, normalized area

43
Q

Which approaches do you know to describe image/region texture?

A

Statistical approaches
▪ Smooth, coarse, grainy, regularity …
Spectral approaches
▪ Global periodicity
▪ Fourier spectrum: High energy, narrow peaks, …

44
Q

What is the biggest limitation of textural features based on histograms only? How can this limitation be overcome?

A

The histogram does not carry information about the spatial relationship of pixels (räuml. Beziehung)

Need to take into account relative positions of pixels => Haralick features
 Grey level co-occurrence matrix (GLCM)

  • Matrix G were measured the number of times intensities occur in a specified position
  • Position specified with a position operator Q
    ▪ E.g. Q = “one pixel immediately to the right”
    ▪ E.g. Q = “one pixel to the right and one pixel above”
45
Q

What are Haralick features?

A

= Set of 14 textural features which can be extracted from the co-occurrence matrix
* Contain information about image texture characteristics such as homogeneity, linearity, contrast
* Four directions of adjacency, calculate GLCM for each direction

46
Q

How do local binary patterns work?

A
  • Operator returns a discrete value at each pixel that characterizes the local texture partially invariant to luminance changes
  • Compares eight neighbourhood pixel intensities to the centre pixel intensity:
    ▪ 0 if intensity is less then centre pixel
    ▪ 1 if intensity is greater or equal than centre pixel
  • Orientation invariant
    ▪ Shift to minimum
  • Reduce classes by aggregation
    ▪ E.g. uniform vs. Non-uniform: 00000000 vs 01010101
47
Q

What is a design matrix?

A

A design matrix is a matrix containing data about multiple characteristics of several individuals or objects. Each row corresponds to an individual and each column to a characteristic.
Gather the feature descriptors for a single image which then can be used for further processing (e.g. kmeans) => here just one feature vector for each object (4x5x1)
Features on one axis, classes/samples on other axis

48
Q

What is the principal component analysis? What is it used for?

A

Given example vectors (design matrix)
2. Calculate covariance matrix, and its eigenvector decomposition
3. Represent data in the new space with reduced dimensionality (also Autoencoder)

 Reduce the dimensions of data in space => large feature vectors (Get rid of useless/redundant information)
▪ Feature selection (Visualization)
▪ Shape normalization
▪ Model shape
▪ Classification
* To visualize high dimensional datasets
* Each example is represented by a vector containing the values for all features (E.g. 3 for colour vectors)

49
Q

What are whole-image features? What are their typical applications?

A

Features to be extracted from whole images
* Corner detection
Corners = Junctions of contours / Large local gradient in multiple directions
Change of intensity for the displacement of a patch (Intensitätsänderung für die Verschiebung)
Harris corner detection:
▪ Look for local patch that produces noticeable difference when moved around
▪ Change of intensity for the displacement [u,v]
* SIFT features
Applications
▪ Image matching purposes (registration, mosaicking): patches containing a corner has distinct local features
▪ Seed point selection

50
Q

What are SIFT features? What are their properties? (Funktionen + Eigenschaften)

A

= Scale-Invariant Feature Transform
= Local histogram-based descriptor
Outline: Input = Image, Output = feature vector
* Detection of keypoints (= SIFT features)
* Refine selection of keypoints
* Describe each keypoint using a feature vector
Properties of SIFT feature descriptor:
* Invariant to uniform scaling, orientation, illumination changes, and partially invariant to affine distortion
* Various feature often combined
* Useful in the application of image registration/matching, mosaicking
Preferred method for:
* Scale changes, rotations, changes in illumination and/or viewpoint
* Applications: object recognition, image stitching, 3D modeling, gesture recognition, video tracking, individual identification, …

51
Q

What are the limitations of hand-designed programs? How can machine learning overcome these limitations?

A

Limitation: hard to maintain for complex problems
Machine learning:
 Allow to learn from data and improve the program in fluctuating environments
 Can help humans learn, by inspecting to see what they have learned (although this can be tricky for some algorithms)
 Getting insights about complex problems and large amounts of data

52
Q

What is the difference between supervised and unsupervised learning? Give a machine learning algorithm for each of the types.

A

Unsupervised learning: No need for human supervision, Training data is unlabelled, model identifies structures like clusters
E.g. k-Means, principal component analysis PCA, association rule learning, GAN
Supervised learning: Human supervision, Training data fed to algorithm includes known labels, model learns decision boundarys and replicates labelling
E.g. k-Nearest neighbours, linear and logistic regression, support vector machines (SVM), decision trees and random forests, neural networks

53
Q

What is the definition of learning algorithms? Give an example for experience, task and performance.

A

… are able to learn from data.
„A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P improves with experience E.“
Task: how the system should process an example
E.g. Classification, Regression, Denoising, Anomaly detection
Performance: evaluate the abilities of the ML algorithm
E.g. ROC curve, Confusion matrix (with FP, FN, TP, TN, accuracy, specificity, sensitivity)
Experience: kind of experience during learning
E.g. supervised, unsupervised, design matrix, …

54
Q

Which datasets are usually involved to find the parameters of a machine learning model? What is a useful split for these datasets?

A

Distinct datasets:
* Training set: Data used to train the model
* Tuning set: Used to evaluate the performance of different models and used to find the hyperparameter values
* (Validation) test set: Used to evaluate the performance with unseen data
* (External validation test set)

55
Q

Assuming only a rather small dataset is available for your classification task. How can you evaluate the model throughly? How can you increase the dataset?

A

Cross-validation: If small dataset => splitting the dataset into training and test set randomly multiple times
Most common: k-fold cross validation
* Dataset is split into k non-overlapping subsets
* k trials are performed
* Average test error of k trials is taken

56
Q

What is the curse of dimensionality?

A

= Performance can be reduced when the amount of features is increased (small number of examples)

 Rule of thumb: number of examples/number of features > 10

57
Q

What is a classifier? What is over- and under fitting?

A

A classifier in machine learning is an algorithm that automatically orders or categorizes data into one or more of a set of “classes.”[Link]
Different classifiers with different approaches to get boundary from training data
* Capture distribution of training data faithfully
* Be fast during training
* Be fast during classification of new data
* Deal with noise in the training data
* Deal with imperfect feature extractors
Over-fitting: the classifier is too faithful to the training data, but doesn’t reflect the underlying distribution anymore (Low bias - high variance)
 Regularization
Under-fitting: the classifier doesn’t reflect the training data or the underlying distribution (High bias - low variance)
 Possibly we need a classifier with more parameters

58
Q

How do nearest neighbour classifiers work?

A

Assign class label based on training example with lowest distance: NN classifier (also k-nearest neighbour, minimum distance classifiers)
Principle:
* Training set of labelled prototypes and classes
* An example to be classified
* Find prototype which is closest to this
* Example gets label of prototype

59
Q

. What are the advantages of decision trees? What are the design considerations when constructing decision trees?

A

Feature vectors mainly numerical vectors (E.g. length, width, intensity)
Advantages:
* Easy to understand and interpret
* Fast
* Versatile and powerful: Minor changes in training set can cause major changes in the classification
Design considerations CART:
1. Branching factor: How many splits should be attached to a node, two (binary) or more?
2. Query selection: Which feature should be tested at a node?
3. Stopping criteria: When does a node should be transferred into a leave?
4. Pruning: If tree gets too big, how can it be made smaller (pruning)?
5. Classification: If a leaf is not pure, how is a decision made?

60
Q

What is ensemble learning? Give an example of a machine learning algorithm based on ensemble learning.

A
  • Aggregate predictions of group of predictors (regressors, classifiers) are often superior than best individual predictor
  • Group of predictors is called ensemble → ensemble learning
    Example: Group of decision trees – random forest
61
Q

Which dimensionality reduction methods do you know?

A

Autoencoders, PCA Principal component analysis

62
Q

How is Bayes’ rule applied in pattern recognition?

A

X is a feature, Y is a class membership
* We want to know P(X|Y) => Probability: feature X belongs to class Y
* Conditional probability: If we know P(X) we can compute P(X|Y) using Bayes’ rule

63
Q

What is a perceptron and how does it work? What are the limitations?

A

= Special case of a linear, binary classifier
* Assuming a n-dimensional feature vector and two classes
* Find a discriminant function which codes the class labels
* Goal: Find weights and bias

  • Is guaranteed to converge if data are linearly separable

Limitations: Good for linearly separable classes  Fails for typical XOR problem (not linear separable), Solution – multiple perceptrons

64
Q

What is the definition of a neural network? Explain the basic architecture of feedforward neural networks.

A

= Interconnected perceptron-like computing elements called artificial neurons
* Interconnected neurons are organized in layers where the output of one layer provides the input of the following layer (Hence, a sensitive activation function would affect all subsequent layers)
* Activation function make a neuron to “fire” and must be differentiable
“Feedforward to next layer”: neuron in hidden layer l goes to all neurons in layer l+1

65
Q

Which parameters define a neural network? How are neural networks trained?

A
  • Weights W
  • Biases B
  • Activation functions h
    Training: use sets of training patterns to estimate these parameters
    Tool “backpropagation”:
    1. Inputting the pattern vectors
    2. Forward pass through the network and determine the classification error
    3. A backward (backpropagation) pass that feeds output error back to estimate required changes
    4. Updating the weights and biases in the network
66
Q

What are convolutional neural networks and what are the differences to feedforward neural networks?

A

Convolutional neural networks:
* Have additional convolutional layers and pooling layers before the NN => Purpose: Automatically detecting features

  • Accept (training) images directly as inputs
  • No pre-defined features by human experts needed  CNNs learn directly from raw image data
    FFN Networks: Convert images to vectors = patterns organized in feature vectors, loss of spatial relationship
67
Q

What is a convolutional layer? What is pooling and what are common pooling methods?

A

CNN layer: Collection of feature maps (One feature map: Out of receptive field = Neighbourhoods of image; with same weights and single bias)

Pooling = Downsampling to reduce spatial resolution (for memory + comp. power)
* Typically 2x2 (no overlap)
* Pooling methods: Average, Max-pooling, L2 pooling

68
Q

What are the main differences between different deep CNN networks? Name an architecture mentioned in the lecture which you would use for the task of biomedical image segmentation?

A

Deep CNN networks: Use of pre-trained models
U-Net: Convolutional network for biomedical image segmentation
 Works with only a few annotated images

69
Q

What is generative modelling?

A

Goal: Given training data set, generate new samples from same distribution
* We train a model from samples drawn from a distribution
* It learns an estimate of this distribution

70
Q

What is the goal of unsupervised learning? What are the applications of unsupervised methods?

A

Learn some underlying structure of data
Examples: Clustering dimensionality reduction, density estimation,…

71
Q

Compare autoencoders against general adversarial networks. What are the main differences?

A

Autoencoders
= Neural networks capable of learning efficient representations of the input data
* Encoder (recognition network) > Decoder (generative network)
* Used for: Dimensionality reduction, Feature, detection, generative modelling
GAN
= Make neural networks to compete each other and pushing them to excel (=übertreffen)
* Generator and Discriminator
* Used for generative modelling: Anomaly detection, data augmentation, super-resolution

72
Q

What are the principal steps in image-guided therapy? Which imaging modalities and which image processing steps are mainly involved?

A

= Aims to use imaging to improve the localization and targeting of diseased tissue, Monitor and control treatments
In image-guided surgery and interventional radiology
Steps:
* Prior to procedure diagnostic imaging is performed
* Images are converted and modelled, e.g. in 3D, to represent the patient’s anatomy
* Information is used for
Pre-operative surgical planning
Intraoperative surgical decision making

73
Q

How can medical imaging improve personalized medicine?

A

Individual treatment check and endpoints – image guided therapy (E.g. Macular Edema Recurrence)
* Dose optimization, optimal efficacy
* Timely treatment
* Prediction of treatment response

Main challenge: find predictive markers for future disease progression and treatment response

74
Q

What is radiomics? What are the main steps involved? What is currently the bottleneck in radiomics?

A

= High-throughput extraction and analysis (mining) of radiological image data
* Extraction of quantitative features from imaging data such as MR or CT
* Bridge between medical imaging and personalized medicine
* Associate features with predictive goals: Diagnosis, Prognosis
Motivation: Image guidance to support clinical decisions (radiations, surgery, treatment)