TG273- recommendations on AI Flashcards
issues of most of the studied sone so far
high risk for bias
goal of TG273
bring attention to proper training and validation of machine learning algorithms
tips for data collection
-data collected should include the intended use popluation
-use public databases so there is more data
data augmentation
create alterations of the training data or to create synthetic data to increase the training set size
data harmonization
Data may include images obtained at different sites, acquired with different equipment and image-acquisition parameters, and reconstructed and/or post-processed using different algorithms. These differences may result in systematic variations across images. Data harmonization aims to reduce these variations retrospectively after acquisition while preserving the biological variability captured in the images
reference standard
defining reference standard can be subjective- increasing number of doctors who define the standard helps to reduce subjectivity
annotations
level of annotation granularity or details, depends on the task
-entire image
-region based
-pixel based
methods for acquiring annotations
expert labels: subjective reference standard from human domain experts
-electronic health record
-crowd sourcing: switch from domain experts to many, less experienced users- not recommended for general use
-phantoms
defining true positive is difficult
3 partitions for dataset
-training, validation, and test sets
-validation is part of training
-testing is done on completely independent data set
types of supervised learning
In supervised learning, a model is trained to map input data to output data based on explicit examples of the desired input-output pairs, as provided by the user.
Semi-supervised learning algorithms exploit a combination of labeled and unlabeled data. In this case, the model is given some guidance about the desired outcome, but the annotations do not need to be as detailed or extensive as those used with supervised learning.
Self-supervised learning can exploit large unlabeled datasets for feature representation and has a regularizing effect on the learned features
Unsupervised learning refers to a class of algorithms that can autonomously learn from data without reference to any labels or any instruction from the user.
transfer learning
Transfer learning in DCNNs is commonly implemented by training a network on one task and then “transferring” the parameters (or weights) from the trained model to initialize the network for a new task, rather than randomly initializing it (also known as “training from scratch”).
multi-task learning
Multi-task learning is a special type of transfer learning in which a DCNN is trained to jointly learn interrelated tasks, as opposed to addressing each task sequentially
federated learning
Federated learning is a distributed machine learning approach that enables collaborative training on decentralized datasets.121-124 Each site trains the model locally with its own dataset and then only the trained model parameters are shared, thus producing a global model benefiting from access to a large corpus of data without requiring data sharing and without posing risks to patient privacy. There are, however, several open-ended questions with regard to federated learning that are relevant to medical imaging.125, 126 In particular, there is no formalized training protocol yet to guarantee that the performance of a model trained with federated learning is comparable to that of a centralized trained model with access to all the data.127 Also unknown is (1) the extent to which local model overfitting negatively impacts the global model, and (2) the tradeoff between access to more data through a federated process versus traditional learning with a fully controlled dataset.
continuous learning
Continuous or “life-long” learning emulates the human ability to continuously learn and adapt as new data are presented.128, 129 Theoretically, continuously learning AI systems can accelerate model optimization and continuously improve their performance by taking advantage of new data presented during clinical use. In practice, adaptive training of shallow and deep neural networks using incrementally available data generally results in rapid overriding of their weights, a phenomenon known as “interference” or “catastrophic forgetting.