Summary
A receiver operating characteristic curve, or ROC curve, is a graphical plot that illustrates the diagnostic ability of a binary classifier system as its discrimination threshold is varied. The ROC curve is the plot of the true positive rate (TPR) against the false positive rate (FPR), at various threshold settings. The ROC can also be thought of as a plot of the power as a function of the Type I Error of the decision rule (when the performance is calculated from just a sample of the population, it can be thought of as estimators of these quantities). The ROC curve is thus the sensitivity or recall as a function of fall-out. Given the probability distributions for both true positive and false positive are known, the ROC curve is obtained as the cumulative distribution function (CDF, area under the probability distribution from to the discrimination threshold) of the detection probability in the y-axis versus the CDF of the false positive probability on the x-axis. ROC analysis provides tools to select possibly optimal models and to discard suboptimal ones independently from (and prior to specifying) the cost context or the class distribution. ROC analysis is related in a direct and natural way to cost/benefit analysis of diagnostic decision making. There are a large number of synonyms for components of a ROC curve. They are tabulated on the right. The true-positive rate is also known as sensitivity, recall or probability of detection. The false-positive rate is also known as probability of false alarm and equals (1 − specificity). The ROC is also known as a relative operating characteristic curve, because it is a comparison of two operating characteristics (TPR and FPR) as the criterion changes. The ROC curve was first developed by electrical engineers and radar engineers during World War II for detecting enemy objects in battlefields, starting in 1941, which led to its name ("receiver operating characteristic"). It was soon introduced to psychology to account for perceptual detection of stimuli.
About this result
This page is automatically generated and may contain information that is not correct, complete, up-to-date, or relevant to your search query. The same applies to every other page on this website. Please make sure to verify the information with EPFL's official sources.
Related courses (19)
MICRO-401: Machine learning programming
This is a practice-based course, where students program algorithms in machine learning and evaluate the performance of the algorithm thoroughly using real-world dataset.
PHYS-455: Introduction to medical radiation physics
This course covers the physical principles underlying medical diagnostic imaging (radiography, fluoroscopy, CT, SPECT, PET, MRI), radiation therapy and radiopharmacy. The focus is not only on risk an
ENG-209: Data science for engineers with Python
Ce cours est divisé en deux partie. La première partie présente le langage Python et les différences notables entre Python et C++ (utilisé dans le cours précédent ICC). La seconde partie est une intro
Show more
Related lectures (49)
Model Selection: ROC Curves, Regression Evaluation, Conclusion
Explores thresholding, ROC curves, regression evaluation, naive methods, and model selection in machine learning.
Logistic Regression: Fundamentals and Applications
Explores logistic regression fundamentals, including cost functions, regularization, and classification boundaries, with practical examples using scikit-learn.
Generalized Linear Regression: Classification
Explores Generalized Linear Regression, Classification, confusion matrices, ROC curves, and noise in data.
Show more
Related publications (64)

Computation of sensitivity coefficients in fixed source simulations with SERPENT2

Andreas Pautz, Mathieu Hursin

Within the scope of the implementation of a nuclear data pipeline aiming at producing the best possible evaluated nuclear data files, a major point is the production of relevant sensitivity coefficients when including integral benchmark information. Thanks ...
Lausanne2024

Multimodal fusion of liquid biopsy and CT enhances differential diagnosis of early-stage lung adenocarcinoma

Jiancheng Yang, Li Zhang, Jun Lu, Hua Zhong, Yunhua Zhang

This research explores the potential of multimodal fusion for the differential diagnosis of early-stage lung adenocarcinoma (LUAD) (tumor sizes < 2 cm). It combines liquid biopsy biomarkers, specifically extracellular vesicle long RNA (evlRNA) and the comp ...
Berlin2024

Parental status and markers of brain and cellular age: A 3D convolutional network and classification study

Donal Patrick Hill

Recent research shows prominent effects of pregnancy and the parenthood transition on structural brain characteristics in humans. Here, we present a comprehensive study of how parental status and number of children born/fathered links to markers of brain a ...
Pergamon-Elsevier Science Ltd2024
Show more
Related concepts (16)
Prevalence
In epidemiology, prevalence is the proportion of a particular population found to be affected by a medical condition (typically a disease or a risk factor such as smoking or seatbelt use) at a specific time. It is derived by comparing the number of people found to have the condition with the total number of people studied and is usually expressed as a fraction, a percentage, or the number of cases per 10,000 or 100,000 people. Prevalence is most often used in questionnaire studies. Incidence (epidemiology)#Incidence vs.
Sensitivity and specificity
Sensitivity and specificity mathematically describe the accuracy of a test that reports the presence or absence of a condition. If individuals who have the condition are considered "positive" and those who do not are considered "negative", then sensitivity is a measure of how well a test can identify true positives and specificity is a measure of how well a test can identify true negatives: Sensitivity (true positive rate) is the probability of a positive test result, conditioned on the individual truly being positive.
Confusion matrix
In the field of machine learning and specifically the problem of statistical classification, a confusion matrix, also known as error matrix, is a specific table layout that allows visualization of the performance of an algorithm, typically a supervised learning one; in unsupervised learning it is usually called a matching matrix. Each row of the matrix represents the instances in an actual class while each column represents the instances in a predicted class, or vice versa – both variants are found in the literature.
Show more