Abstract
Presented at ENAR Conference
Recently, Kullback-Leibler divergence measure (KL), which captures the disparity between two distributions, has been considered as an index for determining the diagnostic performance of biomarkers. Our study investigates variety of applications of the generalized KL divergence (GKL) in medical diagnostics, including overall measures of rule-in and rule-out potential and proposes an optimization criterion based on KL divergence for cut-point selection for k-stage disease. Moreover, the study links the KL divergence with some common Receiver Operating Characteristic (ROC) measures and presents analytically and numerically the relations in situations of one cut-point as well as multiple cut-points. Furthermore, the graphical application and interpretation of GKL divergence, which is referred as the information graph, is discussed. A comprehensive data analysis of the real data example to illustrate the proposed applications is provided.
| Original language | American English |
|---|---|
| State | Published - Mar 1 2019 |
| Event | Eastern North American Region International Biometric Society Conference - Duration: Mar 1 2019 → … |
Conference
| Conference | Eastern North American Region International Biometric Society Conference |
|---|---|
| Period | 03/1/19 → … |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 3 Good Health and Well-being
Disciplines
- Biostatistics
- Environmental Public Health
- Epidemiology
- Public Health
Fingerprint
Dive into the research topics of 'Applications of Generalized Kullback-Leibler Divergence as a Measure of Medical Diagnostic and Cut-Point Criterion for K-Stages Diseases'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver