RJafroc: Artificial Intelligence Systems and Observer Performance
Analyzing the performance of artificial intelligence
 (AI) systems/algorithms characterized by a 'search-and-report'
 strategy. Historically observer performance has dealt with
 measuring radiologists' performances in search tasks, e.g., searching
 for lesions in medical images and reporting them, but the implicit
 location information has been ignored. The implemented methods apply
 to analyzing the absolute and relative performances of AI systems,
 comparing AI performance to a group of human readers or optimizing the
 reporting threshold of an AI system. In addition to performing historical
 receiver operating receiver operating characteristic (ROC) analysis
 (localization information ignored), the software also performs
 free-response receiver operating characteristic (FROC)
 analysis, where lesion localization information is used. A book
 using the software has been published: Chakraborty DP: Observer
 Performance Methods for Diagnostic Imaging - Foundations, Modeling,
 and Applications with R-Based Examples, Taylor-Francis LLC; 2017:
 <https://www.routledge.com/Observer-Performance-Methods-for-Diagnostic-Imaging-Foundations-Modeling/Chakraborty/p/book/9781482214840>.
 Online updates to this book, which use the software, are at
 <https://dpc10ster.github.io/RJafrocQuickStart/>,
 <https://dpc10ster.github.io/RJafrocRocBook/> and at
 <https://dpc10ster.github.io/RJafrocFrocBook/>. Supported data
 collection paradigms are the ROC, FROC and the location ROC (LROC).
 ROC data consists of single ratings per images, where a rating is
 the perceived confidence level that the image is that of a diseased
 patient. An ROC curve is a plot of true positive fraction vs. false
 positive fraction. FROC data consists of a variable number (zero or
 more) of mark-rating pairs per image, where a mark is the location
 of a reported suspicious region and the rating is the confidence
 level that it is a real lesion. LROC data consists of a rating and a
 location of the most suspicious region, for every image. Four models
 of observer performance, and curve-fitting software, are implemented:
 the binormal model (BM), the contaminated binormal model (CBM), the
 correlated contaminated binormal model (CORCBM), and the radiological
 search model (RSM). Unlike the binormal model, CBM, CORCBM and RSM
 predict 'proper' ROC curves that do not inappropriately cross the
 chance diagonal. Additionally, RSM parameters are related to search
 performance (not measured in conventional ROC analysis) and
 classification performance. Search performance refers to finding
 lesions, i.e., true positives, while simultaneously not finding false
 positive locations. Classification performance measures the ability to
 distinguish between true and false positive locations. Knowing these
 separate performances allows principled optimization of reader or AI
 system performance. This package supersedes Windows JAFROC (jackknife
 alternative FROC) software V4.2.1,
 <https://github.com/dpc10ster/WindowsJafroc>. Package functions are
 organized as follows. Data file related function names are preceded
 by 'Df', curve fitting functions by 'Fit', included data sets by 'dataset',
 plotting functions by 'Plot', significance testing functions by 'St',
 sample size related functions by 'Ss', data simulation functions by
 'Simulate' and utility functions by 'Util'. Implemented are figures of
 merit (FOMs) for quantifying performance and functions for visualizing
 empirical or fitted operating characteristics: e.g., ROC, FROC, alternative
 FROC (AFROC) and weighted AFROC (wAFROC) curves. For fully crossed study
 designs significance testing of reader-averaged FOM differences between
 modalities is implemented via either Dorfman-Berbaum-Metz or the
 Obuchowski-Rockette methods. Also implemented is single treatment analysis,
 which allows comparison of performance of a group of radiologists to a
 specified value, or comparison of AI to a group of radiologists interpreting
 the same cases. Crossed-modality analysis is implemented wherein there are
 two crossed treatment factors and the aim is to determined performance in
 each treatment factor averaged over all levels of the second factor. Sample
 size estimation tools are provided for ROC and FROC studies; these use
 estimates of the relevant variances from a pilot study to predict required
 numbers of readers and cases in a pivotal study to achieve the desired power.
 Utility and data file manipulation functions allow data to be read in any of
 the currently used input formats, including Excel, and the results of the
 analysis can be viewed in text or Excel output files. The methods are
 illustrated with several included datasets from the author's collaborations.
 This update includes improvements to the code, some as a result of
 user-reported bugs and new feature requests, and others discovered during
 ongoing testing and code simplification.
| Version: | 2.1.2 | 
| Depends: | R (≥ 3.5.0) | 
| Imports: | bbmle, binom, dplyr, ggplot2, mvtnorm, numDeriv, openxlsx, readxl, Rcpp, stats, stringr, tools, utils | 
| LinkingTo: | Rcpp | 
| Suggests: | testthat, knitr, kableExtra, rmarkdown | 
| Published: | 2022-11-08 | 
| DOI: | 10.32614/CRAN.package.RJafroc | 
| Author: | Dev Chakraborty [cre, aut, cph],
  Peter Phillips [ctb],
  Xuetong Zhai [aut] | 
| Maintainer: | Dev Chakraborty  <dpc10ster at gmail.com> | 
| License: | GPL-3 | 
| URL: | https://dpc10ster.github.io/RJafroc/ | 
| NeedsCompilation: | yes | 
| Materials: | NEWS | 
| CRAN checks: | RJafroc results | 
Documentation:
Downloads:
Linking:
Please use the canonical form
https://CRAN.R-project.org/package=RJafroc
to link to this page.