Automated En Masse Machine Learning Model Generation Shows Comparable Performance as Classic Regression Models for Predicting Delayed Graft Function in Renal Allografts.

2021 
BACKGROUND Several groups have previously developed logistic regression models for predicting delayed graft function (DGF). In this study, we used an automated machine learning (ML) modeling pipeline to generate and optimize DGF prediction models en masse. METHODS Deceased donor renal transplants at our institution from 2010-2018 were included. Input data consisted of 21 donor features from United Network for Organ Sharing. A training set composed of ~50%/50% split in DGF-positive and DGF-negative cases was used to generate 400,869 models. Each model was based on one of seven ML algorithms (gradient boosting machine, k-nearest neighbor, logistic regression, neural network, naive Bayes, random forest, support vector machine) with various combinations of feature sets and hyperparameter values. Each model's performance was based on a separate secondary test dataset and assessed by common statistical metrics. RESULTS The best performing models were based on neural network algorithms, with the highest area under the receiver operating characteristic (AUROC) curve of 0.7595. This model used 10 out of the original 21 donor features, including age, height, weight, ethnicity, serum creatinine, blood urea nitrogen, hypertension history, donation after cardiac death status, cause of death, and cold ischemia time. With the same donor data, the highest AUROC curve for logistic regression models was 0.7484, using all donor features. CONCLUSION Our automated en masse ML modeling approach was able to rapidly generate ML models for DGF prediction. The performance of the ML models was comparable to classic logistic regression models.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    14
    References
    3
    Citations
    NaN
    KQI
    []