diff --git "a/20240123/2209.07805v4.json" "b/20240123/2209.07805v4.json" new file mode 100644--- /dev/null +++ "b/20240123/2209.07805v4.json" @@ -0,0 +1,838 @@ +{ + "title": "A Comprehensive Benchmark for COVID-19 Predictive Modeling Using Electronic Health Records in Intensive Care", + "abstract": "The COVID-19 pandemic highlighted the need for predictive deep learning models in healthcare. However, practical prediction task design, fair comparison and model selection for clinical applications remain a challenge. To address this, we introduced and evaluated two new prediction tasks - Outcome-specific length-of-stay and Early mortality prediction for COVID-19 patients in intensive care - which better reflect clinical realities. We developed evaluation metrics, model adaptation designs, and open-source data preprocessing pipelines for these tasks, while also evaluating 18 predictive models, including clinical scoring methods, traditional machine learning, basic deep learning, and advanced deep learning models tailored for EHR data. Benchmarking results from two real-world COVID-19 EHR datasets are provided, and all results and trained models are released on an online platform for use by clinicians and researchers. Our efforts contribute to the advancement of deep learning and machine learning research in pandemic predictive modeling.", + "sections": [ + { + "section_id": "1", + "parent_section_id": null, + "section_name": "Introduction", + "text": "The COVID-19 pandemic has undoubtedly left an indelible impact on the global community. Although recent studies suggest that new COVID-19 variants are less lethal, their heightened transmissibility contributes to a continual surge in cases worldwide [1 ###reference_1###]. Given these circumstances, the critical need for early risk prediction and disease progression estimation, especially for COVID-19 patients in intensive care units (ICUs), is essential to optimally allocate medical resources and alleviate the strain on our healthcare system in the post-pandemic era.\nDespite the commendable performance of existing COVID-19 predictive modeling works on specific datasets [2 ###reference_2###, 3 ###reference_3###, 4 ###reference_4###, 5 ###reference_5###, 6 ###reference_6###, 7 ###reference_7###, 8 ###reference_8###, 9 ###reference_9###, 10 ###reference_10###], researchers and clinicians often encounter difficulties when attempting to apply state-of-the-art prediction models to new data. Questions arise regarding whether to employ deep learning models or traditional machine learning models, and how different models compare in terms of prediction performances. Though there are a few descriptive review works [11 ###reference_11###, 12 ###reference_12###], there is still a lack of a fair quantitative comparative framework for existing models. Various models have been applied to different datasets, many of which are either not publicly available or entail strict access restrictions. Direct comparison of their results is therefore problematic, and model selection for new data is equally challenging as re-implementing all existing models on new data is resource-intensive, particularly for clinicians. This limitation curtails the practical usage of these models and hinders further research, underlining the necessity for a benchmark for COVID-19 predictive modeling that facilitates model comparison using identical data and evaluation strategies.\nExisting electronic health record (EHR) prediction benchmarks [13 ###reference_13###, 14 ###reference_14###, 15 ###reference_15###, 16 ###reference_16###] predominantly rely on publicly available ICU datasets such as MIMIC-III [17 ###reference_17###] and MIMIC-II [18 ###reference_18###]. These benchmarks compare various machine learning and deep learning models across multiple standardized prediction tasks, including mortality prediction, patient phenotyping, length-of-stay (LOS) prediction, etc. However, these efforts primarily evaluate the performance of basic machine learning and deep learning models like recurrent neural networks (RNN) or long-short term memory networks (LSTM), despite the presence of numerous advanced deep learning models specifically designed for EHR data and clinical tasks. Notably, comprehensive benchmarking results for machine learning, deep learning and more advanced models on publicly available datasets remain absent for COVID-19 predictive tasks.\nAnother concern lies in the fact that current COVID-19 prediction works largely mirror the task settings of previous EHR data-mining endeavors. Risk or mortality prediction models, for instance, typically output the predicted risk score at the final timestep of available EHR sequences[2 ###reference_2###, 10 ###reference_10###]. This methodology can be problematic as it may be too late to initiate life-saving treatments for high-risk COVID-19 patients (especially ICU patients) by the time their status at the last timestep is identified as critical [19 ###reference_19###, 20 ###reference_20###]. Simultaneously, the length-of-stay (LOS) prediction task is commonly formulated as a regression task[15 ###reference_15###], with the model outputting the predicted remaining LOS at each timestep. This approach presents a notable issue where high-risk and low-risk patients may exhibit the same LOS value, as both death and ICU discharge are defined as the end of the stay in previous benchmarking works [15 ###reference_15###, 21 ###reference_21###, 22 ###reference_22###, 16 ###reference_16###, 14 ###reference_14###]. Although the model can theoretically learn non-linearity and map these statuses to distinct locations in the latent embedding space, the practical application may prove challenging for clinicians. Adaptations are thus required for both tasks to cater to the unique intensive care setting associated with COVID-19. Consequently, it is imperative to compare existing deep learning and machine learning models using more clinically practical prediction tasks, identical publicly available data, and consistent evaluation settings.\nIn this study, we aim to bridge these gaps by proposing a standardized and comprehensive benchmark for COVID-19 prediction tasks in intensive care units. This benchmark enables comparison among machine learning, basic deep learning, and state-of-the-art EHR-specific models. Our contributions are three-fold:\nTask, metric and model designing: We introduce two tasks based on the clinical practice for COVID-19 patients in ICUs: Outcome-specific length-of-stay prediction and Early mortality prediction. Distinct from existing length-of-stay prediction frameworks, the outcome-specific length-of-stay prediction is formulated as a multi-target task. This task simultaneously outputs the patient outcome and corresponding length-of-stay, enabling clinicians to distinguish progression statuses between low-risk and high-risk patients. The early mortality prediction task is designed to alert clinicians to high-risk patients as early as possible, thereby preempting potential treatment delays. We design specific evaluation metrics for both tasks, namely the early mortality score (ES) and the outcome-specific mean absolute error (OSMAE), to assess model performance. We have also designed the multi-task training architecture for the LOS prediction task and a time-aware loss term that can significantly improve deep learning models\u2019 early prediction performances.\nData preparation and preprocessing pipelines: We have established data preprocessing pipelines that include cleaning, filtering, missing value imputation, and cohort construction for two real-world COVID-19 EHR datasets. Both datasets consist of heterogeneous longitudinal EHR data from ICUs. Features include lab tests, vital signs, diagnoses, and static demographic information.\nModeling and benchmarking: We have implemented and evaluated 18 state-of-the-art predictive models across the two tasks, including 1 clinical scoring method, 5 machine learning models, 6 basic deep learning models, and 6 deep learning predictive models specifically designed for multivariate time-series EHR data. We conduct fair and reproducible comparisons and provide detailed benchmarking results to foster further research in this field.\nTo the best of our knowledge, this is the first benchmarking effort for patient-level COVID-19 prediction tasks in ICUs. We have made our code publicly available, enabling others to build complete benchmarks and reproduce all results. Our well-structured data preprocessing and modeling modules can also be easily applied to generate customized tasks and results. The benchmark code and documentations can be accessed at https://github.com/yhzhu99/pyehr ###reference_github.com/yhzhu99/pyehr###. Moreover, we have released all the benchmark experiment results and trained models on an online platform, which includes model performances with all hyperparameter combinations for both tasks and makes the results easy to query and download. The platform can be accessed at https://pyehr.netlify.app ###reference_pyehr.netlify.app###." + }, + { + "section_id": "2", + "parent_section_id": null, + "section_name": "Dataset Description and Problem Formulation", + "text": "The datasets used in this study and the proposed prediction tasks are as follows:" + }, + { + "section_id": "2.1", + "parent_section_id": "2", + "section_name": "EHR datasets for COVID-19 patients in intensive care", + "text": "In this work, we utilize two COVID-19 EHR datasets to conduct benchmark experiments. After performing an exhaustive search for publicly available COVID-19 EHR datasets worldwide, we selected two datasets based on ease of access and absence of regional restrictions:\n[2 ###reference_2###]\nThe TJH dataset comprises anonymized EHR data from 485 COVID-19 patients admitted to Tongji Hospital, China, between January 10 and February 24, 2020. The dataset includes 74 lab tests and vital signs, all of which are numerical features, as well as 2 demographic features (age and gender). This dataset is publicly accessible via GitHub (https://github.com/HAIRLAB/Pre_Surv_COVID_19 ###reference_19###). The dataset download script is included in our benchmark code.\n[23 ###reference_23###]\nThe CDSL dataset originates from the HM Hospitales EHR system and contains anonymized records of 4,479 patients admitted with a diagnosis of COVID-19 or suspected COVID-19 infection. The dataset includes heterogeneous medical features such as detailed information on diagnoses, treatments, admissions, ICU admissions, diagnostic imaging tests, laboratory results, and patient discharge or death status. The dataset is open to global researchers and can be accessed upon request (https://www.hmhospitales.com/prensa/notas-de-prensa/comunicado-covid-data-save-lives ###reference_de-prensa/comunicado-covid-data-save-lives###). Prospective users must complete a request form.\nThese datasets have been utilized in several mortality prediction studies [2 ###reference_2###, 24 ###reference_24###]. However, we found no existing well-organized data preprocessing pipeline code or benchmarking results for them, which prompted us to provide a comprehensive benchmark analysis and data-processing suite for these publicly available datasets, thereby facilitating related research." + }, + { + "section_id": "2.2", + "parent_section_id": "2", + "section_name": "Problem formulation and evaluation metrics", + "text": "In this study, we formulate two tasks\u2014Outcome-specific length-of-stay prediction and Early mortality prediction\u2014to evaluate and compare the performances of various machine learning and deep learning models. These tasks, which are adaptations of common length-of-stay and mortality prediction models, are tailored to suit the requirements of COVID-19 intensive care settings.\nLength-of-stay prediction is an important task in clinical practice, which better facilitates clinical resource management.\nWe define outcome-specific length-of-stay prediction as a multi-target prediction task, which encompasses a binary outcome classification task and a length-of-stay regression task. At each time step , the model generates two outputs: a predicted outcome (i.e., 1 for mortality and 0 for survival) and a predicted length-of-stay (LOS) indicating the remaining days corresponding to the predicted outcome or the end of the ICU stay. Compared to traditional length-of-stay prediction models that output only , our model provides a more comprehensive evaluation of a patient\u2019s progression, allowing clinicians to differentiate length-of-stay according to different outcomes. The model learning process may also benefit from this setting as patients with varying health statuses are explicitly modeled in the latent space.\n###figure_1### ###figure_2### We use mean absolute error (MAE) and mean squared error (MSE) as evluation metrics. Furthermore, to evaluate this multi-target task comprehensively, we propose a new metric, Outcome-specific Mean Absolute Error (OSMAE), formulated as follows:\nHere, represents the timestep, is the total length-of-stay, is the maximum value at the 95% percentile of the length-of-stay of all patients. is a penalty term represented by a piecewise function. A lower OSMAE indicates better model performance. Excluding the penalty term , OSMAE is equivalent to the conventional MAE for true positive and true negative predictions, i.e., . However, for false positive and false negative patients, OSMAE is significantly higher than the original MAE, since the model\u2019s prediction for the patient\u2019s outcome deviates completely from the actual outcome. We introduce the penalty term to prevent the model from incurring a high OSMAE during the early timesteps, as a patient\u2019s status may be uncertain during the initial ICU-stay phase. serves as a penalty threshold, indicating that we expect the model to make accurate predictions within the final days. The default value of is , where is the average value of all patients\u2019 total LOS. We provide a sensitivity analysis for in the discussion section. Fig. 0(a) ###reference_f1### illustrates the calculation of OSMAE and the penalty term using three examples: true positive, false negative, and true negative cases.\nEarly mortality prediction is defined as a binary sequential prediction task. For each timestep in the longitudinal EHR input sequence, the model will predict the mortality risk for this ICU stay, signifying whether the patient will succumb by the end of the ICU stay. An optimal early prediction model should raise an alert as soon as possible during the stay. Many existing mortality prediction models either feed the entire EHR sequence into the model and predict the mortality risk within a short future time window (e.g., 24h or 12h) [25 ###reference_25###, 26 ###reference_26###, 16 ###reference_16###, 15 ###reference_15###]. However, this approach restricts the clinical applicability of these models. Particularly for COVID-19 patients in the ICU, their health statuses may have severely deteriorated by the final timestep, making the risks evident to clinicians. At this advanced stage, an alarm indicating high risk may be too late to initiate life-saving treatments or procedures. To address this limitation, we propose to assess early prediction performance using weighted metrics.\nTo evaluate models\u2019 early prediction performances, we not only utilize traditional measures such as Area Under the Receiver Operating Characteristic (AUROC) and Area Under the Precision-Recall Curve (AUPRC), but we also introduce the Early Prediction Score (ES) to specifically assess the early prediction performance of the models. Fig.0(b) ###reference_f2### illustrates the ES. Drawing from previous studies[20 ###reference_20###], the idea is to assign a full score to early true positive predictions while penalizing late false negative predictions, i.e., the model fails to predict patients\u2019 outcomes accurately even in the final moments. We also give false alarms a small penalty (-0.1). Different penalty scores will not affect the relative comparison between different models. We provide details in the discussion section. We normalize the total ES for one patient as follows:\nThis normalization ensures that the ground truth (i.e., the highest possible score) receives a normalized score of 1 and the worst algorithm, which only outputs negative predictions, receives a normalized score of -1." + }, + { + "section_id": "3", + "parent_section_id": null, + "section_name": "Pipeline Design", + "text": "In this section, we introduce our benchmarking pipeline design, including data preprocessing, baseline selection, training and evaluation strategies." + }, + { + "section_id": "3.1", + "parent_section_id": "3", + "section_name": "Data preprocessing", + "text": "Our benchmark data processing pipeline comprises four stages: data cleaning, merging, normalization, and imputation. Although both datasets are well-formatted data tables, they still contain artifacts that necessitate preprocessing. Previous machine learning and deep learning works on COVID-19 predictive modeling lack a uniform setting for these preprocessing details, leading to subtle performance differences. In this study, we propose a reproducible and reliable preprocessing pipeline to establish a fair comparison base for various models.\nData cleaning and merging: Our first step in preprocessing the dataset into feature matrices is to structure the EHR data format for use in machine learning (ML) or deep learning (DL) models. Each patient\u2019s longitudinal EHR records are represented as a matrix, with each column representing a specific feature or label, and each row representing a record. We also extract demographic information for each patient as static data. During this step, we calculate ground truth labels, such as lengths of stay and mortality outcomes.\nFor error record cleaning, we first compute each feature\u2019s statistics, including the mean, standard deviation, minimum, maximum, median, and missing rate. For instance, in the TJH dataset, we observed that all entries for the 2019-nCoV nucleic acid detection are identical, and some other feature values are negative. We remove these clear error records. For example, in the CDSL dataset, we remove obvious errors such as oxygen saturation values above 100 and maximum blood pressure exceeding 220. All filtered values are replaced with NaN (Not a Number). Detailed feature statistics can be found in Appendix I.\nThe raw datasets have record-level missing rates of 84.25% for the TJH dataset and 98.38% for the CDSL dataset. High missing rates may impair models\u2019 prediction performance. Hence, following the previous benchmark preprocessing settings [15 ###reference_15###], we merge data into day-level for the TJH dataset and hour-level for the CDSL dataset, which reduces the missing rate of recorded physiological characteristics. It is worth noting that for the day/hour that does not have any record, we do not duplicate records from previous time steps. This data merging will not result in significant information loss since most records (over 95%) do not record different values for the same feature on the same day or at the same hour. If there are multiple records for the same feature in the same time slot, we record their mean value. For the CDSL dataset, the feature dimensions are much larger and features are sparser than in the TJH dataset: nearly 85% of the features have a missing rate beyond 90%, signifying that over 90% of patients never have records for these features. Therefore, we remove features with a missing rate higher than 90% among all patients.\nData normalization and imputation: We apply Z-score normalization to all demographic features, vital signs, lab test features, and the length-of-stays. Here, we calculate the mean and standard deviation based on the data in the 5% to 95% quantile range. Preventing future information leakage is crucial for maintaining the fairness of a benchmark. To address this, we implemented stringent strategies: (1) For missing values, we employed a forward-filling imputation method. Specifically, we replaced missing values with the most recent ones. If a patient lacked a prior record for a specific feature, we filled the missing data with the median value from all patients in the training set. (2) We utilized the means and standard deviations from the training set to normalize the entire dataset. This practice ensures no leakage regarding test set data distributions, a detail often overlooked in many previous benchmarking studies.\nThe data statistics of two processed datasets are shown in Table 1 ###reference_### and 2 ###reference_###.\n###table_1### ###table_2###" + }, + { + "section_id": "3.2", + "parent_section_id": "3", + "section_name": "Benchmarking experiment settings", + "text": "To provide a comprehensive benchmark comparison between existing models, we categorize the baseline models into four categories: clinical scoring models, machine learning models, basic deep learning models, and EHR-specific predictive models. Machine learning and basic deep learning models are popular for general classification and regression tasks. EHR-specific predictive models are designed explicitly for clinical predictive tasks with EHR data. We provide a list of baseline models below, and detailed descriptions of these models can be found in Appendix A:\nClinical scoring model: 4C mortality score [27 ###reference_27###].\nMachine learning models: Decision tree (DT), Random forest (RF), Gradient Boosting Decision Tree (GBDT), XGBoost [28 ###reference_28###], CatBoost [29 ###reference_29###].\nBasic deep learning models: Multi-layer perceptron (MLP), Recurrent neural network (RNN) [30 ###reference_30###], Long-short term memory network (LSTM) [31 ###reference_31###], Gated recurrent units (GRU) [32 ###reference_32###], Temporal convolutional networks (TCN) [33 ###reference_33###], Transformer[34 ###reference_34###].\nEHR-specific predictive models: RETAIN [35 ###reference_35###], StageNet [36 ###reference_36###], Dr.Agent [22 ###reference_22###], AdaCare [25 ###reference_25###], ConCare [26 ###reference_26###], GRASP [37 ###reference_37###].\nMachine learning models and some deep learning models (e.g., MLP) cannot handle sequential data as input. When training these models, we use the feature values at the current timestep as input and predict the target. The clinical scoring method is only applicable to the mortality prediction task. To minimize the randomness of the test sample selection process, we employ the stratified 10-fold cross-validation strategy to train and evaluate the models. We randomly divide patients into 10 groups or \"folds\". Then, we repeat the training and evaluation process 10 times. In the -th iteration, we select the -th fold as the test set and divide the remaining folds into training and validation sets at an 8:1 ratio. An illustration of this process is provided in Figure 2 ###reference_###. Finally, we report the means and standard deviations of the model performances on all test sets from the 10 iterations. Moreover, to evaluate the model\u2019s generalizability over time more comprehensively, we have also conducted a standard holdout experiment.\nWe divided the dataset into training, validation, and test sets in a 7:1:2 ratio, based on admission times (using the latest 20% of patients as the holdout dataset). We retrained all models five times with different random seeds and report the standard deviations. The results for the hold-out performance are shown in Appendix D.\nExperiments are conducted on a server equipped with dual Intel Xeon Silver 4210R CPUs, each with 10 cores supporting 20 threads, two NVIDIA RTX 3090 GPUs and 62GB RAM. Notably, all experiments are executed solely on one of the GPUs for consistent results. The model parameter size and average training time are shown in Figure 3 ###reference_###.\n###figure_3### ###table_3### Hyperparameters for the models are determined using a grid-search strategy on the validation set for each task. Hyperparameter settings can be found in Appendix B, and the grid search results are available on our online platform. All deep learning models are trained on the training set for 100 epochs, using an early stopping strategy. During each epoch, we assess the model\u2019s performance on the validation set of each fold. After training, we load the model parameters that yield the best performance on the validation set and then evaluate the model on the test set. For the outcome prediction task, we select models based on the AUPRC score. For the length-of-stay prediction task, we use the MAE score as the criteria. All metrics are computed on a per-record basis." + }, + { + "section_id": "3.3", + "parent_section_id": "3", + "section_name": "Model adjustments for the proposed tasks", + "text": "This research additionally introduces model modifications tailored to the intended tasks in terms of both the loss function and the model architecture.\nOutcome-specific length-of-stay prediction: The task of outcome-specific length-of-stay prediction is framed as a multi-target prediction task, for which we employ two distinct strategies. The first is end-to-end multi-task learning, whereby a singular model backbone and two MLP prediction heads are trained - one for outcome prediction and the other for length-of-stay prediction. The backbone and prediction heads are co-trained in an end-to-end fashion. The second approach is the two-stage training, in which two separate models, each with identical structures, are trained to predict the outcome and length-of-stay independently. Depictions of these two training scenarios can be found in Figure 3 ###reference_###. We include performance metrics for both approaches. It is important to note that traditional machine learning models only accommodate two-stage setting, hence we only present the performance metrics of the two-stage training approach for such models.\nEarly mortality prediction: Given the nature of the early mortality prediction task, we devised a straightforward yet potent loss function to guide deep learning models about the decision time threshold. The suggested time-aware loss, is formulated as follows:\nIn this formula, is the cross-entropy loss, is the time to outcome at the -th timestep, denotes the decay rate, denotes the reward factor, and denotes the reward term. The second part confers rewards on earlier correct predictions and penalizes late incorrect predictions by either decreasing or increasing the loss values. The first part of the proposed loss function works to prevent early incorrect predictions from incurring excessive penalties through the application of an exponential decay function. This proposed time-aware loss function can be readily integrated into any deep learning model, thereby improving early mortality prediction performance. It should be noted that as the loss term is only employed during the training phase, the term is not required for the inference stage. Consequently, this ensures that there is no future information leakage at the inference stage.\n###figure_4###" + }, + { + "section_id": "4", + "parent_section_id": null, + "section_name": "Results", + "text": "In this section, we present comprehensive benchmarking results of all predictive models across all datasets and tasks. To assess the statistical significance of our model\u2019s performance, we performed t-tests on all results. The bootstrap T-test [38 ###reference_38###] was employed to calculate p-values. We set the sample number for the bootstrap at 1000 and the number of sampling processes at 50. The comparisons were made between different variations of the model \u2014 for instance, comparing the model with and without Time-Aware (TA) loss in the early mortality prediction task, and between two-stage and multi-task settings in the LOS prediction task. The p-values are reported in Appendix G." + }, + { + "section_id": "4.1", + "parent_section_id": "4", + "section_name": "Benchmarking performance of outcome-specific length-of-stay prediction", + "text": "The benchmarking performances for the task of outcome-specific length-of-stay prediction, under both multi-task and two-stage settings, are presented in Table 4 ###reference_###. On the TJH dataset, the GRU model with two-stage learning achieves the lowest MAE and MSE. Dr.Agent with multi-task learning demonstrates the lowest OSMAE. On the CDSL dataset, StageNet with two-stage learning achieves the lowest MAE and MSE, while the TCN model with multi-task learning achieves the lowest OSMAE.\n. The reported score is in the form of . Subscript signifies a multi-task learning strategy, while subscript indicates a two-stage learning strategy. Bold denotes the best performance. Underline indicates that the multi-task setting outperforms the two-stage learning strategy. The asterisk * denotes that the performance improvement against the two-stage model is statistically significant (p-value < 0.05).\n\n\n\n\nDataset\nTJH\nCDSL\n\nMetric\nMAE ()\nMSE ()\nOSMAE ()\nMAE ()\nMSE ()\nOSMAE ()\n\n\n4.83 0.53\n40.94 8.77\n6.14 0.99\n4.05 0.13\n31.30 4.42\n4.90 0.16\n\n\n5.07 0.53\n50.16 12.06\n7.02 1.33\n4.11 0.15\n32.23 4.81\n4.99 0.14\n\n\n4.79 0.53\n42.03 9.58\n5.92 0.82\n4.01 0.14\n30.74 4.55\n4.82 0.19\n\n\n4.71 0.53\n39.72 9.23\n5.70 0.85\n4.01 0.14\n30.62 4.61\n4.79 0.17\n\n\n4.78 0.55\n41.76 9.47\n5.77 0.92\n4.02 0.13\n31.10 4.42\n4.81 0.15\n\n\n5.08 0.48\n47.21 11.10\n6.34 0.97\n4.05 0.15\n32.08 4.96\n4.80 0.17\n\n\n4.90 0.52*\n44.35 14.00*\n7.26 2.81\n4.08 0.16\n31.54 4.53*\n4.86 0.21\n\n\n4.57 0.48\n39.82 10.02\n5.54 1.30\n3.90 0.22\n30.66 5.58\n4.37 0.26\n\n\n4.68 0.49\n40.21 9.29\n5.46 1.43*\n3.87 0.14*\n29.52 4.98*\n4.43 0.18\n\n\n4.49 0.63\n37.38 12.64\n4.88 1.30\n3.81 0.16\n29.29 4.88\n4.34 0.23\n\n\n4.46 0.60*\n39.24 10.30\n5.53 1.86\n3.85 0.13\n30.13 4.22\n4.40 0.19\n\n\n4.33 0.49\n35.10 9.61\n5.16 1.31\n3.75 0.18\n28.90 4.68\n4.34 0.32\n\n\n4.80 0.48\n41.78 9.49\n5.38 1.09\n3.98 0.18\n32.14 4.31\n4.55 0.22\n\n\n4.69 0.60\n43.29 13.27\n5.77 0.92\n3.75 0.17\n29.48 4.78\n4.31 0.15\n\n\n4.79 0.55\n41.56 9.73*\n5.52 0.85*\n3.80 0.16\n29.45 4.47*\n4.22 0.18*\n\n\n5.04 0.43\n43.88 7.22\n5.96 1.62\n3.98 0.20\n32.35 5.61\n5.19 0.19\n\n\n5.06 0.46\n46.02 11.63\n6.61 1.57\n4.00 0.20\n32.02 4.68*\n5.13 0.19*\n\n\n4.64 0.61\n41.22 14.45\n4.77 1.18\n3.83 0.16\n30.56 5.99\n4.41 0.26\n\n\n4.62 0.55*\n40.06 10.34*\n5.03 0.96\n3.84 0.20\n29.92 4.48*\n4.41 0.30\n\n\n4.60 0.76\n41.70 14.21\n5.59 1.33\n3.72 0.15\n28.81 4.57\n4.33 0.22\n\n\n4.49 0.42*\n39.55 8.52*\n7.10 1.74\n3.78 0.18\n29.93 4.50\n4.28 0.19*\n\n\n4.61 0.58\n40.85 12.12\n4.93 1.14\n3.80 0.18\n29.70 4.96\n4.46 0.27\n\n\n4.41 0.58*\n37.55 11.38*\n4.75 1.11*\n3.82 0.19\n29.45 4.70*\n4.49 0.28\n\n\n4.55 0.61\n39.67 11.35\n5.08 0.62\n3.80 0.15\n29.07 4.57\n4.42 0.15\n\n\n4.41 0.63*\n38.86 10.75*\n5.24 0.98\n3.81 0.16\n32.05 4.27\n4.43 0.18\n\n\n4.57 0.43\n40.65 11.46\n5.51 1.13\n3.84 0.16\n29.95 5.06\n4.43 0.25\n\n\n4.44 0.50*\n37.79 10.27*\n5.31 1.31*\n3.89 0.12\n30.01 4.33\n4.49 0.28\n\n\n4.69 0.52\n40.26 11.13\n5.29 1.31\n3.81 0.17\n30.16 5.03\n4.29 0.23\n\n\n4.67 0.56*\n40.42 10.77\n5.27 1.35*\n3.85 0.11\n29.15 3.90*\n4.46 0.24\nWe observe that the multi-task strategy generally outperforms the two-stage strategy on the TJH dataset. On the TJH dataset, we find that EHR-specific models can better benefit from multi-task learning settings, suggesting that the multi-task setting may better facilitate these models\u2019 ability to map patients\u2019 diverse statuses in the latent space. We visualize the learned embeddings in the latent space using t-SNE in Appendix F. Additionally, it\u2019s important to note the discrepancy between OSMAE and MAE. Despite both metrics assessing absolute error, the OSMAE values are notably larger than the MAE values on both datasets. On average, the OSMAE is 22% higher than the MAE on the TJH dataset and 17% higher on the CDSL dataset. This suggests that MAE might potentially distort the evaluation of model performance, as it cannot measure the error for false negative and false positive predictions, despite its widespread use in previous LOS prediction studies. We also conduct the error analysis for the MAE in Appendix H. In our current setting, we utilize two identical model structures for two-stage training. However, future work could potentially involve the use of distinct models for the two tasks. Further research could also focus on designing a more sophisticated multi-task architecture or incorporating the principles of transfer learning." + }, + { + "section_id": "4.2", + "parent_section_id": "4", + "section_name": "Benchmarking performance of early mortality prediction", + "text": "The comparative performances of the models in early mortality prediction are presented in Table 5 ###reference_###. We also introduced the time-aware loss, labeling these adjusted models with a \u2018-TA\u2019 suffix for all deep learning models. For the TJH dataset, the AdaCare-TA model delivered the highest AUPRC and AUROC, while Dr. Agent provided the highest ES. In contrast, for the CDSL dataset, ConCare-TA achieved the highest AUPRC, while StageNet-TA produced the highest AUROC and ES. Among traditional machine learning and basic deep learning models, CatBoost and GRU showed better performance on both datasets.\nThe results indicate that the proposed time-aware loss term can enhance the performance of most models, particularly in terms of the early prediction score. The average improvement of the ES is 1.93% for the TJH dataset and 33.58% for the CDSL dataset. This underscores that this simplistic loss term can effectively aid models in making correct early decisions. It\u2019s worth noting that our goal is to evaluate the direct effectiveness of the proposed time-aware loss term, so we did not specifically tune the hyper-parameters of all -TA models. So even under this adverse condition, most -TA models can still outperform the original version. Additionally, we observe that for the TJH dataset, the performance of all models was higher, especially for the ES, suggesting that the task is more straightforward on this dataset. Patients\u2019 initial status was also more severe in the TJH dataset, resulting in a much higher ES than in the CDSL dataset. This might be attributable to the fact that the TJH data was collected during the initial months of the pandemic when the virus was significantly more lethal. The performance of EHR-specific models generally surpasses that of basic deep learning models, which in turn outperform traditional machine learning models. This is to be expected, as most EHR-specific models are better equipped to utilize characteristics present in EHR data, such as disease progression, while compared with deep learning models, traditional machine learning models are unable to leverage temporal relationships in sequential data.\n###table_4###" + }, + { + "section_id": "5", + "parent_section_id": null, + "section_name": "Discussions", + "text": "In this section, we further explore models\u2019 performances in terms of the proposed metrics ES and OSMAE. These metrics show models\u2019 performances on the early and outcome-specifc perspectives of two tasks. We also discuss the limitations and future directions of this work." + }, + { + "section_id": "5.1", + "parent_section_id": "5", + "section_name": "Analysis of early prediction performance", + "text": "By enhancing early prediction performance, models can discern patients\u2019 health risks in the early stages. This can address the \u2018cold start\u2019 issue of predictive models, which is particularly beneficial during the early phases of a pandemic when predictions must rely on sparse data. To assess the efficacy of our proposed early prediction loss term in improving early-stage predictions, we simulated a scenario where the model generates forecasts using only the first half of a patient\u2019s visit records. The ES scores for the top five best-performing models (both with and without the time-aware loss) on the CDSL dataset are illustrated in Figure 4 ###reference_###. The findings demonstrate that models incorporating the time-aware loss consistently achieve superior ES scores in the early stages, which directly assesses the true positive rate of predictions. This underscores the adaptability of our proposed time-aware loss in enhancing various models\u2019 early prediction capabilities.\nTo offer a more granular view of the performance gains attributed to the time-aware loss, we\u2019ve charted the AUROC at different visits (days to outcome) for the Dr. Agent model on the CDSL dataset as a case study in Figure 5 ###reference_###. The figure shows that most patients\u2019 LOS are less than 30 days, and the AUROC of the model with the time-aware loss is significantly higher. This suggests that, for most patients, the time-aware loss is more effective in identifying early health risks.\n###figure_5### ###figure_6###" + }, + { + "section_id": "5.2", + "parent_section_id": "5", + "section_name": "Case Study of Prediction Metrics", + "text": "The proposed ES and OSMAE metrics are inherently interpretable as they are based solely on straightforward score calculations, independent of the models. This makes the calculation process transparent and easy to understand. To further elucidate our metrics, we have included a case study here, illustrated in Figure 6 ###reference_###.\nIn the left figure, we plot the predicted risk probability over time alongside the ES score at each timestep for a patient whose actual outcome was mortality. Initially, the model accurately identifies early risk by predicting a risk probability above 0.5, earning a full ES score (1.0) for these early correct predictions. However, over time, as the model incorrectly assesses the patient\u2019s condition as improving, it receives progressively lower ES scores. This decline reflects the model\u2019s erroneous predictions as the patient approaches the final outcome. Near the end, even though the model predicts high risks again, the rewards for these correct predictions are lower due to their proximity to the final outcome.\nIn the right figure, we show the predicted Length of Stay (LOS), the actual LOS, the MAE, and the OSMAE for the same patient. At first glance, the predicted LOS (dotted blue line) closely matches the actual LOS (solid blue line), resulting in a consistently low MAE (dotted green line). However, the higher OSMAE (red line) indicates that the model is making incorrect outcome predictions, suggesting a recovery when the patient\u2019s condition is actually deteriorating. When outcomes are predicted accurately, OSMAE aligns with MAE. Using traditional MAE as a sole metric could lead to biased decisions, as it may not reflect the patient\u2019s actual health trajectory.\nThis case study demonstrates that our proposed ES and OSMAE metrics can more accurately represent a model\u2019s ability to assess early patient risk and understand disease progression, aspects often overlooked in traditional classification and regression metrics.\n###figure_7### ###figure_8###" + }, + { + "section_id": "5.3", + "parent_section_id": "5", + "section_name": "Outcome-specific prediction performance", + "text": "Another primary objective of our study is to determine how the proposed outcome-specific LOS prediction task can enhance clinical decision-making. In Table 6 ###reference_###, we display the prediction discrepancies for the top 5 models with the highest MAE (OSMAE-MAE). To demonstrate how MAE can retain information about the correctness of outcome predictions while computing the MAE, we also include the number and percentage of patients with incorrect outcome predictions.\n###table_5### The results indicate that as MAE increases, the percentage of incorrect outcome predictions also rises. This suggests that an increase in outcome prediction error contributes to the discrepancy between our new OSMAE metric and the traditional MAE. This conclusion further implies that in traditional Length of Stay (LOS) prediction tasks, relying solely on MAE to assess prediction accuracy can lead to erroneous conclusions and potentially compromise clinical decision-making. Consequently, refining the OSMAE metric\u2014rather than solely concentrating on the conventional MAE\u2014is crucial for future predictive modeling endeavors, possibly through a more judicious multi-task learning framework or regularizer." + }, + { + "section_id": "5.4", + "parent_section_id": "5", + "section_name": "Sensitivity analysis of OSMAE and ES", + "text": "We investigate the influence of the threshold parameter, , on the OSMAE and the ES. The OSMAE and ES values for various thresholds, using the same GRU multi-task model on the CDSL dataset, are plotted in Figure 7 ###reference_###. Evidently, as increases, OSMAE increases and ES decreases. This occurs because a larger necessitates the model to accurately predict the target earlier, thereby increasing the task\u2019s difficulty. In practical clinical settings, clinicians can choose an appropriate threshold based on their specific requirements. Another advantage of this setting is making both metrics invariant to the record length, which means patients with too many visits bring less bias to the proposed metrics compared to traditional AUROC and AUPRC. This is because both metrics are normalized by the visit time, ensuring that the length of a patient\u2019s record does not disproportionately influence the metric values, regardless of the duration of the patient record. No matter how long the patient record is, the visits before the penalty time are regularized to have less effect on the metric values.\n###figure_9### ###figure_10### We also aim to explore the effect of penalty score value in ES for FN and FP cases. We selected -0.1 as the penalty score for false positive (FP) cases to mirror scenarios in real-world clinical settings, where the costs associated with false negative predictions are typically higher than those for false positives, particularly in intensive care environments [39 ###reference_39###, 40 ###reference_40###]. Thus, a lower penalty score is assigned to false positive predictions. However, clinicians may adjust this penalty score to suit their specific needs in various applications. For example, in a less critical clinical setting, clinicians may set higher penalty scores to select the model that produces fewer false alarms. Since the Effectiveness Score (ES) solely functions as a metric, it does not influence model performance. Figure 8 ###reference_### illustrates how the penalty score impacts the ES value. As demonstrated in the figure, while the penalty score influences the absolute value of ES, it does not alter the underlying trend. True negative cases receive a score of zero, aligning with the ES metric\u2019s aim to evaluate models for early risk detection in patients. Since true negative predictions do not trigger alarms, the model is neither rewarded nor penalized for these outcomes. It is important to note that the number of true positive predictions is 649, compared to 10,381 true negative predictions. Awarding points for true negative predictions would introduce significant bias into the final score.\n###figure_11### ###figure_12###" + }, + { + "section_id": "5.5", + "parent_section_id": "5", + "section_name": "Model interpretability, limitations and future works", + "text": "Interpretability is an important aspect of models, especially in clinical settings. With model interpretability, the clinicians can understand the model\u2019s prediction process and make reliable clinical decisions. Many models benchmarked in this work claim they have different levels of interpretability. We group them into four groups: 1. visit-level; 2. feature-level; 3. others. Visit-level interpretability indicates that the model can identify the most important visit from longitudinal EHR sequences and feature-level interpretability means that the model can identify the most important risk factors from a single visit data. Some models can provide other information to aid clinical decisions such as disease progression score. We summarize the characteristics of all interpretable models in Table 7 ###reference_###.\n###table_6### We find most machine learning models are interpretable as they are tree-based models. These models can only provide feature-level interpretability (i.e. feature importance score) because they cannot process temporal dependencies. All naive deep learning models are black-box models and they are not interpretable. However, all EHR-specific deep learning models studied in this work are designed to provide some interpretability. RETAIN can provide both visit-level and feature-level interpretabilities. StageNet, GRASP and ConCare can provide additional information including disease progression score, patient clustering information and feature correlations.\nWe aim to establish a comprehensive benchmark for COVID-19 predictive modeling. However, our study has certain limitations. Firstly, within our two-stage approach, the models predicting mortality and Length of Stay (LOS) might have distinct architectures. For comparison purposes, due to the inherent complexity, we only assessed scenarios where both models are identical. Additionally, the multi-task setting may benefit from more intricate multi-task learning structures. Exploring various model combinations and learning structures is a promising avenue for future work. Secondly, during data preprocessing, we imputed missing values using the most recent measurement. While this method has been adopted in numerous studies [15 ###reference_15###, 25 ###reference_25###, 26 ###reference_26###], it is essential to investigate how different imputation techniques influence the outcomes in a clinical context. Lastly, when employing the OSMAE metric, other consequential clinical outcomes, like re-admission, besides mortality and recovery, should be considered. Future studies should account for these competing factors by devising more inclusive tasks and metrics. Our proposed tasks align with clinical practice requirements. Beyond task and metric design, the introduced early prediction loss and multi-task learning framework are innovative approaches yielding encouraging results for the tasks presented. We anticipate further refinements in future research endeavors.\nWe envision that our proposed benchmark analysis pipeline and coding framework can be extended to a wider range of clinical tasks and datasets. In this work, to bolster the comprehensiveness and validity of our evaluation, we incorporated an expansive selection of ICU datasets, namely MIMIC-III [17 ###reference_17###] and MIMIC-IV [41 ###reference_41###]. We undertook the in-hospital mortality prediction task using these two datasets, with results detailed in Appendix E. While our efforts remain exploratory in nature, we aspire for our proposed tasks to be integrated across diverse clinical datasets, thereby enhancing the quality of healthcare delivery. Another line for future research pertains to benchmarking the fairness of model predictions. We have undertaken preliminary experiments using the GRU model with four fairness metrics, with results detailed in Appendix C. Subsequent efforts might focus on a thorough analysis aimed at minimizing bias and enhancing the fairness across various predictive models." + }, + { + "section_id": "6", + "parent_section_id": null, + "section_name": "Conclusions", + "text": "This paper presents a fair, comprehensive, and open-source benchmark for predictive modeling of COVID-19, using Electronic Health Records (EHR) data from ICUs. We introduce two innovative clinical prediction tasks \u2013 early mortality prediction and outcome-specific length-of-stay prediction \u2013 based on the clinical practice of COVID-19 prediction scenarios on two real-world datasets. Our approach ensures fairness and reproducibility through the implementation of meticulous data processing and model training pipelines. We provide benchmarking performances for a total of 17 predictive models, including 5 traditional machine learning models, 6 basic deep learning models, and 6 state-of-the-art EHR-specific deep learning models. To improve accessibility, we have made our benchmarking results and trained models readily available online. This not only facilitates quick access to prediction results using our trained models, but also grants clinicians and researchers the ability to obtain swift prediction results using our trained models. Our endeavors aim to spur continuous development and advancement in the field of deep learning and machine learning, particularly in the context of pandemic predictive modeling." + }, + { + "section_id": "7", + "parent_section_id": null, + "section_name": "Data and Code Availability", + "text": "This research did not involve the collection of new patient EHR data. The TJH EHR dataset [2 ###reference_2###] utilized in this study is publicly available on GitHub (https://github.com/HAIRLAB/Pre_Surv_COVID_19 ###reference_19###). The CDSL dataset [23 ###reference_23###] is open to global researchers and can be accessed on request (https://hmhospitales.com/prensa/notas-de-prensa/comunicado-covid-data-save-lives ###reference_rensa/comunicado-covid-data-save-lives###). To gain access, applicants should fill out the request form. We use these datasets under their respective licenses.\nFurthermore, we have made our benchmarking system available online at https://pyehr.netlify.app ###reference_pyehr.netlify.app###, as illustrated in Figure 9 ###reference_###. All model performances with their associated hyperparameter combinations on both tasks are accessible in the system. These results were used to generate all the performance analysis figures presented in this work. The tables are designed for easy querying, comparison, and sorting. Users can select specific rows and download the CSV files or LaTeX codes. Additionally, all trained models are also released online, enabling clinicians and researchers to conveniently use these models to obtain prediction results with their own test samples. We have also made the source code of this online system publicly available at https://github.com/yhzhu99/pyehr-playground ###reference_d### so that researchers can easily deploy an offline version.\nThe code for this benchmarking work can be accessed on GitHub (https://github.com/yhzhu99/pyehr ###reference_github.com/yhzhu99/pyehr###). The code structure of this benchmarking pipeline is depicted in Figure 10 ###reference_###. The diagram is essentially self-explanatory. All runtime configurations for model parameters are stored in configs/ directory. All deep learning and machine learning models are implemented with various modules. This clearly structured and modular design allows users to effortlessly extend the existing models, design new models and tasks, or even introduce new datasets by adding corresponding components without impacting the downstream benchmark prediction and evaluation process. The analytical files generated in this work are also publicly available online. These data analysis files are housed in the datasets/ folder of the GitHub repository.\n###figure_13### ###figure_14###" + } + ], + "appendix": [ + { + "section_id": "Appendix 1", + "parent_section_id": null, + "section_name": "Appendix A Baseline Models Descriptions", + "text": "Clinical Scoring model\n4C mortality score [27 ###reference_27###] is a risk stratification tool that predict in-hospital mortality or in-hospital clinical deterioration (defined as any requirement of ventilatory support or critical care, or death) for hospitalised COVID-19 patients. They are designed to require only parameters that are commonly available at hospital presentation.\nMachine learning models\nDecision tree (DT) is a non-parametric supervised learning algorithm with a hierarchical tree structure. DT has been widely applied in many clinical predictive tasks, such as mortality prediction for peritoneal dialysis patients[42 ###reference_42###].\nRandom forest (RF) is a supervised ensemble learning method for classification, regression and other tasks. It constructs a multitude of decision trees during the training period. RF has been used to predict the patients\u2019 severity of the COVID-19 case and possible mortality outcome, recovery or death [43 ###reference_43###].\nGradient Boosting Decision Tree (GBDT) is a widely-used algorithm for solving prediction problems in both classification and regression tasks. GBDT takes decision trees as the base learner and determines final prediction results based on a series of DTs\u2019 outputs. GBDT exhibits good performance in performing COVID-19 mortality prediction tasks with structured clinical data [44 ###reference_44###].\nXGBoost [28 ###reference_28###] is a recursive tree-based supervised machine learning classifier. XGBoost has been used to predict the mortality for COVID-19 infected patients[2 ###reference_2###].\nCatBoost [29 ###reference_29###] is also a gradient boost machine learning framework. It builds a symmetric tree and uses an ordered boosting strategy with a faster training and prediction speed compared to XGBoost. Catboost has been used to predict and detect the number of confirmed and death cases of COVID-19 [45 ###reference_45###].\nBasic deep learning models\nMulti-layer perceptron (MLP) is the feed-forward-based neural network. MLP has been used to predict acute kidney injury[46 ###reference_46###].\nRecurrent neural network (RNN) [30 ###reference_30###] is the most popular framework to learn the abstract embedding of variable-length time series.\nRNN has been widely used to predict the risk of the first episode of psychosis or heart failure[47 ###reference_47###, 48 ###reference_48###].\nLong-short term memory network (LSTM) [31 ###reference_31###] is a variant of the Recurrent Neural Network (RNN), capable of learning long-term dependencies. LSTM has been used to perform the 90-day all-cause mortality in the intensive care unit (ICU), based on the concatenated static features and dynamic features[49 ###reference_49###].\nGated recurrent units (GRU) [32 ###reference_32###] embeds the time series as the input to perform the target prediction. It is a widely applied variant of the Recurrent Neural Network (RNN), which improves the capability to maintain historical memories and reduces parameters in the update and reset gates. GRU has been used to predict several severe complications (mortality, renal failure with a need for renal replacement therapy, and postoperative bleeding leading to operative revision) in post-cardiosurgical care in real-time[50 ###reference_50###].\nTemporal convolutional networks (TCN) [33 ###reference_33###] is a generic temporal convolutional network architecture for sequence modeling.\nTCN has been used to forecast hospital resource utilization (i.e., the number of hospital beds and ventilators) for COVID-19 patients [51 ###reference_51###].\nTransformer [34 ###reference_34###] in this experiment is the encoder part of the original Transformer, which comprises the positional encoding module and the self-attention module.\nThe transformer has been used to perform the mortality risk analysis for liver transplant recipients [52 ###reference_52###].\nEHR-specific predictive models\nRETAIN [35 ###reference_35###] is the deep-based Reverse Time Attention model for analyzing EHR data. It utilizes a two-level neural attention module to attend important clinical visits and features. RETAIN has been used to predict heart failure by taking previous diagnoses (categorical variables) as features. In our benchmark, we modify the input layer of RETAIN with multi-layer perceptron (MLP) to make it capable of using numerical variables as features.\nStageNet [36 ###reference_36###] is the Stage-aware neural Network, which extracts disease stage information from EHR and integrates it into risk prediction.\nThis model comprises a stage-aware LSTM module that extracts health stage variations unsupervisedly and a stage-adaptive convolutional module that incorporates stage-related progression patterns.\nStageNet has been used to perform the decompensation prediction for ICU patients in the MIMIC-III dataset and the mortality prediction of End-Stage Renal Disease (ESRD) patients in Peking University Third Hospital.\nDr. Agent [22 ###reference_22###] augments RNN with 2 policy gradient agents.\nIt learns a dynamic skip connection to focus on the relevant information over time.\nDr. Agent has been used to perform the in-hospital mortality prediction, acute care phenotype classification, physiologic decompensation prediction and length of stay forecasting task on the MIMIC-III (Medical Information Mart for Intensive Care) dataset.\nAdaCare [25 ###reference_25###] employs a multi-scale adaptive dilated convolutional module to capture the long and short-term variations of biomarkers to depict the health status in multiple time scales.\nAdaCare has been used to perform decompensation prediction on the MIMIC-III dataset and the mortality prediction on the End-Stage Renal Disease (ESRD) dataset.\nConCare [26 ###reference_26###] employs multi-channel embedding architecture and self-attention mechanism to model the relationship of feature sequences and build the health representation.\nConCare has been used to perform the morality prediction on the MIMIC-III and ESRD datasets.\nGRASP [37 ###reference_37###] is a generic framework for healthcare models, which leverages the information extracted from patients with similar conditions to enhance the cohort representation learning results.\nGRASP has been used to perform the sepsis prediction on the cardiology dataset and mortality prediction on the ESRD dataset." + }, + { + "section_id": "Appendix 2", + "parent_section_id": null, + "section_name": "Appendix B Experiment Environments and Model Hyperparameters", + "text": "For DT, RF, GBDT models, we use the scikit-learn package (version 1.2.2), XGBoost (version 1.7.5), CatBoost (version 1.2) to construct and train the models. For deep learning models, we use the PyTorch (version 2.0.1) and PyTorch Lighting (version 2.0.2) to implement the models. We use the mini-batch gradient descent to train the models and the batch size is set to 64 for TJH dataset and 128 for CDSL dataset. The and are set to . For the metric calculations, is the maximum value at the 95% percentile of the length-of-stay of all patient in both CDSL and TJH datasets. In the hold-out set, is 27.25 in the CDSL dataset and 21.44 in the TJH dataset. We use AdamW optimizer with tuned learning rate to train the models over 100 epochs. The model training process will be stopped if the score on the validation set do not improve over 10 consecutive iterations. We set the random seed to 0 in all computational runs.\nExperiments are conducted on a server equipped with dual Intel Xeon Silver 4210R CPUs, each with 10 cores supporting 20 threads, two NVIDIA RTX 3090 GPUs and 64GB RAM. Notably, all experiments are executed solely on one of the GPUs for consistent results.\nThe model hyperparameters are shown in Table 8 ###reference_###, Table 9 ###reference_###, and 10 ###reference_###. To perform the 4C calculations, we followed the guidelines and methodology provided at https://isaric4c.net/risk/4c/ ###reference_isaric4c.net/risk/4c/###. Also, we have incorporated additional details about the 4C mortality score in the main text. The computation of the 4C mortality score is a straightforward process. Each patient in our study is evaluated based on a set of variables \u2014 four variables in the TJH dataset and five in the CDSL dataset. For each variable, a specific score is assigned, reflecting the patient\u2019s status in that variable. The total 4C mortality score for a patient is the sum of these individual scores. Each level of the accumulated score correlates with a distinct probability of mortality. We have presented the variables used for calculating the 4C mortality scores in both datasets in Table 11 ###reference_###.\n###table_7###" + }, + { + "section_id": "Appendix 3", + "parent_section_id": null, + "section_name": "Appendix C Evaluation of Prediction Fairness", + "text": "To evaluate the models\u2019 fairness, we test the GRU model using fairness experiments. The fairness metrics include disparate impact (DI) [53 ###reference_53###], average odds difference (AOD) [54 ###reference_54###], equal opportunity difference (EOD) [54 ###reference_54###] and statistical parity difference (SPD) [55 ###reference_55###]. The results are shown in Table 12 ###reference_###.\nDisparate Impact (DI) [53 ###reference_53###]: DI measures the ratio of the positive outcome rates between the unprivileged and privileged groups. The ideal value is 1.0, indicating equal positive outcome rates for both groups. A value less than 1 indicates that the unprivileged group is less likely to receive the positive outcome than the privileged group. This metric is given by:\nWhere: is the outcome variable (e.g., dead or alive) and represents the group (privileged or unprivileged). For the purpose of our study, the privileged groups are identified as male and younger patients.\nAverage Odds Difference (AOD) [54 ###reference_54###]: AOD measures the average difference of FPR and TPR for the unprivileged and privileged groups. The ideal value is 0, indicating both groups have equal FPR and TPR. The formula for AOD is:\nEqual Opportunity Difference (EOD) [54 ###reference_54###]: EOD is similar to AOD but only considers the TPR. The ideal value is 0, indicating both groups have equal TPR. EOD is computed as:\nStatistical Parity Difference (SPD) [55 ###reference_55###]: SPD measures the difference in the probability of positive decisions between the unprivileged and privileged groups. The ideal value is 0, indicating both groups have equal probabilities of receiving a positive decision. SPD can be calculated using:\n###table_8### The results in Table 12 ###reference_### show that most model predictions are more fair on the CDSL dataset than the TJH dataset, especially in terms of sex. This may be because the population size is limited in the TJH dataset and thus has induced population bias. Future work may include conducting a more comprehensive analysis to improve the bias and fairness of different predicting models." + }, + { + "section_id": "Appendix 4", + "parent_section_id": null, + "section_name": "Appendix D Benchmarking Experiment under Hold-Out Data Split Settings", + "text": "We employ a 10-fold cross-validation strategy to mitigate the bias associated with small sample sizes. Randomly splitting the test set can lead to unstable model performance due to its limited size. Cross-validation allows the model to be tested on all samples, thereby yielding more consistent performance conclusions. However, to evaluate the model\u2019s generalizability over time more comprehensively, we conduct a standard holdout experiment. We divide the dataset into training, validation, and test sets in a 7:1:2 ratio, based on admission times (using the latest 20% of patients as the holdout dataset, as recommended). We retrain all models five times with different random seeds and report the standard deviations. The model performances are detailed in Table 13 ###reference_### and Table 14 ###reference_###.\n###table_9### ###table_10### For outcome-specific LOS prediction on the TJH dataset, AdaCare achieves the lowest MAE and MSE. On the CDSL dataset, TCN, AdaCare, and RNN exhibit the lowest MAE, MSE, and OSMAE. In the early mortality prediction task, the results align closely with those from the cross-validation setting. AdaCare and ConCare demonstrate superior performance in AUROC, AUPRC, and ES on the TJH dataset. Meanwhile, ConCare, StageNet, and GBDT achieve the best performance on the CDSL dataset. We observe significant variability in some model performances, particularly in MSE on the TJH dataset, attributable mainly to its smaller size and consequent less stable conclusions compared to the cross-validation setting. Furthermore, we note that overall model performance on the hold-out dataset is inferior to that in the cross-validation setting, suggesting a shift in dataset distribution over time and a corresponding decline in the models\u2019 predictive capabilities. Overall, AdaCare, Dr. Agent and StageNet are top-performing models for both tasks on both datasets. For a light-weight choice, GBDT outperformed other machine learning models and could be considered as the choice for machine learning models." + }, + { + "section_id": "Appendix 5", + "parent_section_id": null, + "section_name": "Appendix E Benchmarking Experiment on MIMIC Datasets", + "text": "To enhance the comprehensiveness and validity of the evaluation, we include a broader array of ICU datasets (MIMIC-III [17 ###reference_17###] and MIMIC-IV [41 ###reference_41###]). We conduct the in-hospital mortality prediction task on the two datasets. The label definition and data preprocessing are following previous benchmark works [15 ###reference_15###], but we provide a more comprehensive baseline comparison. The results are shown in Table 15 ###reference_###.\nWe find that Dr. Agent and StageNet achieve better performance on the MIMIC-III dataset, while Dr. Agent and GRU achieve better performance on the MIMIC-IV dataset. This conclusion is similar to the CDSL and TJH dataset, which proves the generalizability of these models. Since the LOS information in the two MIMIC datasets has high variance, we do not report the OSMAE on these datasets. In future works, these metrics can be applied to more specific cohorts that have lower LOS variance.\nSimilar to the TJH and CDSL datasets, we also explore the early prediction setting on the MIMIC-III and MIMIC-IV datasets. As shown in Figure 11 ###reference_###, the time-aware loss can effectively improve various models\u2019 prediction performance on more general datasets. The MIMIC dataset statistics used in our experiments are shown in Table 16 ###reference_### and 17 ###reference_###.\n###table_11### ###figure_15### ###figure_16### ###table_12### ###table_13###" + }, + { + "section_id": "Appendix 6", + "parent_section_id": null, + "section_name": "Appendix F Model Embedding Visualization for Two-Stage and Multi-task Settings", + "text": "To more effectively illustrate the learned embeddings from both model types, we employ t-SNE for visualizing the hidden states of patient embeddings. As depicted in Figure 12 ###reference_###, the left figure presents the embeddings of the TCN model under a two-stage setting, while the right figure displays the embeddings from a multi-task setting. The embeddings derived under the multi-task setting appear more clustered, suggesting that the model is capable of learning more consistent embeddings in the latent space. This clustering is likely due to the model being tasked with optimizing for two objectives simultaneously.\n###figure_17### ###figure_18###" + }, + { + "section_id": "Appendix 7", + "parent_section_id": null, + "section_name": "Appendix G P-Values of T-Tests", + "text": "To assess the statistical significance of our model\u2019s performance, we performed t-tests on all results. The bootstrap T-test [38 ###reference_38###] was employed to calculate p-values. We set the sample number for the bootstrap at 1000 and the number of sampling processes at 50, including all results from the 10-fold prediction process. The comparisons were made between different variations of the model \u2014 for instance, comparing the model with and without Time-Aware (TA) loss in the early mortality prediction task, and between two-stage and multi-task settings in the LOS prediction task. We report the p-value of AUPRC in the early mortality prediction task and MAE in the LOS prediction task in Table 18 ###reference_### and 19 ###reference_###.\n###table_14### ###table_15###" + }, + { + "section_id": "Appendix 8", + "parent_section_id": null, + "section_name": "Appendix H Error Analysis", + "text": "To more effectively analyze the sources of MAE, we plotted the MAE distributions for groups of patients who were alive and deceased. Additionally, we segmented each patient\u2019s record into two halves: the initial half and the latter half. As depicted in Figure 13 ###reference_###, we observe that the MAE distributions for the first halves of the records are similar across both patient groups, with many records of MAE greater than 8. This is primarily because, in the initial stages of their hospital stay, patients\u2019 health statuses are often unstable, making it challenging for the model to accurately predict the remaining LOS. Conversely, during the latter half of the patients\u2019 stay, their health status tends to be more defined, resulting in a left-skewed MAE distribution for both groups. Notably, the MAE for deceased patients is predominantly less than 3, reflecting the more apparent nature of their health status and easier classification. In contrast, predicting LOS for alive patients is more complex, as their discharge from the ICU can be influenced by various external factors. Hence, their MAE values are larger (mostly clustered between 3 and 4).\nThese findings suggest that while the average MAE for the total cohort may be as large as the true LOS of patients, it can still effectively identify patients in critical condition, as indicated by the generally low MAE values for such cases. On the CDSL dataset, the similarity in MAE values is attributed to the fact that 87% of the patients are alive, with their MAE predominantly ranging between 2 and 4, leading to closely aligned average values.\n###figure_19### ###figure_20### ###figure_21### ###figure_22###" + }, + { + "section_id": "Appendix 9", + "parent_section_id": null, + "section_name": "Appendix I Feature Statistics and Distributions", + "text": "We provide statistics of all features which are used in the modeling process in Table 20 ###reference_### and 21 ###reference_###. The data preprocessing details are shown in Figure 14 ###reference_###. The length of stay distributions are shown in Figure 15 ###reference_###. We plot the distributions of 16 features with the lowest missing rates in two datasets in Figure 16 ###reference_### and Figure 17 ###reference_###.\n###figure_23### ###figure_24### ###figure_25### ###figure_26###" + } + ], + "tables": { + "1": { + "table_html": "
\n
Table 1: Statistics of the TJH dataset. The reported statistics are of the form .
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
Mortality OutcomeTotalAliveDead
# Patients361195 (54.02%)166 (45.98%)
# Records17041050 (61.62%)654 (38.38%)
# Avg. records5.0 [3, 6]5.0 [4, 7]3.0 [2, 5]
Age62.0 [46.0, 70.0]51.0 [37.0, 62.0]69.0 [62.25, 77.0]
Age > Avg. (58)205 (56.79 %)68 (34.87 %)137 (82.53 %)
\nAge Avg. (58)\n156 (43.21 %)127 (65.13 %)29 (17.47 %)
Gender58.7% Male47.2% Male72.3% Male
Male212 (58.73 %)92 (47.18 %)120 (72.29 %)
Female149 (41.27 %)103 (52.82 %)46 (27.71 %)
# Features75
Length of stay10.0 [4.0, 15.0]14.0 [9.0, 17.0]5.0 [3.0, 10.0]
\n
", + "capture": "Table 1: Statistics of the TJH dataset. The reported statistics are of the form ." + }, + "2": { + "table_html": "
\n
Table 2: Statistics of the CDSL dataset. The reported statistics are of the form .
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
Mortality OutcomeTotalAliveDead
# Patients42553715 (87.31%)540 (12.69%)
# Records123044108142 (87.89%)14902 (12.11%)
# Avg. records24.0 [15, 39]25.0 [15, 39]22.5 [11, 37]
Age67.2 [56.0, 80.0]65.1 [54.0, 77.0]81.6 [75.0, 89.0]
Age > Avg. (67)2228 (52.36 %)1748 (47.05 %)480 (88.89 %)
\nAge Avg. (58)\n2027 (47.64 %)1967 (52.95 %)60 (11.11 %)
Gender59.1% Male58.5% Male63.3% Male
Male2515 (59.11 %)2173 (58.49 %)342 (63.33 %)
Female1740 (40.89 %)1542 (41.51 %)198 (36.67 %)
# Features99
Length of stay6.4 [4.0, 11.0]6.1 [4.0, 11.0]6.0 [3.0, 10.0]
\n
", + "capture": "Table 2: Statistics of the CDSL dataset. The reported statistics are of the form ." + }, + "3": { + "table_html": "
\n
Table 3: Models\u2019 parameter size and training time (seconds).
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTJHCDSL
Model# ParameterRuntime (1 epoch)# ParameterRuntime (1 epoch)
RF/1.71 s/26.69 s
DT/1.30 s/8.10 s
GBDT/4.04 s/128.55 s
CatBoost/1.54 s/7.40 s
XGBoost/1.50 s/9.40 s
MLP38.0 K1.91 s39.5 K6.27 s
RNN13.9 K1.91 s17.0 K6.10 s
LSTM41.0 K1.89 s48.6 K6.30 s
GRU31.9 K1.85 s38.1 K6.84 s
TCN55.8 K3.31 s60.4 K31.60 s
Transformer73.0 K2.29 s124 K17.37 s
RETAIN79.0 K2.38 s135 K60.30 s
StageNet603 K2.31 s622 K17.81 s
Dr.Agent41.9 K2.89 s48.1 K15.81 s
AdaCare114 K3.05 s130 K21.38 s
GRASP35.3 K2.41 s39.9 K37.65 s
ConCare66.4 K2.59 s89.1 K11.98 s
\n
", + "capture": "Table 3: Models\u2019 parameter size and training time (seconds)." + }, + "4": { + "table_html": "
\n
Table 4: Benchmarking performance of outcome-specific length-of-stay prediction
\n

. The reported score is in the form of . Subscript signifies a multi-task learning strategy, while subscript indicates a two-stage learning strategy. Bold denotes the best performance. Underline indicates that the multi-task setting outperforms the two-stage learning strategy. The asterisk * denotes that the performance improvement against the two-stage model is statistically significant (p-value < 0.05).\n\n\n\n\nDataset\nTJH\nCDSL\n\nMetric\nMAE ()\nMSE ()\nOSMAE ()\nMAE ()\nMSE ()\nOSMAE ()\n\n\n4.83 0.53\n40.94 8.77\n6.14 0.99\n4.05 0.13\n31.30 4.42\n4.90 0.16\n\n\n5.07 0.53\n50.16 12.06\n7.02 1.33\n4.11 0.15\n32.23 4.81\n4.99 0.14\n\n\n4.79 0.53\n42.03 9.58\n5.92 0.82\n4.01 0.14\n30.74 4.55\n4.82 0.19\n\n\n4.71 0.53\n39.72 9.23\n5.70 0.85\n4.01 0.14\n30.62 4.61\n4.79 0.17\n\n\n4.78 0.55\n41.76 9.47\n5.77 0.92\n4.02 0.13\n31.10 4.42\n4.81 0.15\n\n\n5.08 0.48\n47.21 11.10\n6.34 0.97\n4.05 0.15\n32.08 4.96\n4.80 0.17\n\n\n4.90 0.52*\n44.35 14.00*\n7.26 2.81\n4.08 0.16\n31.54 4.53*\n4.86 0.21\n\n\n4.57 0.48\n39.82 10.02\n5.54 1.30\n3.90 0.22\n30.66 5.58\n4.37 0.26\n\n\n4.68 0.49\n40.21 9.29\n5.46 1.43*\n3.87 0.14*\n29.52 4.98*\n4.43 0.18\n\n\n4.49 0.63\n37.38 12.64\n4.88 1.30\n3.81 0.16\n29.29 4.88\n4.34 0.23\n\n\n4.46 0.60*\n39.24 10.30\n5.53 1.86\n3.85 0.13\n30.13 4.22\n4.40 0.19\n\n\n4.33 0.49\n35.10 9.61\n5.16 1.31\n3.75 0.18\n28.90 4.68\n4.34 0.32\n\n\n4.80 0.48\n41.78 9.49\n5.38 1.09\n3.98 0.18\n32.14 4.31\n4.55 0.22\n\n\n4.69 0.60\n43.29 13.27\n5.77 0.92\n3.75 0.17\n29.48 4.78\n4.31 0.15\n\n\n4.79 0.55\n41.56 9.73*\n5.52 0.85*\n3.80 0.16\n29.45 4.47*\n4.22 0.18*\n\n\n5.04 0.43\n43.88 7.22\n5.96 1.62\n3.98 0.20\n32.35 5.61\n5.19 0.19\n\n\n5.06 0.46\n46.02 11.63\n6.61 1.57\n4.00 0.20\n32.02 4.68*\n5.13 0.19*\n\n\n4.64 0.61\n41.22 14.45\n4.77 1.18\n3.83 0.16\n30.56 5.99\n4.41 0.26\n\n\n4.62 0.55*\n40.06 10.34*\n5.03 0.96\n3.84 0.20\n29.92 4.48*\n4.41 0.30\n\n\n4.60 0.76\n41.70 14.21\n5.59 1.33\n3.72 0.15\n28.81 4.57\n4.33 0.22\n\n\n4.49 0.42*\n39.55 8.52*\n7.10 1.74\n3.78 0.18\n29.93 4.50\n4.28 0.19*\n\n\n4.61 0.58\n40.85 12.12\n4.93 1.14\n3.80 0.18\n29.70 4.96\n4.46 0.27\n\n\n4.41 0.58*\n37.55 11.38*\n4.75 1.11*\n3.82 0.19\n29.45 4.70*\n4.49 0.28\n\n\n4.55 0.61\n39.67 11.35\n5.08 0.62\n3.80 0.15\n29.07 4.57\n4.42 0.15\n\n\n4.41 0.63*\n38.86 10.75*\n5.24 0.98\n3.81 0.16\n32.05 4.27\n4.43 0.18\n\n\n4.57 0.43\n40.65 11.46\n5.51 1.13\n3.84 0.16\n29.95 5.06\n4.43 0.25\n\n\n4.44 0.50*\n37.79 10.27*\n5.31 1.31*\n3.89 0.12\n30.01 4.33\n4.49 0.28\n\n\n4.69 0.52\n40.26 11.13\n5.29 1.31\n3.81 0.17\n30.16 5.03\n4.29 0.23\n\n\n4.67 0.56*\n40.42 10.77\n5.27 1.35*\n3.85 0.11\n29.15 3.90*\n4.46 0.24\n

\n
", + "capture": "Table 4: Benchmarking performance of outcome-specific length-of-stay prediction" + }, + "5": { + "table_html": "
\n
Table 5: Benchmarking performance on the task of early mortality prediction. The reported score is of the form . \u2018TA\u2019 denotes the model trained with the time-aware loss. Bold denotes the best performance. Underline indicates that the model with time-aware loss outperforms the original model. The asterisk * denotes that the performance improvement against the model without TA version is statistically significant (p-value < 0.05). All three metrics are multiplied by 100 for readability purposes.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTJHCDSL
MetricAUPRC()AUROC()ES()AUPRC()AUROC()ES()
4C89.84 4.4694.16 2.57-23.93 2.9976.15 4.06-
RF95.56 2.6996.58 2.2072.63 9.1549.48 3.7984.35 2.66-10.54 3.57
DT80.48 7.2687.41 3.9967.86 11.2438.27 5.2179.67 4.61-8.23 4.17
GBDT95.13 3.5196.41 2.3576.45 7.1350.32 5.1585.15 2.832.25 4.86
CatBoost95.99 2.6197.14 1.8174.91 9.0050.86 4.3485.09 2.86-3.94 3.89
XGBoost95.70 2.9896.84 2.0976.41 9.9049.70 5.0684.59 3.09-3.12 4.44
MLP93.78 2.8095.95 1.6272.87 8.2248.60 3.5784.15 2.72-0.76 3.95
MLP-TA93.21 2.7895.65 1.8573.94 8.9648.67 3.0784.24 2.560.55 3.71
RNN96.03 3.4297.41 2.1978.33 11.3557.57 6.5587.81 2.5519.66 8.86
RNN-TA95.97 3.6597.38 2.29\n79.79 11.06*\n58.21 6.34*\n88.01 2.36*\n20.15 11.25*
LSTM94.82 7.6597.08 3.6584.80 10.6855.53 8.5487.02 3.0119.63 11.41
LSTM-TA\n95.40 5.53*\n97.10 3.05*83.78 8.3056.89 7.3487.76 2.5020.92 13.10
GRU96.33 3.1897.59 2.1378.33 15.4556.78 8.0287.93 2.8921.66 8.86
GRU-TA96.50 3.0497.70 2.06\n80.93 13.84*\n57.75 5.36*\n87.98 2.62*\n23.54 8.10*
TCN93.13 5.1096.39 1.8974.66 13.0357.09 5.8487.40 2.6713.48 12.04
TCN-TA93.31 4.97\n96.74 1.87*76.47 12.2757.71 5.7287.59 2.60\n21.59 12.68*
Transformer93.47 6.7396.86 2.1379.21 13.0538.34 5.0781.32 3.46-3.96 13.23
Transformer-TA93.86 6.4297.01 2.01\n80.73 12.47*40.18 5.00\n82.36 3.61*\n-0.78 11.42*
RETAIN96.49 2.0597.84 1.4482.59 12.2954.54 7.9785.02 3.937.32 10.67
RETAIN-TA95.99 2.4297.82 1.51\n82.85 11.36*54.30 6.5085.43 2.819.00 7.66
StageNet95.71 3.7797.27 2.3877.44 12.5256.57 6.8287.09 3.5423.93 9.26
StageNet-TA\n95.79 3.86*\n97.32 2.41*\n78.88 12.29*\n58.19 6.48*\n88.18 2.82*24.55 9.32
Dr.Agent97.22 2.4498.00 1.7585.80 7.9754.17 8.9286.76 3.6417.53 11.25
Dr.Agent-TA97.16 2.5698.00 1.9786.01 7.6653.06 7.34\n87.06 3.30*\n19.35 9.60*
AdaCare97.86 1.0998.53 0.7778.39 7.5155.32 4.4586.59 1.9917.46 8.66
AdaCare-TA98.11 1.1398.64 0.89\n84.51 8.17*55.62 5.4487.00 2.18\n20.09 9.86*
GRASP96.04 3.1597.30 2.3081.63 9.7153.95 8.3784.44 4.4015.27 15.17
GRASP-TA96.16 3.1797.40 2.2182.82 9.0853.50 7.89\n84.91 3.73*\n15.61 13.76*
ConCare97.01 2.4697.89 1.6280.97 13.6658.59 6.4987.76 3.0217.58 12.34
ConCare-TA97.04 2.5197.94 1.62\n82.31 13.08*58.68 7.0487.96 3.2320.90 10.43
\n
", + "capture": "Table 5: Benchmarking performance on the task of early mortality prediction. The reported score is of the form . \u2018TA\u2019 denotes the model trained with the time-aware loss. Bold denotes the best performance. Underline indicates that the model with time-aware loss outperforms the original model. The asterisk * denotes that the performance improvement against the model without TA version is statistically significant (p-value < 0.05). All three metrics are multiplied by 100 for readability purposes." + }, + "6": { + "table_html": "
\n
Table 6: Prediction discrepancy of 5 models on the CDSL dataset.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
ModelType\nMAE\n\n\n\n\n
Outcome Wrong #
\n
\n\n\n\n\n
Outcome Wrong %
\n
TransformerTwo-stage1.21146911.89%
TransformerMulti-task1.13138911.24%
MLPMulti-task0.78132210.70%
MLPTwo-stage0.75129510.48%
Dr.AgentMulti-task0.6711088.97%
\n
", + "capture": "Table 6: Prediction discrepancy of 5 models on the CDSL dataset." + }, + "7": { + "table_html": "
\n
Table 7: Interpretability characteristics of all interpretable models
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
ModelVisit-levelFeature-levelOthers
4C-\u2713-
RF-\u2713-
DT-\u2713-
GBDT-\u2713-
CatBoost-\u2713-
XGBoost-\u2713-
RETAIN\u2713\u2713-
StageNet--disease progression score
Dr. Agent\u2713--
AdaCare-\u2713-
GRASP--patient clusters
ConCare-\u2713feature correlations
\n
", + "capture": "Table 7: Interpretability characteristics of all interpretable models" + }, + "8": { + "table_html": "
\n
Table 8: Hyperparameter settings of machine learning models. All hyperparameters are obtained using grid-search on the validation set.
\n
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTaskModel\n \n\n\nDetail\n \n
TJHoutcomeRFmax depth 10, n_estimators 100, learning rate 0.1, criterion gini
length of stayRFmax depth 15, n_estimators 50, learning rate 1.0, criterion squared_error
outcomeDTmax depth 10, min_samples_split 2, min_samples_leaf 1
length of stayDTmax depth 5, criterion squared_error, min_samples_split 2
outcomeGBDTmax depth 5, n_estimators 100, learning rate 0.1, subsample 1.0
length of stayGBDTmax depth 5, n_estimators 100, learning rate 0.1, subsample 1.0
outcomeXGBoostmax depth 5, n_estimators 100, learning rate 0.1, objective binary:logistic
length of stayXGBoostmax depth 5, n_estimators 100, learning rate 0.1, objective reg:squarederror
outcomeCatBoostmax depth 5, n_estimators 100, learning rate 0.1, loss CrossEntropy
length of stayCatBoostmax depth 5, n_estimators 100, learning rate 0.1, loss RMSE
CDSLoutcomeRFmax depth 15, n_estimators 100, learning rate 1.0, criterion gini
length of stayRFmax depth 15, n_estimators 100, learning rate 1.0, criterion squared_error
outcomeDTmax depth 5, min_samples_split 2, min_samples_leaf 1
length of stayDTmax depth 5, criterion squared_error, min_samples_split 2
outcomeGBDTmax depth 5, n_estimators 100, learning rate 0.1, subsample 1.0
length of stayGBDTmax depth 5, n_estimators 100, learning rate 0.1, subsample 1.0
outcomeXGBoostmax depth 5, n_estimators 50, learning rate 0.1, objective binary:logistic
length of stayXGBoostmax depth 10, n_estimators 50, learning rate 0.1, objective reg:squarederror
outcomeCatBoostmax depth 5, n_estimators 50, learning rate 0.1, loss CrossEntropy
length of stayCatBoostmax depth 10, n_estimators 100, learning rate 0.1, loss RMSE
\n
\n
", + "capture": "Table 8: Hyperparameter settings of machine learning models. All hyperparameters are obtained using grid-search on the validation set." + }, + "9": { + "table_html": "
\n
Table 9: Hyperparameter settings of deep learning models in the TJH dataset. All hyperparameters are obtained using grid-search on the validation set.
\n
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
TaskModel\n \n\n\nDetail\n \n
outcomeMLPhidden dim 32, learning rate 0.01, feed-forward dim 128
length of stayMLPhidden dim 128, learning rate 0.001, feed-forward dim 512
multi-taskMLPhidden dim 32, learning rate 0.001, feed-forward dim 128
outcomeRNNhidden dim 64, learning rate 0.001, layers 1
length of stayRNNhidden dim 64, learning rate 0.01, layers 1
multi-taskRNNhidden dim 64, learning rate 0.01, layers 1
outcomeLSTMhidden dim 128, learning rate 0.01, layers 1
length of stayLSTMhidden dim 128, learning rate 0.01, layers 1
multi-taskLSTMhidden dim 128, learning rate 0.01, layers 1
outcomeGRUhidden dim 32, learning rate 0.01, layers 1
length of stayGRUhidden dim 128, learning rate 0.01, layers 1
multi-taskGRUhidden dim 32, learning rate 0.0001, layers 1
outcomeTCNhidden dim 64, learning rate 0.0001, kernel size 2, dropout rate 0.2
length of stayTCNhidden dim 64, learning rate 0.01, kernel size 2, dropout rate 0.2
multi-taskTCNhidden dim 128, learning rate 0.01, kernel size 2, dropout rate 0.2
outcomeTransformerhidden dim 64, learning rate 0.0001, layers 1, heads 1
length of stayTransformerhidden dim 32, learning rate 0.001, layers 1, heads 1
multi-taskTransformerhidden dim 32, learning rate 0.001, layers 1, heads 1
outcomeRETAINhidden dim 64, learning rate 0.001, dropout rate 0.1
length of stayRETAINhidden dim 64, learning rate 0.001, dropout rate 0.1
multi-taskRETAINhidden dim 64, learning rate 0.001, dropout rate 0.1
outcomeStageNethidden dim 64, learning rate 0.0001, levels 3, chunk size 64
length of stayStageNethidden dim 128, learning rate 0.01, levels 3, chunk size 128
multi-taskStageNethidden dim 32, learning rate 0.001, levels 3, chunk size 32
outcomeDr. Agenthidden dim 32, learning rate 0.01, actions 10, units 64
length of stayDr. Agenthidden dim 64, learning rate 0.01, actions 10, units 64
multi-taskDr. Agenthidden dim 64, learning rate 0.01, actions 10, units 64
outcomeAdaCarehidden dim 32, learning rate 0.0001, kernel size 2, kernel num 64, RNN type GRU
length of stayAdaCarehidden dim 64, learning rate 0.01, kernel size 2, kernel num 64, RNN type GRU
multi-taskAdaCarehidden dim 64, learning rate 0.01, kernel size 2, kernel num 64, RNN type GRU
outcomeGRASPhidden dim 128, learning rate 0.01, block GRU, clusters 12, dropout rate 0.5
length of stayGRASPhidden dim 64, learning rate 0.01, block GRU, clusters 12, dropout rate 0.5
multi-taskGRASPhidden dim 32, learning rate 0.01, block GRU, clusters 12, dropout rate 0.5
outcomeConCarehidden dim 64, learning rate 0.001, GRU layers 1, attention None
length of stayConCarehidden dim 64, learning rate 0.001, GRU layers 1, attention None
multi-taskConCarehidden dim 64, learning rate 0.001, GRU layers 1, attention None
\n
\n
", + "capture": "Table 9: Hyperparameter settings of deep learning models in the TJH dataset. All hyperparameters are obtained using grid-search on the validation set." + }, + "10": { + "table_html": "
\n
Table 10: Hyperparameter settings of deep learning models in the CDSL dataset. All hyperparameters are obtained using grid-search on the validation set.
\n
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
TaskModel\n \n\n\nDetail\n \n
outcomeMLPhidden dim 32, learning rate 0.001, feed-forward dim 128
length of stayMLPhidden dim 64, learning rate 0.001, feed-forward dim 256
multi-taskMLPhidden dim 128, learning rate 0.001, feed-forward dim 512
outcomeRNNhidden dim 128, learning rate 0.001, layers 1
length of stayRNNhidden dim 128, learning rate 0.01, layers 1
multi-taskRNNhidden dim 32, learning rate 0.01, layers 1
outcomeLSTMhidden dim 64, learning rate 0.001, layers 1
length of stayLSTMhidden dim 32, learning rate 0.0001, layers 1
multi-taskLSTMhidden dim 32, learning rate 0.001, layers 1
outcomeGRUhidden dim 64, learning rate 0.01, layers 1
length of stayGRUhidden dim 64, learning rate 0.01, layers 1
multi-taskGRUhidden dim 32, learning rate 0.01, layers 1
outcomeTCNhidden dim 128, learning rate 0.01, kernel size 2, dropout rate 0.2
length of stayTCNhidden dim 128, learning rate 0.001, kernel size 2, dropout rate 0.2
multi-taskTCNhidden dim 128, learning rate 0.001, kernel size 2, dropout rate 0.2
outcomeTransformerhidden dim 32, learning rate 0.01, layers 1, heads 1
length of stayTransformerhidden dim 128, learning rate 0.01, layers 1, heads 1
multi-taskTransformerhidden dim 128, learning rate 0.01, layers 1, heads 1
outcomeRETAINhidden dim 64, learning rate 0.01, dropout rate 0.1
length of stayRETAINhidden dim 128, learning rate 0.001, dropout rate 0.1
multi-taskRETAINhidden dim 128, learning rate 0.01, dropout rate 0.1
outcomeStageNethidden dim 32, learning rate 0.001, levels 3, chunk size 32
length of stayStageNethidden dim 128, learning rate 0.001, levels 3, chunk size 128
multi-taskStageNethidden dim 32, learning rate 0.0001, levels 3, chunk size 32
outcomeDr. Agenthidden dim 64, learning rate 0.001, actions 10, units 64
length of stayDr. Agenthidden dim 32, learning rate 0.01, actions 10, units 64
multi-taskDr. Agenthidden dim 32, learning rate 0.01, actions 10, units 64
outcomeAdaCarehidden dim 32, learning rate 0.01, kernel size 2, kernel num 64, RNN type GRU
length of stayAdaCarehidden dim 128, learning rate 0.001, kernel size 2, kernel num 64, RNN type GRU
multi-taskAdaCarehidden dim 64, learning rate 0.01, kernel size 2, kernel num 64, RNN type GRU
outcomeGRASPhidden dim 32, learning rate 0.01, block GRU, clusters 12, dropout rate 0.5
length of stayGRASPhidden dim 32, learning rate 0.001, block GRU, clusters 12, dropout rate 0.5
multi-taskGRASPhidden dim 128, learning rate 0.001, block GRU, clusters 12, dropout rate 0.5
outcomeConCarehidden dim 64, learning rate 0.001, GRU layers 1, attention None
length of stayConCarehidden dim 64, learning rate 0.001, GRU layers 1, attention None
multi-taskConCarehidden dim 64, learning rate 0.001, GRU layers 1, attention None
\n
\n
", + "capture": "Table 10: Hyperparameter settings of deep learning models in the CDSL dataset. All hyperparameters are obtained using grid-search on the validation set." + }, + "11": { + "table_html": "
\n
Table 11: Variables employed in the 4C mortality score calculation for both TJH and CDSL datasets.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
TJHCDSL4C Mortality Score Component
SexSexSex at birth
AgeAgeAge
UreaUreaUrea
CRPCRPC reactive protein
-SO2CPeripheral oxygen saturation on room air
\n
", + "capture": "Table 11: Variables employed in the 4C mortality score calculation for both TJH and CDSL datasets." + }, + "12": { + "table_html": "
\n
Table 12: Comprehensive fairness assessment across deep learning models on TJH and CDSL datasets. The table presents a detailed evaluation of fairness metrics, including Disparate Impact (DI), Average Odds Difference (AOD), Equal Opportunity Difference (EOD), and Statistical Parity Difference (SPD), for multiple deep learning models (e.g., MLP) across different privileged groups (Sex, Age).
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTJHCDSL
Metric\nDI ()\n\nAOD ()\n\nEOD ()\n\nSPD ()\n\nDI ()\n\nAOD ()\n\nEOD ()\n\nSPD ()\n
\nMLP\nSex\n0.800.07\n\n0.010.02\n\n0.010.04\n\n0.090.04\n\n0.840.04\n\n0.020.01\n\n0.040.02\n\n0.010.00\n
Age\n0.780.05\n\n0.020.01\n\n0.040.02\n\n0.090.02\n\n0.820.02\n\n0.020.01\n\n0.030.01\n\n0.010.00\n
\nRNN\nSex\n0.810.09\n\n0.010.03\n\n0.000.05\n\n0.090.06\n\n0.880.05\n\n0.020.01\n\n0.040.02\n\n0.010.01\n
Age\n0.760.07\n\n0.030.02\n\n0.050.03\n\n0.100.03\n\n0.800.04\n\n0.030.01\n\n0.060.02\n\n0.020.00\n
\nLSTM\nSex\n0.820.09\n\n0.000.03\n\n0.020.03\n\n0.090.06\n\n0.890.06\n\n0.020.01\n\n0.030.03\n\n0.010.01\n
Age\n0.790.06\n\n0.020.02\n\n0.030.02\n\n0.090.02\n\n0.810.03\n\n0.030.01\n\n0.060.02\n\n0.020.00\n
\nGRU\nSex\n0.810.10\n\n0.010.04\n\n0.010.06\n\n0.100.07\n\n0.890.06\n\n0.010.01\n\n0.020.03\n\n0.010.01\n
Age\n0.760.07\n\n0.030.02\n\n0.050.03\n\n0.100.02\n\n0.820.05\n\n0.030.01\n\n0.060.02\n\n0.020.00\n
\nTCN\nSex\n0.810.09\n\n0.000.03\n\n0.000.05\n\n0.090.06\n\n0.880.07\n\n0.020.01\n\n0.050.03\n\n0.010.01\n
Age\n0.760.06\n\n0.030.01\n\n0.060.03\n\n0.100.02\n\n0.790.03\n\n0.030.01\n\n0.060.02\n\n0.020.00\n
\nTransformer\nSex\n0.830.09\n\n0.000.03\n\n0.020.03\n\n0.090.06\n\n0.860.05\n\n0.010.01\n\n0.010.03\n\n0.000.01\n
Age\n0.800.06\n\n0.010.01\n\n0.020.02\n\n0.090.02\n\n0.790.05\n\n0.020.01\n\n0.040.02\n\n0.010.01\n
\nRETAIN\nSex\n0.790.09\n\n0.010.02\n\n0.000.04\n\n0.100.06\n\n0.850.08\n\n0.020.01\n\n0.050.03\n\n0.010.01\n
Age\n0.760.06\n\n0.020.02\n\n0.050.03\n\n0.100.02\n\n0.770.02\n\n0.030.01\n\n0.060.02\n\n0.010.00\n
\nStageNet\nSex\n0.800.10\n\n0.010.03\n\n0.000.06\n\n0.100.07\n\n0.890.06\n\n0.020.01\n\n0.030.02\n\n0.010.01\n
Age\n0.760.07\n\n0.030.02\n\n0.060.03\n\n0.100.03\n\n0.810.04\n\n0.030.01\n\n0.060.02\n\n0.020.01\n
\nDr. Agent\nSex\n0.800.09\n\n0.010.02\n\n0.010.03\n\n0.100.06\n\n0.860.06\n\n0.020.02\n\n0.030.04\n\n0.010.01\n
Age\n0.790.05\n\n0.010.01\n\n0.030.01\n\n0.090.02\n\n0.820.06\n\n0.030.01\n\n0.060.02\n\n0.020.01\n
\nAdaCare\nSex\n0.790.10\n\n0.010.02\n\n0.010.04\n\n0.100.06\n\n0.890.06\n\n0.020.01\n\n0.040.02\n\n0.010.01\n
Age\n0.770.06\n\n0.020.01\n\n0.030.03\n\n0.090.03\n\n0.830.03\n\n0.030.01\n\n0.050.02\n\n0.010.00\n
\nGRASP\nSex\n0.810.09\n\n0.010.03\n\n0.010.04\n\n0.100.06\n\n0.900.08\n\n0.010.01\n\n0.020.02\n\n0.010.01\n
Age\n0.790.06\n\n0.020.01\n\n0.040.02\n\n0.090.02\n\n0.790.04\n\n0.030.01\n\n0.060.02\n\n0.020.01\n
\nConCare\nSex\n0.790.08\n\n0.010.02\n\n0.000.04\n\n0.100.05\n\n0.880.06\n\n0.030.01\n\n0.050.03\n\n0.010.01\n
Age\n0.770.06\n\n0.020.01\n\n0.040.02\n\n0.100.02\n\n0.780.05\n\n0.030.01\n\n0.060.02\n\n0.020.00\n
\n
", + "capture": "Table 12: Comprehensive fairness assessment across deep learning models on TJH and CDSL datasets. The table presents a detailed evaluation of fairness metrics, including Disparate Impact (DI), Average Odds Difference (AOD), Equal Opportunity Difference (EOD), and Statistical Parity Difference (SPD), for multiple deep learning models (e.g., MLP) across different privileged groups (Sex, Age)." + }, + "13": { + "table_html": "
\n
Table 13: Benchmarking performance of outcome-specific length-of-stay prediction on TJH and CDSL hold-out test sets.. The reported score is in the form of . Subscript signifies a multi-task learning strategy, while subscript indicates a two-stage learning strategy. Bold denotes the best performance. Underline indicates that the multi-task setting outperforms the two-stage learning strategy. The asterisk * denotes that the performance improvement against the two-stage model is statistically significant (p-value < 0.05).
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTJHCDSL
Metric\nMAE()\n\nMSE()\n\nOSMAE()\n\nMAE()\n\nMSE()\n\nOSMAE()\n
\n5.81 0.18\n\n45.89 2.03\n\n8.18 0.19\n\n4.21 0.00\n\n40.47 0.05\n\n4.24 0.01\n
\n5.00 0.06\n\n51.22 1.83\n\n9.34 0.54\n\n4.34 0.00\n\n41.82 0.00\n\n4.43 0.00\n
\n5.98 0.06\n\n50.39 1.40\n7.21 0.16\n4.24 0.00\n\n40.58 0.02\n\n4.27 0.01\n
\n5.51 0.14\n\n41.25 2.12\n\n7.71 0.33\n\n4.22 0.00\n\n40.49 0.08\n\n4.24 0.01\n
\n5.84 0.00\n\n46.86 0.00\n\n7.87 0.00\n\n4.22 0.00\n\n40.57 0.00\n\n4.23 0.00\n
\n5.98 0.27\n\n51.09 5.21\n\n9.77 1.18\n\n4.22 0.02\n\n40.52 0.04\n\n4.20 0.02\n
\n6.35 0.24\n\n71.81 17.97\n\n9.19 0.27*\n\n4.22 0.04\n\n40.14 0.27*\n4.19 0.07
\n5.56 0.11\n\n41.27 1.42\n\n8.73 0.45\n\n4.08 0.03\n\n41.10 1.42\n4.00 0.05
\n7.81 0.33\n\n87.99 9.73\n\n12.14 0.78\n\n4.11 0.06\n\n39.25 0.61*\n\n4.15 0.15\n
\n5.28 0.57\n\n40.00 11.06\n\n8.65 0.60\n\n4.06 0.02\n\n39.71 0.14\n\n4.06 0.04\n
\n6.71 0.48\n\n70.72 9.57\n\n10.94 1.12\n\n4.18 0.07\n\n40.33 0.60\n\n4.25 0.10\n
\n5.87 0.31\n\n47.79 5.05\n\n8.80 0.70\n\n4.08 0.07\n\n40.57 2.20\n\n4.03 0.09\n
\n5.01 0.29*\n\n32.03 3.06*\n\n9.44 1.47\n\n4.17 0.03\n\n40.28 1.63*\n\n4.29 0.09\n
\n6.29 0.84\n\n77.96 34.06\n\n9.53 0.93\n\n4.08 0.06\n\n40.38 0.27\n\n4.12 0.24\n
\n7.50 1.16\n\n162.34 108.99\n\n11.93 1.23\n\n4.04 0.03*\n\n39.39 0.41*\n\n4.03 0.07*\n
\n6.38 0.20\n\n62.65 12.11\n\n11.11 0.91\n\n4.14 0.03\n\n41.23 0.55\n\n4.60 0.25\n
\n6.46 0.19\n\n74.44 26.43\n\n11.27 1.35\n\n4.16 0.06\n\n43.27 1.13\n4.54 0.27
\n5.36 0.33\n\n39.42 5.26\n\n12.60 2.19\n\n4.17 0.08\n\n40.87 0.35\n\n4.24 0.09\n
\n6.30 0.57\n\n115.36 72.40\n\n12.63 1.56\n\n4.11 0.02*\n\n39.41 0.49*\n\n4.35 0.15\n
\n5.49 0.85\n\n47.51 22.86\n\n9.45 0.46\n\n4.08 0.03\n\n40.75 1.04\n\n4.14 0.08\n
\n5.10 0.43*\n\n33.51 6.01*\n\n15.09 7.87\n\n4.09 0.02\n40.40 0.43\n4.16 0.02\n
\n6.36 0.29\n\n57.96 3.57\n\n9.69 0.84\n\n4.09 0.03\n\n41.17 1.03\n\n4.28 0.20\n
\n7.06 0.43\n\n72.02 2.90\n\n10.15 1.17\n\n4.13 0.09\n\n39.83 0.95*\n4.22 0.19
3.58 0.6619.65 5.81\n9.26 1.58\n4.04 0.01\n40.37 0.86\n\n4.10 0.08\n
\n5.13 0.76\n\n34.44 8.48\n\n16.89 8.79\n\n4.12 0.19\n\n43.40 4.03\n\n4.11 0.12\n
\n5.40 0.53\n\n40.75 10.74\n\n8.94 0.74\n\n4.20 0.07\n\n41.00 0.20\n\n4.29 0.10\n
\n5.80 0.90\n\n47.69 18.62\n\n11.91 4.34\n4.18 0.06\n41.03 0.93\n\n4.31 0.09\n
\n4.79 0.44\n\n29.92 5.63\n\n8.15 0.22\n\n4.16 0.02\n\n41.12 0.45\n\n4.08 0.03\n
\n7.62 0.20\n\n78.77 6.50\n\n10.24 0.43\n4.13 0.16\n39.36 0.62*\n\n4.12 0.23\n
\n
", + "capture": "Table 13: Benchmarking performance of outcome-specific length-of-stay prediction on TJH and CDSL hold-out test sets.. The reported score is in the form of . Subscript signifies a multi-task learning strategy, while subscript indicates a two-stage learning strategy. Bold denotes the best performance. Underline indicates that the multi-task setting outperforms the two-stage learning strategy. The asterisk * denotes that the performance improvement against the two-stage model is statistically significant (p-value < 0.05)." + }, + "14": { + "table_html": "
\n
Table 14: Benchmarking performance on the task of early mortality prediction on TJH and CDSL hold-out test sets. The reported score is of the form . \u2018TA\u2019 denotes the model trained with the time-aware loss. Bold denotes the best performance. Underline indicates that the model with time-aware loss outperforms the original model. The asterisk * denotes that the performance improvement against the model without TA version is statistically significant (p-value < 0.05). All three metrics are multiplied by 100 for readability purposes.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTJHCDSL
Metric\nAUPRC()\n\nAUROC()\n\nES()\n\nAUPRC()\n\nAUROC()\n\nES()\n
RF\n99.48 0.04\n\n98.13 0.19\n\n52.16 2.32\n\n52.16 2.32\n\n83.28 0.21\n\n-8.08 0.30\n
DT\n94.31 0.80\n\n85.68 1.80\n\n44.84 3.33\n\n27.71 0.00\n\n76.58 0.00\n\n-4.47 0.00\n
GBDT\n99.34 0.03\n\n97.57 0.11\n\n50.80 0.62\n\n41.91 0.28\n84.71 0.04\n3.55 0.35\n
CatBoost\n99.38 0.11\n\n97.74 0.39\n\n67.62 6.16\n\n40.00 0.36\n\n84.17 0.23\n\n-6.91 0.78\n
XGBoost\n99.46 0.00\n\n98.17 0.00\n\n39.92 0.00\n\n40.32 0.00\n\n84.44 0.00\n\n-1.22 0.00\n
MLP\n99.03 0.82\n\n97.63 0.98\n\n69.16 8.37\n\n39.63 0.79\n\n83.41 0.36\n\n-3.68 1.24\n
MLP-TA\n98.75 0.74\n\n96.94 1.01\n69.44 6.61\n40.46 0.89*\n83.49 0.32\n-2.26 2.14*\n
RNN\n99.69 0.14\n\n98.94 0.46\n\n75.56 4.28\n\n41.91 1.20\n\n83.36 0.90\n\n5.44 3.40\n
RNN-TA\n99.71 0.13*\n\n98.99 0.44*\n75.78 4.04\n42.65 1.17*\n\n83.35 0.63\n\n7.00 5.62*\n
LSTM\n99.14 0.14\n\n97.50 0.24\n\n73.70 1.79\n\n39.76 1.40\n\n81.73 0.77\n\n10.56 1.76\n
LSTM-TA99.16 0.1497.56 0.23\n74.15 2.61*\n\n40.77 1.32*\n\n82.39 0.53*\n\n13.03 2.66*\n
GRU\n99.43 0.25\n\n98.07 0.78\n\n75.74 3.88\n\n41.43 1.93\n\n83.28 0.90\n\n5.38 5.13\n
GRU-TA99.48 0.20\n98.21 0.64*\n\n76.42 2.06*\n\n42.18 1.35*\n\n83.64 0.82*\n\n9.01 5.59*\n
TCN\n98.71 0.23\n\n96.84 0.25\n\n74.38 1.29\n\n39.64 1.48\n\n83.50 0.44\n\n4.17 8.71\n
TCN-TA98.72 0.20\n96.84 0.22\n\n73.70 1.79\n\n40.61 1.68*\n\n83.46 0.44\n\n3.45 3.26\n
Transformer\n99.06 0.86\n\n97.84 1.23\n\n64.22 6.73\n\n31.18 4.86\n\n80.04 1.15\n\n-1.93 12.86\n
Transformer-TA99.53 0.2298.52 0.56\n64.22 6.81\n\n32.31 2.44*\n\n80.89 1.22*\n\n9.48 6.30*\n
RETAIN\n97.23 1.21\n\n92.70 2.90\n\n50.39 15.40\n\n37.65 2.17\n\n81.84 1.33\n\n1.92 9.56\n
RETAIN-TA97.50 1.1193.14 2.8252.66 16.36\n38.83 2.30*\n\n82.42 1.65*\n\n5.31 9.79*\n
StageNet\n99.35 0.26\n\n97.94 0.64\n\n70.07 2.59\n\n42.08 0.53\n\n83.09 0.36\n\n12.88 3.08\n
StageNet-TA\n99.38 0.25*\n\n98.06 0.63*\n\n70.98 3.98*\n\n42.51 0.52*\n83.17 0.32\n14.06 2.80*\n
Dr. Agent\n99.35 0.24\n\n97.85 0.57\n\n72.57 3.84\n\n40.83 1.18\n\n82.98 0.78\n\n11.18 6.92\n
Dr. Agent-TA99.37 0.2497.92 0.57\n73.47 2.93*\n\n41.87 1.45*\n\n83.07 1.12*\n\n12.46 6.58*\n
AdaCare\n99.85 0.13\n\n99.49 0.41\n\n62.32 10.08\n\n40.78 0.94\n\n83.46 0.12\n\n8.27 1.81\n
AdaCare-TA\n99.87 0.09*\n\n99.53 0.32*\n\n69.35 4.41*\n\n41.90 1.15*\n\n83.72 0.28*\n\n9.70 1.59*\n
GRASP\n99.42 0.16\n\n98.19 0.39\n\n72.79 4.63\n\n33.13 1.10\n\n75.04 0.74\n\n-2.32 6.00\n
GRASP-TA99.49 0.22\n98.40 0.57*\n\n72.79 4.90\n\n34.57 1.12*\n75.43 1.201.92 2.39
ConCare\n99.50 0.20\n\n98.24 0.55\n\n77.33 2.82\n\n43.12 0.09\n\n84.39 0.08\n\n-1.23 5.16\n
ConCare-TA\n99.49 0.24\n\n98.22 0.66\n\n78.91 1.72*\n\n43.47 0.14*\n\n84.28 0.17\n\n2.55 3.38*\n
\n
", + "capture": "Table 14: Benchmarking performance on the task of early mortality prediction on TJH and CDSL hold-out test sets. The reported score is of the form . \u2018TA\u2019 denotes the model trained with the time-aware loss. Bold denotes the best performance. Underline indicates that the model with time-aware loss outperforms the original model. The asterisk * denotes that the performance improvement against the model without TA version is statistically significant (p-value < 0.05). All three metrics are multiplied by 100 for readability purposes." + }, + "15": { + "table_html": "
\n
Table 15: Benchmarking performance on the task of early mortality prediction on MIMIC-III and MIMIC-IV hold-out test sets. The reported score is of the form . \u2018TA\u2019 denotes the model trained with the time-aware loss. Bold denotes the best performance. Underline indicates that the model with time-aware loss outperforms the original model. The asterisk * denotes that the performance improvement against the model without TA version is statistically significant (p-value < 0.05). All three metrics are multiplied by 100 for readability purposes.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetMIMIC-IIIMIMIC-IV
Metric\nAUPRC()\n\nAUROC()\n\nES()\n\nAUPRC()\n\nAUROC()\n\nES()\n
RF\n43.56 0.24\n\n81.94 0.08\n\n-1.58 0.31\n\n39.46 0.21\n\n80.58 0.06\n\n0.50 0.23\n
DT\n31.35 0.16\n\n70.65 0.17\n\n7.45 0.20\n\n29.56 0.22\n\n74.04 0.22\n\n9.49 0.16\n
GBDT\n45.73 0.02\n\n83.67 0.00\n\n11.80 0.03\n\n42.06 0.04\n\n81.80 0.01\n\n13.57 0.05\n
CatBoost\n45.44 0.34\n\n83.50 0.08\n\n6.72 0.14\n\n41.44 0.11\n\n81.66 0.04\n\n8.67 0.27\n
XGBoost\n45.73 0.20\n\n83.65 0.04\n\n10.68 0.43\n\n41.54 0.00\n\n81.76 0.00\n\n12.29 0.00\n
MLP\n45.01 0.28\n\n83.04 0.07\n\n12.73 1.51\n\n41.39 0.54\n\n81.35 0.17\n\n10.66 1.29\n
MLP-TA\n44.92 0.57\n\n82.86 0.18\n\n13.68 1.70*\n\n40.97 0.35\n\n81.18 0.14\n\n13.80 1.22*\n
RNN\n45.87 0.61\n\n83.34 0.25\n\n13.45 0.96\n\n44.08 0.21\n\n82.85 0.10\n\n15.93 2.35\n
RNN-TA\n45.38 0.52\n\n82.99 0.20\n13.69 1.12\n43.59 0.34\n\n82.59 0.16\n\n17.88 1.81*\n
LSTM\n47.46 0.34\n\n83.66 0.06\n\n15.35 1.44\n\n44.52 0.49\n\n83.14 0.21\n\n15.27 1.73\n
LSTM-TA\n47.14 0.35\n\n83.29 0.15\n15.70 1.32\n44.34 0.44\n\n82.86 0.15\n\n17.37 1.65*\n
GRU\n47.19 0.45\n\n83.80 0.18\n\n14.05 0.70\n45.58 0.18\n83.50 0.09\n\n15.43 1.60\n
GRU-TA\n46.87 0.45\n\n83.39 0.19\n\n15.35 2.62*\n\n45.33 0.22\n\n83.18 0.14\n\n17.44 2.03*\n
TCN\n47.71 0.46\n\n84.17 0.13\n18.72 1.85\n45.00 0.51\n83.53 0.23\n16.42 2.34\n
TCN-TA47.73 0.24\n84.04 0.13\n\n18.10 2.31\n45.03 0.59\n83.47 0.20\n\n17.74 1.75*\n
Transformer\n42.51 0.49\n\n82.43 0.16\n\n8.55 1.88\n\n42.32 0.24\n\n82.54 0.16\n\n10.12 3.95\n
Transformer-TA\n42.38 0.54\n\n82.24 0.18\n\n9.68 2.31*\n\n42.24 0.36\n\n82.54 0.14\n\n13.77 4.17*\n
RETAIN\n46.72 0.44\n\n83.61 0.23\n\n6.72 1.94\n\n44.89 0.54\n\n83.34 0.26\n\n6.65 0.59\n
RETAIN-TA\n46.49 0.35\n\n83.21 0.15\n\n8.27 1.86*\n\n44.48 0.25\n\n82.68 0.22\n\n7.33 1.19*\n
StageNet\n47.33 0.38\n\n83.60 0.16\n\n16.66 2.23\n\n44.79 0.26\n\n83.21 0.35\n\n15.62 0.77\n
StageNet-TA\n46.81 0.41\n\n83.22 0.19\n\n16.49 1.26\n\n44.22 0.44\n\n82.70 0.33\n\n17.68 3.12*\n
Dr. Agent\n47.94 0.33\n84.30 0.10\n15.71 2.32\n\n44.53 0.69\n\n83.15 0.24\n\n13.54 1.53\n
Dr. Agent-TA\n48.10 0.23*\n\n84.17 0.13\n\n16.04 2.59*\n\n44.00 0.29\n\n82.67 0.15\n13.84 1.64
AdaCare\n47.32 0.37\n\n83.62 0.33\n\n15.90 1.16\n\n44.41 0.13\n\n82.74 0.13\n\n15.33 2.10\n
AdaCare-TA\n47.29 0.24\n\n83.49 0.23\n\n15.45 1.85\n\n44.19 0.16\n\n82.42 0.11\n\n17.78 2.22*\n
GRASP\n46.96 0.19\n\n83.55 0.11\n\n13.95 1.66\n\n44.22 0.31\n\n82.87 0.12\n\n14.54 0.81\n
GRASP-TA\n46.64 0.33\n\n83.13 0.18\n14.13 1.56\n44.03 0.41\n\n82.49 0.15\n\n16.03 1.33*\n
ConCare\n46.40 0.30\n\n83.03 0.10\n\n15.30 1.25\n\n44.10 0.47\n\n82.70 0.22\n\n13.92 3.04\n
ConCare-TA\n45.91 0.49\n\n82.74 0.26\n\n15.29 0.91\n\n43.92 0.46\n\n82.48 0.18\n\n15.62 1.55*\n
\n
", + "capture": "Table 15: Benchmarking performance on the task of early mortality prediction on MIMIC-III and MIMIC-IV hold-out test sets. The reported score is of the form . \u2018TA\u2019 denotes the model trained with the time-aware loss. Bold denotes the best performance. Underline indicates that the model with time-aware loss outperforms the original model. The asterisk * denotes that the performance improvement against the model without TA version is statistically significant (p-value < 0.05). All three metrics are multiplied by 100 for readability purposes." + }, + "16": { + "table_html": "
\n
Table 16: Statistics of the MIMIC-III dataset. The reported statistics are of the form .
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
Mortality OutcomeTotalAliveDead
# Patients4151737108 (89.38 %)4409 (10.62 %)
# Records35090052952672 (84.15 %)556333 (15.85 %)
# Avg. records46.0 [26.0, 88.0]45.0 [26.0, 80.0]66.0 [26.0, 154.0]
Age65.5 [52.4, 77.9]64.5 [51.6, 77.0]74.6 [60.7, 83.3]
Age > Avg. (75.1)12852(30.96 %)10703(28.84 %)2149(48.74 %)
\nAge Avg. (75.1)\n28665(69.04 %)26405(71.16 %)2260(51.26 %)
Gender55.9% Male56.3% Male52.8% Male
Male23214(55.91 %)20885(56.28 %)2329(52.82 %)
Female18303(44.09 %)16223(43.72 %)2080(47.18 %)
# Features61
Length of stay46.9 [26.5, 88.5]45.9 [26.3, 80.5]68.6 [29.3, 158.8]
\n
", + "capture": "Table 16: Statistics of the MIMIC-III dataset. The reported statistics are of the form ." + }, + "17": { + "table_html": "
\n
Table 17: Statistics of the MIMIC-IV dataset. The reported statistics are of the form .
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
Mortality OutcomeTotalAliveDead
# Patients5688851451 (90.44 %)5437 (9.56 %)
# Records40555193466575 (85.48 %)588944 (14.52 %)
# Avg. records42.0 [24.0, 75.0]41.0 [24.0, 72.0]59.0 [24.0, 137.0]
Age65.0 [53.0, 76.0]64.0 [52.0, 75.0]72.0 [60.0, 82.0]
Age > Avg. (63.1)29950(52.65 %)26232(50.98 %)3718(68.38 %)
\nAge Avg. (63.1)\n26938(47.35 %)25219(49.02 %)1719(31.62 %)
Gender55.7% Male55.8% Male54.1% Male
Male31669(55.67 %)28726(55.83 %)2943(54.13 %)
Female25219(44.33 %)22725(44.17 %)2494(45.87 %)
# Features61
Length of stay42.1 [24.0, 75.2]41.0 [24.0, 71.9]61.2[25.2, 138.6]
\n
", + "capture": "Table 17: Statistics of the MIMIC-IV dataset. The reported statistics are of the form ." + }, + "18": { + "table_html": "
\n
Table 18: P-values of two-stage vs. multi-task performance in LOS predictions using t-test. P-values are rounded to three decimal places; a value of 0.000 indicates a rounded p-value < 0.0005.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTJHCDSL
ModelMAEMSEOSMAEMAEMSEOSMAE
MLP0.0000.0000.7380.0000.0000.000
RNN0.0000.0000.0000.0000.0000.000
LSTM0.0000.0000.0000.0000.0000.000
GRU0.0000.0110.0000.0000.0000.000
TCN0.0000.0000.0000.0000.0000.000
Transformer0.0000.3490.0000.0000.0000.000
RETAIN0.0000.0000.0000.0000.0000.000
StageNet0.0000.0000.0000.0000.0000.000
Dr. Agent0.0000.0000.0000.0000.0000.000
AdaCare0.0000.0000.0000.0000.0000.000
GRASP0.0000.0000.0000.0000.0000.000
ConCare0.0000.0000.0000.0000.0000.000
\n
", + "capture": "Table 18: P-values of two-stage vs. multi-task performance in LOS predictions using t-test. P-values are rounded to three decimal places; a value of 0.000 indicates a rounded p-value < 0.0005." + }, + "19": { + "table_html": "
\n
Table 19: P-values of naive models vs. models with the TA loss in early mortality predictions using t-test. P-values are rounded to three decimal places; a value of 0.000 indicates a rounded p-value < 0.0005.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
DatasetTJHCDSL
ModelAUPRCAUROCESAUPRCAUROCES
MLP0.0000.0000.1950.8960.5700.430
RNN0.1360.2740.0010.0000.0000.000
LSTM0.0000.0000.0421.0000.3781.000
GRU0.5840.1760.0000.0000.0000.000
TCN0.1320.0020.0920.9720.5400.005
Transformer0.8170.8810.0000.9660.0000.000
RETAIN0.0000.0000.0001.0001.0001.000
StageNet0.0000.0000.0030.0100.0101.000
Dr. Agent0.0300.0000.2360.0000.0000.000
AdaCare0.5860.1180.0000.7000.9640.007
GRASP0.1510.1760.1350.0000.0000.000
ConCare0.3800.9530.0010.8150.7551.000
\n
", + "capture": "Table 19: P-values of naive models vs. models with the TA loss in early mortality predictions using t-test. P-values are rounded to three decimal places; a value of 0.000 indicates a rounded p-value < 0.0005." + }, + "20": { + "table_html": "
\n
Table 20: (Continued) Statistics of lab test features in the TJH dataset.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
Feature nameStatisticsMissing Rate
Hypersensitive cardiac troponinI19.80 [4.20, 216.25]29.23%
hemoglobin125.00 [113.00, 137.00]55.52%
Serum chloride102.00 [98.95, 105.50]55.58%
Prothrombin time14.70 [13.60, 16.60]38.44%
procalcitonin0.10 [0.04, 0.40]26.94%
eosinophils(%)0.10 [0.00, 0.90]55.40%
Interleukin 2 receptor680.50 [460.00, 1169.75]15.38%
Alkaline phosphatase69.00 [54.00, 95.00]53.81%
albumin32.40 [27.50, 36.70]53.99%
basophil(%)0.20 [0.10, 0.30]55.40%
Interleukin 105.80 [5.00, 12.50]15.32%
Total bilirubin10.60 [7.30, 16.70]53.81%
Platelet count181.00 [112.75, 250.25]55.16%
monocytes(%)5.70 [2.90, 8.70]55.46%
antithrombin86.00 [74.00, 97.00]19.19%
Interleukin 815.95 [8.62, 33.58]15.38%
indirect bilirubin5.40 [3.80, 8.00]52.46%
Red blood cell distribution width12.60 [12.00, 13.60]53.40%
neutrophils(%)82.05 [65.07, 92.20]55.40%
total protein66.00 [61.10, 70.40]53.81%
Treponema pallidum antibodies0.05 [0.04, 0.07]16.31%
Prothrombin activity81.50 [65.00, 95.00]38.26%
HBsAg0.01 [0.00, 0.01]16.31%
mean corpuscular volume90.00 [86.90, 93.80]55.40%
hematocrit36.70 [33.50, 39.90]55.40%
White blood cell count7.60 [5.08, 12.52]60.09%
Tumor necrosis factor\n8.60 [6.70, 11.58]15.38%
corpuscular hemoglobin concentration343.00 [333.75, 350.00]55.40%
fibrinogen4.12 [3.06, 5.51]32.86%
Interleukin 1\n5.00 [5.00, 5.00]15.38%
Urea5.90 [4.00, 11.10]53.99%
lymphocyte count0.80 [0.47, 1.32]55.40%
PH value6.50 [6.00, 7.00]21.01%
Red blood cell count4.16 [3.67, 4.69]60.09%
Eosinophil count0.01 [0.00, 0.06]55.40%
Corrected calcium2.36 [2.27, 2.44]52.82%
Serum potassium4.41 [3.95, 4.86]55.75%
glucose6.98 [5.53, 10.15]45.07%
neutrophils count5.82 [3.08, 10.82]55.40%
Direct bilirubin4.80 [3.20, 8.00]53.81%
Mean platelet volume10.80 [10.10, 11.50]50.12%
ferritin711.60 [385.80, 1425.30]16.49%
RBC distribution width SD40.90 [38.50, 44.68]53.40%
Thrombin time16.80 [15.60, 18.30]32.86%
lymphocyte(%)11.70 [4.00, 25.00]55.46%
HCV antibody quantification0.06 [0.04, 0.09]16.31%
D-D dimer2.12 [0.60, 21.00]36.74%
Total cholesterol3.63 [3.01, 4.27]53.87%
aspartate aminotransferase27.00 [20.00, 42.00]53.93%
Uric acid244.00 [184.00, 332.60]53.87%
HCO3-23.50 [21.00, 25.90]53.87%
calcium2.09 [1.98, 2.19]55.75%
NT-proBNP571.50 [147.00, 2589.00]27.58%
Lactate dehydrogenase339.00 [217.00, 596.25]53.87%
platelet large cell ratio30.80 [25.50, 37.10]50.12%
Interleukin 619.56 [4.66, 61.12]15.61%
Fibrin degradation products17.80 [4.00, 150.00]19.19%
monocytes count0.41 [0.27, 0.58]55.40%
PLT distribution width12.40 [11.10, 14.30]50.12%
globulin32.70 [29.70, 36.50]53.81%
glutamyl transpeptidase34.00 [22.00, 58.00]53.81%
International standard ratio1.14 [1.03, 1.33]38.26%
basophil count(#)0.01 [0.01, 0.02]55.40%
mean corpuscular hemoglobin30.90 [29.70, 32.20]55.40%
Activation of partial thromboplastin time39.20 [35.50, 44.10]32.92%
Hypersensitive c-reactive protein50.50 [5.35, 118.50]42.66%
HIV antibody quantification0.09 [0.07, 0.11]16.26%
serum sodium140.30 [137.70, 143.30]55.58%
thrombocytocrit0.21 [0.15, 0.27]50.12%
ESR28.00 [14.00, 45.50]22.48%
glutamic-pyruvic transaminase24.00 [16.00, 40.00]53.87%
eGFR88.10 [64.70, 104.20]53.99%
creatinine76.00 [58.00, 98.00]53.99%
\n
", + "capture": "Table 20: (Continued) Statistics of lab test features in the TJH dataset." + }, + "21": { + "table_html": "
\n
Table 21: (Continued) Statistics of lab test features in the CDSL dataset.
\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n
Feature nameStatisticsMissing Rate
ADW \u2013 Coeficiente de anisocitosis13.20 [12.10, 14.55]8.34%
ADW \u2013 SISTEMATICO DE SANGRE13.85 [12.30, 15.30]2.68%
ALB \u2013 ALBUMINA3.10 [2.70, 3.50]1.27%
AMI \u2013 AMILASA69.00 [46.92, 111.90]1.25%
AP \u2013 ACTIVIDAD DE PROTROMBINA79.00 [70.00, 87.00]5.86%
APTT \u2013 TIEMPO DE CEFALINA (APTT)30.80 [28.10, 33.70]5.37%
AU \u2013 ACIDO URICO4.70 [3.50, 6.50]0.53%
BAS \u2013 Bas\u00a2filos0.02 [0.01, 0.04]8.31%
BAS \u2013 SISTEMATICO DE SANGRE0.02 [0.01, 0.04]2.70%
BAS% \u2013 Bas\u00a2filos %0.30 [0.10, 0.50]8.32%
BAS% \u2013 SISTEMATICO DE SANGRE0.30 [0.10, 0.54]2.70%
BD \u2013 BILIRRUBINA DIRECTA0.26 [0.18, 0.40]1.76%
BE(b) \u2013 BE(b)2.80 [0.00, 5.90]2.57%
BE(b)V \u2013 BE (b)1.90 [-0.60, 4.50]1.00%
BEecf \u2013 BEecf3.20 [-0.30, 6.78]2.57%
BEecfV \u2013 BEecf2.50 [-0.60, 5.40]1.00%
BT \u2013 BILIRRUBINA TOTAL0.40 [0.27, 0.57]1.68%
BT \u2013 BILIRRUBINA TOTAL0.52 [0.36, 0.77]3.64%
CA \u2013 CALCIO8.30 [8.00, 8.70]0.77%
CA++ \u2013 Ca++ Gasometria4.41 [4.21, 4.64]2.30%
CHCM \u2013 Conc. Hemoglobina Corpuscular Media33.40 [32.50, 34.20]8.39%
CHCM \u2013 SISTEMATICO DE SANGRE33.10 [32.10, 34.00]2.70%
CK \u2013 CK (CREATINQUINASA)66.10 [36.68, 140.00]3.10%
CL \u2013 CLORO102.30 [98.90, 106.43]2.19%
CREA \u2013 CREATININA0.82 [0.64, 1.06]8.28%
DD \u2013 DIMERO D985.00 [525.75, 2138.75]6.31%
EOS \u2013 Eosin\u00a2filos0.03 [0.00, 0.13]8.31%
EOS \u2013 SISTEMATICO DE SANGRE0.10 [0.01, 0.23]2.70%
EOS% \u2013 Eosin\u00a2filos %0.50 [0.00, 1.90]8.32%
EOS% \u2013 SISTEMATICO DE SANGRE1.30 [0.10, 3.20]2.70%
FA \u2013 FOSFATASA ALCALINA72.70 [55.50, 104.90]3.29%
FER \u2013 FERRITINA908.55 [445.92, 1645.00]1.91%
FIB \u2013 FIBRIN\u00e0GENO562.00 [405.00, 724.00]2.95%
FOS \u2013 FOSFORO3.28 [2.70, 3.90]2.15%
G-CORONAV (RT-PCR)1.00 [0.00, 1.00]0.49%
GGT \u2013 GGT65.00 [30.00, 140.50]5.53%
GLU \u2013 GLUCOSA111.20 [94.00, 144.20]7.67%
GOT \u2013 GOT (AST)30.30 [20.80, 48.50]7.42%
GPT \u2013 GPT (ALT)32.60 [18.90, 61.00]7.23%
HCM \u2013 Hemoglobina Corpuscular Media29.80 [28.60, 30.80]8.39%
HCM \u2013 SISTEMATICO DE SANGRE29.90 [28.70, 30.90]2.70%
HCO3 \u2013 HCO3-27.40 [24.10, 31.10]2.57%
HCO3V \u2013 HCO3-27.20 [24.05, 29.90]1.00%
HCTO \u2013 Hematocrito38.30 [33.60, 42.00]8.39%
HCTO \u2013 SISTEMATICO DE SANGRE34.40 [29.60, 39.20]2.70%
HEM \u2013 Hemat\u00a1es4.33 [3.75, 4.78]8.39%
HEM \u2013 SISTEMATICO DE SANGRE3.83 [3.25, 4.45]2.70%
HGB \u2013 Hemoglobina12.80 [11.10, 14.10]8.39%
HGB \u2013 SISTEMATICO DE SANGRE11.30 [9.70, 13.10]2.70%
INR \u2013 INR1.17 [1.09, 1.27]5.87%
K \u2013 POTASIO4.25 [3.86, 4.67]8.10%
LAC \u2013 LACTATO1.50 [1.10, 2.10]2.46%
LDH \u2013 LDH532.00 [410.05, 704.00]7.29%
LEUC \u2013 Leucocitos7.25 [5.34, 10.17]8.39%
LEUC \u2013 SISTEMATICO DE SANGRE7.70 [5.74, 10.85]2.70%
LIN \u2013 Linfocitos1.08 [0.72, 1.55]8.39%
LIN \u2013 SISTEMATICO DE SANGRE1.24 [0.79, 1.78]2.70%
LIN% \u2013 Linfocitos %15.50 [8.80, 24.50]8.38%
LIN% \u2013 SISTEMATICO DE SANGRE16.33 [9.06, 26.30]2.70%
MG \u2013 MAGNESIO2.05 [1.81, 2.31]2.40%
MONO \u2013 Monocitos0.55 [0.36, 0.77]8.32%
MONO \u2013 SISTEMATICO DE SANGRE0.60 [0.42, 0.82]2.70%
MONO% \u2013 Monocitos %7.60 [4.80, 10.40]8.32%
MONO% \u2013 SISTEMATICO DE SANGRE7.80 [5.15, 10.30]2.70%
NA \u2013 SODIO138.00 [135.60, 140.70]8.10%
NEU \u2013 Neutr\u00a2filos5.23 [3.46, 8.10]8.39%
NEU \u2013 SISTEMATICO DE SANGRE5.39 [3.53, 8.43]2.70%
NEU% \u2013 Neutr\u00a2filos %74.10 [62.90, 84.60]8.39%
NEU% \u2013 SISTEMATICO DE SANGRE71.70 [59.20, 83.40]2.70%
PCO2 \u2013 pCO241.50 [35.50, 48.50]2.57%
PCO2V \u2013 pCO244.00 [38.00, 50.00]1.00%
PCR \u2013 PROTEINA C REACTIVA34.75 [8.72, 93.02]8.03%
PH \u2013 pH7.44 [7.39, 7.47]2.57%
PHV \u2013 pH7.40 [7.36, 7.43]1.00%
PLAQ \u2013 Recuento de plaquetas241.00 [176.00, 326.00]8.39%
PLAQ \u2013 SISTEMATICO DE SANGRE238.00 [174.00, 315.00]2.70%
PO2 \u2013 pO282.80 [63.00, 106.00]2.57%
PO2V \u2013 pO242.00 [27.00, 59.00]1.00%
PROCAL \u2013 PROCALCITONINA0.14 [0.08, 0.30]0.77%
PT \u2013 PROTEINAS TOTALES5.62 [5.10, 6.10]0.84%
SO2C \u2013 sO2c (Saturaci\u00a2n de ox\u00a1geno)95.67 [92.00, 97.67]2.57%
SO2CV \u2013 sO2c (Saturaci\u00a2n de ox\u00a1geno)76.00 [48.00, 90.00]1.00%
TCO2 \u2013 tCO2(B)c28.70 [25.20, 32.53]2.57%
TCO2V \u2013 tCO2 (B)28.60 [25.10, 31.49]1.00%
TP \u2013 TIEMPO DE PROTROMBINA13.00 [12.10, 14.10]5.86%
TROPO \u2013 TROPONINA15.16 [7.95, 34.92]0.99%
U \u2013 UREA41.60 [29.00, 61.10]8.06%
VCM \u2013 SISTEMATICO DE SANGRE90.00 [87.00, 93.40]2.70%
VCM \u2013 Volumen Corpuscular Medio88.90 [86.00, 92.30]8.39%
VPM \u2013 SISTEMATICO DE SANGRE10.50 [9.80, 11.30]2.67%
VPM \u2013 Volumen plaquetar medio10.40 [9.70, 11.10]8.30%
VSG \u2013 VSG47.50 [16.00, 75.00]0.47%
\n
", + "capture": "Table 21: (Continued) Statistics of lab test features in the CDSL dataset." + } + }, + "image_paths": { + "1(a)": { + "figure_path": "2209.07805v4_figure_1(a).png", + "caption": "(a) Illustrations of the OSMAE metric.\nFigure 1: Illustrations of the proposed OSMAE and ES metrics.", + "url": "http://arxiv.org/html/2209.07805v4/x1.png" + }, + "1(b)": { + "figure_path": "2209.07805v4_figure_1(b).png", + "caption": "(b) Illustration of the early mortality prediction score.\nFigure 1: Illustrations of the proposed OSMAE and ES metrics.", + "url": "http://arxiv.org/html/2209.07805v4/x2.png" + }, + "2": { + "figure_path": "2209.07805v4_figure_2.png", + "caption": "Figure 2: The K-fold cross-validation strategy. We take 4-fold as an example in the figure. We use a stratified shuffle split to ensure the proportions of alive and dead patients on all folds are the same as the total cohort.", + "url": "http://arxiv.org/html/2209.07805v4/x3.png" + }, + "3": { + "figure_path": "2209.07805v4_figure_3.png", + "caption": "Figure 3: Illustrations of the two-stage training and multi-task training settings", + "url": "http://arxiv.org/html/2209.07805v4/x4.png" + }, + "4": { + "figure_path": "2209.07805v4_figure_4.png", + "caption": "Figure 4: Early prediction performance of 5 models with the highest ES on the CDSL dataset. All models are trained using the first half of patient records. Error bars are standard deviations. All performance improvements are statistically significant (p-value < 0.05).", + "url": "http://arxiv.org/html/2209.07805v4/x5.png" + }, + "5": { + "figure_path": "2209.07805v4_figure_5.png", + "caption": "Figure 5: AUROC of Dr. Agent and Dr. Agent-TA at each visit", + "url": "http://arxiv.org/html/2209.07805v4/x6.png" + }, + "6(a)": { + "figure_path": "2209.07805v4_figure_6(a).png", + "caption": "(a) ES and risk curve\nFigure 6: Case prediction plot on the CDSL dataset. We use the TCN model with the multi-task setting to generate the prediction results.", + "url": "http://arxiv.org/html/2209.07805v4/x7.png" + }, + "6(b)": { + "figure_path": "2209.07805v4_figure_6(b).png", + "caption": "(b) OSMAE, MAE, predicted LOS and ground truth LOS curve\nFigure 6: Case prediction plot on the CDSL dataset. We use the TCN model with the multi-task setting to generate the prediction results.", + "url": "http://arxiv.org/html/2209.07805v4/x8.png" + }, + "7(a)": { + "figure_path": "2209.07805v4_figure_7(a).png", + "caption": "(a) OSMAE with \u03b3\ud835\udefe\\gammaitalic_\u03b3 from 0-10\nFigure 7: OSMAE and ES values with different \u03b3\ud835\udefe\\gammaitalic_\u03b3. Scores are calculated using the same TCN multi-task model on CDSL dataset.", + "url": "http://arxiv.org/html/2209.07805v4/x9.png" + }, + "7(b)": { + "figure_path": "2209.07805v4_figure_7(b).png", + "caption": "(b) ES with \u03b3\ud835\udefe\\gammaitalic_\u03b3 from 0-10\nFigure 7: OSMAE and ES values with different \u03b3\ud835\udefe\\gammaitalic_\u03b3. Scores are calculated using the same TCN multi-task model on CDSL dataset.", + "url": "http://arxiv.org/html/2209.07805v4/x10.png" + }, + "8(a)": { + "figure_path": "2209.07805v4_figure_8(a).png", + "caption": "(a) FP case penalty term\nFigure 8: ES values with different penalty terms for TN case and FP case ranging from \u221222-2- 2 to 2222. Scores are calculated using the same TCN multi-task model on the CDSL dataset.", + "url": "http://arxiv.org/html/2209.07805v4/x11.png" + }, + "8(b)": { + "figure_path": "2209.07805v4_figure_8(b).png", + "caption": "(b) TN case penalty term\nFigure 8: ES values with different penalty terms for TN case and FP case ranging from \u221222-2- 2 to 2222. Scores are calculated using the same TCN multi-task model on the CDSL dataset.", + "url": "http://arxiv.org/html/2209.07805v4/x12.png" + }, + "9": { + "figure_path": "2209.07805v4_figure_9.png", + "caption": "Figure 9: Screenshot of the online benchmark results visualization system.", + "url": "http://arxiv.org/html/2209.07805v4/extracted/5364382/figures/screenshot.png" + }, + "10": { + "figure_path": "2209.07805v4_figure_10.png", + "caption": "Figure 10: Benchmark code structure.", + "url": "http://arxiv.org/html/2209.07805v4/x13.png" + }, + "11(a)": { + "figure_path": "2209.07805v4_figure_11(a).png", + "caption": "(a) MIMIC-III\nFigure 11: Early prediction performance of 5 models with the highest ES on the MIMIC-III and MIMIC-IV datasets. All models are trained using the first half of patient records. Error bars are standard deviations.", + "url": "http://arxiv.org/html/2209.07805v4/x14.png" + }, + "11(b)": { + "figure_path": "2209.07805v4_figure_11(b).png", + "caption": "(b) MIMIC-IV\nFigure 11: Early prediction performance of 5 models with the highest ES on the MIMIC-III and MIMIC-IV datasets. All models are trained using the first half of patient records. Error bars are standard deviations.", + "url": "http://arxiv.org/html/2209.07805v4/x15.png" + }, + "12(a)": { + "figure_path": "2209.07805v4_figure_12(a).png", + "caption": "(a) TCN model\u2019s embedding under two-stage setting\nFigure 12: Visualization of TCN model\u2019s embedding for LOS prediction under two settings using t-SNE. The embeddings are extracted at each patient\u2019s last visit on the CDSL dataset.", + "url": "http://arxiv.org/html/2209.07805v4/x16.png" + }, + "12(b)": { + "figure_path": "2209.07805v4_figure_12(b).png", + "caption": "(b) TCN model\u2019s embedding under multi-task setting\nFigure 12: Visualization of TCN model\u2019s embedding for LOS prediction under two settings using t-SNE. The embeddings are extracted at each patient\u2019s last visit on the CDSL dataset.", + "url": "http://arxiv.org/html/2209.07805v4/x17.png" + }, + "13(a)": { + "figure_path": "2209.07805v4_figure_13(a).png", + "caption": "(a) The first half of records (alive patients)\nFigure 13: MAE score distribution on the CDSL test set. The figure compares the discrepancies in MAE performance distributions between the first and last half of patient records. The analysis utilizes the StageNet model with the multi-task setting.", + "url": "http://arxiv.org/html/2209.07805v4/x18.png" + }, + "13(b)": { + "figure_path": "2209.07805v4_figure_13(b).png", + "caption": "(b) The last half of records (alive patients)\nFigure 13: MAE score distribution on the CDSL test set. The figure compares the discrepancies in MAE performance distributions between the first and last half of patient records. The analysis utilizes the StageNet model with the multi-task setting.", + "url": "http://arxiv.org/html/2209.07805v4/x19.png" + }, + "13(c)": { + "figure_path": "2209.07805v4_figure_13(c).png", + "caption": "(c) The first half of records (deceased patients)\nFigure 13: MAE score distribution on the CDSL test set. The figure compares the discrepancies in MAE performance distributions between the first and last half of patient records. The analysis utilizes the StageNet model with the multi-task setting.", + "url": "http://arxiv.org/html/2209.07805v4/x20.png" + }, + "13(d)": { + "figure_path": "2209.07805v4_figure_13(d).png", + "caption": "(d) The last half of records (deceased patients)\nFigure 13: MAE score distribution on the CDSL test set. The figure compares the discrepancies in MAE performance distributions between the first and last half of patient records. The analysis utilizes the StageNet model with the multi-task setting.", + "url": "http://arxiv.org/html/2209.07805v4/x21.png" + }, + "14": { + "figure_path": "2209.07805v4_figure_14.png", + "caption": "Figure 14: Data preprocessing details of two datasets.", + "url": "http://arxiv.org/html/2209.07805v4/x22.png" + }, + "15": { + "figure_path": "2209.07805v4_figure_15.png", + "caption": "Figure 15: Length of stay distributions in two datasets. To keep the figure informative, we only show the statistics in the 0%-95% range for the CDSL dataset.", + "url": "http://arxiv.org/html/2209.07805v4/extracted/5364382/figures/tjh_cdsl_los.png" + }, + "16": { + "figure_path": "2209.07805v4_figure_16.png", + "caption": "Figure 16: Distributions of 16 features with the lowest missing rates in the TJH dataset. We plot the data distributions of 16 features with the lowest missing rates for the overall, alive, and dead patients. The blue bars are distributions of total patients. The green and pink curves are distributions for alive and dead patients, respectively. To keep the figure informative, we only show the statistics in the 0%-95% range.", + "url": "http://arxiv.org/html/2209.07805v4/extracted/5364382/figures/tjh_feat.png" + }, + "17": { + "figure_path": "2209.07805v4_figure_17.png", + "caption": "Figure 17: Distributions of 16 features with the lowest missing rates in the CDSL dataset. We plot the data distributions of 16 features with the lowest missing rates for the overall, alive, and dead patients. The blue bars are distributions of total patients. The green and pink curves are distributions for alive and dead patients, respectively. To keep the figure informative, we only show the statistics in the 0%-95% range.", + "url": "http://arxiv.org/html/2209.07805v4/extracted/5364382/figures/cdsl_feat.png" + } + }, + "validation": true, + "references": [ + { + "1": { + "title": "How bad is omicron? what scientists know so far.", + "author": "Callaway, E., Ledford, H. et al.", + "venue": "\\JournalTitleNature 600, 197\u2013199 (2021).", + "url": null + } + }, + { + "2": { + "title": "An interpretable mortality prediction model for covid-19 patients.", + "author": "Yan, L. et al.", + "venue": "\\JournalTitleNature machine intelligence 2, 283\u2013288 (2020).", + "url": null + } + }, + { + "3": { + "title": "Using machine learning to predict mortality for covid-19 patients on day 0 in the icu.", + "author": "Jamshidi, E. et al.", + "venue": "\\JournalTitleFrontiers in digital health 3 (2021).", + "url": null + } + }, + { + "4": { + "title": "Characteristics, outcomes, and severity risk factors associated with sars-cov-2 infection among children in the us national covid cohort collaborative.", + "author": "Martin, B. et al.", + "venue": "\\JournalTitleJAMA network open 5, e2143151\u2013e2143151 (2022).", + "url": null + } + }, + { + "5": { + "title": "Assessment of clinical outcomes among children and adolescents hospitalized with covid-19 in 6 sub-saharan african countries.", + "author": "Nachega, J. B. et al.", + "venue": "\\JournalTitleJAMA pediatrics 176, e216436\u2013e216436 (2022).", + "url": null + } + }, + { + "6": { + "title": "A bayesian model to predict covid-19 severity in children.", + "author": "Dom\u00ednguez-Rodr\u00edguez, S. et al.", + "venue": "\\JournalTitleThe Pediatric Infectious Disease Journal 40, e287\u2013e293 (2021).", + "url": null + } + }, + { + "7": { + "title": "Comparison of the first and second waves of the coronavirus disease 2019 pandemic in children and adolescents in a middle-income country: Clinical impact associated with severe acute respiratory syndrome coronavirus 2 gamma lineage.", + "author": "Oliveira, E. A. et al.", + "venue": "\\JournalTitleThe Journal of pediatrics 244, 178\u2013185 (2022).", + "url": null + } + }, + { + "8": { + "title": "Clinical characterization and prediction of clinical severity of sars-cov-2 infection among us adults using data from the us national covid cohort collaborative.", + "author": "Bennett, T. D. et al.", + "venue": "\\JournalTitleJAMA network open 4, e2116901\u2013e2116901 (2021).", + "url": null + } + }, + { + "9": { + "title": "Covid-19 mortality in the uk biobank cohort: revisiting and evaluating risk factors.", + "author": "Elliott, J. et al.", + "venue": "\\JournalTitleEuropean journal of epidemiology 36, 299\u2013309 (2021).", + "url": null + } + }, + { + "10": { + "title": "Medml: Fusing medical knowledge and machine learning models for early pediatric covid-19 hospitalization and severity prediction.", + "author": "Gao, J. et al.", + "venue": "\\JournalTitleIscience 104970 (2022).", + "url": null + } + }, + { + "11": { + "title": "Prediction models for diagnosis and prognosis of covid-19: systematic review and critical appraisal.", + "author": "Wynants, L. et al.", + "venue": "\\JournalTitlebmj 369 (2020).", + "url": null + } + }, + { + "12": { + "title": "Covid-19 prediction models: a systematic literature review.", + "author": "Shakeel, S. M., Kumar, N. S., Madalli, P. P., Srinivasaiah, R. & Swamy, D. R.", + "venue": "\\JournalTitleOsong public health and research perspectives 12, 215 (2021).", + "url": null + } + }, + { + "13": { + "title": "Mortality prediction in the icu based on mimic-ii results from the super icu learner algorithm (sicula) project.", + "author": "Pirracchio, R.", + "venue": "\\JournalTitleSecondary Analysis of Electronic Health Records 295\u2013313 (2016).", + "url": null + } + }, + { + "14": { + "title": "Benchmarking deep learning models on large healthcare datasets.", + "author": "Purushotham, S., Meng, C., Che, Z. & Liu, Y.", + "venue": "\\JournalTitleJournal of biomedical informatics 83, 112\u2013134 (2018).", + "url": null + } + }, + { + "15": { + "title": "Multitask learning and benchmarking with clinical time series data.", + "author": "Harutyunyan, H., Khachatrian, H., Kale, D. C., Ver Steeg, G. & Galstyan, A.", + "venue": "\\JournalTitleScientific data 6, 1\u201318 (2019).", + "url": null + } + }, + { + "16": { + "title": "Hirid-icu-benchmark\u2013a comprehensive machine learning benchmark on high-resolution icu data.", + "author": "Y\u00e8che, H. et al.", + "venue": "\\JournalTitlearXiv preprint arXiv:2111.08536 (2021).", + "url": null + } + }, + { + "17": { + "title": "Mimic-iii, a freely accessible critical care database.", + "author": "Johnson, A. E. et al.", + "venue": "\\JournalTitleScientific data 3, 1\u20139 (2016).", + "url": null + } + }, + { + "18": { + "title": "Open-access mimic-ii database for intensive care research.", + "author": "Lee, J. et al.", + "venue": "In 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, 8315\u20138318 (IEEE, 2011).", + "url": null + } + }, + { + "19": { + "title": "Early prediction of the risk of icu mortality with deep federated learning.", + "author": "Rand, K., Armengol, N. L., Mondrejevski, L. & Miliou, I.", + "venue": "\\JournalTitlearXiv preprint arXiv:2212.00554 (2022).", + "url": null + } + }, + { + "20": { + "title": "Early prediction of sepsis from clinical data: the physionet/computing in cardiology challenge 2019.", + "author": "Reyna, M. A. et al.", + "venue": "In 2019 Computing in Cardiology (CinC), Page\u20131 (IEEE, 2019).", + "url": null + } + }, + { + "21": { + "title": "Covidcare: Transferring knowledge from existing emr to emerging epidemic for interpretable prognosis.", + "author": "Ma, L. et al.", + "venue": "\\JournalTitlearXiv preprint arXiv:2007.08848 (2020).", + "url": null + } + }, + { + "22": { + "title": "Dr. agent: Clinical predictive model via mimicked second opinions.", + "author": "Gao, J., Xiao, C., Glass, L. M. & Sun, J.", + "venue": "\\JournalTitleJournal of the American Medical Informatics Association 27, 1084\u20131091 (2020).", + "url": null + } + }, + { + "23": { + "title": "Covid data save lives.", + "author": "Hospitales, H.", + "venue": "https://www.hmhospitales.com/prensa/notas-de-prensa/comunicado-covid-data-save-lives (2020).", + "url": null + } + }, + { + "24": { + "title": "Distilling knowledge from publicly available online emr data to emerging epidemic for prognosis.", + "author": "Ma, L. et al.", + "venue": "In Proceedings of the Web Conference 2021, 3558\u20133568 (2021).", + "url": null + } + }, + { + "25": { + "title": "Adacare: Explainable clinical health status representation learning via scale-adaptive feature extraction and recalibration.", + "author": "Ma, L. et al.", + "venue": "\\JournalTitleProceedings of the AAAI Conference on Artificial Intelligence 34, 825\u2013832, 10.1609/aaai.v34i01.5427 (2020).", + "url": null + } + }, + { + "26": { + "title": "Concare: Personalized clinical feature embedding via capturing the healthcare context.", + "author": "Ma, L. et al.", + "venue": "\\JournalTitleProceedings of the AAAI Conference on Artificial Intelligence 34, 833\u2013840, 10.1609/aaai.v34i01.5428 (2020).", + "url": null + } + }, + { + "27": { + "title": "Risk stratification of patients admitted to hospital with covid-19 using the isaric who clinical characterisation protocol: development and validation of the 4c mortality score.", + "author": "Knight, S. R. et al.", + "venue": "\\JournalTitlebmj 370 (2020).", + "url": null + } + }, + { + "28": { + "title": "Xgboost: A scalable tree boosting system.", + "author": "Chen, T. & Guestrin, C.", + "venue": "In Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, 785\u2013794 (2016).", + "url": null + } + }, + { + "29": { + "title": "Catboost: gradient boosting with categorical features support.", + "author": "Dorogush, A. V., Ershov, V. & Gulin, A.", + "venue": "\\JournalTitlearXiv preprint arXiv:1810.11363 (2018).", + "url": null + } + }, + { + "30": { + "title": "Learning representations by back-propagating errors.", + "author": "Rumelhart, D. E., Hinton, G. E. & Williams, R. J.", + "venue": "\\JournalTitlenature 323, 533\u2013536 (1986).", + "url": null + } + }, + { + "31": { + "title": "Long short-term memory.", + "author": "Hochreiter, S. & Schmidhuber, J.", + "venue": "\\JournalTitleNeural computation 9, 1735\u20131780 (1997).", + "url": null + } + }, + { + "32": { + "title": "Empirical evaluation of gated recurrent neural networks on sequence modeling.", + "author": "Chung, J., Gulcehre, C., Cho, K. & Bengio, Y.", + "venue": "\\JournalTitlearXiv preprint arXiv:1412.3555 (2014).", + "url": null + } + }, + { + "33": { + "title": "An empirical evaluation of generic convolutional and recurrent networks for sequence modeling.", + "author": "Bai, S., Kolter, J. Z. & Koltun, V.", + "venue": "\\JournalTitlearXiv preprint arXiv:1803.01271 (2018).", + "url": null + } + }, + { + "34": { + "title": "Attention is all you need.", + "author": "Vaswani, A. et al.", + "venue": "\\JournalTitleAdvances in neural information processing systems 30 (2017).", + "url": null + } + }, + { + "35": { + "title": "Retain: An interpretable predictive model for healthcare using reverse time attention mechanism.", + "author": "Choi, E. et al.", + "venue": "\\JournalTitleAdvances in neural information processing systems 29 (2016).", + "url": null + } + }, + { + "36": { + "title": "Stagenet: Stage-aware neural networks for health risk prediction.", + "author": "Gao, J. et al.", + "venue": "In Proceedings of The Web Conference 2020, 530\u2013540 (2020).", + "url": null + } + }, + { + "37": { + "title": "Grasp: Generic framework for health status representation learning based on incorporating knowledge from similar patients.", + "author": "Zhang, C. et al.", + "venue": "\\JournalTitleProceedings of the AAAI Conference on Artificial Intelligence 35, 715\u2013723, 10.1609/aaai.v35i1.16152 (2021).", + "url": null + } + }, + { + "38": { + "title": "Analysis of small sample size studies using nonparametric bootstrap test with pooled resampling method.", + "author": "Dwivedi, A. K., Mallawaarachchi, I. & Alvarado, L. A.", + "venue": "\\JournalTitleStatistics in medicine 36, 2187\u20132205 (2017).", + "url": null + } + }, + { + "39": { + "title": "Interpreting diagnostic accuracy studies for patient care.", + "author": "Mallett, S., Halligan, S., Thompson, M., Collins, G. S. & Altman, D. G.", + "venue": "\\JournalTitleBmj 345 (2012).", + "url": null + } + }, + { + "40": { + "title": "Reducing false arrhythmia alarms in the icu: The physionet/computing in cardiology challenge 2015.", + "author": "Clifford, G. D., Silva, I., Moody, B. & Mark, R.", + "venue": "\\JournalTitleComputing in Cardiology 42, 273\u2013276 (2015).", + "url": null + } + }, + { + "41": { + "title": "Mimic-iv.", + "author": "Johnson, A. et al.", + "venue": "\\JournalTitlePhysioNet. Available online at: https://physionet. org/content/mimiciv/1.0/(accessed August 23, 2021) (2020).", + "url": null + } + }, + { + "42": { + "title": "Prediction of the mortality risk in peritoneal dialysis patients using machine learning models: a nation-wide prospective cohort in korea.", + "author": "Noh, J. et al.", + "venue": "\\JournalTitleScientific reports 10, 1\u201311 (2020).", + "url": null + } + }, + { + "43": { + "title": "Covid-19 patient health prediction using boosted random forest algorithm.", + "author": "Iwendi, C. et al.", + "venue": "\\JournalTitleFrontiers in public health 8, 357 (2020).", + "url": null + } + }, + { + "44": { + "title": "Development and external evaluation of predictions models for mortality of covid-19 patients using machine learning method.", + "author": "Li, S. et al.", + "venue": "\\JournalTitleNeural Computing and Applications 1\u201310 (2021).", + "url": null + } + }, + { + "45": { + "title": "Covid-19 prediction and detection using machine learning algorithms: Catboost and linear regression.", + "author": "Kim, J. S.", + "venue": "\\JournalTitleAmerican Journal of Theoretical and Applied Statistics 10, 208\u2013215 (2021).", + "url": null + } + }, + { + "46": { + "title": "A clinically applicable approach to continuous prediction of future acute kidney injury.", + "author": "Toma\u0161ev, N. et al.", + "venue": "\\JournalTitleNature 572, 116\u2013119 (2019).", + "url": null + } + }, + { + "47": { + "title": "Dynamic electronic health record detection (detect) of individuals at risk of a first episode of psychosis: a case-control development and validation study.", + "author": "Raket, L. L. et al.", + "venue": "\\JournalTitleThe Lancet Digital Health 2, e229\u2014e239 (2020).", + "url": null + } + }, + { + "48": { + "title": "Doctor ai: Predicting clinical events via recurrent neural networks.", + "author": "Choi, E., Bahadori, M. T., Schuetz, A., Stewart, W. F. & Sun, J.", + "venue": "In Machine learning for healthcare conference, 301\u2013318 (PMLR, 2016).", + "url": null + } + }, + { + "49": { + "title": "Dynamic and explainable machine learning prediction of mortality in patients in the intensive care unit: a retrospective study of high-frequency data in electronic patient records.", + "author": "Thorsen-Meyer, H.-C. et al.", + "venue": "\\JournalTitleThe Lancet Digital Health 2, e179\u2013e191 (2020).", + "url": null + } + }, + { + "50": { + "title": "Machine learning for real-time prediction of complications in critical care: a retrospective study.", + "author": "Meyer, A. et al.", + "venue": "\\JournalTitleThe Lancet Respiratory Medicine 6, 905\u2013914 (2018).", + "url": null + } + }, + { + "51": { + "title": "Learning models for forecasting hospital resource utilization for covid-19 patients in canada.", + "author": "Zhang, J., Pathak, H. S., Snowdon, A. & Greiner, R.", + "venue": "\\JournalTitleScientific reports 12, 1\u201314 (2022).", + "url": null + } + }, + { + "52": { + "title": "Long-term mortality risk stratification of liver transplant recipients: real-time application of deep learning algorithms on longitudinal data.", + "author": "Nitski, O. et al.", + "venue": "\\JournalTitleThe Lancet Digital Health 3, e295\u2013e305 (2021).", + "url": null + } + }, + { + "53": { + "title": "Certifying and removing disparate impact.", + "author": "Feldman, M., Friedler, S. A., Moeller, J., Scheidegger, C. & Venkatasubramanian, S.", + "venue": "In proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, 259\u2013268 (2015).", + "url": null + } + }, + { + "54": { + "title": "Equality of opportunity in supervised learning.", + "author": "Hardt, M., Price, E. & Srebro, N.", + "venue": "\\JournalTitleAdvances in neural information processing systems 29 (2016).", + "url": null + } + }, + { + "55": { + "title": "Three naive bayes approaches for discrimination-free classification.", + "author": "Calders, T. & Verwer, S.", + "venue": "\\JournalTitleData mining and knowledge discovery 21, 277\u2013292 (2010).", + "url": null + } + } + ], + "url": "http://arxiv.org/html/2209.07805v4" +} \ No newline at end of file