This story draft by @escholar has not been reviewed by an editor, YET.

Results and discussion

EScholar: Electronic Academic Papers for Scholars HackerNoon profile picture

Authors:

(1) Rasoul Samani, School of Electrical and Computer Engineering, Isfahan University of Technology and this author contributed equally to this work;

(2) Mohammad Dehghani, School of Electrical and Computer Engineering, University of Tehran, Tehran, Iran and this author contributed equally to this work ([email protected]);

(3) Fahime Shahrokh, School of Electrical and Computer Engineering, Isfahan University of Technology.

Table of Links

Abstract and 1. Introduction

2. Related Works

3. Methodology and 3.1 Data

3.2 Data preprocessing

3.3. Predictive models

4. Evaluation

4.1. Evaluation metrics

4.2. Results and discussion

5. Conclusion and References

4.2. Results and discussion

The dataset initially comprised 51,113 records, which underwent preprocessing resulting in 49,083 records. All these records were utilized to construct classification models. It's noteworthy that the dataset exhibits a high degree of imbalance, and to maintain realism and promote better generalization, no balancing techniques were applied. The data was then partitioned into three subsets: 70% for training, amounting to 34,358 records, 15% for validation, containing 7,363 records, and the remaining 15% for testing, also consisting of 7,362 records. Table 1 provide the distribution of each class.


Table 1: statics of dataset.


Table 2 presents the results obtained from various classifiers employed in our study. Notably, the Final Method, which combines the BDSS model with MLP, outperformed the state-of-the-art models in terms of AUC. Furthermore, this model, along with logistic regression, achieved the highest accuracy, recall, and F1-score, underscoring the continued relevance of machine learning models. In Figure 5, the ROC curve illustrates the performance of different models, with the Final Method achieving an impressive AUC of 75%, surpassing all other models. Remarkably, logistic regression exhibited superior performance with a rate of 73.2%, outperforming alternative machine learning techniques.


In the medical domain, metrics like recall and AUC play a crucial role in evaluating AI models. Recall, which measures the ability of a model to correctly identify positive cases, is particularly important in healthcare settings where identifying all potential cases is paramount. Similarly, AUC provides an overall measure of model performance and is widely used for assessing predictive models in medical applications. The Final Method, leveraging the BDSS model, is considered the best model due to its superior performance in terms of recall and AUC. This model is trained on discharge summaries data and harnesses the power of BDSS, which is pre-trained on a large corpus of text data and is adept at understanding the semantic nuances of text.


Table 2: Results of proposed models.


Figure 5: ROC curve.


One advantage of the logistic regression model is its clarity and interpretability. To gain insights into the model's decision-making process, we extracted and presented the features that exerted the most significant impact on the outcomes, as shown in Figure 6. As observed, words such as "milliliter," "mg," and "chronic" had the greatest influence on categorizing patients as readmitted. This can be attributed to the prescription of various drugs with specific doses by the medical practitioners during the patient's discharge. The higher the number of prescribed drugs, the higher the likelihood of patient readmission. Conversely, the presence of words like "without," "family," "negative," "normal," and "transferred" in the patient's discharge text had the most substantial impact on categorizing patients as non-returning to the hospital.


Figure 6: The most effective words in the classification of patients in the logistic regression model.


Several previous studies have investigated models for predicting ICU readmission, with logistic regression consistently demonstrating favorable results, achieving AUC rates of 65% [49], 66% [50], and 70% [51]. However, a recent study by Orangi-Fard et al. [41], utilized various machine learning techniques on the MIMIC-III dataset to predict patient readmission. Their SVM-RBF model achieved an AUC rate of 74%. It's worth noting that Orangi-Fard et al. only utilized a portion of the dataset (4000 for training and 6000 for validation), balanced data, and employed 825 features. In contrast to previous approaches, our study took a comprehensive approach by utilizing the entire dataset, including imbalanced data. Additionally, we focused solely on textual features, omitting other factors such as demographics. This deliberate choice allowed us to gain deeper insights into the specific aspects we aimed to explore. Furthermore, while previous studies solely relied on machine learning models, our study also incorporated deep learning methods. This highlights the novelty and potential advantages of leveraging deep learning techniques in predicting ICU readmission. Table 3 provide a comparison with existing methods based on AUC metric.



This paper is available on arxiv under CC BY-NC-ND 4.0 DEED license.


L O A D I N G
. . . comments & more!

About Author

EScholar: Electronic Academic Papers for Scholars HackerNoon profile picture
EScholar: Electronic Academic Papers for Scholars@escholar
We publish the best academic work (that's too often lost to peer reviews & the TA's desk) to the global tech community

Topics

Around The Web...

Trending Topics

blockchaincryptocurrencyhackernoon-top-storyprogrammingsoftware-developmenttechnologystartuphackernoon-booksBitcoinbooks