Sensors 23 00970 v2
Sensors 23 00970 v2
Article
FedSepsis: A Federated Multi-Modal Deep Learning-Based
Internet of Medical Things Application for Early Detection of
Sepsis from Electronic Health Records Using Raspberry Pi and
Jetson Nano Devices
Mahbub Ul Alam * and Rahim Rahmani
Department of Computer and Systems Sciences, Stockholm University, 16407 Stockholm, Sweden
* Correspondence: [email protected]
Abstract: The concept of the Internet of Medical Things brings a promising option to utilize various
electronic health records stored in different medical devices and servers to create practical but secure
clinical decision support systems. To achieve such a system, we need to focus on several aspects, most
notably the usability aspect of deploying it using low-end devices. This study introduces one such
application, namely FedSepsis, for the early detection of sepsis using electronic health records. We
incorporate several cutting-edge deep learning techniques for the prediction and natural-language
processing tasks. We also explore the multimodality aspect for the better use of electronic health
records. A secure distributed machine learning mechanism is essential to building such a practical
internet of medical things application. To address this, we analyze two federated learning techniques.
Moreover, we use two different kinds of low-computational edge devices, namely Raspberry Pi
and Jetson Nano, to address the challenges of using such a system in a practical setting and report
the comparisons. We report several critical system-level information about the devices, namely
CPU utilization, disk utilization, process CPU threads in use, process memory in use (non-swap),
process memory available (non-swap), system memory utilization, temperature, and network traffic.
Citation: Alam, M.U.; Rahmani, R. We publish the prediction results with the evaluation metrics area under the receiver operating
FedSepsis: A Federated Multi-Modal characteristic curve, the area under the precision–recall curve, and the earliness to predict sepsis in
Deep Learning-Based Internet of
hours. Our results show that the performance is satisfactory, and with a moderate amount of devices,
Medical Things Application for Early
the federated learning setting results are similar to the single server-centric setting. Multimodality
Detection of Sepsis from Electronic
provides the best results compared to any single modality in the input features obtained from the
Health Records Using Raspberry Pi
electronic health records. Generative adversarial neural networks provide a clear superiority in
and Jetson Nano Devices. Sensors
2023, 23, 970. https://doi.org/
handling the sparsity of electronic health records. Multimodality with the generative adversarial
10.3390/s23020970 neural networks provides the best result: the area under the precision–recall curve is 96.55%, the area
under the receiver operating characteristic curve is 99.35%, and earliness is 4.56 h. FedSepsis suggests
Academic Editors: Francesco
that incorporating such a concept together with low-end computational devices could be beneficial
Picariello and Francesco Lamonaca
for all the medical sector stakeholders and should be explored further.
Received: 14 December 2022
Revised: 6 January 2023 Keywords: Internet of Medical Things; smart healthcare; clinical decision support system; deep
Accepted: 11 January 2023 learning; federated learning; multi-modality; natural language processing; electronic health records;
Published: 14 January 2023 early sepsis detection
the opportunity to help the healthcare system and medical facilities improve their overall
aspects. We refer to this relatively new structural and usability paradigm shift as the Inter-
net of Medical Things (IoMT). Healthcare professionals, patients, and other stakeholders
can be interconnected through different electronic devices to exchange valuable related
information with each other. Considering this, the structural (connection, communication,
and information generation) and the usability (application) aspects are vital components in
IoMT [2–4]. IoMT successfully utilizes the new wave of information explosion, the scale
of inter-connectivity among electronic devices that we have never seen before, and the
healthcare revolution transforming our notion of traditional healthcare systems. Effective
prevention of diseases, patient-centric healthcare, real-time distance-based monitoring with
automatic diagnosis support tools, improved collaboration among caregivers and patients,
sustainable health and longevity, and low-cost healthcare for everyone are critical examples
of IoMT applications [5]. We hope that this patient-centric approach will help reduce
hospital costs, errors, and visit times. It will ensure effective distance-based management
and remote monitoring of the patient’s health which will also help achieve sustainable
development by reducing the urbanization trend as remote areas tend not to have better
medical facilities. One of the keys to ensuring this is by employing effective clinical decision
support systems to detect the disease using patient data early [6].
Clinical decision support systems (CDSSs) accommodate medical professionals in their
decision-making process by providing intelligent and relevant information to diagnose,
monitor, or manage patients’ health. A well-designed CDSS can enhance the overall
performance by providing valid results, which should be convenient and confident in
its prediction [7]. The notable examples of CDSS are disease detection/classification,
sequential prediction of clinical events, concept embedding, data augmentation, and de-
identification [8].
One way to ensure the excellent performance of a CDSS is by efficiently utilizing the
medical data. IoMT can generate vast amounts of medical data. A crucial example of this
type of data is electronic health records (EHRs). EHRs can be described as the digital records
of all types of medical information of patients stored in different medical infrastructures.
Therefore, it is ideal to utilize EHRs to create a robust CDSS using the IoMT setup. EHR
can be extraordinarily sparse and heterogeneous, consisting of prescription notes about
medicines; diagnosis history; clinical text from doctors and nurses; image-based, numeric,
and categorical information from body vitals and laboratory tests, and demographic infor-
mation about the patient [9,10]. This exceptional heterogeneity of the data and sensitivity
of the information introduces some fascinating but challenging problems to consider before
utilizing it in a CDSS. The effective utilization of EHRs should ensure improved predictability
in terms of diagnosis. The heterogeneity aspect should also be handled carefully to obtain
the optimum performance from the CDSS. The data should be explored with the utmost
security and privacy to avoid information breaches.
Machine learning (ML) [11] is a solid choice to ensure improved predictability in CDSSs;
which is a compendium of probabilistic computational methods to predict a decision using
data for a particular task. Usually, several parameter values are fine-tuned to achieve such a
high-accuracy prediction. These parameters are primarily specific to the data that have been
used in an ML method. This effective fine-tuning of the parameters is known as ’learning
the model’. Deep learning (DL) [12] is a section of ML. The input feature information
is transformed in a hierarchical fashion so that a DL model can learn the representation
by organising the lower-level information to a gradually higher-level one. The key idea
is that this in-between transformation that bridges the map between input and output
prediction should have minimal human feature engineering involvements from the input
to the final prediction. It can be composed using deep neural networks, which can use a
considerable amount of hyper-parameters to learn the useful features from the input data
directly as a complex function of functions. It can be argued that it mimics the human
learning process as we also gradually develop our knowledge over the iteration of different
Sensors 2023, 23, 970 3 of 28
research on this. Section 3 provides an overview of the different methods and data pro-
cessing we used for implementing FedSepsis. We also provide a detailed description of the
sepsis-detection problem here. In Section 4, we discuss the experimental setup of FedSepsis
application and explain the evaluation metrics used in different experiments. Section 5
reports the results of different experiments. Here, we also report critical system-level
information about the Raspberry Pi and Jetson Nano devices. Finally, in Section 6, we
discuss the results and their importance in a broader context.
The code implementation of FedSepsis can be obtained using the following link (ac-
cessed on 6 January 2023): https://github.com/anondo1969/FedSepsis.
2. Related Works
The primary focus of our study is to investigate the usability of the low-end edge
devices in a federated IoMT-based CDSS application for CDSS tasks. In this section, we
summarized the importance of low-end edge devices by analyzing the recent related works.
Kairouz et al. [17] provided an in-depth overview of the advances and open problems
in FL. One key challenge identified in the platform development and deployment area is the
difference in hardware among different devices. They mainly focused on the performance
and device stability and mentioned that running computation must not affect it. Communi-
cation, storage, and computational capabilities of different devices are categorized as a core
challenge in an FL setting by Li et al. [28]. Their primary concerns were the power (bat-
tery level) and hardware (CPU (central processing unit), memory). Scarce computational
resources and relatively small storage capacity of the devices are mentioned by Bonawitz
et al. [29], discussing the scaling issue of federated learning with the practical system
design case. Gao et al. [30] used Raspberry Pi devices and investigated the performance of
speech commands and electrocardiogram-related tasks in the FL setup. Orescanin et al. [31]
investigated the fine-tuning performance in FL setting using Raspberry Pi devices using
a pre-trained MobileNetV2 model on the CelebA dataset. The comparison of different
edge devices was not performed in these works. The abovementioned discussions indicate
the challenges to using low-end computational devices and highlight why it should be
explored more in an FL setting to obtain the best performance.
3.1. Dataset
We used EHRs from the MIMIC-III (Multiparameter Intelligent Monitoring in Intensive
Care) database, version 1.4 [32]. It comprises over 58,000 hospital admissions of over
45,000 patients between June 2001 and October 2012.
Figure 1. A care episode having with 48 time-windows (care episode length = 48).
Figure 2. Histogram of different care episode lengths (left) and missing rates (right).
Non-Text
Clinical Text
Vital Data Demographic Data Laboratory Data
Systolic Blood Pressure Gender Albumin Nursing Notes
Diastolic Blood Pressure Admission Age Bands (Immature Neutrophils) Physician Notes
Mean Blood Pressure Ethnicity Bicarbonate
Respiratory Rate Admission Type Bilirubin
Heart Rate Admission Location Creatinine
SpO2 (Pulsoxymetry) Chloride
Temperature Celsius Sodium
Cardiac Output Potassium
Tidal Volume Set Lactate
Tidal Volume Observed Hematocrit
Tidal Volume Spontaneous Hemoglobin
Peak Inspiratory Pressure Platelet Count
Total Peep Level Partial Thromboplastin Time
O2 flow INR (Standardized Quick)
FiO2 (Fraction of Inspired Oxygen) Blood Urea Nitrogen
White Blood Cells
Creatine Kinase
Creatine Kinase MB
Fibrinogen
Lactate Dehydrogenase
Magnesium
Calcium (free)
pO2 Bloodgas
pH Bloodgas
pCO2 Bloodgas
SO2 Bloodgas
Glucose
Troponin T
Prothrombin Time (Quick)
h = f ( Xi ) (1)
Usually, a hidden state is needed to preserve the previous information from a care
episode. This state can be interpreted as additional features from the previous time-windows.
If we want to predict whether the current stage is positive or not at a given time-window
then we need to consider both the input Xt and the hidden state from the previous time
step ht−1 , as can be seen in the following equation:
h t = f ( x t , h t −1 ) (2)
h is the hidden states (1) composed from the previous time-windows. We can use it to
(2) predict the subsequent outcome. Usually, LSTM sub-divides these two usages into two
Sensors 2023, 23, 970 8 of 28
distinct variables ht and C. C symbolizes the LSTM hidden cell. Now we can control the
information flow using the three gates we mentioned above,
1
σ( x) = (9)
(1 + exp(− x ))
If the output sum of a sigmoid function is equal to one, then it is known as the soft-max
function. Usually, a log form of soft-max function is used for prediction for multi-label
classification because we can then consider the outputs as probability scores. Therefore it
can be regarded as a probabilistic classification. tanh( x ) also takes a real-valued input and
provides the output in a range between [−1, 1],
Huang [40]. Both ClinicalBERT-Alsentzer and ClinicalBERT-Huang used the clinical texts from
the MIMIC-III v1.4 database [32].
ClinicalBERT-Alsentzer aggregates the 15 types of clinical texts into discharge sum-
mary type or non-discharge summary type data and then tokenized them using the Scis-
paCy [45]. These tokenized sentences were used to fine-tune the original BERT base model.
ClinicalBERT-Huang additionally fine-tuned the model for the hospital-readmission predic-
tion task. In BERT, the maximum token length is 512 which can be limited for long clinical
texts. ClinicalBERT-Huang tackles it by splitting the long text and combining these multiple
predictions for one global prediction in an efficient way.
4. Experimental Setup
In this section, we discuss several aspects of the experimental setup to address the
research question described in Section 1.
final text embeddings. First, the feature vector was generated from the average of all the
layers (with a total length of (768). This is referred to as short-text-embedding. Second, the
feature vector was generated from the concatenation of all the layers (with a total length of
768*4=3072). This is referred to as long-text-embedding.
Name Values
hidden layers 1, 2, 3
neurons 512, 800, 1024, 2048
drop out 0, 10, 20, 30
epochs 1, 2
classification function log-softmax
optimizer Adam optimizer
GAIN alpha 0.1, 1, 10, 100
GAIN mini-batch 512
GAIN hint rate 0.9
GAIN total iterations 15,000
Figure 4. Raspberry Pi (left) and Jetson Nano (right) edge devices used in the experiments.
Sensors 2023, 23, 970 11 of 28
Socket programming [46] was used to establish the federated learning (discussed in
Section 3.2) setup. In socket programming we can connect two nodes on a network to
establish a communication channel. A server usually creates the listener socket where
multiple clients can connect. PyTorch [47] framework was used to develop the deep
learning models.
4.6. Experiments
We conducted several experiments to evaluate the performance of early detection
of sepsis using multi-modal data. We compared the performance in a federated learning
based setting using two different kinds of edge devices. The experiments are classified into
following categories.
Figure 5. Schematic diagram of the proposed model for the single-server setting. For details please
see Section 4.6.1.
limitation constrain compared to the server, the mini-batch size used for training multi-
modal and clinical-text models was 500, and for non-text model it was 1000. The details about
the devices and federated learning setting is discussed in Section 4.5. Both simple and opt
(described in Section 3.2) aggregation techniques were used to create the global RNN-LSTM
model. Due to time-constraint issue we only conducted the experiments compared to the
best three models (based on AUPRC score) obtained from Experiment A (Section 4.6.1). As
the training dataset is relatively small, therefore to overcome the overfitting issue we used
a total of five rounds for each experiment. Along with the evaluation results we also report
the system information of the client devices.
TP
Precision = (11)
TP + FP
TP
Recall = (12)
TP + FN
Area under the receiver operating characteristic curve (AUROC) and area under the
precision–recall curve (AUPRC) were used to evaluate the model performance.
For a classification task with a positive and a negative class, a model will estimate the
probability for an example belonging to the positive class, and if this value surpasses a
threshold, the example will be classified as positive. The AUROC score is based on the true
positive rate (the proportion of correct classifications of positive examples) and the false
positive rate (the proportion of negative examples classified as belonging to the positive
class). By varying the threshold for when an example is considered as belonging to the
positive class, pairs of true positive rates and false positives rate will be obtained. These
pairs can be plotted as a curve, with the false positive rate on the x-axis, and the true
positive rate on they y-axis. This curve is known as the receiver operating characteristic
and the AUROC value corresponds to the area under this curve.
Reviewing both precision and recall is useful in cases where there is an imbalance
in the observations between the two classes. Specifically, there are many examples of no
event (class 0) and only a few examples of an event (class 1). The reason for this is that
typically the large number of class 0 examples means we are less interested in the skill of
the model at predicting class 0 correctly, e.g., high true negatives. The key to the calculation
of precision and recall is that the calculations do not make use of the true negatives. It is
only concerned with the correct prediction of the minority class, class 1. A precision–recall
curve is a plot of the precision (y-axis) and the recall (x-axis) for different thresholds.
The earliness in predicting sepsis is calculated based on the median prediction time
from the sepsis onset in hours for all the true positive cases. A prediction score threshold
of >0.5 is used to determine the correct prediction, and we retained the first positive time-
window value of the care episode to calculate earliness. Therefore, the larger value of earliness
represents better prediction in general. As earliness is calculated based on true positives
only, we should consider it together with AUROC and AUPRC values to obtain a more
balanced overview of the model’s performance.
Table 4 shows the client device system information topics that we reported in
Section 5.3.
Sensors 2023, 23, 970 13 of 28
5. Results
In this section, we present the results obtained by conducting different experiments
discussed in Section 4.6.
Table 6 shows the results obtained using only the clinical-text data. We can see that
the result is comparatively worse and it is quite challenging to select a better model as
different evaluation metrics provides best results in different rounds. The best evaluation
scores obtained as AUPRC: 14.06%, AUROC: 59.99%, and earliness: 7.28 h.
Text
AUPRC AUROC Earliness
clinicalBERT Embeddings Missingness Representation
(%) (%) (h)
Representation
Huang short GAIN-imputation 14.06 57.51 6.35
Huang long distinct-value 13.97 58.50 6.53
Alsentzer short carry-forward-mean-imputation 13.70 59.48 7.15
Alsentzer short distinct-value 13.46 59.99 6.49
Alsentzer long carry-forward-mean-imputation 13.44 56.31 5.77
Alsentzer long distinct-value 13.35 56.80 7.28
Alsentzer short GAIN-imputation 13.06 57.11 6.70
Huang short carry-forward-mean-imputation 12.91 56.93 6.68
Alsentzer long GAIN-imputation 12.36 55.72 6.26
Huang long GAIN-imputation 12.32 55.17 6.43
Huang short distinct-value 12.30 54.91 7.00
Huang long carry-forward-mean-imputation 11.69 53.07 6.62
Table 7 shows the results obtained using both the clinical-text and non-text data. It can
be seen that using multi-modal setting provides the best results. The best evaluation scores
obtained as AUPRC: 96.55%, AUROC: 99.35%, and earliness: 4.58 h.
Sensors 2023, 23, 970 14 of 28
Table 7. Evaluation results using multi-modal (both clinical-text and non-text) input features in single-
server setting.
Text
AUPRC AUROC Earliness
clinicalBERT Embeddings Missingness Representation
(%) (%) (h)
Representation
Huang short GAIN-imputation 96.55 99.35 4.56
Huang long GAIN-imputation 96.09 98.90 4.53
Alsentzer long GAIN-imputation 94.04 97.90 4.45
Alsentzer short GAIN-imputation 93.95 98.61 4.58
Huang long distinct-value 31.95 82.90 3.71
Huang short distinct-value 31.31 81.74 3.86
Alsentzer short distinct-value 31.06 82.68 3.94
Huang long carry-forward-mean-imputation 28.98 81.25 3.87
Huang short carry-forward-mean-imputation 28.61 80.57 3.71
Alsentzer long carry-forward-mean-imputation 28.37 81.85 3.58
Alsentzer long distinct-value 24.33 79.60 3.90
Alsentzer short carry-forward-mean-imputation 24.23 77.71 3.98
5.2.1. Raspberry Pi
Table 8 shows the results obtained using non-text data in a federated learning setting
using 10 Raspberry Pi devices. We can see that federated aggregation technique simple
performed better over opt. The best evaluation scores obtained as AUPRC: 71.39%, AUROC:
97.60%, and earliness: 4.54 h.
Table 9 shows the results obtained using clinical-text data in a federated learning setting
using 10 Raspberry Pi devices. Although the result is comparatively worse, but similar to
before the result improves after each round and aggregation technique simple performed
better over opt. The best evaluation scores obtained as AUPRC: 14.62%, AUROC: 58.91%,
and earliness: 7.73 h.
Table 10 shows the results obtained using multi-modal data in a federated learning
setting using 10 Raspberry Pi devices. multi-modal model provides the best results. It is
worth noting that the best AUPRC score was obtained after the round 4. The best evaluation
scores obtained as AUPRC: 84.80%, AUROC: 98.49%, and earliness: 4.54 h.
Table 8. Evaluation results using non-text input features in a federated learning setting using 10
Raspberry Pi devices.
Table 9. Evaluation results using clinical-text input features in a federated learning setting using 10
Raspberry Pi devices.
Table 10. Evaluation results using multi-modal (both clinical-text and non-text) input features in a
federated learning setting using 10 Raspberry Pi devices.
Figure 6 provides the loss comparison for different federated aggregation techniques
using 10 Raspberry Pi devices for the training of the multi-modal model. We can see that the
loss value gradually decreased after each round and it is much smoother in the case of the
aggregation technique simple over opt.
Figure 6. Loss comparison for different federated aggregation techniques using 10 Raspberry Pi
devices. The x-axis represents the incremental number of training mini-batches, the y-axis represents
loss-values.
Sensors 2023, 23, 970 16 of 28
Table 11 shows the results obtained using multi-modal data in a federated learning
setting using different number of Raspberry Pi devices. It provides insights into the
empirical settings regarding data and the number of devices. We can see that the best
results were obtained using only one or two devices, and after increasing it to six and more
devices the AUPRC score decreased. This suggests that using a relatively small amount of
data can increase the chance of overfitting, and we have to be careful in that case. Choosing
less number of devices may be a walk around in that case. The best evaluation scores
obtained using two devices as AUPRC: 99.20%, and AUROC: 99.91%, and using eight
devices the earliness score is 4.53 h.
Table 11. Evaluation results using multi-modal (both clinical-text and non-text) input features in a
federated learning setting using a different number of Raspberry Pi devices.
Table 12. Evaluation results using non-text input features in a federated learning setting using three
Jetson Nano devices.
Table 13 shows the results obtained using clinical-text data in a federated learning
setting using three Jetson Nano devices. We can observe that the performance is significantly
worse, similar to the Raspberry Pi cases. It suggests that it may be possible that the text
modality alone sometimes may not be a wise choice as input features. The best evaluation
scores obtained as AUPRC: 15.12%, AUROC: 63.58%, and earliness: 7.73 h.
Sensors 2023, 23, 970 17 of 28
Table 13. Evaluation results using clinical-text input features in a federated learning setting using
three Jetson Nano devices.
Table 14 shows the results obtained using multi-modal data in a federated learning
setting using three Jetson Nano devices. Once again, the multi-modal representation of
the input features provided the best results. Compared to Table 11 the results with three
devices provides a better AUPRC score. The best evaluation scores obtained as AUPRC:
98.99%, AUROC: 99.89%, and earliness: 4.56 h.
Table 14. Evaluation results using multi-modal (both clinical-text and non-text) input features in a
federated learning setting using three Jetson Nano devices.
Figure 7 provides the loss comparison for different federated aggregation techniques
using three Jetson Nano devices for the training of multi-modal model. We can see that the
loss value was gradually decreased after each round and it is much smoother in the case of
the aggregation technique simple over opt.
Figure 7. Loss comparison for different federated aggregation methods using Jetson Nano devices.
x-axis represents the incremental number of training mini-batches, y-axis represents loss-values.
Sensors 2023, 23, 970 18 of 28
Table 15 shows the results obtained using multi-modal data in a federated learning
setting using different Jetson Nano devices. As it can be observed, with two devices,
the best AUPRC score was obtained. The best evaluation scores obtained as AUPRC:
99.36%, AUROC: 99.93% with two devices, and earliness: 4.51 h with three devices. The best
performance here is slightly worse than the best performance obtained with the similar
experiments using Raspberry Pi devices, as can be compared with Table 11.
Table 15. Evaluation results using multi-modal (both clinical-text and non-text) input features in a
federated learning setting using different number of Jetson Nano devices.
Figure 8. CPU utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis
represents the time in minutes, the y-axis represents the CPU utilization percentage (%).
Figure 9 provides the disk utilization comparison for different Raspberry Pi and Jetson
Nano devices. The total storage memory capacity for 8 of the 10 Raspberry Pi devices was
40 Gigabytes (GB). The rest had a capacity of 128 GB. All three Jetson Nano devices had a
storage memory capacity of 128 GB. We can see that the disk utilization is also consistent
Sensors 2023, 23, 970 19 of 28
in all the relatively low-storage memory devices. Even with a 40 GB total capacity, the
disk utilization is below 60% for the Raspberry Pi devices. This suggests that we need to
optimize the disk storage given the dataset size.
Figure 9. Disk utilization comparison for different Raspberry Pi and Jetson Nano devices. The x-axis
represents the time in minutes, the y-axis represents the disk utilization percentage (%).
Figure 10 provides the comparison regarding the number of ’CPU process threads
in use’ for different Raspberry Pi and Jetson Nano devices. We can see that it is entirely
consistent across all the devices, and in almost all the cases, it is close to 12. It suggests that
if we can effectively utilize all the cores of the CPU, then we may obtain a faster and better
parallel performance, as can be seen in these graphs.
Figure 10. Process CPU threads in use comparison for different Raspberry Pi and Jetson Nano devices.
The x-axis represents the time in minutes, the y-axis represents the number of CPU threads in use.
Sensors 2023, 23, 970 20 of 28
Figure 11 provides the comparison of process memory in use for different Raspberry
Pi and Jetson Nano devices. Process memory is the collection of processes responsible
for obtaining, preserving, and retrieving information. We can see that process memory
utilization is relatively low in Raspberry Pi devices, and it is even much less (with a bit of
irregularity) in Jetson Nano devices.
Figure 11. Process memory in use comparison for different Raspberry Pi and Jetson Nano de-
vices. The x-axis represents the time in minutes, the y-axis represents the used process memory in
megabytes (MB).
Figure 12 provides the process memory available comparison for different Raspberry
Pi and Jetson Nano devices. If we observe this figure jointly with Figure 11 then we can see
that Raspberry Pi uses significantly less process memory compared to Jetson Nano devices.
Even with the small dataset and relatively small DL models used in FedSepsis experiments,
it takes almost all the process memory in Jetson Nano. Therefore, Raspberry Pi is suited for
more data and more complex training models.
Figure 13 provides the system memory utilization comparison for different Raspberry
Pi and Jetson Nano devices. System memory or random access memory (RAM) stores
data and instructions temporarily so that the CPU can use it later by accessing it randomly.
Similar to process memory, system memory utilization is also much higher in Jetson
Nano devices than Raspberry Pi devices. We need to be very cautious about the dataset
size (mini-batch size) and the complexity of the DL training model while utilizing Jetson
Nano devices.
Figure 14 provides the network traffic comparison for different Raspberry Pi and
Jetson Nano devices. The data-size pattern is very similar as we exchanged the same types
of information for every round. If the model is more complex, the data size will be bigger.
Therefore, network usage should be carefully constructed and monitored.
Figure 15 provides the temperature comparison for different Raspberry Pi and Jetson
Nano devices. We can see that the temperature obtains significantly high for Raspberry
Pi devices. As we used a relatively small dataset and simple DL training model there, on
average, it took 40 to 50 min to complete one multi-modal training. Therefore, we could
provide sufficient time-gap to cool down the devices. In a more practical scenario, using a
cooling fan is advisable due to this high temperature during the training process.
Sensors 2023, 23, 970 21 of 28
Figure 12. Process memory available comparison for different Raspberry Pi and Jetson Nano devices.
The x-axis represents the time in minutes, the y-axis represents the available process memory in
megabytes (MB).
Figure 13. System memory utilization comparison for different Raspberry Pi and Jetson Nano devices.
The x-axis represents the time in minutes, the y-axis represents the system memory utilization in
percentage (%).
Sensors 2023, 23, 970 22 of 28
Figure 14. Network traffic comparison for different Raspberry Pi and Jetson Nano devices. The x-axis
represents the time in minutes, the y-axis represents network traffic in bytes.
Sensors 2023, 23, 970 23 of 28
Figure 15. Temperature comparison for different Raspberry Pi and Jetson Nano devices. The x-axis
represents the incremental number of training mini-batches, the y-axis represents temperature in
Celsius (◦ C).
6. Discussion
In this study, we introduced FedSepsis, an FL-based IoMT application for the early
prediction of sepsis using DL and EHRs. We utilized two recent federated aggregation
algorithms for the FL setup. We incorporated several cutting-edge concepts and methods,
such as multimodality, ClinicalBERT for the NLP, RNN-LSTM neural networks, and GAN to
implement FedSepsis. Moreover, to investigate the aspects of practical usability, we deployed
two popular low-end edge devices, namely Raspberry Pi and Jetson Nano. To provide
a more balanced and multifaceted evaluation, we presented the results using AUPRC,
AUROC, and earliness in terms of hours. We also reported system-level information on the
low-end devices to provide insights into using such devices in a practical setting. Instead of
focusing on the model improvement only, we also emphasized the applicability aspect by
integrating cutting-edge techniques with a keen interest in low-edge devices. Therefore, we
heavily invested in analyzing the system-level aspects of these low-edge devices. We expect
that by taking this combined approach, our FedSepsis application will be more effective and
ensure far-reaching benefits.
We used genuine (de-identified) EHRs to extract the input features. EHRs are relatively
sparse, or the missingness in EHR is generally quite high, which was in our case also. To
tackle this issue, we used a GAN-based technique which proved to be successful as the best
results were obtained based on this technique to handle missingness.
Representing the clinical text from the EHRs for prediction tasks is one of the most
challenging problems in the NLP domain. To investigate this issue in-depth, we used two
latest ClinicalBERT models and two different representations from each. The comparison
in terms of best performance is inconclusive as both models performed relatively well
along with different representations. This study concatenated the text representations (text
embeddings) with the non-text input features as a combined multi-modal input feature. In
the future, we will explore different training combinations to attach it to different layers
of RNN-LSTM.
Incorporating multimodality provided a precise and superior performance over any
form of single-modal input features. It also supports the real-time diagnosis procedure
Sensors 2023, 23, 970 24 of 28
and suggests that in future, we should investigate further in utilizing this concept more
efficiently. Effective utilization of the image data is one possible option here. The overall
optimization in the training model should also be fine-tuned according to different modali-
ties, as we can see that different modalities may carry different degrees of importance in
terms of the overall prediction.
Compared to the FL setting we used, the single-server setting’s storage (one terabyte)
and computational capacity were considerably superior to the FedSepsis application per-
formance. Therefore, a bit of degrading performance in the FL setup should be taken for
granted compared to the single-server setup. We also showed that the performance is close
to the single-server setup for this task with a smaller number of devices (five, for example).
It suggests that we need to be careful about the device choice and the data distribution in
an FL setting not to overfit the model. The model weight aggregation techniques also need
to be explored further so that the effective loss optimization can be fine-tuned according to
the model or the data we have for a particular task.
For the early sepsis prediction task, we reported AUROC, AUPRC, and earliness prediction
in hours. AUROC and AUPRC provide a more balanced and clinically relevant evaluation
of the prediction as we should be more careful about the true positives. Overall the AUPRC
score is satisfactory. However, it should be considered together with earliness to understand the
effectiveness of the prediction. On average, the earliness is close to 4.5 h which is satisfactory
given the sparsity of the data and the variance in the care episode lengths.
The key aim of this study is to evaluate the performance difference of the low-end edge
devices as the research question mentioned in section 1. We can categorize this performance
into model-prediction-level and system-level perspectives.
There is a clear performance difference in model prediction while incorporating GAIN
for handling missingness. On average, the AUPRC improved 23.17%, and the AUROC
improved 6.89% compared to the next imputation strategy across all non-text, clinical-text,
and multi-modal cases in the single-server setting. The combination of clinical-text with
non-text or the multi-modal case always provided superior performance in all three settings
(single-server, Raspberry Pi, and Jetson Nano). In the single-server setting, for the multi-
modal case, the AUPRC improvement is 9.25%, and the AUROC improvement is 1.04%,
compared to the non-text modality. Using Raspberry Pi, the AUPRC improvement is 11.06%,
and the AUROC improvement is 0.89%. Using Jetson Nano the AUPRC improvement
is 4.56%, and the AUROC improvement is 0.39%. The federated aggregation technique
(discussed in Section 3.2) simple performed slightly better than the technique opt. For the
multi-modal case, compared to opt technique, using Raspberry Pi the AUPRC improvement
is 1.86%, and the AUROC improvement is 0.43% in using the simple technique. Using Jetson
Nano, the AUPRC improvement is 3.99%, and the AUROC improvement is 0.54%. Overall,
Jetson Nano provided better performance than Raspberry Pi. For the multi-modal case,
compared to Raspberry Pi, using Jetson Nano, the AUPRC improvement is 16.55%, and the
AUROC improvement is 1.40%. Using fewer devices (three) compared to Raspberry Pi (10)
and the GPU could be one reason for this improvement. It is worth noting that different
combinations of ClinicalBERT-based text embedding representation (discussed in Section 4.3)
did not provide significant performance differences. It should be investigated further in
the future.
Section 5.3 provides an in-depth overview of the performance difference from a system-
level perspective. For a faster implementation, Jetson Nano can be a viable choice; on the
other hand, Raspberry Pi is less memory exhaustive, given that the temperature issue
is effectively handled. Effective CPU utilization can also be achieved by exploring the
feature importance issue instead of using all the features. Reducing unimportant features
could enhance the speed and efficient CPU utilization. The device system information
provides valuable insights regarding the practical usability of FedSepsis. CPU utilization,
process memory, and system memory are crucial to ensure optimum performance in these
low-computational devices, and, therefore, the training model or the data to create input
features should be fine-tuned accordingly. Device temperature is also a critical issue to
Sensors 2023, 23, 970 25 of 28
consider to ensure that it runs effectively. Network traffic could be vital, and effective
allocation needs to be ensured, especially when we have low power and lossy network
(LLN), typical for such IoMT infrastructures.
EHRs are either stored or generated in different medical equipment and servers across
different hospital physical locations. We should be very cautious about combining or
storing these precious and copious amounts of data in a single place from cost–benefit and
security–privacy aspects. As CDSSs can be built with these EHRs, which can substantially
positively impact the healthcare system, effective interconnection among the medical
equipment and servers should be paramount. IoMT can be one excellent solution for this
distributed platform. It can be expanded further with a minimal additional interconnection
cost to connect it with other hospitals, even across different countries, while ensuring the
utmost security and privacy to create a global impact. This has the potential to revolutionize
the healthcare industry with a long-lasting impact to help every stakeholder.
We must be cautious about the security and privacy issue as the EHRs are extraor-
dinarily vulnerable, and breaching the information could bring disastrous after-effects.
Therefore, the IoMT platform and applications should ensure the reliability, security, and
safety aspects as much as possible [48]. Advanced level data intelligence, supported by
cutting-edge DL and ML techniques, should be deployed to enable the IoMT platform and
applications to be interactive, stateful, iterative, adaptive, and contextual. Along with NLP,
we should also focus on other cross-disciplinary domains to make it more effective, such
as speech recognition [49], computer vision [50], dialogue and narrative generation, and
human–computer interaction.
Effective data aggregation with the help of metadata aggregation [51], and medical
sensors [52] can provide us with new insights. As the inter-relations of different medical
conditions or diseases are ubiquitous phenomena among patients (comorbidity), incorpo-
rating this aspect (for example, general infection detection from various input sources and
outcomes) could be an exciting option to explore further. The optimum use of the low-end
edge devices will ensure that these new generation computer systems are cost-effective,
patient-centric, robust, and are here to create a long-lasting impact, which will help all
parties involved, including the patients and the healthcare professionals.
One potential limitation regarding the acceptance of FedSepsis as a CDSS to the medical
community is the lack of interpretability of the black box model. The medical significance
should be investigated [53] as well while predicting the model to make it more acceptable.
We can denote the medical significance of a symptom or disease process as a type of clinical
manifestation affecting the regular functionalities of organ systems, tissues, living cells, or
organs, for example, in the chest area [54]. We will explore this issue in the future.
7. Conclusions
In this study, we presented an Internet of Medical Things-based application, namely
FedSepsis, for early detection of sepsis as an example of a promising clinical decision
support system. We used cutting-edge deep learning methods and relevant concepts, such
as multimodality, natural-language processing and federated learning to make a robust
and practical application. Moreover, we utilized low-end edge devices, such as Raspberry
Pi and Jetson Nano, to show the usability aspects of such applications. Multimodality
incorporating clinical-text data provided improved performance. Jetson Nano yielded better
results from both the prediction-level and system-level perspectives. The results suggest
that FedSepsis is viable to use and can be an ideal choice given that the interpretability
aspect is also ensured to a degree as the prediction scores are entirely satisfactory and the
possibility of relatively low-cost device usage can benefit all the stakeholders. In the future,
sensor integration, interpreting the medical significance of the prediction, and enhancing
the security aspects will be explored.
Sensors 2023, 23, 970 26 of 28
References
1. Plsek, P.E.; Greenhalgh, T. The challenge of complexity in health care. Bmj 2001, 323, 625–628. [CrossRef] [PubMed]
2. Irfan, M.; Ahmad, N. Internet of Medical Things: Architectural model, motivational factors and impediments. In Proceedings of
the 2018 15th Learning and Technology Conference (L&T), Jeddah, Saudi Arabia, 25–26 February 2018; pp. 6–13.
3. Joyia, G.J.; Liaqat, R.M.; Farooq, A.; Rehman, S. Internet of Medical Things (IoMT): Applications, benefits and future challenges
in healthcare domain. J. Commun. 2017, 12, 240–247. [CrossRef]
4. Manogaran, G.; Chilamkurti, N.; Hsu, C.H. Emerging trends, issues, and challenges in Internet of Medical Things and wireless
networks. Pers. Ubiquitous Comput. 2018, 22, 879–882. [CrossRef]
5. Goyal, S.; Sharma, N.; Bhushan, B.; Shankar, A.; Sagayam, M. Iot enabled technology in secured healthcare: applications,
challenges and future directions. In Cognitive Internet of Medical Things for Smart Healthcare; Springer International Publishing:
Cham, Switzerland, 2021; pp. 25–48.
6. Kukhtevich, I.; Goryunova, V.; Goryunova, T.; Zhilyaev, P. Medical Decision Support Systems and Semantic Technologies
in Healthcare. In Proceedings of the Russian Conference on Digital Economy and Knowledge Management (RuDEcK 2020),
Voronezh, Russia, 27–29 February 2020; Atlantis Press: Amsterdam, The Netherlands, 2020; pp. 370–375. [CrossRef]
7. Souza-Pereira, L.; Ouhbi, S.; Pombo, N. Quality-in-use characteristics for clinical decision support system assessment. Comput.
Methods Programs Biomed. 2021, 207, 106169. [CrossRef] [PubMed]
8. Xiao, C.; Choi, E.; Sun, J. Opportunities and challenges in developing deep learning models using electronic health records data:
A systematic review. J. Am. Med. Inform. Assoc. 2018, 25, 1419–1428. [CrossRef]
9. Müller, H.; Unay, D. Retrieval From and Understanding of Large-Scale Multi-modal Medical Datasets: A Review. IEEE Trans.
Multimed. 2017, 19, 2093–2104. [CrossRef]
10. Itani, S.; Lecron, F.; Fortemps, P. Specifics of medical data mining for diagnosis aid: A survey. Expert Syst. Appl. 2019, 118, 300–314.
[CrossRef]
11. Molnar, C. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable. 2022. Available online:
https://christophm.github.io/interpretable-ml-book/ (accessed on 6 January 2023).
12. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016.
13. Cai, Q.; Wang, H.; Li, Z.; Liu, X. A survey on multimodal data-driven smart healthcare systems: approaches and applications.
IEEE Access 2019, 7, 133583–133599. [CrossRef]
14. Huang, S.C.; Pareek, A.; Seyyedi, S.; Banerjee, I.; Lungren, M.P. Fusion of medical imaging and electronic health records using
deep learning: A systematic review and implementation guidelines. NPJ Digit. Med. 2020, 3, 1–9. [CrossRef]
15. Baltrušaitis, T.; Ahuja, C.; Morency, L.P. Multimodal machine learning: A survey and taxonomy. IEEE Trans. Pattern Anal. Mach.
Intell. 2018, 41, 423–443. [CrossRef]
16. Ramachandram, D.; Taylor, G.W. Deep multimodal learning: A survey on recent advances and trends. IEEE Signal Process. Mag.
2017, 34, 96–108. [CrossRef]
17. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings,
R.; et al. Advances and open problems in federated learning. arXiv 2019, arXiv:1912.04977.
18. Briggs, C.; Fan, Z.; Andras, P. A Review of Privacy Preserving Federated Learning for Private IoT Analytics. arXiv 2020,
arXiv:2004.11794.
19. Xu, J.; Glicksberg, B.S.; Su, C.; Walker, P.; Bian, J.; Wang, F. Federated learning for healthcare informatics. J. Healthc. Inform. Res.
2021, 5, 1–19. [CrossRef]
20. Singer, M.; Deutschman, C.S.; Seymour, C.W.; Shankar-Hari, M.; Annane, D.; Bauer, M.; Bellomo, R.; Bernard, G.R.; Chiche,
J.D.; Coopersmith, C.M.; et al. The third international consensus definitions for sepsis and septic shock (Sepsis-3). JAMA 2016,
315, 801–810. [CrossRef]
21. Ferrer, R.; Martin-Loeches, I.; Phillips, G.; Osborn, T.M.; Townsend, S.; Dellinger, R.P.; Artigas, A.; Schorr, C.; Levy, M.M. Empiric
antibiotic treatment reduces mortality in severe sepsis and septic shock from the first hour: Results from a guideline-based
performance improvement program. Crit. Care Med. 2014, 42, 1749–1755. [CrossRef]
22. Kumar, A.; Roberts, D.; Wood, K.E.; Light, B.; Parrillo, J.E.; Sharma, S.; Suppes, R.; Feinstein, D.; Zanotti, S.; Taiberg, L.; et al.
Duration of hypotension before initiation of effective antimicrobial therapy is the critical determinant of survival in human septic
shock. Crit. Care Med. 2006, 34, 1589–1596. [CrossRef]
Sensors 2023, 23, 970 27 of 28
23. Johnston, S.J.; Cox, S.J. The Raspberry Pi: A Technology Disrupter, and the Enabler of Dreams. Electronics 2017, 6, 51. [CrossRef]
24. Cass, S. Nvidia makes it easy to embed AI: The Jetson nano packs a lot of machine-learning power into DIY projects-[Hands on].
IEEE Spectr. 2020, 57, 14–16. [CrossRef]
25. Otter, D.W.; Medina, J.R.; Kalita, J.K. A survey of the usages of deep learning for natural language processing. IEEE Trans. Neural
Netw. Learn. Syst. 2020, 32, 604–624. [CrossRef]
26. Alam, M.U.; Rahmani, R. Cognitive Internet of Medical Things Architecture for Decision Support Tool to Detect Early Sepsis
Using Deep Learning. In Proceedings of the Biomedical Engineering Systems and Technologies, BIOSTEC 2020, Valletta, Malta,
24–26 February 2020; Revised Selected Papers; Ye, X., Soares, F., De Maria, E., Gómez Vilda, P., Cabitza, F., Fred, A., Gamboa, H.,
Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 366–384.
27. Alam, M.U.; Rahmani, R. Federated Semi-Supervised Multi-Task Learning to Detect COVID-19 and Lungs Segmentation Marking
Using Chest Radiography Images and Raspberry Pi Devices: An Internet of Medical Things Application. Sensors 2021, 21, 5025.
[CrossRef] [PubMed]
28. Li, T.; Sahu, A.K.; Talwalkar, A.; Smith, V. Federated Learning: Challenges, Methods, and Future Directions. IEEE Signal Process.
Mag. 2020, 37, 50–60. [CrossRef]
29. Bonawitz, K.; Eichner, H.; Grieskamp, W.; Huba, D.; Ingerman, A.; Ivanov, V.; Kiddon, C.; Konečný, J.; Mazzocchi, S.; McMahan,
H.B.; et al. Towards Federated Learning at Scale: System Design. Proc. Mach. Learn. Syst. 2019, 1, 374–388.
30. Gao, Y.; Kim, M.; Abuadbba, S.; Kim, Y.; Thapa, C.; Kim, K.; Camtepe, S.A.; Kim, H.; Nepal, S. End-to-End Evaluation of
Federated Learning and Split Learning for Internet of Things. arXiv 2020, arXiv:2003.13376.
31. Orescanin, M.; Ergezer, M.; Singh, G.; Baxter, M. Federated Fine-Tuning Performance on Edge Devices. In Proceedings of
the 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA), Online, 13–15 December 2021;
pp. 1174–1181. [CrossRef]
32. Johnson, A.E.; Stone, D.J.; Celi, L.A.; Pollard, T.J. The MIMIC Code Repository: Enabling reproducibility in critical care research.
J. Am. Med. Inform. Assoc. 2018, 25, 32–39. [CrossRef]
33. Seymour, C.W.; Liu, V.X.; Iwashyna, T.J.; Brunkhorst, F.M.; Rea, T.D.; Scherag, A.; Rubenfeld, G.; Kahn, J.M.; Shankar-Hari, M.;
Singer, M.; et al. Assessment of clinical criteria for sepsis: For the Third International Consensus Definitions for Sepsis and Septic
Shock (Sepsis-3). JAMA 2016, 315, 762–774. [CrossRef]
34. Vincent, J.L.; Moreno, R.; Takala, J.; Willatts, S.; De Mendonça, A.; Bruining, H.; Reinhart, C.; Suter, P.; Thijs, L. The SOFA
(Sepsis-related Organ Failure Assessment) score to describe organ dysfunction/failure. Intensive Care Med. 1996, 22, 707–710.
[CrossRef]
35. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from
decentralized data. In Proceedings of the Artificial Intelligence and Statistics, PMLR, Fort Lauderdale, FL, USA, 20–22 April 2017;
pp. 1273–1282.
36. Reddi, S.; Charles, Z.; Zaheer, M.; Garrett, Z.; Rush, K.; Konečnỳ, J.; Kumar, S.; McMahan, H.B. Adaptive federated optimization.
arXiv 2020, arXiv:2003.00295.
37. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536.
[CrossRef]
38. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [CrossRef]
39. Yoon, J.; Jordon, J.; Schaar, M. Gain: Missing data imputation using generative adversarial nets. In Proceedings of the International
Conference on Machine Learning, PMLR, Baltimore, Maryland, USA, 17–23 July 2018; pp. 5689–5698.
40. Huang, K.; Altosaar, J.; Ranganath, R. ClinicalBERT: Modeling Clinical Notes and Predicting Hospital Readmission. arXiv 2019,
arXiv:1904.05342 .
41. Alsentzer, E.; Murphy, J.; Boag, W.; Weng, W.H.; Jin, D.; Naumann, T.; McDermott, M. Publicly Available Clinical BERT
Embeddings. In Proceedings of the 2nd Clinical Natural Language Processing Workshop, Minneapolis, MN, USA, 7–10 June
2019; pp. 72–78. [CrossRef]
42. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding.
arXiv 2018, arXiv:1810.04805.
43. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial
nets. Adv. Neural Inf. Process. Syst. 2014, 27, 139–144.
44. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need.
Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008.
45. Neumann, M.; King, D.; Beltagy, I.; Ammar, W. ScispaCy: Fast and robust models for biomedical natural language processing.
arXiv 2019, arXiv:1902.07669.
46. Goerzen, J. Foundations of Python Network Programming; Apress: New York, NY, USA, 2004.
47. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic
differentiation in PyTorch. In Proceedings of the NIPS-W 2017, Long Beach, CA, USA, 4–9 December 2017.
48. Hatzivasilis, G.; Soultatos, O.; Ioannidis, S.; Verikoukis, C.; Demetriou, G.; Tsatsoulis, C. Review of security and privacy for the
Internet of Medical Things (IoMT). In Proceedings of the 2019 15th International Conference on Distributed Computing in Sensor
Systems (DCOSS), Santorini, Greece, 29–31 May 2019; pp. 457–464.
Sensors 2023, 23, 970 28 of 28
49. Malik, M.; Malik, M.K.; Mehmood, K.; Makhdoom, I. Automatic speech recognition: A survey. Multimed. Tools Appl. 2021,
80, 9411–9457. [CrossRef]
50. Goel, A.; Tung, C.; Lu, Y.H.; Thiruvathukal, G.K. A survey of methods for low-power deep learning and computer vision. In
Proceedings of the 2020 IEEE 6th World Forum on Internet of Things (WF-IoT), New Orleans, LA, USA, 2–16 June 2020; pp. 1–6.
51. Alam, M.U.; Rahmani, R. Intelligent context-based healthcare metadata aggregator in internet of medical things platform. Procedia
Comput. Sci. 2020, 175, 411–418. [CrossRef]
52. Ray, P.P.; Dash, D.; Kumar, N. Sensors for Internet of Medical Things: State-of-the-art, security and privacy issues, challenges and
future directions. Comput. Commun. 2020, 160, 111–131. [CrossRef]
53. Alam, M.U.; Baldvinsson, J.R.; Wang, Y. Exploring LRP and Grad-CAM visualization to interpret multi-label-multi-class pathology
prediction using chest radiography. In Proceedings of the 2022 IEEE 35th International Symposium on Computer-Based Medical
Systems (CBMS), Shenzhen, China, 21–22 July 2022; pp. 258–263. [CrossRef]
54. Goodman, L.R. Felson’s Principles of Chest Roentgenology, a Programmed Text; Elsevier Health Sciences: Amsterdam, The Netherlands, 2014.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.