Next Article in Journal
Altered Treg Infiltration after Discoidin Domain Receptor 1 (DDR1) Inhibition and Knockout Promotes Tumor Growth in Lung Adenocarcinoma
Previous Article in Journal
SOX2-OT Binds with ILF3 to Promote Head and Neck Cancer Progression by Modulating Crosstalk between STAT3 and TGF-β Signaling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Chest Radiograph Simulation Technique with Deep Convolutional Neural Networks: A Proof-of-Concept Study

1
Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong
2
Department of Radiation Oncology, Duke University Medical Center, Durham, NC 27708, USA
3
The Affiliated Cancer Hospital of Zhengzhou University & Henan Cancer Hospital, Zhengzhou 450008, China
4
The Hong Kong Polytechnic University Shenzhen Research Institute, Shenzhen 518000, China
*
Authors to whom correspondence should be addressed.
Cancers 2023, 15(24), 5768; https://doi.org/10.3390/cancers15245768
Submission received: 9 November 2023 / Revised: 30 November 2023 / Accepted: 4 December 2023 / Published: 8 December 2023

Abstract

:

Simple Summary

Dynamic chest radiographs offer a distinct advantage over traditional chest radiographs by integrating motion and functional data, elevating their significance in clinical diagnostics. This study introduces a pioneering technique employing deep neural networks to simulate respiratory lung motion and extract local functional details from single-phase chest X-rays, thereby enhancing lung cancer clinical diagnostic capabilities. Our research establishes the viability of generating patient-specific respiratory motion profiles from single-phase chest radiographs. The evaluation of results from the network developed here underscores its substantial accuracy and fidelity, affirming its robustness in providing valuable supplementary insights into pulmonary function.

Abstract

In this study, we present an innovative approach that harnesses deep neural networks to simulate respiratory lung motion and extract local functional information from single-phase chest X-rays, thus providing valuable auxiliary data for early diagnosis of lung cancer. A novel radiograph motion simulation (RMS) network was developed by combining a U-Net and a long short-term memory (LSTM) network for image generation and sequential prediction. By utilizing a spatial transformer network to deform input images, our proposed network ensures accurate image generation. We conducted both qualitative and quantitative assessments to evaluate the effectiveness and accuracy of our proposed network. The simulated respiratory motion closely aligns with pulmonary biomechanics and reveals enhanced details of pulmonary diseases. The proposed network demonstrates precise prediction of respiratory motion in the test cases, achieving remarkable average Dice scores exceeding 0.96 across all phases. The maximum variation in lung length prediction was observed during the end-exhale phase, with average deviation of 4.76 mm (±6.64) for the left lung and 4.77 mm (±7.00) for the right lung. This research validates the feasibility of generating patient-specific respiratory motion profiles from single-phase chest radiographs.

1. Introduction

Chest radiography is one of the most frequently used modalities for routine pulmonary disease diagnosis and is commonly utilized as the first exploratory study, as it has the advantages of high availability, low radiation dose, and an efficient workflow. It can suffice as a preliminary examination for conditions including pneumonia, lung cancer, emphysema, and pulmonary fibrosis [1]. In the clinical process, radiologists manually locate anomalies on the radiograph and diagnose the pathology after the picture has been acquired. Then, depending on the circumstances, further diagnostic imaging procedures may be recommended to obtain specialized data for a more precise diagnosis.
Dynamic chest X-ray (DXR) is a functional imaging technique that uses a flat-panel detector (FPD) to generate sequential images. The large field of view and real-time observation can provide information about diaphragm kinetics, pulmonary ventilation, and circulation [2]. It can be deployed as a simple and rapid means of functional imaging, but currently DXR machines have not been widely deployed in clinical practice. Furthermore, the total patient dose of a DXR image is 0.23 mSv, about twice that of a conventional chest X-ray (CXR) [3]. Although dynamic chest X-rays may provide extra functional information, the high dose and scarcity of clinical deployment make it difficult to apply them in routine medical practice. In contrast, conventional chest X-ray machines are widely deployed and expose patients to a lower radiation dose [4].
In recent years, advances in engineering and computer science have enabled researchers to reveal a wealth of hidden information encoded in chest radiographs [5], and thereby improve clinical diagnosis. A decrease in X-ray pixel value suggests the presence of a localized air space, which might indicate abnormalities such as a lung cyst, emphysema, bulla, or pneumothorax. Conversely, an increase in the pixel value may indicate a reduction in pulmonary air or intensified tissues, pointing to conditions such as pulmonary inflammation, fibrosis, edema, or sclerosis [6]. By leveraging the capabilities of artificial intelligence (AI), researchers are able to extract a significantly greater amount of information compared to relying solely on observations with the human eye [7]. Approaches like image enhancement [8,9,10,11], organ segmentation [12,13,14], and feature extraction [15,16,17] have been investigated and have achieved promising results. Previous studies have considered whether chest X-rays contain rich three-dimensional and functional information that could be extracted by AI to provide extra information that would improve clinical diagnosis [18]. One previous study tried to predict patients’ lung function to evaluate the need for mechanical ventilation in hospitalized patients based on chest X-ray images and achieved 90.06% accuracy [19]. Liyue et al. explored a deep learning approach for tomographic X-ray imaging with a single-view projection data input, introducing the concept of dimension transformation in the image feature domain to enable volumetric imaging using either single or multiple 2D projections [18]. The feasibility of transforming inter-modality medical imaging with deep learning has also been studied by earlier investigators [20,21].
Motivated by such earlier research, in this study, we utilize the power of deep learning and convolutional neural networks to simulate respiratory lung motion and relative functional information from single-phase digital reconstructed radiographs (DRRs). This technique could be used on conventional chest X-rays to provide auxiliary information for clinical lung cancer diagnosis like dynamic chest X-rays while maintaining the convolutional chest X-ray’s low dosage and high availability.

2. Methods

2.1. Datasets

In this study, 60 patients’ 4D CT sequences were obtained from several online and local datasets [22,23,24]. The distribution of cases from each dataset is summarized in Table 1. The present study was conducted with the approval of the Institutional Review Board of the Affiliated Cancer Hospital of Zhengzhou University and Henan Cancer Hospital. Due to the scarcity of DXR FPD and DXR data, we generated sequential DRRs from 4D CT to simulate the DXR. During the preprocessing stage, thresholding was first applied to segment the lung volume from the patients’ bodies. With less interference from the contiguous bone components, abnormalities could be readily visualized, allowing the information to be interpreted more precisely. After that, beam eye view (BEV) ray-casting was utilized to generate the DRR. Then, minimum bounding boxes were generated by detecting the boundary of the projected lung volume, and these served as the basis for DRR segmentation. To simulate the screening procedure of chest X-rays, the end-of-inhale (EOI) phase served as the input to the network and the subsequent phases were set as targets. We randomly selected 40 out of 60 patients as the training set, while the remaining 20 patients served as the test set to evaluate the effectiveness of the model.

2.2. Network Design

In this study, a novel deep neural network was developed to predict the entire sequence of phases of a complete respiration cycle and the corresponding motion pattern from a single-phase chest X-ray. This technique could be applied to generate auxiliary diagnostic information from a chest X-ray without unnecessary radiation doses or screening time.
The general workflow is shown in Figure 1. The development of the proposed radiograph motion simulation (RMS) network mainly comprised two steps: image generation and sequential prediction. To predict the succeeding phases of the respiration cycle p 1 , p 2 , p 9 from the EOI phase p 0 , we adopted a recurrent neural network (RNN) embedded U-Net structure [25]. First, the input image p 0   was down-sampled by the encoder block of the U-Net to a one-dimensional feature vector x 0 , which was fed into the RNN for sequential prediction. The decoder block then up-sampled the predicted vector to generate the subsequent phases. Instead of explicitly predicting the images, we used the decoder to generate deformation vector fields (DVFs) and a spatial transformer network to deform the input image to generate accurate medical imaging without sacrificing anatomical detail.

2.2.1. Medical Imaging Generation Model

The schematic of the proposed RMS network is shown in Figure 2. The U-Net contracting path consists of four down-sampling blocks. Each block comprises the repeated application of two 3 × 3 convolutions (unpadded convolutions) followed by a rectified linear unit (ReLU) and a 2 × 2 max-pooling operation with stride 2 for down-sampling. Each convolutional operation involves a doubling of the feature channels. After the convolution and max-pooling operation in the contracting path, the down-sampled feature maps are flattened and sent to a fully connected layer to generate a 1 × 512 latent representation. A one-to-many long short-term memory (LSTM) [26] network is then utilized to predict the sequential latent representations. As a kind of recurrent neural network that uses a single input to generate a sequence of outputs, it utilizes the LSTM cells’ ability to maintain an internal state across time steps to selectively retain or discard information. The two-layer network takes as input the previous phase’s latent vector x t 1 , and outputs the latent vector of the next time step x t , which is then fed into the network as the input for the next time step until the maximum time step t = 9 is reached. From the LSTM, a sequence of latent representations with 9 phases was generated.
After sequential prediction by the LSTM, the latent vectors were recovered to the same size as the input image through a symmetrically expanding path. Skip-connections between the encoders and decoders ensured information fusion and precise prediction. The expanding path of the U-Net output 9 DVFs ( D 0 , D 1 , D 2 D 9 ) for the subsequent phases of the respiration cycle. A differentiable spatial transformer module [27] was concatenated after the U-Net to deform the initial phase image p 0 to p 1 9 according to the predicted DVFs. For each pixel V i t in the target phase image p t ,
V i t = n H m W p 0 n , m k h i t m ; Φ x k w i t n ; Φ y       i [ 1 H W ]
where H and W are the height and width of the initial phase image, and p 0 n , m is the value at location n , m in the input image. Each h i t , w i t coordinate in the corresponding phase’s DVF defines the spatial location in the input image at which a sampling kernel is applied, to obtain the value at a specific output pixel. The Φ x and Φ y are parameters of the sampling kernel; we use bilinear interpolation in this study.
After spatial transformation, the discrepancy between predicted phases and the ground truth images was computed and backpropagated to optimize the network parameter.

2.2.2. Loss Function Design

During the optimization, three loss functions were combined to promote precise motion simulation. The first loss function L s m o o t h penalizes the gradient in the predicted DVFs, which aims to foster smoother spatial variations, aligning with the principles of physical feasibility and promoting a more realistic representation of the lung motion.
L s m o o t h = 1 2 N t = 1 9 i = 0 N 1 ( ( D t i h ) 2 + ( D t i w ) 2 )
This computes the L2 loss of the DVF’s gradient in both directions. N denotes all the pixels in the DVFs and D t i is the i th pixel value in the phase t DVF.
The second loss function is the mean squared error (MSE) loss, measuring the discrepancy between the predicted phase image and the ground truth phase image. It can be expressed as
L M S E = t = 1 9 p t g t p t p r e 2
where p t g t is the ground truth image for phase t while p t p r e represents the predicted image for phase t . Beyond the MSE loss, we also utilized the local cross-correlation loss to improve model robustness, which is computed as
L N C C = i p t , i g t p ^ t , i g t p t , i p r e p ^ t , i p r e 2 ( i p t , i g t p ^ t , i g t 2 ) ( i p t , i p r e p ^ t , i p r e 2 )
where p t , i g t , p t , i p r e denote the ground truth and predicted value of pixel i in phase t , and p ^ t , i g t , p ^ t , i p r e are the corresponding pixel values calculated using the sliding window technique. The final loss was computed as the sum of these three loss functions:
L = L s m o o t h + L M S E + L N C C

2.2.3. Evaluation Metrics

The accuracy and authenticity of the simulated motion were evaluated using both qualitative and quantitative methods. For qualitative evaluation, the Jacobian determinant of the DVFs was computed. The Jacobian determinant for each pixel shows how much the pixel was stretched or compressed during the deformation, which provides regional information regarding the expansion and contraction of the lung tissue during the respiration cycle to help further identify regions with inadequate or excessive ventilation and assess overall lung function.
  v x , y = J x , y 1 = 1 + u x ( x , y ) x u x ( x , y ) y u y ( x , y ) x 1 + u y ( x , y ) y
where v(x, y) is the ventilation of a volume at the point (x, y); J(x, y) is the Jacobian of the volume at the point (x, y); and ux, uy correspond to the x , y components of the DVF, respectively.
For quantitative evaluation, the Dice similarity coefficient (DSC) was computed between the predicted images and the ground truth images for each phase.
D S C = 2 × | p t g t p t p r e | | p t g t | + | p t p r e |
Diaphragmatic positions (i.e., the difference between the lung apex and diaphragm dome) were also measured to further analyze the accuracy of the predicted respiratory movements.

2.2.4. Experiment Setup

In this study, we implemented the proposed RMS framework using Python 3.8 and Pytorch 1.12.1 with CUDA 11.6 on a workstation equipped with an NVIDIA GeForce RTX 2080 Ti GPU. The training took approximately two hours. After training, it took less than three seconds to generate the simulated motion per patient.

3. Results

For qualitative evaluation, Figure 3 displays the predicted respiratory phases’ images and corresponding Jacobian determinant distribution for one test patient. The movement of the diaphragm can be interpreted by observing the distance between the diaphragm dome and the reference line. During the EOI phase ( p 0 ) and the end of exhale (EOX) phase ( p 4 ), as the lung muscles contract, the diaphragm moves upwards, resulting in a significant increase in the distance between the diaphragm dome and the reference. Subsequently, as air is inhaled and the lungs expand, the diaphragm gradually moves downwards, and the distance progressively decreases.
When analyzing the Jacobian determinant distribution, it becomes evident that all phases exhibit a higher contraction factor in the lower lung region compared to the upper lung region. Moreover, phases 4 and 5 demonstrate an overall higher contraction factor compared to the other phases. This observation highlights the regional variations in lung expansion and contraction throughout the respiratory cycle, providing valuable insights into the dynamics of lung function.
The phase-based Dice coefficient distribution of the test set is shown in Figure 4. For each phase, a normal distribution curve of best fit, with the corresponding mean and standard deviation (std), is included. As shown in the figure, the Dice scores of the phases close to EOI are generally better than those of the EOX phases. Meanwhile, the minimum Dice score of all phases is above 0.93, which indicates that the RMS framework is able to accurately predict the respiratory motion based on the input EOI image.
The variation in predicted and ground truth diaphragmatic positions is presented in Figure 5. It can be seen that for all phases, the majority of the cases’ variation is less than 10 mm for both lungs. The trend of inter-phase variation is also consistent with the Dice score; the prediction of the end of the sequence is more accurate and precise than the middle of the sequence.
Table 2 presents the mean and std of the Dice score and diaphragmatic position variation for all phases.
Given the absence of dynamic chest X-ray data, we conducted our network training using digital reconstructed radiograph (DRR) images to establish proof of concept. Subsequently, we sourced chest X-rays from the JSRT dataset [28] for external validation. These chest X-ray images were processed and integrated into the RMS framework for motion synthesis without further training. A representative example of the motion simulation results from one chest X-ray is illustrated in Figure 6. Notably, the relative motion of a nodule in the right upper lobe becomes readily apparent, alerting radiologists to the presence of abnormal lung function in this specific region.

4. Discussions

The chest radiograph continues to be one of the most widely used diagnostic tools for routine pulmonary diseases, for which there has traditionally been a reliance on manual interpretation by radiologists. With the advent of artificial intelligence and computer-assisted diagnosis (CAD) systems, there is persuasive evidence that a wealth of valuable information may be retrieved to improve diagnosis and clinical efficiency [29,30]. While numerous investigations have been made into develo** automated classification models based on extracted deep features, the application of AI for providing lung functional information to clinicians remains relatively unexplored. In contrast to directly generating diagnostic decisions, generative models offer additional information that can be harnessed by human radiologists, fostering a more human-centered clinical decision-making process. In this study, we aimed to develop an innovative generative Respiratory Motion Synthesis (RMS) system to generate patient-specific respiratory motion patterns and corresponding deformation vector fields using chest radiographs. The primary objective of this system is to provide clinicians with supplementary functional information, enhancing their understanding of patients’ respiratory dynamics.
The backbone of the RMS framework is a four-layer U-Net. Initially, the input chest radiographs undergo sequential convolution and max-pooling operations in the encoder blocks, resulting in down-sampled one-dimensional vectors with a length of 512. Subsequently, these vectors are fed into a one-to-many LSTM network, predicting the latent representation of subsequent phases. Then, they are up-sampled by the decoder blocks to generate DVFs with the same size as the input radiographs. Throughout the up-sampling process, skip connections are incorporated to pass features from the encoder path to the decoder path, enabling the recovery of lost spatial information and enhancing model performance. A subsequent spatial transformer network is utilized to deform the input radiograph to the target phases with the predicted DVFs. Compared with directly reconstructing the images from the decoder block, this prevents the loss of anatomical details in the images.
The prediction results of one patient randomly selected from the test set are visualized in Figure 3. It can be quantitively observed that the anatomical details of the lung are well-preserved in the subsequent phases. And the diaphragm motion is shown by the distance between the diaphragm dome and the reference line. The lung tissue’s two-dimensional contraction is visualized by calculating the Jacobian determinant. In the process of exhalation, as respiratory gases are gradually expelled from the lungs, a progressive diminution of the lung volume is observed. This progress of contraction was visualized by sequentially plotting the increasing value of the Jacobian determinant. It can also be observed that the Jacobian determinant of the lower lung region is larger than the upper lung region, representing a higher expansion or contraction factor during respiration, which is consistent with pulmonary biomechanics. The distribution appears smooth without significant fluctuations. The simulation results presented in Figure 6 further substantiate the practicality of integrating this technique into clinical workflows, offering radiologists valuable supplementary information to enhance diagnosis.
For quantitative analysis, the statistics of Dice coefficients between the ground truth and predicted phase images are presented in Figure 4 and Table 2. In the test set, our model consistently attained an average Dice score exceeding 0.96 across all phases, demonstrating its capability to accurately predict overall pulmonary deformation during the respiratory cycle. Notably, when comparing different phases, the inhale phases exhibited generally higher Dice coefficients than the exhale phases. Figure 5 further illustrates a similar pattern in the measurements of diaphragmatic positions. The observed trend could be attributed to increased variation in pulmonary motion and inconsistencies in phase gating. This issue may potentially be mitigated with sequences having a higher imaging frequency and a larger dataset. It also reveals that while most of the test data exhibited errors conforming to a normal distribution with modest mean values, a few data points displayed significant variation across all phases. This discrepancy is likely due to the presence of severe pulmonary diseases, where patients demonstrate relatively extreme motion dynamics that differ from the training dataset, and consequently elude accurate capture by the model. In conclusion, the developed model demonstrates the ability to effectively generate respiratory motion based on single-phase DRR, thereby providing valuable supplementary functional information for diagnosis. This capability holds the potential to enhance clinical efficiency, reduce costs, and facilitate improved patient outcomes.
However, it is important to acknowledge several limitations of this proof-of-concept study. Firstly, the model was trained using digitally reconstructed radiographs (DRRs) generated from 4D CT sequences, rather than sequential chest radiograph datasets. This disparity between the chest radiograph and DRRs may have resulted in a potential decrease in model performance. Secondly, the current dataset size is limited, which could potentially hinder the model’s generalizability across a wide range of clinical conditions. To address these limitations, future efforts will involve acquiring sequential chest radiograph data to fine-tune the model and comprehensively evaluate its performance. By incorporating sequential chest radiographs and expanding the dataset size, we aim to improve the model’s accuracy, robustness, and applicability in real-world clinical scenarios. Another potential area for enhancement is to include a nodule segmentation module in the proposed framework. Currently, in the absence of nodule-specific ground truth annotations within our dataset, our assessment relies on the aggregate motion of the entire lung. Consequently, this approach does not capture the distinct motion characteristics of the nodules. To address this, we intend to collect a dataset with nodule annotations and integrate a dedicated nodule segmentation module to refine the simulation’s precision and improve the accuracy and robustness of the simulation framework.

5. Conclusions

This study validated the feasibility of synthesizing patient-specific respiratory motion from a single-phase chest radiograph. Specifically, a novel deep convolutional neural network (RMS) was developed to predict the deformation vector fields of the subsequent phases and deform the initial radiograph with a spatial transformer. The test results of the proposed network have shown promising accuracy and authenticity, confirming its reliability as a tool for extracting additional pulmonary function information. Efficiently providing this information in a human-centered manner, without incurring any additional costs, gives this network the potential to significantly augment the clinical decision-making process.

Author Contributions

Conceptualization, D.Y., J.C. and G.R.; Methodology, D.Y., B.L. and Y.H.; Software, D.Y. and Y.H.; Validation, Y.H.; Resources, B.L.; Data curation, B.L.; Writing—original draft, D.Y.; Writing—review & editing, J.C. and G.R.; Supervision, J.C. and G.R.; Funding acquisition, B.L., J.C. and G.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly supported by the General Research Fund (15103520) of the University Research Committee, by the Health and Medical Research Fund (07183266, 09200576) of the Health Bureau, PolyU (UGC), by the RI-IWEAR Seed Project (P0044802), by the Shenzhen Science and Technology Program (JCYJ 20230807140403007),by the Henan Provincial Medical Science and Technology Research Project (SBGJ202103038), and by the Natural Science Foundation of Henan Province of China (222300420575 and 232300420231).

Institutional Review Board Statement

The present study was conducted with the approval of the Institutional Review Board of the Affiliated Cancer Hospital of Zhengzhou University and Henan Cancer Hospital.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets generated or analyzed during this study are not publicly available due to restrictions related to patient privacy but are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wielputz, M.O.; Heussel, C.P.; Herth, F.J.; Kauczor, H.U. Radiological diagnosis in lung disease: Factoring treatment options into the choice of diagnostic modality. Dtsch. Arztebl. Int. 2014, 111, 181–187. [Google Scholar] [CrossRef] [PubMed]
  2. Tanaka, R.; Sanada, S.; Okazaki, N.; Kobayashi, T.; Suzuki, M.; Matsui, T.; Matsui, O. Detectability of regional lung ventilation with flat-panel detector-based dynamic radiography. J. Digit. Imaging 2008, 21, 109–120. [Google Scholar] [CrossRef] [PubMed]
  3. Tamura, M.; Matsumoto, I.; Saito, D.; Yoshida, S.; Takata, M.; Tanaka, R.; Takemura, H. Dynamic chest radiography: Novel and less-invasive imaging approach for preoperative assessments of pleural invasion and adhesion. Radiol. Case Rep. 2020, 15, 702–704. [Google Scholar] [CrossRef]
  4. Johnston, D.A.; Brennan, P.C. Reference dose levels for patients undergoing common diagnostic X-ray examinations in Irish hospitals. Br. J. Radiol. 2000, 73, 396–402. [Google Scholar] [CrossRef] [PubMed]
  5. van Ginneken, B.; Hogeweg, L.; Prokop, M. Computer-aided diagnosis in chest radiography: Beyond nodules. Eur. J. Radiol. 2009, 72, 226–230. [Google Scholar] [CrossRef]
  6. Fraser, R.S.; Müller, N.L.; Colman, N.; Pare, P. Fraser and Paré’s Diagnosis of Diseases of the Chest; Wb Saunders: Philadelphia, PA, USA, 1999; Volumes 1–4. [Google Scholar]
  7. Qin, C.; Yao, D.; Shi, Y.; Song, Z. Computer-aided detection in chest radiography based on artificial intelligence: A survey. Biomed. Eng. Online 2018, 17, 113. [Google Scholar] [CrossRef] [PubMed]
  8. Anand, S.; Roshan, R.K.; Deiveega, S.M. Chest X ray image enhancement using deep contrast diffusion learning. Optik 2023, 279, 170751. [Google Scholar] [CrossRef]
  9. Lin, C.-H.; Wu, J.-X.; Li, C.-M.; Chen, P.-Y.; Pai, N.-S.; Kuo, Y.-C. Enhancement of Chest X-ray Images to Improve Screening Accuracy Rate Using Iterated Function System and Multilayer Fractional-Order Machine Learning Classifier. IEEE Photonics J. 2020, 12, 1–18. [Google Scholar] [CrossRef]
  10. Xu, L.; Zeng, X.; Huang, Z.; Li, W.; Zhang, H. Low-dose chest X-ray image super-resolution using generative adversarial nets with spectral normalization. Biomed. Signal Process. Control 2020, 55, 101600. [Google Scholar] [CrossRef]
  11. Ghosh, S.K.; Ghosh, A. ENResNet: A novel residual neural network for chest X-ray enhancement based COVID-19 detection. Biomed. Signal Process. Control 2022, 72, 103286. [Google Scholar] [CrossRef]
  12. Rahman, T.; Khandakar, A.; Kadir, M.A.; Islam, K.R.; Islam, K.F.; Mazhar, R.; Hamid, T.; Islam, M.T.; Kashem, S.; Mahbub, Z.B.; et al. Reliable Tuberculosis Detection Using Chest X-ray With Deep Learning, Segmentation and Visualization. IEEE Access 2020, 8, 191586–191601. [Google Scholar] [CrossRef]
  13. Souza, J.C.; Diniz, J.O.B.; Ferreira, J.L.; Da Silva, G.L.F.; Silva, A.C.; de Paiva, A.C. An automatic method for lung segmentation and reconstruction in chest X-ray using deep neural networks. Comput. Methods Programs Biomed. 2019, 177, 285–296. [Google Scholar]
  14. Stirenko, S.; Kochura, Y.; Alienin, O.; Rokovyi, O.; Gordienko, Y.; Gang, P.; Zeng, W. Chest X-ray Analysis of Tuberculosis by Deep Learning with Segmentation and Augmentation. In Proceedings of the 2018 IEEE 38th International Conference on Electronics and Nanotechnology (ELNANO), Kyiv, Ukraine, 24–26 April 2018; pp. 422–428. [Google Scholar]
  15. Yang, D.; Ren, G.; Ni, R.; Huang, Y.H.; Lam, N.F.D.; Sun, H.; Wan, S.B.N.; Wong, M.F.E.; Chan, K.K.; Tsang, H.C.H.; et al. Deep learning attention-guided radiomics for COVID-19 chest radiograph classification. Quant. Imaging Med. Surg. 2023, 13, 572–584. [Google Scholar] [CrossRef] [PubMed]
  16. Shankar, K.; Perumal, E.; Tiwari, P.; Shorfuzzaman, M.; Gupta, D. Deep learning and evolutionary intelligence with fusion-based feature extraction for detection of COVID-19 from chest X-ray images. Multimed. Syst. 2022, 28, 1175–1187. [Google Scholar] [CrossRef] [PubMed]
  17. Sharma, H.; Jain, J.S.; Bansal, P.; Gupta, S. Feature Extraction and Classification of Chest X-ray Images Using CNN to Detect Pneumonia. In Proceedings of the 2020 10th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, 29–31 January 2020; pp. 227–231. [Google Scholar]
  18. Shen, L.; Zhao, W.; **ng, L. Patient-specific reconstruction of volumetric computed tomography images from a single projection view via deep learning. Nat. Biomed. Eng. 2019, 3, 880–888. [Google Scholar] [CrossRef] [PubMed]
  19. Kulkarni, A.R.; Athavale, A.M.; Sahni, A.; Sukhal, S.; Saini, A.; Itteera, M.; Zhukovsky, S.; Vernik, J.; Abraham, M.; Joshi, A.; et al. Deep learning model to predict the need for mechanical ventilation using chest X-ray images in hospitalised patients with COVID-19. BMJ Innov. 2021, 7, 261–270. [Google Scholar] [CrossRef]
  20. Li, W.; Li, Y.; Qin, W.; Liang, X.; Xu, J.; **ong, J.; **e, Y. Magnetic resonance image (MRI) synthesis from brain computed tomography (CT) images based on deep learning methods for magnetic resonance (MR)-guided radiotherapy. Quant. Imaging Med. Surg. 2020, 10, 1223–1236. [Google Scholar] [CrossRef] [PubMed]
  21. Boulanger, M.; Nunes, J.C.; Chourak, H.; Largent, A.; Tahri, S.; Acosta, O.; De Crevoisier, R.; Lafond, C.; Barateau, A. Deep learning methods to generate synthetic CT from MRI in radiotherapy: A literature review. Phys. Med. 2021, 89, 265–281. [Google Scholar] [CrossRef]
  22. Castillo, R.; Castillo, E.; Guerra, R.; Johnson, V.E.; McPhail, T.; Garg, A.K.; Guerrero, T. A framework for evaluation of deformable image registration spatial accuracy using large landmark point sets. Phys. Med. Biol. 2009, 54, 1849–1870. [Google Scholar] [CrossRef] [PubMed]
  23. Kipritidis, J.; Tahir, B.A.; Cazoulat, G.; Hofman, M.S.; Siva, S.; Callahan, J.; Hardcastle, N.; Yamamoto, T.; Christensen, G.E.; Reinhardt, J.M.; et al. The VAMPIRE challenge: A multi-institutional validation study of CT ventilation imaging. Med. Phys. 2019, 46, 1198–1217. [Google Scholar] [CrossRef]
  24. Vandemeulebroucke, J.; Rit, S.; Kybic, J.; Clarysse, P.; Sarrut, D. Spatiotemporal motion estimation for respiratory-correlated imaging of the lungs. Med. Phys. 2011, 38, 166–178. [Google Scholar] [CrossRef] [PubMed]
  25. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015, Munich, Germany, 5–9 October 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  26. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  27. Jaderberg, M.; Simonyan, K.; Zisserman, A.; Kavukcuoglu, K. Spatial transformer networks. Adv. Neural Inf. Process. Syst. 2015, 2015, 2017–2025. [Google Scholar]
  28. Shiraishi, J.; Katsuragawa, S.; Ikezoe, J.; Matsumoto, T.; Kobayashi, T.; Komatsu, K.-I.; Matsui, M.; Fujita, H.; Kodera, Y.; Doi, K. Development of a Digital Image Database for Chest Radiographs With and Without a Lung Nodule. Am. J. Roentgenol. 2000, 174, 71–74. [Google Scholar] [CrossRef] [PubMed]
  29. Singh, R.; Kalra, M.K.; Nitiwarangkul, C.; Patti, J.A.; Homayounieh, F.; Padole, A.; Rao, P.; Putha, P.; Muse, V.V.; Sharma, A.; et al. Deep learning in chest radiography: Detection of findings and presence of change. PLoS ONE 2018, 13, e0204155. [Google Scholar] [CrossRef] [PubMed]
  30. Ahn, J.S.; Ebrahimian, S.; McDermott, S.; Lee, S.; Naccarato, L.; Di Capua, J.F.; Wu, M.Y.; Zhang, E.W.; Muse, V.; Miller, B.; et al. Association of Artificial Intelligence–Aided Chest Radiograph Interpretation with Reader Performance and Efficiency. JAMA Netw. Open 2022, 5, e2229289. [Google Scholar] [CrossRef]
Figure 1. Illustration of the general workflow. (a) Preprocessing steps: The lung volumes were initially segmented from the 4D CT body volumes using thresholding. Subsequently, ray-casting was employed to generate the sequential DRR images, which were then cropped to fit the minimum bounding box. (b) Sequential prediction steps: The preprocessed DRR images were down-sampled by the encoder to 1D latent vectors. The vectors were then fed into the RNN for sequential prediction. The predicted latent vectors for nine phases were decoded by the decoder to predict the corresponding DVFs. These DVFs were registered to the initial DRR to generate target phase images using a spatial transformer. Abbreviations: DRR—digital reconstructed radiograph; Bbox—bounding box; RNN—recurrent neural network; DVF—deformation vector field.
Figure 1. Illustration of the general workflow. (a) Preprocessing steps: The lung volumes were initially segmented from the 4D CT body volumes using thresholding. Subsequently, ray-casting was employed to generate the sequential DRR images, which were then cropped to fit the minimum bounding box. (b) Sequential prediction steps: The preprocessed DRR images were down-sampled by the encoder to 1D latent vectors. The vectors were then fed into the RNN for sequential prediction. The predicted latent vectors for nine phases were decoded by the decoder to predict the corresponding DVFs. These DVFs were registered to the initial DRR to generate target phase images using a spatial transformer. Abbreviations: DRR—digital reconstructed radiograph; Bbox—bounding box; RNN—recurrent neural network; DVF—deformation vector field.
Cancers 15 05768 g001
Figure 2. Schematic representation of the advanced RMS framework. The encoding path comprises four successive down-sampling blocks utilizing convolution and max-pooling operations. Sequential prediction is accomplished through a two-layer one-to-many LSTM network. The decoding path is constructed with four symmetrical up-sampling blocks employing de-convolution operations. Additionally, motion registration is achieved simultaneously through a differentiable spatial transformer.
Figure 2. Schematic representation of the advanced RMS framework. The encoding path comprises four successive down-sampling blocks utilizing convolution and max-pooling operations. Sequential prediction is accomplished through a two-layer one-to-many LSTM network. The decoding path is constructed with four symmetrical up-sampling blocks employing de-convolution operations. Additionally, motion registration is achieved simultaneously through a differentiable spatial transformer.
Cancers 15 05768 g002
Figure 3. Example test case of respiratory motion simulation. The first and third rows represent the phase images, while the second and final rows display corresponding Jacobian determinant visualizations. Reference lines are shown in the phase images to facilitate a comparison of diaphragm movement between different phases. Abbreviations: P—phase image; J—Jacobian determinant distribution.
Figure 3. Example test case of respiratory motion simulation. The first and third rows represent the phase images, while the second and final rows display corresponding Jacobian determinant visualizations. Reference lines are shown in the phase images to facilitate a comparison of diaphragm movement between different phases. Abbreviations: P—phase image; J—Jacobian determinant distribution.
Cancers 15 05768 g003
Figure 4. Histogram of phase-based Dice coefficient distribution. For each subfigure, the x-axis represents the Dice score of the predicted image, and the y-axis represents the probability density.
Figure 4. Histogram of phase-based Dice coefficient distribution. For each subfigure, the x-axis represents the Dice score of the predicted image, and the y-axis represents the probability density.
Cancers 15 05768 g004
Figure 5. Violin plot of phase-based diaphragmatic position variation. Blue denotes left lung and orange denotes right lung.
Figure 5. Violin plot of phase-based diaphragmatic position variation. Blue denotes left lung and orange denotes right lung.
Cancers 15 05768 g005
Figure 6. Simulated phase images of a JSRT chest X-ray sample. The region around the nodule was enlarged to improve the visibility of the relative motion.
Figure 6. Simulated phase images of a JSRT chest X-ray sample. The region around the nodule was enlarged to improve the visibility of the relative motion.
Cancers 15 05768 g006
Table 1. Dataset information.
Table 1. Dataset information.
Dataset ResourceNumber of Patients
DIR Lab 4D-CT dataset [22]10
POPI dataset [24]5
VAMPIRE challenge 4D CT dataset [23]12
Henan Cancer Hospital 33
Table 2. Phase-based prediction statistics.
Table 2. Phase-based prediction statistics.
Phase #123456789
Dice: m e a n   ( ± s t d ) 0.975
( ± 0.0075)
0.969
( ± 0.0112)
0.968
( ± 0.0112)
0.966
( ± 0.0108)
0.964
( ± 0.0114)
0.966
( ± 0.0137)
0.970
( ± 0.0146)
0.972
( ± 0.0122)
0.977
( ± 0.0070)
LLE (mm):
m e a n   ( ± s t d )
4.00
( ± 2.60)
4.73
( ± 3.29 )
5.80
( ± 6.42 )
4.77
( ± 6.64 )
4.36
( ± 7.72 )
3.09
( ± 7.98 )
1.30
( ± 7.00 )
0.28
( ± 4.93 )
0.36
( ± 2.89 )
RLE (mm):
m e a n   ( ± s t d )
3.93
( ± 2.59 )
4.62
( ± 3.47 )
5.75
( ± 6.42 )
4.76
( ± 7.00 )
4.29
( ± 7.92 )
3.03
( ± 7.95 )
1.35
( ± 7.07 )
0.18
( ± 5.00 )
0.31
( ± 2.98 )
Abbreviations: LLE—left lung diaphragmatic position variation; RLE—right lung diaphragmatic position variation.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, D.; Huang, Y.; Li, B.; Cai, J.; Ren, G. Dynamic Chest Radiograph Simulation Technique with Deep Convolutional Neural Networks: A Proof-of-Concept Study. Cancers 2023, 15, 5768. https://doi.org/10.3390/cancers15245768

AMA Style

Yang D, Huang Y, Li B, Cai J, Ren G. Dynamic Chest Radiograph Simulation Technique with Deep Convolutional Neural Networks: A Proof-of-Concept Study. Cancers. 2023; 15(24):5768. https://doi.org/10.3390/cancers15245768

Chicago/Turabian Style

Yang, Dongrong, Yuhua Huang, Bing Li, **g Cai, and Ge Ren. 2023. "Dynamic Chest Radiograph Simulation Technique with Deep Convolutional Neural Networks: A Proof-of-Concept Study" Cancers 15, no. 24: 5768. https://doi.org/10.3390/cancers15245768

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop