Next Article in Journal
Machine Learning Predicts Decompression Levels for Lumbar Spinal Stenosis Using Canal Radiomic Features from Computed Tomography Myelography
Next Article in Special Issue
Diabetic Macular Edema Optical Coherence Tomography Biomarkers Detected with EfficientNetV2B1 and ConvNeXt
Previous Article in Journal
A Modified Corona Score Using Lung Ultrasound to Identify COVID-19 Patients
Previous Article in Special Issue
Effect of Punctal Occlusion on Blinks in Eyes with Severe Aqueous Deficient Dry Eye
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dry Eye Subtype Classification Using Videokeratography and Deep Learning

1
Department of Ophthalmology, Kyoto Prefectural University of Medicine, Kyoto 602-0841, Japan
2
Department of Optics and Spectroscopy, Faculty of Physics, St. Kliment Ohridski University of Sofia, 1164 Sofia, Bulgaria
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(1), 52; https://doi.org/10.3390/diagnostics14010052
Submission received: 17 October 2023 / Revised: 16 December 2023 / Accepted: 20 December 2023 / Published: 26 December 2023
(This article belongs to the Special Issue Diagnosis, Treatment and Management of Eye Diseases)

Abstract

:
We previously reported on ‘Tear Film Oriented Diagnosis’ (TFOD), a method for the dry eye (DE) subtype classification using fluorescein staining and an examination of fluorescein breakup patterns via slit-lamp biomicroscopy. Here, we report ‘AI-supported TFOD’, a novel non-invasive method for DE subtype classification using videokeratography (VK) and “Blur Value” (BV), a new VK indicator of the extent of blur in Meyer-ring images and deep learning (DL). This study involved 243 eyes of 243 DE cases (23 males and 220 females; mean age: 64.4 ± 13.9 (SD) years)—i.e., 31 severe aqueous-deficient DE (sADDE) cases, 73 mild-to-moderate ADDE (m/mADDE) cases, 84 decreased wettability DE (DWDE) cases, and 55 increased evaporation DE (IEDE) cases diagnosed via the fluorescein-supported TFOD pathway. For DL, a 3D convolutional neural network classification model was used (i.e., the original image and BV data of eyes kept open for 7 s were randomly divided into training data (146 cases) and the test data (97 cases), with the training data increased via data augmentation and corresponding to 2628 cases). Overall, the DE classification accuracy was 78.40%, and the accuracies for the subtypes sADDE, m/mADDE, DWDE, and IEDE were 92.3%, 79.3%, 75.8%, and 72.7%, respectively. ‘AI-supported TFOD’ may become a useful tool for DE subtype classification.

1. Introduction

Due to the worldwide increase in elderly populations, the use of visual display terminals, contact lens wearers, etc., dry eye (DE), which is one of the most common ocular surface (OS) disorders (OSDs) frequently encountered in the clinical setting, is increasing [1]. Moreover, DE reportedly has a significant impact on the economy as a result of reduced work efficiency due to DE-related symptoms such as ocular discomfort and/or visual disturbance [2]. According to the Asia Dry Eye Society, DE is defined as “a multifactorial disease characterized by unstable tear film causing a variety of symptoms and/or visual impairment, potentially accompanied by ocular surface damage” [3]. Similarly, the Dry Eye Workshop II (DEWS II) report defined DE as “a multifactorial disease of the ocular surface characterized by a loss of homeostasis of the tear film and accompanied by ocular symptoms, in which tear film instability and hyperosmolarity, OS inflammation and damage, and neurosensory abnormalities play etiological roles” [4]. Based on those definitions, tear film (TF) instability is an essential abnormality for the pathophysiology of DE, and proper evaluation of TF instability is vital for a correct diagnosis of the disease.
Recently, a new diagnostic pathway for DE, termed ‘TF oriented diagnosis’ (TFOD), has advanced in Japan [5,6,7], in which the differential diagnosis of DE includes (1) severe aqueous deficient DE (ADDE) (sADDE), (2) mild-to-moderate ADDE (m/mADDE), (3) decreased wettability dry eye (DWDE), and (4) increased evaporation DE (IEDE). Via the TFOD pathway, a proper diagnosis of those DE subtypes can be achieved based on fluorescein-stained aqueous TF dynamics and fluorescein breakup (BU) patterns (BUPs) (FBUPs). In addition, based on FBUPs, not only the DE subtype but also the insufficient component of the OS responsible for the TF BU in each DE subtype can be suggested, and the most effective topical eye therapy currently available in each country to treat TF instability can be proposed. As FBUPs, eight representative FBUPs can be differentiated, in which area break (AB), partial AB (pAB), line break (LB), spot break (SB), dimple break (DB), random break (RB), LB with rapid expansion (RE), and RB with RE, respectively, correspond to sADDE (AB and pAB), m/mADDE (LB), and DWDE (SB, DB, LB with RE, and RB with RE, precisely). Although RB with RE corresponds to IEDE with DWDE, DWDE is thought to be the predominant pathophysiology of RB with RE [5,6,7]. In addition, the respective insufficient components for sADDE, m/mADDE, DWDE, and IEDE are severely diminished aqueous tears, mildly to moderately diminished aqueous tears, membrane-associated mucins, especially the longest, MUC16, which can maintain corneal surface wettability [8,9], and TF lipid layer or secretory mucins, which are thought to suppress the evaporation of the aqueous tears in the TF aqueous layer [10,11]. It has been reported that among those DE subtypes, DWDE presenting the BUs SB, DB, or RE might be involved in short BU time type DE, which is characterized by relatively severe symptoms despite normal tear volume and minimal involvement of corneal epithelial damage (CED) [3,5,12,13]. Furthermore, DWDE is often overlooked and is not included in the DE classification proposed in the DEWSII report [4], although it is included in that of the Asia Dry Eye Society [14]. Thus, to appropriately treat DWDE, a proper diagnosis is important.
When implementing TFOD based on the classification of FBUPs, the examination of tears by slit-lamp microscopy after staining the tears with sodium fluorescein via a fluorescein test strip, etc., is essential. However, fluorescein staining is a somewhat invasive procedure, and the results can differ (i.e., be non-reproducible, inaccurate, or inconsistent) due to the method being used to apply the strip to the eye, which can vary among examiners [5]. To overcome those intrinsic fluorescein staining problems, various more advanced non-invasive methods have been developed for the evaluation of TF instability in DE, including videointerferometry [7,15], videothermography [16], wavefront analysis [17,18], videokeratoscopy [19], and videokeratography (VK) [20,21,22]. However, to the best of our knowledge, no non-invasive methods have been developed to simultaneously assess both TF stability and CED in any type of DE and indicate a subtype classification. However, due to our recent advancements in VK involving the use of a “Blur Value” (BV), a new indicator for the assessment of the extent of blur in Meyer-ring (MR) images, we are now able to successfully differentiate sADDE, m/mADDE, DWDE, and IEDE non-invasively [21,22].
There have been numerous attempts to diagnose and assess DE via the use of artificial intelligence (AI), in which machine learning, especially deep learning (DL), is applied for the evaluation of DE and other OSDs. For example, for the screening of characteristic abnormalities in DE and other OSDs, DL has been adopted for the examination of DE aspects such as tear meniscus height [23], fluorescein-stained corneal photographs [24], interferometry [25], in vivo laser confocal microscopy [26], and meibography [27,28], to detect and/or assess those abnormalities. However, to date, there have been no published studies reporting a successful use of AI for the classification of the DE subtype. Thus, if AI could be successfully adopted into the TFOD pathway in combination with TF-oriented therapy (TFOT), it might be an ideal application in the clinical setting, as it is theorized that in the fluorescein-staining-supported TFOD/TFOT pathway [5,6,7], AI-supported non-invasive TFOD would be useful.
Hence, the aim of this present study was to adopt our newly developed BV-supported VK system to AI using DL in order to classify the DE subtype [21,22].

2. Methods

2.1. Subjects

This retrospective study involved 243 eyes that were followed at the Dry Eye Outpatient Clinic of the Department of Ophthalmology, Kyoto Prefectural University of Medicine, Kyoto, Japan, between 4 January 2019 and 8 October 2021. Details of the patient data are shown in Table 1.
All patients were diagnosed as DE based on the current Japanese diagnostic criteria [29], i.e., exhibiting the DE-related symptoms of eye discomfort and/or visual disturbance and a fluorescein BU time of ≤5 s, and the eyes deemed eligible for inclusion were those with more severe symptoms. In cases in which the severity of the symptoms was the same in both eyes, right-eye data were used. Prior to enrollment in the study, confirmation was obtained from all patients that no eye drop medications had been used for at least 1 h prior to the initial examination.
Patients with any of the following conditions were excluded from the study: (1) patients with an eyelid disease such as blepharoptosis, lagophthalmos, blepharospasm, entropion, or ectropion, (2) patients with severe conjunctivochalasis, and (3) patients with any history of undergoing eye surgery, such as for the puncta, an OSD other than DE, the eyelid, and glaucoma. All patients deemed inappropriate for involvement in this study were excluded via a consensus by three ophthalmologists (N.Y., N.K., and H.K.).

2.2. Study Protocol

On the day of examination, FBUPs were assessed via the following 4 steps [5,6,7]:
(1)
Two drops of saline were instilled onto a fluorescein test strip (Ayumi Pharmaceutical Corporation, Tokyo, Japan) that was then vigorously shaken to minimize, as much as possible, the amount of aqueous fluid on the strip. Using this procedure, no significant difference in meniscus height, implying no significant tear volume increase, was noted between fluorescein staining and without fluorescein staining (unpublished data);
(2)
The test strip was then gently placed a bit temporal to the center of the patient’s lower lid margin, with the patient then being verbally instructed to gently blink several times to mix the fluorescein with the aqueous tears;
(3)
The patient was then instructed to briskly open the eye 3 consecutive times after gently closing the eye, followed by kee** the eye open until fluorescein BU appears for as long as 10 s at each blink. Those procedures were observed 3 times with a slit-lamp biomicroscope using the appropriate filters for observing the upward movement of the fluorescein-stained aqueous tears and FBUPs and were video-recorded;
(4)
The FBUPs and corresponding DE subtypes were then classified.
The eyes that repeatedly presented typical FBUPs, including AB, partial AB, LB, SB, DB, RB, LB with RE, and RB with RE at each of the 3 successive blinks, were enrolled in this study. They were, respectively, categorized as either ADDE (i.e., sADDE (when AB or partial AB was observed) or m/mADDE (when LB was observed)), DWDE (when SB, DB, LB with RE, or RB with RE was observed), or IEDE (when RB was observed). Table 2 shows the relationship between DE subtypes and FBUPs. As a data sample, eight typical example images (150 original images acquired by VK) of 8 types of BUPs including DB, LB with RE, RB with RE and SB for DWDE, RB for IEDE, LB for mild to moderate ADDE, and AB and partial AB for severe ADDE are available in the Supplementary files.
More than 10 min after the classification of the FBUPs and corresponding DE subtype, a time-dependent change in precorneal TF behavior when the eye is kept open was acquired through the use of a VK instrument (RET-700; Rexxam Co., Ltd., Osaka, Japan), in which two data sets, including the original VK data and the VK data with BV [21,22], the new indicator that enables assessment of the corneal-surface abnormalities from the point of the extent of the disturbance of the MRs, were obtained using the software incorporated in the instrument [21,22]. Those two data sets during 7 s of eye opening were used for DL.

2.3. Clinical Assessment

As shown in Figure 1, the original video data of the subject’s eye were acquired via the VK instrument for 15 [s] (150 [frame]) at 0.1 [s] intervals. The resolution of the video data is 640 × 480 (pixel), and each pixel is an 8-bit grayscale. The VK data with BV was obtained using the original video data. Then, the video data obtained from 243 eyes in total were divided into independent open-eye data sets comprising the original VK data and the VK data with BV, and a single data set was used for the analysis. If the eye was kept open only once or was opened for the entire 15 s, the data were selected as they were.
The video data of all 243 eyes was cropped in the range of 200 (pixel) in the upper, lower, right, and left directions from the center of the MRs, with a resolution of 400 × 400 (pixel), in order to analyze only the data within which MRs are included. To keep the data size consistent, the time for data acquisition was adjusted to 7 (s) (70 (frames)) from the start of the eye opening (data for time over 7 (s) was deleted). If the eye could not be kept open for as long as 7 (s), 0 data points (black data) were added to the missing frames. For example, if an eye was kept open for as long as only 5 s, with data of 0–2 s being missed, 0 data points for 2 s (20 frames) were added as the missing data for the 5 s open eye data.
To prepare the data for DL, video data of all 243 eyes were randomly divided into a ratio of training data 6 (146 eyes)/test data 4 (97 eyes) using the stratified hold-out method for learning model evaluation. The BUP data were also divided into a ratio of 6:4. As for the training data and the test data, in order to reduce the training and inference time and the number of parameters, for all 97 video data, 400 × 400 (pixel) video data were cropped in the range of 150 (pixel) in the upper, lower, right, and left directions from the center of the video data, with a resolution of 300 × 300 (pixel), and its resolution was further lowered by half, 150 × 150 (pixel).
To improve generalization performance, data augmentation was performed on the training data so that the amount of data was increased for all 146 video data points. Data augmentation was performed as follows (1)–(3):
(1)
Two Rotations: +5 [deg.] and −5 [deg.];
(2)
Two Rescalings: 3 [%] enlarging and 3 [%] reducing;
(3)
Horizontal flip**.
Increased the total data by a factor of 18, bringing the total data count to 2628. The breakdown of the number of test data and training data is shown in Table 3.
In creating the rotated and scaled images, sub-pixel interpolation was performed by the bicubic interpolation method [30]. The resolution was reduced to 150 × 150 (pixel) in the same way as the test data for 2628 increased video data. The reason for crop** 640 × 480 (pixel) video data at 400 × 400 (pixel) and then crop** it again at 300 × 300 (pixel) was to prevent the edges of the video data from being lost when the image is reduced or rotated in augmentation.
A subtype classification of DE was conducted in a standard way in training and inference processing, as shown in Figure 2, using the video data created by the above-mentioned method. First, the training processing was performed using the training data, and the trained models of 3 classifications and 2 classifications were prepared as follows:
(1)
Three Classifications: (sADDE, m/mADDE), DWDE, and IEDE data were used as inputs for the trained model data;
(2)
Two Classifications: sADDE and m/mADDE data were used as inputs for the trained model data.
Second, inference processing was carried out by inputting the test data into the newly created trained model. It was first classified into the ADDE, DWDE, or IEDE subtypes described in (1) above, and then classified further into either sADDE or m/mADDE, as described in (2) above, to differentiate between sADDE and m/mADDE when the data are classified as ADDE. Based on the above results, the probability of 4 classifications was calculated from the classification probabilities of 3 classifications and 2 classifications.
The network structure of the DL model is shown in Figure 3, with further details shown in Table 4. In this analysis, a model of DL with a structure consisting of 2-dimensional (D) and 3-D convolutional deep neural networks (2D-CNN and 3D-CNN, respectively) was created. First, the input dimension was set to 5, including 3638 of training data, original video data with the resolution of 150 × 150 [pixel], 2 channels of the BV video data (8-bit for each pixel), and 70 [frames], and the features were extracted from CNN1 Block1 (2D-CNN) based on ResNet-D with a 2-D convolution layer [31], 2 layers of the 2-D convolution layer, a 2-D max pooling layer, and a 2-D average pooling layer. At the end of CNN Block 1, training stabilization was performed by disabling the neurons with a constant probability in the dropout layer. In this method, the probability is set to 25%. Second, a 3-D (vertical, horizontal, and temporal) structure was created with a time-distributed layer for the obtained features. The features were obtained 3-dimensionally with a 3-D convolution layer and 2-D average pooling layer by entering the result into CNN Block2 (3D-CNN) based on 3D-ResNet [32]. For both CNN1 Block 1 and CNN1 Block 2, a batch-normalized layer was placed after the convolution layer to stabilize the training by batch normalization. ReLU (Rectified Linear Units) was used for the activation function. Then, the features thus far obtained were made to be one-dimensional by a Global Max Pooling layer, and the probability that 3 or 2 classifications became the largest in the Dense layer with Softmax function as activation function was classified as the subtype. The mini-batch size for training was 64, an Adam optimizer was used as the optimization algorithm, the learning rate was set to 0.001, and categorical cross entropy was used as the loss function. The number of epochs was set to 60. In addition, the trained models of both 3 classifications and 2 classifications with the same structure as that in Figure 3 were used.

3. Results

The learning curves for 3-class and 2-class classifications are shown in Figure 4. Based on the learning curves, the 60th epoch, where the changes in loss and accuracy converged, was adopted as the trained model.
As shown in Figure 2, the inference processing was performed by entering the test data into the newly created trained model data to calculate the possibility of four classes for all 97 data points, and the class of the maximum probability was classified as a subtype. A confusion matrix was then created from the result (Figure 5). The following are shown in Figure 5, respectively: (A) the table for the confusion matrix of the four classifications and the precision of each subtype, and the table for recall and f1-score (harmonic average of precision and recall), their simple average (macro avg), and weighted average (weighted avg) of the number of data; (B) the result of three classifications; and (C) the result of two classifications. It was found that the accuracy of the final four classifications was 0.784.
For each subtype, sADDE was less prone to being misinterpreted than the other subtypes, with recall exceeding 0.9 and a precision of 0.857, which were both higher than those in the other subtypes. m/mADDE, DWDE, and IEDE were prone to misinterpret each other with lower recall and precision compared to sADDE, but the recall was above 0.7 in all cases. However, DWDE had lower precision, and it was more frequently prone to being misinterpreted as other types, such as m/mADDE and IEDE. m/mADDE and IEDE had precision equal to or greater than 0.8 and were less prone to being misinterpreted. m/mADDEs and IEDEs had a precision of 0.8 or higher and were less misinterpreted as other types than DWDEs. m/mADDE, DWDE, and IEDE had lower BVs, and they were prone to misinterpret each other due to the presence of hard-to-distinguish data.
The receiver operating characteristic analysis curves for each subtype are shown in Figure 6. The receiver operating characteristic curve was plotted graphically by calculating the true-positive rate and false-positive rate. Since this study employed multiple classification methods, it was calculated as the sum of the rate of true positive rate as a correct subtype and the rate of false positive rate as an incorrect subtype (1 minus true positive rate). The classification accuracy for each subtype was relatively good, as our findings showed that the area under the curve of sADDE, IEDE, m/mADDE, and DWDE was 0.985 (very high), 0.928, 0.860, and 0.830, respectively, with all being over 0.8.
A heat map that can visually confirm the judgmental basis of the CNN model was created by the class activation map (CAM). Grad-CAM++ was used for the algorithm of the heat map preparation [33], and the gradient was calculated for the weight of CNN Block1, as shown in Figure 3. This heat map is called a Saliency Map [32], which emphasizes the pixels that were used as decision-making evidence during DL, with blue being weak and green to yellow to red indicating increasingly strong features. Figure 7 shows one example for sADDE, m/mADDE, DWDE, and IEDE each. The BUP in the example in Figure 7 shows AB for sADDE, LB for m/mADDE, SB for DWDE, and RB for IEDE. As an overall trend, the areas with strong disturbance or blur of the MRs were shown in yellow to red, and the areas with weak disturbance or blur, no disturbance or blur, or no MR were shown in green to blue. In the case of sADDE (AB), the heat map shows a wide area of red, with no upward movement of the aqueous layer and greater disturbance of the MR pattern over the entire corneal surface. In the case of m/mADDE (LB), the heat map shows a long red area extending vertically and disturbance of the lines. In the case of DWDE (SB), the heat map shows a circular red-to-yellow area near the center of the cornea. IEDE (RB) was a relatively mild case, and even in this case, there are few red areas; the yellow-to-light-blue areas were generally faint on the heat map, and the disturbance of the ring pattern was relatively small.

4. Discussion

For the diagnosis of DE in this study, FBUPs were used to classify three DE subtypes, including ADDE, DWDE, and IEDE, with ADDE further classified into sADDE and m/mADDE [5,6,7]. In addition, for the recruitment of typical DE patients, representative cases presenting typical FBUPs, with those FBUPs appearing repeatedly in three successive blink times, were enrolled. Thus, we deemed it reasonable to consider that the same BUPs as those of FBUPs must be reflected upon the distortion of the MRs when the eyes were evaluated via VK at more than 10 min after the classification of FBUPs and DE subtypes based on the FBUPs. Moreover, the DE subtype classified by FBUPs could be appropriately used as the ground truth, i.e., the correct answer for DL. In fact, our findings demonstrated that the accuracy of DL for calculating the overall DE classification was 78.4% and that the accuracies for DE subtypes sADDE, m/mADDE, DWED, and IEDE were 92.3%, 79.3%, 75.8%, and 72.7%, respectively. Moreover, it was demonstrated that a higher accuracy of 97.2% was found in the two classifications for ADDE, in which differential diagnosis between sADDE and m/mADDE was achievable with very high accuracy. This was also found to be true for the greater accuracy of sADDE in the four classifications. However, that is what we expected, considering that the subtype sADDE presents a remarkably greater MR disturbance, which was not detected in the other subtypes. In addition, although the accuracy of the three classifications was 79.4%, it was just slightly higher than that of the final four classifications, and there was only one incorrect interpretation. Therefore, the performance of the algorithm for the four classifications is deemed to be determined by the three classifications. Considering that subtype sADDE had the lowest number of test data (i.e., 13), while DWDE had the highest number (i.e., 33), and that the weighted average of the f1-score was the same as that of accuracy and there was no event in which only accuracy was higher, this suggests that the bias in the amount of data per subtype had almost no effect.
A heat map can tell how and from what aspect AI evaluates the disturbance of the MR within the corneal surface in DE, and in the case of sADDE corresponding to AB, the heat map showed red in a wider area with no time-dependent change. That finding was interpreted as precorneal TF comprising the lipid layer and aqueous layer presenting no upward movement after the eye was opened and only the resultant CED due to the lack of TF being reflected upon the distortion of the MRs [5,6,7]. In contrast, in m/mADDE corresponding to LB, the heat map showed a long red area extending vertically while accompanying the time-dependent enhancement of disturbance within the lower cornea, which may reflect the counteraction between the meniscus-induced thinning and upward movement of the aqueous layer simultaneously occurring within the lower part of the cornea in m/mADDE after the eye is opened [5,6,7,11,34].
In the case of DWDE, in which a representative BUP is SB, the heat map showed a spot-like smaller area with red to yellow appearing typically at the central part of the cornea immediately after eye opening, which is recognized as remarkable spot-like disturbances with time-dependent lessening when the eye is kept open [5,6,7,21].
Viewing the time-dependent distortional changes of the corneal surface, which are expected to be derived from TF instability and/or BU and CED [22], we considered that the use of the CAM heat map would help to visualize the basis for AI judgment via DL to some extent and that the combination of the CAM heat map with the subtype classification results via AI would be of great value for ophthalmologists to diagnose DE subtype via the TFOD pathway [5,6,7]. In addition, considering that VK is non-invasive with no aid of fluorescein, it is possible that even medical staff members could play some roles in DE diagnosis in the examination room before ophthalmologists examine the patient’s eyes using fluorescein in the consultation room with reference to the AI judgment, and this could become a possible realization of non-invasive TFOD by VK that we previously reported or AI-supported TFOD [21,22].
This study adopted the original VK data and BV data obtained from the patients’ eyes for DL. Compared to fluorescein, which can assess not only the dynamics of fluorescein-stained aqueous TF and fluorescein BU but also fluorescein-stained areas of CED over the entire corneal surface within the palpebral zone, the information obtainable through VK is somewhat limited, as it can only detect the disturbance of the MR associated with TF instability and CED in DE. In sADDE, MR disturbance is derived chiefly from more severe CED, and in DWDE and IEDE, the disturbance comes only from TF instability because of its sufficiency of tear volume and minimal involvement of CED, and such abnormalities are only detected by the MR disturbance [22]. Therefore, there are some intrinsic limitations for VK to detect corneal surface abnormalities in DE, even with the aid of AI. Considering those limitations in VK applied to the present study, the classification of DE via VK seems to be satisfactory.
As for the feasibility of the present AI-supported TFOD using VK being applied in the clinical setting, there are some important points that need to be considered. First, the TFOD pathway needs to be adopted in VK in the same manner in which it is conducted in fluorescein-supported TFOD using slit-lamp biomicroscopy. Second, verbal instruction to the patient of the required blink is essential, i.e., “gently close your eye”, “now rapid open your eye”, and “now keep your eye open” [5], as it greatly enhances the detection of the breakup pattern. Third, it is important that each patient be instructed to not use any eye drops for at least one hour prior to the test, as was done in this present study, as it makes for an accurate diagnosis of the DE subtype without any interference by the eye drops that the patient routinely uses. If those specific rules are strictly kept, the diagnosis of the DE subtype can effectively be performed with AI-supported TFOD using VK. However, in each examination before and after treatment, AI-supported TFOD using VK should be conducted at a similar time of day, i.e., either in the morning or in the afternoon, in consideration of the patient’s daily activities.
The limitation of this AI-supported TFOD using VK is that there is a hurdle to overcome for the differentiation of mild DE subtypes, including mild ADDE with mild CED, IEDE, and DWDE presenting RE of the BU [6,7], as those DE subtypes may show an MR disturbance that is too minor for AI to differentiate. However, mild ADDE is likely to manifest LB and mild CED, especially at the lower region of the cornea, which would be differentially diagnosed with IDDE presenting RB using the semantic segmentation method, which can together be detected by the CAM heat map [35]. In DWDE, BUPs other than SB, such as DB, LB with RE, and RB with RE, tend to take longer to appear and are followed by TF BU, leading to an MR disturbance after the eye is kept open, which sometimes takes more than 7 s to occur [7]. That is because non-invasive BU, i.e., full-thickness TF BU, takes a longer time to appear compared to fluorescein BU, merely corresponding to a thinning of the aqueous layer [11]. In this study, the AI algorithm was made by the images until 7 s after eye opening, and that might have been attributed to this limitation. However, semantic segmentation with the CAM heat map may be useful to detect the RE of the BU and diagnose DWDE properly. In addition, the possibility of AI-supported TFOD discovered through this study should be applied to the real world of DE to test its effectiveness in diagnosing the disease in a large number of DE patients in future studies. Moreover, the number of patients and eyes (243 eyes out of 243 DE cases) used in the current study is comparable with the sample size (350 eyes out of 178 subjects) used in a recently published report [25], which attempted to implement the classification of TF patterns suggested by us for specular microscopy images.
In conclusion, the findings in this study indicated the possibility that DL via VK with BV may be useful to DE diagnosis as a new non-invasive tool for DE classification. In future studies, that possibility would be enhanced by the combination of this method with semantic segmentation. Since some clue to the diagnosis of the DE subtype in the examination room could be possible via this present method, it will be of great value in the realization of shifting the diagnosis of DE from the consultation room to the examination room and before the accurate diagnosis using fluorescein in the consultation room.

Supplementary Materials

The following supporting information can be downloaded at https://mdpi.longhoe.net/article/10.3390/diagnostics14010052/s1. A total of 8 typical example images (150 original images acquired by VK) of 8 types of BUPs can be downloaded, including dimple break, line break with rapid expansion, random break with rapid expansion, and spot break for DWDE, as well as random break for IEDE, line break for mild-to-moderate ADDE, and area break and partial area break for severe ADDE.

Author Contributions

Conceptualization and methodology, N.Y. and N.K.; formal analysis, N.Y., N.K. and H.K.; resources, N.Y., N.K., H.K. and Y.F.; data curation, N.Y., N.K., H.K. and Y.F.; writing (original draft), N.Y.; writing (review and editing), N.Y., N.K., H.K., C.S. and G.A.G.; visualization, N.Y.; supervision, G.A.G.; project administration, N.Y.; funding acquisition, N.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by JSPS KAKENHI for Scientific Research from the Japanese Ministry of Education, Culture, Sports, Science and Technology (grant number: JP20K09794). The funding organization had no role in the design or conduct of this research. G. As. Georgiev acknowledges the support of the European Union NextGenerationEU, through the National Recovery and Resilience Plan of the Republic of Bulgaria (Project No.: BG-RRP-2.004-0008-C01).

Institutional Review Board Statement

This study was approved by the Institutional Review Board of the Kyoto Prefectural University of Medicine, Kyoto, Japan (project identification code ERB-C-1233-4 and 20 August 2018 of approval), and was conducted in accordance with the tenets set forth in the Declaration of Helsinki.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data that support the findings in this study are available on reasonable request from the corresponding author.

Acknowledgments

The authors wish to thank Yuji Iseki, Jun Kawai, Reiji Yoshioka, Ken-ichi Yoshida, and Daichi Yamamoto of Rexxam Co., Ltd., Osaka, Japan, for their technical assistance in the development of the novel deep learning algorithm applied for this research (Patent Classification No.: 2021-064437), and John Bush for reviewing the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

ABarea break
ADDEaqueous-deficient dry eye
AIartificial intelligence
BVblur value
DBdimple break
BUbreakup
BUPBU pattern
CAMclass activation map
CEDcorneal epithelial damage
CNNconvolutional neural network
Ddimension
DEdry eye
DEWSIIDry Eye Workshop II
DLdeep learning
DWDEdecreased wettability DE
FBUPfluorescein BUP
IEDEincreased evaporation DE
LBline break
m/mADDEmild-to-moderate ADDE
MRMeyer-ring
OSocular surface
OSDsOS disorders
pABpartial AB
RBrandom break
RErapid expansion
sADDEsevere ADDE
SBspot break
SDstandard deviation
TFtear film
TFODTF-oriented diagnosis
TFOTTF-oriented therapy
VKvideokeratography

References

  1. Stapleton, F.; Alves, M.; Bunya, V.Y.; Jalbert, I.; Lekhanont, K.; Malet, F.; Na, K.S.; Schaumberg, D.; Uchino, M.; Vehof, J.; et al. TFOS DEWS II Epidemiology Report. Ocul. Surf. 2017, 15, 334–365. [Google Scholar] [CrossRef] [PubMed]
  2. Uchino, M.; Uchino, Y.; Dogru, M.; Kawashima, M.; Yokoi, M.; Komuro, A.; Sonomura, Y.; Kato, H.; Kinoshita, S.; Schaumberg, D.A.; et al. Dry eye disease and work productivity loss in visual display users: The Osaka Study. Am. J. Ophthalmol. 2014, 157, 294–300. [Google Scholar] [CrossRef] [PubMed]
  3. Tsubota, K.; Yokoi, N.; Shimazaki, J.; Watanabe, H.; Dogru, M.; Yamada, M.; Kinoshita, S.; Kim, H.M.; Tchah, H.M.; Hyon, J.Y.; et al. New perspectives on dry eye definition and diagnosis: A consensus report by the Asia Dry Eye Society. Ocul. Surf. 2017, 15, 65–76. [Google Scholar] [CrossRef] [PubMed]
  4. Craig, J.P.; Nichols, K.K.; Akpek, E.K.; Caffery, B.; Dua, H.S.; Joo, C.K.; Liu, Z.; Nelson, J.D.; Nichols, J.J.; Tsubota, K.; et al. TFOS DEWS II definition and classification report. Ocul. Surf. 2017, 15, 276–283. [Google Scholar] [CrossRef] [PubMed]
  5. Yokoi, N.; Georgiev, G.A.; Kato, H.; Komuro, A.; Sonomura, Y.; Sotozono, C.; Tsubota, K.; Kinoshita, S. Classification of fluorescein breakup patterns: A novel method of differential diagnosis for dry eye. Am. J. Ophthalmol. 2017, 180, 72–85. [Google Scholar] [CrossRef] [PubMed]
  6. Yokoi, N.; Georgiev, G.A. Tear-film-oriented diagnosis and therapy for dry eye based on tear film dynamics. Investig. Ophthalmol. Vis. Sci. 2018, 59, DES13–DES22. [Google Scholar] [CrossRef] [PubMed]
  7. Yokoi, N.; Georgiev, G.A. Tear-film-oriented diagnosis for dry eye. Jpn. J. Ophthalmol. 2019, 63, 127–136. [Google Scholar] [CrossRef] [PubMed]
  8. Ablamowicz, A.F.; Nichols, J.J. Ocular surface membrane-associated mucins. Ocul. Surf. 2016, 14, 331–341. [Google Scholar] [CrossRef]
  9. Georgiev, G.A.; Eftimov, P.; Yokoi, N. Contribution of mucins towards the physical properties of the tear film: A modern update. Int. J. Mol. Sci. 2019, 20, 6132. [Google Scholar] [CrossRef]
  10. Georgiev, G.A.; Eftimov, P.; Yokoi, N. Structure-function relationship of tear film lipid layer: A contemporary perspective. Exp. Eye Res. 2017, 163, 17–28. [Google Scholar] [CrossRef]
  11. Yokoi, N.; Bron, A.J.; Georgiev, G.A. The precorneal tear film as a fluid shell: The effect of blinking and saccades on tear film distribution and dynamics. Ocul. Surf. 2014, 12, 252–266. [Google Scholar] [CrossRef] [PubMed]
  12. Toda, I.; Shimazaki, J.; Tsubota, K. Dry eye with only decreased tear break-up time is sometimes associated with allergic conjunctivitis. Ophthalmology 1995, 102, 302–309. [Google Scholar] [CrossRef] [PubMed]
  13. Yokoi, N.; Uchino, M.; Uchino, Y.; Dogru, M.; Kawashima, M.; Komuro, A.; Sonomura, Y.; Kato, H.; Tsubota, K.; Kinoshita, S.; et al. Importance of tear film instability in dry eye disease in office workers using visual display terminals: The Osaka study. Am. J. Ophthalmol. 2015, 159, 748–754. [Google Scholar] [CrossRef] [PubMed]
  14. Tsubota, K.; Yokoi, N.; Watanabe, H.; Dogru, M.; Kojima, T.; Yamada, M.; Kinoshita, S.; Kim, H.M.; Tchah, H.W.; Hyon, J.Y.; et al. A new perspective on dry eye classification: Proposal by the Asia Dry Eye Society. Eye Contact Lens 2020, 46, S2–S13. [Google Scholar] [CrossRef] [PubMed]
  15. Ishibashi, T.; Yokoi, N.; Kinoshita, S. Comparison of the short-term effects on the human corneal surface of topical timolol maleate with and without benzalkonium chloride. J. Glaucoma 2003, 12, 486–490. [Google Scholar] [CrossRef] [PubMed]
  16. Su, T.Y.; Chang, S.W. Normalized ocular surface temperature models for tear film characteristics and dry eye disease evaluation. Ocul. Surf. 2021, 19, 74–82. [Google Scholar] [CrossRef] [PubMed]
  17. Koh, S.; Maeda, N.; Hirohara, Y.; Mihashi, T.; Bessho, K.; Hori, Y.; Inoue, T.; Watanabe, H.; Fujikado, T.; Tano, Y.; et al. Serial measurements of higher-order aberrations after blinking in patients with dry eye. Investig. Ophthalmol. Vis. Sci. 2008, 49, 133–138. [Google Scholar] [CrossRef] [PubMed]
  18. Koh, S.; Maeda, N.; Kuroda, T.; Hori, Y.; Watanabe, H.; Fujikado, T.; Tano, Y.; Hirohara, Y.; Mihashi, T. Effect of tear film break-up on higher-order aberrations measured with wavefront sensor. Am. J. Ophthalmol. 2002, 134, 115–117. [Google Scholar] [CrossRef]
  19. Llorens-Quintana, C.; Iskander, D.R. Assessment of tear film using videokeratoscopy based on fractal dimension. Optom. Vis. Sci. 2018, 95, 32–42. [Google Scholar] [CrossRef]
  20. Goto, T.; Zheng, X.; Klyce, S.D.; Kataoka, H.; Uno, T.; Karon, M.; Tatematsu, Y.; Bessyo, T.; Tsubota, K.; Ohashi, Y.; et al. A new method for tear film stability analysis using videokeratography. Am. J. Ophthalmol. 2003, 135, 607–612. [Google Scholar] [CrossRef]
  21. Yokoi, N.; Kusada, N.; Kato, H.; Furusawa, Y.; Sotozono, C.; Georgiev, G.A. Successful detection of the characteristics of tear film breakup appearing immediately after eye opening by videokeratography with a newly-developed indicator. Diagnostics 2023, 13, 240. [Google Scholar] [CrossRef] [PubMed]
  22. Kusada, N.; Yokoi, N.; Kato, H.; Furusawa, Y.; Sakai, R.; Sotozono, C. Evaluation of dry eye with videokeratographer using a newly developed indicator. Am. J. Ophthalmol. 2023, 252, 135–146. [Google Scholar] [CrossRef] [PubMed]
  23. Deng, X.; Tian, L.; Liu, Z.; Zhou, Y.; Jie, Y. A deep learning approach for the quantification of lower tear meniscus height. Biomed. Signal Process. Control 2021, 68, 102655. [Google Scholar] [CrossRef]
  24. Driscoll, T.; Braun, R.J.; Begley, C.G. Automatic detection of the cornea location in video captures of fluorescence. Modeling Artif. Intel. Ophthalmol. 2021, 3, 55–70. [Google Scholar] [CrossRef]
  25. Kikukawa, Y.; Tanaka, S.; Kosugi, T.; Pflugfelder, S.C. Non-invasive and objective tear film breakup detection on interference color images using convolutional neural networks. PLoS ONE 2023, 18, e0282973. [Google Scholar] [CrossRef] [PubMed]
  26. Maruoka, S.; Tabuchi, H.; Nagasato, D.; Masumoto, H.; Chikama, T.; Kawai, A.; Oishi, N.; Maruyama, T.; Kato, Y.; Hayashi, T.; et al. Deep neural network-based method for detecting obstructive meibomian gland dysfunction with in vivo laser confocal microscopy. Cornea 2020, 39, 720–725. [Google Scholar] [CrossRef]
  27. Wang, J.; Yeh, T.N.; Chakraborty, R.; Yu, S.X.; Lin, M.C. A deep learning approach for meibomian gland atrophy evaluation in meibography images. Transl. Vis. Sci. Technol. 2019, 8, 37. [Google Scholar] [CrossRef]
  28. Setu, M.A.K.; Horstmann, J.; Schmidt, S.; Stern, M.E.; Steven, P. Deep learning-based automatic meibomian gland segmentation and morphology assessment in infrared meibography. Sci. Rep. 2021, 11, 7649. [Google Scholar] [CrossRef]
  29. Shimazaki, J.; Yokoi, N.; Watanabe, H.; Amano, S.; Ohashi, Y.; Kinoshita, S.; Shimomura, Y.; Takamura, E.; Hori, Y.; Nishida, K.; et al. Definition and diagnosis of dry eye in Japan, 2016. J. Eye (Atarashii Ganka) 2017, 34, 309–313. (In Japanese) [Google Scholar]
  30. Keys, R. Cubic convolution interpolation for digital image processing. IEEE Trans. Acoust. Speech Signal Process. 1981, 29, 1153–1160. [Google Scholar] [CrossRef]
  31. He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; **e, J.; Li, M. Bag of tricks for image classification with convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 558–567. [Google Scholar] [CrossRef]
  32. Zolfaghari, M.; Singh, K.; Brox, T. Eco: Efficient convolutional network for online video understanding. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 695–712. [Google Scholar] [CrossRef]
  33. Chattopadhyay, A.; Sarkar, A.; Howlader, P.; Balasubramanian, V.N. Grad-CAM++: Improved visual explanations for deep convolutional networks. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 839–847. [Google Scholar] [CrossRef]
  34. McDonald, J.E.; Brubaker, S. Meniscus-induced tear film thinning. Am. J. Ophthalmol. 1971, 72, 139–146. [Google Scholar] [CrossRef] [PubMed]
  35. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional net-works for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
Figure 1. Schema showing an outline of the open-eye data sets acquired from video data.
Figure 1. Schema showing an outline of the open-eye data sets acquired from video data.
Diagnostics 14 00052 g001
Figure 2. Schema showing the outlines of the (A) training processing and the (B) inference processing.
Figure 2. Schema showing the outlines of the (A) training processing and the (B) inference processing.
Diagnostics 14 00052 g002
Figure 3. Schema showing the network structure of DL, i.e., input layer, CNN1 Block1 (2D-CNN) based on ResNet-D [31], time-distributed layer, CNN Block2 (3D-CNN) based on 3D-ResNet, Global Max Pooling layer, and Dense layer that outputs probabilities for each subtype.
Figure 3. Schema showing the network structure of DL, i.e., input layer, CNN1 Block1 (2D-CNN) based on ResNet-D [31], time-distributed layer, CNN Block2 (3D-CNN) based on 3D-ResNet, Global Max Pooling layer, and Dense layer that outputs probabilities for each subtype.
Diagnostics 14 00052 g003
Figure 4. The learning curves for 3 and 2 classifications, (A) 3 classifications of training and validation loss, (B) 2 classifications of training and validation loss, (C) 3 classifications of training and validation accuracy, and (D) 2 classifications of training and validation accuracy.
Figure 4. The learning curves for 3 and 2 classifications, (A) 3 classifications of training and validation loss, (B) 2 classifications of training and validation loss, (C) 3 classifications of training and validation accuracy, and (D) 2 classifications of training and validation accuracy.
Diagnostics 14 00052 g004
Figure 5. Detail of the confusion matrix and 4 classification results, (A) detail of confusion matrices of 4 classifications and matrices for evaluation, (B) confusion matrix of 3 classifications, and (C) confusion matrix of 2 classifications.
Figure 5. Detail of the confusion matrix and 4 classification results, (A) detail of confusion matrices of 4 classifications and matrices for evaluation, (B) confusion matrix of 3 classifications, and (C) confusion matrix of 2 classifications.
Diagnostics 14 00052 g005
Figure 6. Receiver operating characteristic curve of 4 classifications.
Figure 6. Receiver operating characteristic curve of 4 classifications.
Diagnostics 14 00052 g006
Figure 7. Heat map with Grad-CAM++ (left: original image; right: heat map), (A) example of sADDE (5.0 s after opening eye), (B) example of m/mADDE (5.0 s after opening eye), (C) example of DWDE (immediately after opening eye), and (D) example of IEDE (5.0 s after opening eye).
Figure 7. Heat map with Grad-CAM++ (left: original image; right: heat map), (A) example of sADDE (5.0 s after opening eye), (B) example of m/mADDE (5.0 s after opening eye), (C) example of DWDE (immediately after opening eye), and (D) example of IEDE (5.0 s after opening eye).
Diagnostics 14 00052 g007
Table 1. Details of the patient data.
Table 1. Details of the patient data.
Data ConditionsData Number
Number of eyesAll numbers243
Left107
Right136
SexMales23
Females220
AgeMean66.4
SD13.9
Table 2. Relationship between DE subtypes and FBUPs.
Table 2. Relationship between DE subtypes and FBUPs.
DE SubtypeFBUP
ADDEsADDEAB, partial AB
m/mADDELB
DWDE SB, DB, LB with RE, RB with RE
IEDE RB
Table 3. Breakdown of the amount of training data and test data.
Table 3. Breakdown of the amount of training data and test data.
DE SubtypesTraining DataTraining Data
(Augmented)
Test DataFBUPTraining DataTraining Data
(Augmented)
Test Data
sADDE1832413AB71265
Partial AB111988
m/mADDE4479229LB4479229
DWDE5191833SB1832412
DB132348
LB with RE101806
RB with RE101807
IEDE3359422RB3359422
Total146262897Total146262897
DE: dry eye; FBUP: fluorescein breakup pattern; sADDE: severe aqueous-deficient dry eye; AB: area break; m/mADDE: mild-to-moderate aqueous-deficient dry eye; LB: line break; SB: spot break; DB: dimple break; DWDE: decreased wettability dry eye; RE: rapid expansion; RB: random break; and IEDE: increased evaporation dry eye.
Table 4. Details of the DL model.
Table 4. Details of the DL model.
LayersDetailsActivationOutput Dimension
Input (video data)Input-70 × 150 × 150 × 2
CNN Block1 (2D-CNN)Conv2DReLU75 × 75 × 16
Conv2DReLU75 × 75 × 16
Batch Normalization-75 × 75 × 16
Max Pooling 2D-75 × 75 × 16
ResNet-DReLU38 × 38 × 32
Batch Normalization-38 × 38 × 32
ResNet-DReLU19 × 19 × 64
Batch Normalization-19 × 19 × 64
DropoutReLU19 × 19 × 64
Time distributionTime Distribution-70 × 19 × 19 × 64
CNN Block2 (3D-CNN)3D-ResNetReLU35 × 10 × 10 × 128
Batch Normalization-35 × 10 × 10 × 128
3D-ResNetReLU18 × 5 × 5 × 256
Batch Normalization-18 × 5 × 5 × 256
Global Max Pooling 3DGlobal Max Pooling 3D-256
PredictionDenseSoftmax3 or 2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yokoi, N.; Kusada, N.; Kato, H.; Furusawa, Y.; Sotozono, C.; Georgiev, G.A. Dry Eye Subtype Classification Using Videokeratography and Deep Learning. Diagnostics 2024, 14, 52. https://doi.org/10.3390/diagnostics14010052

AMA Style

Yokoi N, Kusada N, Kato H, Furusawa Y, Sotozono C, Georgiev GA. Dry Eye Subtype Classification Using Videokeratography and Deep Learning. Diagnostics. 2024; 14(1):52. https://doi.org/10.3390/diagnostics14010052

Chicago/Turabian Style

Yokoi, Norihiko, Natsuki Kusada, Hiroaki Kato, Yuki Furusawa, Chie Sotozono, and Georgi As. Georgiev. 2024. "Dry Eye Subtype Classification Using Videokeratography and Deep Learning" Diagnostics 14, no. 1: 52. https://doi.org/10.3390/diagnostics14010052

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop