1
|
Woźniak M, Połap D, Capizzi G, Sciuto GL, Kośmider L, Frankiewicz K. Small lung nodules detection based on local variance analysis and probabilistic neural network. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2018; 161:173-180. [PMID: 29852959 DOI: 10.1016/j.cmpb.2018.04.025] [Citation(s) in RCA: 34] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/14/2018] [Revised: 04/10/2018] [Accepted: 04/26/2018] [Indexed: 06/08/2023]
Abstract
BACKGROUND AND OBJECTIVE In medical examinations doctors use various techniques in order to provide to the patients an accurate analysis of their actual state of health. One of the commonly used methodologies is the x-ray screening. This examination very often help to diagnose some diseases of chest organs. The most frequent cause of wrong diagnosis lie in the radiologist's difficulty in interpreting the presence of lungs carcinoma in chest X-ray. In such circumstances, an automated approach could be highly advantageous as it provides important help in medical diagnosis. METHODS In this paper we propose a new classification method of the lung carcinomas. This method start with the localization and extraction of the lung nodules by computing, for each pixel of the original image, the local variance obtaining an output image (variance image) with the same size of the original image. In the variance image we find the local maxima and then by using the locations of these maxima in the original image we found the contours of the possible nodules in lung tissues. However after this segmentation stage we find many false nodules. Therefore to discriminate the true ones we use a probabilistic neural network as classifier. RESULTS The performance of our approach is 92% of correct classifications, while the sensitivity is 95% and the specificity is 89.7%. The misclassification errors are due to the fact that network confuses false nodules with the true ones (6%) and true nodules with the false ones (2%). CONCLUSIONS Several researchers have proposed automated algorithms to detect and classify pulmonary nodules but these methods fail to detect low-contrast nodules and have a high computational complexity, in contrast our method is relatively simple but at the same time provides good results and can detect low-contrast nodules. Furthermore, in this paper is presented a new algorithm for training the PNN neural networks that allows to obtain PNNs with many fewer neurons compared to the neural networks obtained by using the training algorithms present in the literature. So considerably lowering the computational burden of the trained network and at same time keeping the same performances.
Collapse
|
|
7 |
34 |
2
|
Application of breast cancer diagnosis based on a combination of convolutional neural networks, ridge regression and linear discriminant analysis using invasive breast cancer images processed with autoencoders. Med Hypotheses 2019; 135:109503. [PMID: 31760247 DOI: 10.1016/j.mehy.2019.109503] [Citation(s) in RCA: 25] [Impact Index Per Article: 4.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/22/2019] [Revised: 11/09/2019] [Accepted: 11/16/2019] [Indexed: 12/13/2022]
Abstract
Invasive ductal carcinoma cancer, which invades the breast tissues by destroying the milk channels, is the most common type of breast cancer in women. Approximately, 80% of breast cancer patients have invasive ductal carcinoma and roughly 66.6% of these patients are older than 55 years. This situation points out a powerful relationship between the type of breast cancer and progressed woman age. In this study, the classification of invasive ductal carcinoma breast cancer is performed by using deep learning models, which is the sub-branch of artificial intelligence. In this scope, convolutional neural network models and the autoencoder network model are combined. In the experiment, the dataset was reconstructed by processing with the autoencoder model. The discriminative features obtained from convolutional neural network models were utilized. As a result, the most efficient features were determined by using the ridge regression method, and classification was performed using linear discriminant analysis. The best success rate of classification was achieved as 98.59%. Consequently, the proposed approach can be admitted as a successful model in the classification.
Collapse
|
Journal Article |
6 |
25 |
3
|
Karar ME, Hemdan EED, Shouman MA. Cascaded deep learning classifiers for computer-aided diagnosis of COVID-19 and pneumonia diseases in X-ray scans. COMPLEX INTELL SYST 2020; 7:235-247. [PMID: 34777953 PMCID: PMC7507595 DOI: 10.1007/s40747-020-00199-4] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2020] [Accepted: 09/09/2020] [Indexed: 12/18/2022]
Abstract
Computer-aided diagnosis (CAD) systems are considered a powerful tool for physicians to support identification of the novel Coronavirus Disease 2019 (COVID-19) using medical imaging modalities. Therefore, this article proposes a new framework of cascaded deep learning classifiers to enhance the performance of these CAD systems for highly suspected COVID-19 and pneumonia diseases in X-ray images. Our proposed deep learning framework constitutes two major advancements as follows. First, complicated multi-label classification of X-ray images have been simplified using a series of binary classifiers for each tested case of the health status. That mimics the clinical situation to diagnose potential diseases for a patient. Second, the cascaded architecture of COVID-19 and pneumonia classifiers is flexible to use different fine-tuned deep learning models simultaneously, achieving the best performance of confirming infected cases. This study includes eleven pre-trained convolutional neural network models, such as Visual Geometry Group Network (VGG) and Residual Neural Network (ResNet). They have been successfully tested and evaluated on public X-ray image dataset for normal and three diseased cases. The results of proposed cascaded classifiers showed that VGG16, ResNet50V2, and Dense Neural Network (DenseNet169) models achieved the best detection accuracy of COVID-19, viral (Non-COVID-19) pneumonia, and bacterial pneumonia images, respectively. Furthermore, the performance of our cascaded deep learning classifiers is superior to other multi-label classification methods of COVID-19 and pneumonia diseases in previous studies. Therefore, the proposed deep learning framework presents a good option to be applied in the clinical routine to assist the diagnostic procedures of COVID-19 infection.
Collapse
|
research-article |
5 |
23 |
4
|
Application of optical flow algorithms to laser speckle imaging. Microvasc Res 2018; 122:52-59. [PMID: 30414869 DOI: 10.1016/j.mvr.2018.11.001] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2018] [Revised: 11/06/2018] [Accepted: 11/07/2018] [Indexed: 11/22/2022]
Abstract
Since of its introduction in 1980s, laser speckle imaging has become a powerful tool in flow imaging. Its high performance and low cost made it one of the preferable imaging methods. Initially, speckle contrast measurements were the main algorithm for analyzing laser speckle images in biological flows. Speckle contrast measurements, also referred as Laser Speckle Contrast Imaging (LSCI), use statistical properties of speckle patterns to create mapped image of the blood vessels. In this communication, a new method named Laser Speckle Optical Flow Imaging (LSOFI) is introduced. This method uses the optical flow algorithms to calculate the apparent motion of laser speckle patterns. The differences in the apparent motion of speckle patterns are used to identify the blood vessels from surrounding tissue. LSOFI has better spatial and temporal resolution compared to LSCI. This higher spatial resolution enables LSOFI to be used for autonomous blood vessels detection. Furthermore, Graphics Processing Unit (GPU) based LSOFI can be used for quasi real time imaging.
Collapse
|
|
7 |
11 |
5
|
Alhudhaif A, Cömert Z, Polat K. Otitis media detection using tympanic membrane images with a novel multi-class machine learning algorithm. PeerJ Comput Sci 2021; 7:e405. [PMID: 33817048 PMCID: PMC7959604 DOI: 10.7717/peerj-cs.405] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2020] [Accepted: 01/30/2021] [Indexed: 05/10/2023]
Abstract
BACKGROUND Otitis media (OM) is the infection and inflammation of the mucous membrane covering the Eustachian with the airy cavities of the middle ear and temporal bone. OM is also one of the most common ailments. In clinical practice, the diagnosis of OM is carried out by visual inspection of otoscope images. This vulnerable process is subjective and error-prone. METHODS In this study, a novel computer-aided decision support model based on the convolutional neural network (CNN) has been developed. To improve the generalized ability of the proposed model, a combination of the channel and spatial model (CBAM), residual blocks, and hypercolumn technique is embedded into the proposed model. All experiments were performed on an open-access tympanic membrane dataset that consists of 956 otoscopes images collected into five classes. RESULTS The proposed model yielded satisfactory classification achievement. The model ensured an overall accuracy of 98.26%, sensitivity of 97.68%, and specificity of 99.30%. The proposed model produced rather superior results compared to the pre-trained CNNs such as AlexNet, VGG-Nets, GoogLeNet, and ResNets. Consequently, this study points out that the CNN model equipped with the advanced image processing techniques is useful for OM diagnosis. The proposed model may help to field specialists in achieving objective and repeatable results, decreasing misdiagnosis rate, and supporting the decision-making processes.
Collapse
|
research-article |
4 |
11 |
6
|
İçer S, Acer İ, Baş A. Gender-based functional connectivity differences in brain networks in childhood. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2020; 192:105444. [PMID: 32200049 DOI: 10.1016/j.cmpb.2020.105444] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/05/2019] [Revised: 03/02/2020] [Accepted: 03/10/2020] [Indexed: 06/10/2023]
Abstract
BACKGROUND AND OBJECTIVE Understanding the effect of gender differences on the brain can provide important information to characterize normal changes throughout life and to increase the likelihood of sex-specific approaches for neurological and psychiatric diseases. In this study, Functional Connectivity (FC), Amplitude of Low-Frequency Fluctuations (ALFF) and fractional ALFF (fALFF) analyzes will be compared between female and male brains between the ages of 7 and 18 years using resting state-functional magnetic resonance imaging (rs-fMRI). METHODS The rs-fMRI data in this study has been provided by The New York University (NYU) Child Study Center of the publicly shared ADHD200 database. From the NYU dataset, 68 (34 females, 34 males) healthy subjects in the age range of 7-18 years were selected. The female group (mean age: 12.3271±3.1380) and male group (mean age: 11.8766±2.9697) consisted of right-handed, small head motion and similar IQ values. FC was obtained by seed voxel analysis and the effect of low-frequency fluctuations on gender was examined by ALFF and fALFF analyses. Two-sample t-test was used to compare female and male groups with the significance thresholds set to FDR-corrected p<0.05. RESULTS In the results of our study, both in the ALFF, fALFF analyses and the seed regions belonging to many network regions, higher FC rates were found in girls than boys. Our results show that the females' language functions, visual functions such as object detection and recognition, working memory, executive functions, and episodic memory are more developed than males in this age range. In addition, as another result of our study, the seed regions are statistically stronger where the higher activation of female participants than male participants has concentrated in the left hemisphere. CONCLUSIONS Gender differences in brain networks should be taken into consideration when examining childhood cognitive and neuropsychiatric disorders and the results should also be evaluated according to gender. Evaluation of gender differences in childhood can increase the likelihood of early and definitive diagnosis and correct treatment for neurological diseases and can help doctors and scientists find new diagnostic tools to discover brain differences.
Collapse
|
Comparative Study |
5 |
11 |
7
|
Wei J, Li G. Automated Lung Segmentation and Image Quality Assessment for Clinical 3D/4D Computed Tomography. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE-JTEHM 2014; 2. [PMID: 25621194 PMCID: PMC4302269 DOI: 10.1109/jtehm.2014.2381213] [Citation(s) in RCA: 11] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Four-dimensional computed tomography (4DCT) provides not only a new dimension of patient-specific information for radiation therapy planning and treatment but also a challenging scale of data volume to process and analyze. Manual analysis using existing 3D tools is unable to keep up with vastly increased 4D data volume, automated processing and analysis are thus needed to process 4DCT data effectively and efficiently. In this work, we applied ideas and algorithms from image/signal processing, computer vision and machine learning to 4DCT lung data so that lungs can be reliably segmented in a fully-automated manner, lung features can be visualized and measured on-the-fly via user interactions, and data quality classifications can be computed in a robust manner. Comparisons of our results with an established treatment planning system and calculation by experts demonstrated negligible discrepancies (within ±2%) for volume assessment but one to two orders of magnitude performance enhancement. An empirical Fourier-analysis-based quality measure delivered performances closely emulating human experts. Three machine learners are inspected to justify the viability of machine learning techniques used to robustly identify data quality of 4DCT images in the scalable manner. The resultant system provides tools that speeds up 4D tasks in the clinic and facilitates clinical research to improve current clinical practice.
Collapse
|
Journal Article |
11 |
11 |
8
|
A simple and accurate method for computer-aided transapical aortic valve replacement. Comput Med Imaging Graph 2014; 50:31-41. [PMID: 25306532 DOI: 10.1016/j.compmedimag.2014.09.005] [Citation(s) in RCA: 8] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2014] [Revised: 07/21/2014] [Accepted: 09/12/2014] [Indexed: 11/22/2022]
Abstract
BACKGROUND AND PURPOSE Transapical aortic valve replacement (TAVR) is a recent minimally invasive surgical treatment technique for elderly and high-risk patients with severe aortic stenosis. In this paper, a simple and accurate image-based method is introduced to aid the intra-operative guidance of TAVR procedure under 2-D X-ray fluoroscopy. METHODS The proposed method fuses a 3-D aortic mesh model and anatomical valve landmarks with live 2-D fluoroscopic images. The 3-D aortic mesh model and landmarks are reconstructed from interventional X-ray C-arm CT system, and a target area for valve implantation is automatically estimated using these aortic mesh models. Based on template-based tracking approach, the overlay of visualized 3-D aortic mesh model, landmarks and target area of implantation is updated onto fluoroscopic images by approximating the aortic root motion from a pigtail catheter motion without contrast agent. Also, a rigid intensity-based registration algorithm is used to track continuously the aortic root motion in the presence of contrast agent. Furthermore, a sensorless tracking of the aortic valve prosthesis is provided to guide the physician to perform the appropriate placement of prosthesis into the estimated target area of implantation. RESULTS Retrospective experiments were carried out on fifteen patient datasets from the clinical routine of the TAVR. The maximum displacement errors were less than 2.0mm for both the dynamic overlay of aortic mesh models and image-based tracking of the prosthesis, and within the clinically accepted ranges. Moreover, high success rates of the proposed method were obtained above 91.0% for all tested patient datasets. CONCLUSION The results showed that the proposed method for computer-aided TAVR is potentially a helpful tool for physicians by automatically defining the accurate placement position of the prosthesis during the surgical procedure.
Collapse
|
Journal Article |
11 |
8 |
9
|
Automated segmentation of villi in histopathology images of placenta. Comput Biol Med 2019; 113:103420. [PMID: 31514041 DOI: 10.1016/j.compbiomed.2019.103420] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2019] [Revised: 08/12/2019] [Accepted: 08/27/2019] [Indexed: 12/17/2022]
Abstract
PURPOSE Manual analysis of clinical placenta pathology samples under the microscope is a costly and time-consuming task. Computer-aided diagnosis might offer a means to obtain fast and reliable results and also substantially reduce inter- and intra-rater variability. Here, we present a fully automated segmentation method that is capable of distinguishing the complex histological features of the human placenta (i.e., the chorionic villous structures). METHODS The proposed pipeline consists of multiple steps to segment individual placental villi structures in hematoxylin and eosin (H&E) stained placental images. Artifacts and undesired objects in the histological field of view are detected and excluded from further analysis. One of the challenges in our new algorithm is the detection and segmentation of touching villi in our dataset. The proposed algorithm uses the top-hat transformation to detect candidate concavities in each structure, which might represent two distinct villous structures in close proximity. The detected concavities are classified by extracting multiple features from each candidate concavity. Our proposed pipeline is evaluated against manual segmentations, confirmed by an expert pathologist, on 12 scans from three healthy control patients and nine patients diagnosed with preeclampsia, containing nearly 5000 individual villi. The results of our method are compared to a previously published method for villi segmentation. RESULTS Our algorithm detected placental villous structures with an F1 score of 80.76% and sensitivity of 82.18%. These values are substantially better than the previously published method, whose F1 score and sensitivity are 65.30% and 55.12%, respectively. CONCLUSION Our method is capable of distinguishing the complex histological features of the human placenta (i.e., the chorionic villous structures), removing artifacts over a large histopathology sample of human placenta, and (importantly) account for touching adjacent villi structures. Compared to existing methods, our developed method yielded high accuracy in detecting villi in placental images.
Collapse
|
Research Support, Non-U.S. Gov't |
6 |
8 |
10
|
Farsoni S, Astolfi L, Bonfe M, Spadaro S, Volta CA. A Versatile Ultrasound Simulation System for Education and Training in High-Fidelity Emergency Scenarios. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE-JTEHM 2017; 5:1800109. [PMID: 29018630 PMCID: PMC5477762 DOI: 10.1109/jtehm.2016.2635635] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/01/2016] [Revised: 05/19/2016] [Accepted: 11/14/2016] [Indexed: 11/26/2022]
Abstract
Point of care ultrasonography and the related focused assessment with sonography for trauma protocol, if performed by experienced physicians, is a highly sensitive examination, and specific for the detection of free fluids. Different systems and methods have been proposed for training, including simulation as one of the most efficient. This paper presents an ultrasound training system, specifically designed to be used during bedside high fidelity simulation scenarios, that could facilitate the learning process. The development of the proposed system exploited novel rapid prototyping electronic boards as a means to obtain good performances with a low cost. Moreover, the design of the data structure permits the construction of a library that caters for individual needs, with the possibility of adding emergency scenarios, collecting pictures or videos, as well as 3-D volumes. The device has been compared with currently commercial ultrasound simulators and its innovative aspects have been highlighted. Finally, it has been tested during a training session in order to evaluate features, such as realism and user-friendliness.
Collapse
|
Journal Article |
8 |
7 |
11
|
A Computer-Aided Diagnosis System for Measuring Carotid Artery Intima-Media Thickness (IMT) Using Quaternion Vectors. J Med Syst 2016; 40:149. [PMID: 27137786 DOI: 10.1007/s10916-016-0507-4] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2016] [Accepted: 04/19/2016] [Indexed: 10/21/2022]
Abstract
This study aims investigating adjustable distant fuzzy c-means segmentation on carotid Doppler images, as well as quaternion-based convolution filters and saliency mapping procedures. We developed imaging software that will simplify the measurement of carotid artery intima-media thickness (IMT) on saliency mapping images. Additionally, specialists evaluated the present images and compared them with saliency mapping images. In the present research, we conducted imaging studies of 25 carotid Doppler images obtained by the Department of Cardiology at Fırat University. After implementing fuzzy c-means segmentation and quaternion-based convolution on all Doppler images, we obtained a model that can be analyzed easily by the doctors using a bottom-up saliency model. These methods were applied to 25 carotid Doppler images and then interpreted by specialists. In the present study, we used color-filtering methods to obtain carotid color images. Saliency mapping was performed on the obtained images, and the carotid artery IMT was detected and interpreted on the obtained images from both methods and the raw images are shown in Results. Also these results were investigated by using Mean Square Error (MSE) for the raw IMT images and the method which gives the best performance is the Quaternion Based Saliency Mapping (QBSM). 0,0014 and 0,000191 mm(2) MSEs were obtained for artery lumen diameters and plaque diameters in carotid arteries respectively. We found that computer-based image processing methods used on carotid Doppler could aid doctors' in their decision-making process. We developed software that could ease the process of measuring carotid IMT for cardiologists and help them to evaluate their findings.
Collapse
|
Journal Article |
9 |
6 |
12
|
Szostek K, Piórkowski A. Real-time simulation of ultrasound refraction phenomena using ray-trace based wavefront construction method. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2016; 135:187-197. [PMID: 27586490 DOI: 10.1016/j.cmpb.2016.07.034] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/20/2015] [Revised: 07/06/2016] [Accepted: 07/29/2016] [Indexed: 06/06/2023]
Abstract
Ultrasound (US) imaging is one of the most popular techniques used in clinical diagnosis, mainly due to lack of adverse effects on patients and the simplicity of US equipment. However, the characteristics of the medium cause US imaging to imprecisely reconstruct examined tissues. The artifacts are the results of wave phenomena, i.e. diffraction or refraction, and should be recognized during examination to avoid misinterpretation of an US image. Currently, US training is based on teaching materials and simulators and ultrasound simulation has become an active research area in medical computer science. Many US simulators are limited by the complexity of the wave phenomena, leading to intensive sophisticated computation that makes it difficult for systems to operate in real time. To achieve the required frame rate, the vast majority of simulators reduce the problem of wave diffraction and refraction. The following paper proposes a solution for an ultrasound simulator based on methods known in geophysics. To improve simulation quality, a wavefront construction method was adapted which takes into account the refraction phenomena. This technique uses ray tracing and velocity averaging to construct wavefronts in the simulation. Instead of a geological medium, real CT scans are applied. This approach can produce more realistic projections of pathological findings and is also capable of providing real-time simulation.
Collapse
|
|
9 |
3 |
13
|
DSA image registration using non-uniform MRF model and pivotal control points. Comput Med Imaging Graph 2013; 37:323-36. [PMID: 23726230 DOI: 10.1016/j.compmedimag.2013.04.006] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2012] [Revised: 04/02/2013] [Accepted: 04/30/2013] [Indexed: 11/20/2022]
Abstract
In order to reduce the motion artifacts in DSA, non-rigid image registration is commonly used before subtracting the mask from the contrast image. Since DSA registration requires a set of spatially non-uniform control points, a conventional MRF model is not very efficient. In this paper, we introduce the concept of pivotal and non-pivotal control points to address this, and propose a non-uniform MRF for DSA registration. We use quad-trees in a novel way to generate the non-uniform grid of control points. Our MRF formulation produces a smooth displacement field and therefore results in better artifact reduction than that of registering the control points independently. We achieve improved computational performance using pivotal control points without compromising on the artifact reduction. We have tested our approach using several clinical data sets, and have presented the results of quantitative analysis, clinical assessment and performance improvement on a GPU.
Collapse
|
|
12 |
3 |
14
|
Chen F, Jiang Y, Zeng X, Zhang J, Gao X, Xu M. PUB-SalNet: A Pre-trained Unsupervised Self-Aware Backpropagation Network for Biomedical Salient Segmentation. ALGORITHMS 2020; 13:126. [PMID: 34567437 PMCID: PMC8460134 DOI: 10.3390/a13050126] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Salient segmentation is a critical step in biomedical image analysis, aiming to cut out regions that are most interesting to humans. Recently, supervised methods have achieved promising results in biomedical areas, but they depend on annotated training data sets, which requires labor and proficiency in related background knowledge. In contrast, unsupervised learning makes data-driven decisions by obtaining insights directly from the data themselves. In this paper, we propose a completely unsupervised self-aware network based on pre-training and attentional backpropagation for biomedical salient segmentation, named as PUB-SalNet. Firstly, we aggregate a new biomedical data set from several simulated Cellular Electron Cryo-Tomography (CECT) data sets featuring rich salient objects, different SNR settings and various resolutions, which is called SalSeg-CECT. Based on the SalSeg-CECT data set, we then pre-train a model specially designed for biomedical tasks as a backbone module to initialize network parameters. Next, we present a U-SalNet network to learn to selectively attend to salient objects. It includes two types of attention modules to facilitate learning saliency through global contrast and local similarity. Lastly, we jointly refine the salient regions together with feature representations from U-SalNet, with the parameters updated by self-aware attentional backpropagation. We apply PUB-SalNet for analysis of 2D simulated and real images and achieve state-of-the-art performance on simulated biomedical data sets. Furthermore, our proposed PUB-SalNet can be easily extended to 3D images. The experimental results on the 2d and 3d data sets also demonstrate the generalization ability and robustness of our method.
Collapse
|
research-article |
5 |
2 |
15
|
Choi SJ, Kim DK, Kim BS, Cho M, Jeong J, Jo YH, Song KJ, Kim YJ, Kim S. Mask R-CNN based multiclass segmentation model for endotracheal intubation using video laryngoscope. Digit Health 2023; 9:20552076231211547. [PMID: 38025115 PMCID: PMC10631336 DOI: 10.1177/20552076231211547] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2023] [Accepted: 10/17/2023] [Indexed: 12/01/2023] Open
Abstract
Objective Endotracheal intubation (ETI) is critical to secure the airway in emergent situations. Although artificial intelligence algorithms are frequently used to analyze medical images, their application to evaluating intraoral structures based on images captured during emergent ETI remains limited. The aim of this study is to develop an artificial intelligence model for segmenting structures in the oral cavity using video laryngoscope (VL) images. Methods From 54 VL videos, clinicians manually labeled images that include motion blur, foggy vision, blood, mucus, and vomitus. Anatomical structures of interest included the tongue, epiglottis, vocal cord, and corniculate cartilage. EfficientNet-B5 with DeepLabv3+, EffecientNet-B5 with U-Net, and Configured Mask R-Convolution Neural Network (CNN) were used; EffecientNet-B5 was pretrained on ImageNet. Dice similarity coefficient (DSC) was used to measure the segmentation performance of the model. Accuracy, recall, specificity, and F1 score were used to evaluate the model's performance in targeting the structure from the value of the intersection over union between the ground truth and prediction mask. Results The DSC of tongue, epiglottis, vocal cord, and corniculate cartilage obtained from the EfficientNet-B5 with DeepLabv3+, EfficientNet-B5 with U-Net, and Configured Mask R-CNN model were 0.3351/0.7675/0.766/0.6539, 0.0/0.7581/0.7395/0.6906, and 0.1167/0.7677/0.7207/0.57, respectively. Furthermore, the processing speeds (frames per second) of the three models stood at 3, 24, and 32, respectively. Conclusions The algorithm developed in this study can assist medical providers performing ETI in emergent situations.
Collapse
|
research-article |
2 |
1 |
16
|
Haq I, Mazhar T, Asif RN, Ghadi YY, Ullah N, Khan MA, Al-Rasheed A. YOLO and residual network for colorectal cancer cell detection and counting. Heliyon 2024; 10:e24403. [PMID: 38304780 PMCID: PMC10831604 DOI: 10.1016/j.heliyon.2024.e24403] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2023] [Revised: 12/30/2023] [Accepted: 01/08/2024] [Indexed: 02/03/2024] Open
Abstract
The HT-29 cell line, derived from human colon cancer, is valuable for biological and cancer research applications. Early detection is crucial for improving the chances of survival, and researchers are introducing new techniques for accurate cancer diagnosis. This study introduces an efficient deep learning-based method for detecting and counting colorectal cancer cells (HT-29). The colorectal cancer cell line was procured from a company. Further, the cancer cells were cultured, and a transwell experiment was conducted in the lab to collect the dataset of colorectal cancer cell images via fluorescence microscopy. Of the 566 images, 80 % were allocated to the training set, and the remaining 20 % were assigned to the testing set. The HT-29 cell detection and counting in medical images is performed by integrating YOLOv2, ResNet-50, and ResNet-18 architectures. The accuracy achieved by ResNet-18 is 98.70 % and ResNet-50 is 96.66 %. The study achieves its primary objective by focusing on detecting and quantifying congested and overlapping colorectal cancer cells within the images. This innovative work constitutes a significant development in overlapping cancer cell detection and counting, paving the way for novel advancements and opening new avenues for research and clinical applications. Researchers can extend the study by exploring variations in ResNet and YOLO architectures to optimize object detection performance. Further investigation into real-time deployment strategies will enhance the practical applicability of these models.
Collapse
|
research-article |
1 |
|
17
|
Hasebe K, Kojima T, Okanoue Y, Yuki R, Yamamoto H, Otsuki S, Fujimura S, Hori R. Novel evaluation method for facial nerve palsy using 3D facial recognition system in iPhone. Auris Nasus Larynx 2024; 51:460-464. [PMID: 38520978 DOI: 10.1016/j.anl.2024.02.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/02/2023] [Revised: 02/05/2024] [Accepted: 02/07/2024] [Indexed: 03/25/2024]
Abstract
OBJECTIVE While subjective methods like the Yanagihara system and the House-Brackmann system are standard in evaluating facial paralysis, they are limited by intra- and inter-observer variability. Meanwhile, quantitative objective methods such as electroneurography and electromyography are time-consuming. Our aim was to introduce a swift, objective, and quantitative method for evaluating facial movements. METHODS We developed an application software (app) that utilizes the facial recognition functionality of the iPhone (Apple Inc., Cupertino, USA) for facial movement evaluation. This app leverages the phone's front camera, infrared radiation, and infrared camera to provide detailed three-dimensional facial topology. It quantitatively compares left and right facial movements by region and displays the movement ratio of the affected side to the opposite side. Evaluations using the app were conducted on both normal and facial palsy subjects and were compared with conventional methods. RESULTS Our app provided an intuitive user experience, completing evaluations in under a minute, and thus proving practical for regular use. Its evaluation scores correlated highly with the Yanagihara system, the House-Brackmann system, and electromyography. Furthermore, the app outperformed conventional methods in assessing detailed facial movements. CONCLUSION Our novel iPhone app offers a valuable tool for the comprehensive and efficient evaluation of facial palsy.
Collapse
|
|
1 |
|
18
|
Lin C, Chen H, Huang J, Peng J, Guo L, Yang Z, Du J, Li S, Yin A, Zhao G. ChromosomeNet: A massive dataset enabling benchmarking and building basedlines of clinical chromosome classification. Comput Biol Chem 2022; 100:107731. [PMID: 35907293 DOI: 10.1016/j.compbiolchem.2022.107731] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2022] [Accepted: 07/12/2022] [Indexed: 11/30/2022]
Abstract
Chromosome karyotyping analysis is a vital cytogenetics technique for diagnosing genetic and congenital malformations, analyzing gestational and implantation failures, etc. Since the chromosome classification as an essential stage in chromosome karyotype analysis is a highly time-consuming, tedious, and error-prone task, which requires a large amount of manual work of experienced cytogenetics experts. Many deep learning-based methods have been proposed to address the chromosome classification issues. However, two challenges still remain in current chromosome classification methods. First, most existing methods were developed by different private datasets, making these methods difficult to compare with each other on the same base. Second, due to the absence of reproducing details of most existing methods, these methods are difficult to be applied in clinical chromosome classification applications widely. To address the above challenges in the chromosome classification issue, this work builds and publishes a massive clinical dataset. This dataset enables the benchmarking and building chromosome classification baselines suitable for different scenarios. The massive clinical dataset consists of 126,453 privacy preserving G-band chromosome instances from 2763 karyotypes of 408 individuals. To our best knowledge, it is the first work to collect, annotate, and release a publicly available clinical chromosome classification dataset whose data size scale is also over 120,000. Meanwhile, the experimental results show that the proposed dataset can boost performance of existing chromosome classification models at a varied range of degrees, with the highest accuracy improvement by 5.39 % points. Moreover, the best baseline with 99.33 % accuracy reports state-of-the-art classification performance. The clinical dataset and state-of-the-art baselines can be found at https://github.com/CloudDataLab/BenchmarkForChromosomeClassification.
Collapse
|
|
3 |
|
19
|
Farber JM, Totterman SMS, Martinez-Torteya A, Tamez-Peña JG. Scan-rescan precision of subchondral bone curvature maps from routine 3D DESS water excitation sequences: Data from the Osteoarthritis Initiative. Comput Biol Med 2016; 69:83-91. [PMID: 26751403 DOI: 10.1016/j.compbiomed.2015.12.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2015] [Revised: 12/11/2015] [Accepted: 12/12/2015] [Indexed: 10/22/2022]
Abstract
BACKGROUND Subchondral bone (SCB) undergoes changes in the shape of the articulating bone surfaces and is currently recognized as a key target in osteoarthritis (OA) treatment. The aim of this study was to present an automated system that determines the curvature of the SCB regions of the knee and to evaluate its cross-sectional and longitudinal scan-rescan precision METHODS Six subjects with OA and six control subjects were selected from the Osteoarthritis Initiative (OAI) pilot study database. As per OAI protocol, these subjects underwent 3T MRI at baseline and every twelve months thereafter, including a 3D DESS WE sequence. We analyzed the baseline and twenty-four month images. Each subject was scanned twice at these visits, thus generating scan-rescan information. Images were segmented with an automated multi-atlas framework platform and then 3D renderings of the bone structure were created from the segmentations. Curvature maps were extracted from the 3D renderings and morphed into a reference atlas to determine precision, to generate population statistics, and to visualize cross-sectional and longitudinal curvature changes. RESULTS The baseline scan-rescan root mean square error values ranged from 0.006mm(-1) to 0.013mm(-1), and from 0.007mm(-1) to 0.018mm(-1) for the SCB of the femur and the tibia, respectively. The standardized response of the mean of the longitudinal changes in curvature in these regions ranged from -0.09 to 0.02 and from -0.016 to 0.015, respectively. CONCLUSION The fully automated system produces accurate and precise curvature maps of femoral and tibial SCB, and will provide a valuable tool for the analysis of the curvature changes of articulating bone surfaces during the course of knee OA.
Collapse
|
Research Support, N.I.H., Extramural |
9 |
|
20
|
Xiang J, Yang W, Zhang H, Zhu F, Pu S, Li R, Wang C, Yan Z, Li W. Digital signal extraction approach for cardiotocography image. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 225:107089. [PMID: 36058063 DOI: 10.1016/j.cmpb.2022.107089] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/07/2022] [Revised: 07/27/2022] [Accepted: 08/23/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Cardiotocography, commonly called CTG, has become an indispensable auxiliary examination in obstetrics. Generally, CTG is provided in the form of a report, so the fetal heart rate and uterine contraction signals have to be extracted from the CTG images. However, most studies focused on reading data for a single curve, and the influence of complex backgrounds was usually not considered. METHODS An efficient signal extraction method was proposed for the binary CTG images with complex backgrounds. Firstly, the images' background grids and symbol noise were removed by templates. Then a morphological method was used to fill breakpoints of curves. Moreover, the projection map was utilized to localize the area and the starting and ending positions of curves. Subsequently, data of the curves were extracted by column scanning. Finally, the amplitude of the extracted signal was calibrated. RESULTS This study had tested 552 CTG images simulated using the CTU-UHB database. The correlation coefficient between the extracted and original signals was 0.9991 ± 0.0030 for fetal heart rate and 0.9904 ± 0.0208 for uterine contraction, and the mean absolute error of fetal heart rate and uterine contraction were 2.4658 ± 1.8446 and 1.8025 ± 0.6155, and the root mean square error of fetal heart rate and uterine contraction were 4.2930 ± 2.9771 and 2.5214 ± 0.9640, respectively. After being validated using 293 clinical authentic CTG images, the extracted signals were remarkably similar to the original counterparts, and no significant differences were observed. CONCLUSIONS The proposed method could effectively extract the fetal heart rate and uterine contraction signals from the binary CTG images with complex backgrounds.
Collapse
|
|
3 |
|
21
|
Kaplan E, Chan WY, Altinsoy HB, Baygin M, Barua PD, Chakraborty S, Dogan S, Tuncer T, Acharya UR. PFP-HOG: Pyramid and Fixed-Size Patch-Based HOG Technique for Automated Brain Abnormality Classification with MRI. J Digit Imaging 2023; 36:2441-2460. [PMID: 37537514 PMCID: PMC10584767 DOI: 10.1007/s10278-023-00889-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2023] [Revised: 07/18/2023] [Accepted: 07/24/2023] [Indexed: 08/05/2023] Open
Abstract
Detecting neurological abnormalities such as brain tumors and Alzheimer's disease (AD) using magnetic resonance imaging (MRI) images is an important research topic in the literature. Numerous machine learning models have been used to detect brain abnormalities accurately. This study addresses the problem of detecting neurological abnormalities in MRI. The motivation behind this problem lies in the need for accurate and efficient methods to assist neurologists in the diagnosis of these disorders. In addition, many deep learning techniques have been applied to MRI to develop accurate brain abnormality detection models, but these networks have high time complexity. Hence, a novel hand-modeled feature-based learning network is presented to reduce the time complexity and obtain high classification performance. The model proposed in this work uses a new feature generation architecture named pyramid and fixed-size patch (PFP). The main aim of the proposed PFP structure is to attain high classification performance using essential feature extractors with both multilevel and local features. Furthermore, the PFP feature extractor generates low- and high-level features using a handcrafted extractor. To obtain the high discriminative feature extraction ability of the PFP, we have used histogram-oriented gradients (HOG); hence, it is named PFP-HOG. Furthermore, the iterative Chi2 (IChi2) is utilized to choose the clinically significant features. Finally, the k-nearest neighbors (kNN) with tenfold cross-validation is used for automated classification. Four MRI neurological databases (AD dataset, brain tumor dataset 1, brain tumor dataset 2, and merged dataset) have been utilized to develop our model. PFP-HOG and IChi2-based models attained 100%, 94.98%, 98.19%, and 97.80% using the AD dataset, brain tumor dataset1, brain tumor dataset 2, and merged brain MRI dataset, respectively. These findings not only provide an accurate and robust classification of various neurological disorders using MRI but also hold the potential to assist neurologists in validating manual MRI brain abnormality screening.
Collapse
|
research-article |
2 |
|
22
|
Dutta S, Goswami S, Debnath S, Adhikary S, Majumder A. MusicalBSI - musical genres responses to fMRI signals analysis with prototypical model agnostic meta-learning for brain state identification in data scarce environment. Comput Biol Med 2025; 188:109795. [PMID: 39946786 DOI: 10.1016/j.compbiomed.2025.109795] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2024] [Revised: 12/04/2024] [Accepted: 02/02/2025] [Indexed: 03/05/2025]
Abstract
Functional magnetic resonance imaging is a popular non-invasive brain-computer interfacing technique to monitor brain activities corresponding to several physical or neurological responses by measuring blood flow changes at different brain parts. Recent studies have shown that blood flow within the brain can have signature activity patterns in response to various musical genres. However, limited studies exist in the state of the art for automatized recognition of the musical genres from functional magnetic resonance imaging. This is because the feasibility of obtaining these kinds of data is limited, and currently available open-sourced data is insufficient to build an accurate deep-learning model. To solve this, we propose a prototypical model agnostic meta-learning framework for accurately classifying musical genres by studying blood flow dynamics using functional magnetic resonance imaging. A test with open-sourced data collected from 20 human subjects with consent for 6 different mental states resulted in up to 97.25 ± 1.38% accuracy by training with only 30 samples surpassing state-of-the-art methods. Further, a detailed evaluation of the performances confirms the model's reliability.
Collapse
|
|
1 |
|
23
|
Sikder N, Khan MAM, Bairagi AK, Masud M, Tiang JJ, Nahid AA. Heterogeneous virus classification using a functional deep learning model based on transmission electron microscopy images. Sci Rep 2024; 14:28954. [PMID: 39578636 PMCID: PMC11584783 DOI: 10.1038/s41598-024-80013-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2024] [Accepted: 11/14/2024] [Indexed: 11/24/2024] Open
Abstract
Viruses are submicroscopic agents that can infect other lifeforms and use their hosts' cells to replicate themselves. Despite having simplistic genetic structures among all living beings, viruses are highly adaptable, resilient, and capable of causing severe complications in their hosts' bodies. Due to their multiple transmission pathways, high contagion rate, and lethality, viruses pose the biggest biological threat both animal and plant species face. It is often challenging to promptly detect a virus in a host and accurately determine its type using manual examination techniques. However, computer-based automatic diagnosis methods, especially the ones using Transmission Electron Microscopy (TEM) images, have proven effective in instant virus identification. Using TEM images collected from a recent dataset, this article proposes a deep learning-based classification model to identify the virus type within those images. The methodology of this study includes two coherent image processing techniques to reduce the noise present in raw microscopy images and a functional Convolutional Neural Network (CNN) model for classification. Experimental results show that it can differentiate among 14 types of viruses with a maximum of 97.44% classification accuracy and F1-score, which asserts the effectiveness and reliability of the proposed method. Implementing this scheme will impart a fast and dependable virus identification scheme subsidiary to the thorough diagnostic procedures.
Collapse
|
research-article |
1 |
|
24
|
Pande SD, Kalyani P, Nagendram S, Alluhaidan AS, Babu GH, Ahammad SH, Pandey VK, Sridevi G, Kumar A, Bonyah E. Comparative analysis of the DCNN and HFCNN Based Computerized detection of liver cancer. BMC Med Imaging 2025; 25:37. [PMID: 39901085 PMCID: PMC11792691 DOI: 10.1186/s12880-025-01578-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2024] [Accepted: 01/28/2025] [Indexed: 02/05/2025] Open
Abstract
Liver cancer detection is critically important in the discipline of biomedical image testing and diagnosis. Researchers have explored numerous machine learning (ML) techniques and deep learning (DL) approaches aimed at the automated recognition of liver disease by analysing computed tomography (CT) images. This study compares two frameworks, Deep Convolutional Neural Network (DCNN) and Hierarchical Fusion Convolutional Neural Networks (HFCNN), to assess their effectiveness in liver cancer segmentation. The contribution includes enhancing the edges and textures of CT images through filtering to achieve precise liver segmentation. Additionally, an existing DL framework was employed for liver cancer detection and segmentation. The strengths of this paper include a clear emphasis on the criticality of liver cancer detection in biomedical imaging and diagnostics. It also highlights the challenges associated with CT image detection and segmentation and provides a comprehensive summary of recent literature. However, certain difficulties arise during the detection process in CT images due to overlapping structures, such as bile ducts, blood vessels, image noise, textural changes, size and location variations, and inherent heterogeneity. These factors may lead to segmentation errors and subsequently different analyses. This research analysis compares two advanced methodologies, DCNN and HFCNN, for liver cancer detection. The evaluation of DCNN and HFCNN in liver cancer detection is conducted using multiple performance metrics, including precision, F1-score, recall, and accuracy. This comprehensive assessment provides a detailed evaluation of these models' effectiveness compared to other state-of-the-art methods in identifying liver cancer.
Collapse
|
Comparative Study |
1 |
|
25
|
Tarquino J, Rodríguez J, Becerra D, Roa-Peña L, Romero E. Engineered feature embeddings meet deep learning: A novel strategy to improve bone marrow cell classification and model transparency. J Pathol Inform 2024; 15:100390. [PMID: 39712979 PMCID: PMC11662281 DOI: 10.1016/j.jpi.2024.100390] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/24/2024] [Revised: 06/18/2024] [Accepted: 06/29/2024] [Indexed: 12/24/2024] Open
Abstract
Cytomorphology evaluation of bone marrow cell is the initial step to diagnose different hematological diseases. This assessment is still manually performed by trained specialists, who may be a bottleneck within the clinical process. Deep learning algorithms are a promising approach to automate this bone marrow cell evaluation. These artificial intelligence models have focused on limited cell subtypes, mainly associated to a particular disease, and are frequently presented as black boxes. The herein introduced strategy presents an engineered feature representation, the region-attention embedding, which improves the deep learning classification performance of a cytomorphology with 21 bone marrow cell subtypes. This embedding is built upon a specific organization of cytology features within a squared matrix by distributing them after pre-segmented cell regions, i.e., cytoplasm, nucleus, and whole-cell. This novel cell image representation, aimed to preserve spatial/regional relations, is used as input of the network. Combination of region-attention embedding and deep learning networks (Xception and ResNet50) provides local relevance associated to image regions, adding up interpretable information to the prediction. Additionally, this approach is evaluated in a public database with the largest number of cell subtypes (21) by a thorough evaluation scheme with three iterations of a 3-fold cross-validation, performed in 80% of the images (n = 89,484), and a testing process in an unseen set of images composed by the remaining 20% of the images (n = 22,371). This evaluation process demonstrates the introduced strategy outperforms previously published approaches in an equivalent validation set, with a f1-score of 0.82, and presented competitive results on the unseen data partition with a f1-score of 0.56.
Collapse
|
research-article |
1 |
|