RESUMEN
BACKGROUND: Tuberculosis (TB) is a highly infectious disease that mainly affects the human lungs. The gold standard for TB diagnosis is Xpert Mycobacterium tuberculosis/ resistance to rifampicin (MTB/RIF) testing. X-ray, a relatively inexpensive and widely used imaging modality, can be employed as an alternative for early diagnosis of the disease. Computer-aided techniques can be used to assist radiologists in interpreting X-ray images, which can improve the ease and accuracy of diagnosis. OBJECTIVE: To develop a computer-aided technique for the diagnosis of TB from X-ray images using deep learning techniques. METHODS: This research paper presents a novel approach for TB diagnosis from X-ray using deep learning methods. The proposed method uses an ensemble of two pre-trained neural networks, namely EfficientnetB0 and Densenet201, for feature extraction. The features extracted using two CNNs are expected to generate more accurate and representative features than a single CNN. A custom-built artificial neural network (ANN) called PatternNet with two hidden layers is utilized to classify the extracted features. RESULTS: The effectiveness of the proposed method was assessed on two publicly accessible datasets, namely the Montgomery and Shenzhen datasets. The Montgomery dataset comprises 138 X-ray images, while the Shenzhen dataset has 662 X-ray images. The method was further evaluated after combining both datasets. The method performed exceptionally well on all three datasets, achieving high Area Under the Curve (AUC) scores of 0.9978, 0.9836, and 0.9914, respectively, using a 10-fold cross-validation technique. CONCLUSION: The experiments performed in this study prove the effectiveness of features extracted using EfficientnetB0 and Densenet201 in combination with PatternNet classifier in the diagnosis of tuberculosis from X-ray images.
Asunto(s)
Tuberculosis , Humanos , Rayos X , Tuberculosis/diagnóstico por imagen , Redes Neurales de la Computación , Diagnóstico por Computador/métodos , ComputadoresRESUMEN
Video summarization aims to find a compact representation of input videos. The method finds out interesting parts of the video by discarding the remaining parts of the video. The abstracts thus generated enhances browsing and retrieval of video data. The quality of summaries generated by video summarization algorithms can be improved if the redundant frames in the input video are taken care of before summarization. This paper presents a novel domain-independent method for redundancy elimination from input videos before summarization maintaining keyframes in the original video. The frames of input video are first presampled by selecting two frames in one second. The flow vectors between consecutive frames are computed using SIFT Flow algorithm. The magnitude of flow vectors at each pixel position of the frame are summed up to find the displacement magnitude between the consecutive frames. The redundant frames are filtered out based on local averaging of the displacement values. The evaluation of the method is done using two standard datasets namely VSUMM and OVP. The results demonstrate that an average reduction rate of 97.64% is achieved consistently on videos of all categories. The method also gives superior results compared to other state-of-the-art redundancy elimination methods for video summarization.