Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 64
Filtrar
1.
Brief Bioinform ; 24(1)2023 01 19.
Artículo en Inglés | MEDLINE | ID: mdl-36502371

RESUMEN

Deoxyribonucleic acid(DNA) N6-methyladenine plays a vital role in various biological processes, and the accurate identification of its site can provide a more comprehensive understanding of its biological effects. There are several methods for 6mA site prediction. With the continuous development of technology, traditional techniques with the high costs and low efficiencies are gradually being replaced by computer methods. Computer methods that are widely used can be divided into two categories: traditional machine learning and deep learning methods. We first list some existing experimental methods for predicting the 6mA site, then analyze the general process from sequence input to results in computer methods and review existing model architectures. Finally, the results were summarized and compared to facilitate subsequent researchers in choosing the most suitable method for their work.


Asunto(s)
Metilación de ADN , Aprendizaje Automático , Proyectos de Investigación , ADN/genética
2.
J Appl Clin Med Phys ; 25(6): e14358, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38634799

RESUMEN

PURPOSE: We evaluate the performance of a deformable image registration (DIR) software package in registering abdominal magnetic resonance images (MRIs) and then develop a mechanical modeling method to mitigate detected DIR uncertainties. MATERIALS AND METHODS: Three evaluation metrics, namely mean displacement to agreement (MDA), DICE similarity coefficient (DSC), and standard deviation of Jacobian determinants (STD-JD), are used to assess the multi-modality (MM), contour-consistency (CC), and image-intensity (II)-based DIR algorithms in the MIM software package, as well as an in-house developed, contour matching-based finite element method (CM-FEM). Furthermore, we develop a hybrid FEM registration technique to modify the displacement vector field of each MIM registration. The MIM and FEM registrations were evaluated on MRIs obtained from 10 abdominal cancer patients. One-tailed Wilcoxon-Mann-Whitney (WMW) tests were conducted to compare the MIM registrations with their FEM modifications. RESULTS: For the registrations performed with the MIM-CC, MIM-MM, MIM-II, and CM-FEM algorithms, their average MDAs are 0.62 ± 0.27, 2.39 ± 1.30, 3.07 ± 2.42, 1.04 ± 0.72 mm, and average DSCs are 0.94 ± 0.03, 0.80 ± 0.12, 0.77 ± 0.15, 0.90 ± 0.11, respectively. The p-values of the WMW tests between the MIM registrations and their FEM modifications are less than 0.0084 for STD-JDs and greater than 0.87 for MDA and DSC. CONCLUSIONS: Among the three MIM DIR algorithms, MIM-CC shows the smallest errors in terms of MDA and DSC but exhibits significant Jacobian uncertainties in the interior regions of abdominal organs. The hybrid FEM technique effectively mitigates the Jacobian uncertainties in these regions.


Asunto(s)
Algoritmos , Procesamiento de Imagen Asistido por Computador , Imagen por Resonancia Magnética , Planificación de la Radioterapia Asistida por Computador , Radioterapia Guiada por Imagen , Humanos , Imagen por Resonancia Magnética/métodos , Procesamiento de Imagen Asistido por Computador/métodos , Radioterapia Guiada por Imagen/métodos , Planificación de la Radioterapia Asistida por Computador/métodos , Dosificación Radioterapéutica , Radioterapia de Intensidad Modulada/métodos , Programas Informáticos , Incertidumbre , Neoplasias Abdominales/radioterapia , Neoplasias Abdominales/diagnóstico por imagen
3.
Sensors (Basel) ; 24(12)2024 Jun 15.
Artículo en Inglés | MEDLINE | ID: mdl-38931669

RESUMEN

In recent years, with the rapid development of deep learning and its outstanding capabilities in target detection, innovative methods have been introduced for infrared dim small target detection. This review comprehensively summarizes public datasets, the latest networks, and evaluation metrics for infrared dim small target detection. This review mainly focuses on deep learning methods from the past three years and categorizes them based on the six key issues in this field: (1) enhancing the representation capability of small targets; (2) improving the accuracy of bounding box regression; (3) resolving the issue of target information loss in the deep network; (4) balancing missed detections and false alarms; (5) adapting for complex backgrounds; (6) lightweight design and deployment issues of the network. Additionally, this review summarizes twelve public datasets for infrared dim small targets and evaluation metrics used for detection and quantitatively compares the performance of the latest networks. Finally, this review provides insights into the future directions of this field. In conclusion, this review aims to assist researchers in gaining a comprehensive understanding of the latest developments in infrared dim small target detection networks.

4.
BMC Bioinformatics ; 24(1): 388, 2023 Oct 12.
Artículo en Inglés | MEDLINE | ID: mdl-37828466

RESUMEN

BACKGROUND: Image segmentation pipelines are commonly used in microscopy to identify cellular compartments like nucleus and cytoplasm, but there are few standards for comparing segmentation accuracy across pipelines. The process of selecting a segmentation assessment pipeline can seem daunting to researchers due to the number and variety of metrics available for evaluating segmentation quality. RESULTS: Here we present automated pipelines to obtain a comprehensive set of 69 metrics to evaluate segmented data and propose a selection methodology for models based on quantitative analysis, dimension reduction or unsupervised classification techniques and informed selection criteria. CONCLUSION: We show that the metrics used here can often be reduced to a small number of metrics that give a more complete understanding of segmentation accuracy, with different groups of metrics providing sensitivity to different types of segmentation error. These tools are delivered as easy to use python libraries, command line tools, Common Workflow Language Tools, and as Web Image Processing Pipeline interactive plugins to ensure a wide range of users can access and use them. We also present how our evaluation methods can be used to observe the changes in segmentations across modern machine learning/deep learning workflows and use cases.


Asunto(s)
Algoritmos , Procesamiento de Imagen Asistido por Computador , Procesamiento de Imagen Asistido por Computador/métodos , Microscopía , Aprendizaje Automático , Citoplasma
5.
Ecotoxicol Environ Saf ; 255: 114806, 2023 Apr 15.
Artículo en Inglés | MEDLINE | ID: mdl-36948010

RESUMEN

Cancer, the second largest human disease, has become a major public health problem. The prediction of chemicals' carcinogenicity before their synthesis is crucial. In this paper, seven machine learning algorithms (i.e., Random Forest (RF), Logistic Regression (LR), Support Vector Machines (SVM), Complement Naive Bayes (CNB), K-Nearest Neighbor (KNN), XGBoost, and Multilayer Perceptron (MLP)) were used to construct the carcinogenicity triple classification prediction (TCP) model (i.e., 1A, 1B, Category 2). A total of 1444 descriptors of 118 hazardous organic chemicals were calculated by Discovery Studio 2020, Sybyl X-2.0 and PaDEL-Descriptor software. The constructed carcinogenicity TCP model was evaluated through five model evaluation indicators (i.e., Accuracy, Precision, Recall, F1 Score and AUC). The model evaluation results show that Accuracy, Precision, Recall, F1 Score and AUC evaluation indicators meet requirements (greater than 0.6). The accuracy of RF, LR, XGBoost, and MLP models for predicting carcinogenicity of Category 2 is 91.67%, 79.17%, 100%, and 100%, respectively. In addition, the constructed machine learning model in this study has potential for error correction. Taking XGBoost model as an example, the predicted carcinogenicity level of 1,2,3-Trichloropropane (96-18-4) is Category 2, but the actual carcinogenicity level is 1B. But the difference between Category 2 and 1B is only 0.004, indicating that the XGBoost is one optimum model of the seven constructed machine learning models. Besides, results showed that functional groups like chlorine and benzene ring might influence the prediction of carcinogenic classification. Therefore, considering functional group characteristics of chemicals before constructing the carcinogenicity prediction model of organic chemicals is recommended. The predicted carcinogenicity of the organic chemicals using the optimum machine leaning model (i.e., XGBoost) was also evaluated and verified by the toxicokinetics. The RF and XGBoost TCP models constructed in this paper can be used for carcinogenicity detection before synthesizing new organic substances. It also provides technical support for the subsequent management of organic chemicals.


Asunto(s)
Carcinógenos , Sustancias Peligrosas , Aprendizaje Automático , Compuestos Orgánicos , Teorema de Bayes , Carcinogénesis , Carcinógenos/toxicidad , Carcinógenos/química , Sustancias Peligrosas/química , Sustancias Peligrosas/toxicidad , Compuestos Orgánicos/toxicidad , Compuestos Orgánicos/química , Máquina de Vectores de Soporte , Organización Mundial de la Salud , Algoritmos , Estados Unidos , Unión Europea , China , Bases de Datos Factuales
6.
Sensors (Basel) ; 23(4)2023 Feb 13.
Artículo en Inglés | MEDLINE | ID: mdl-36850682

RESUMEN

Parkinson's disease (PD) has become widespread these days all over the world. PD affects the nervous system of the human and also affects a lot of human body parts that are connected via nerves. In order to make a classification for people who suffer from PD and who do not suffer from the disease, an advanced model called Bayesian Optimization-Support Vector Machine (BO-SVM) is presented in this paper for making the classification process. Bayesian Optimization (BO) is a hyperparameter tuning technique for optimizing the hyperparameters of machine learning models in order to obtain better accuracy. In this paper, BO is used to optimize the hyperparameters for six machine learning models, namely, Support Vector Machine (SVM), Random Forest (RF), Logistic Regression (LR), Naive Bayes (NB), Ridge Classifier (RC), and Decision Tree (DT). The dataset used in this study consists of 23 features and 195 instances. The class label of the target feature is 1 and 0, where 1 refers to the person suffering from PD and 0 refers to the person who does not suffer from PD. Four evaluation metrics, namely, accuracy, F1-score, recall, and precision were computed to evaluate the performance of the classification models used in this paper. The performance of the six machine learning models was tested on the dataset before and after the process of hyperparameter tuning. The experimental results demonstrated that the SVM model achieved the best results when compared with other machine learning models before and after the process of hyperparameter tuning, with an accuracy of 92.3% obtained using BO.


Asunto(s)
Enfermedad de Parkinson , Humanos , Enfermedad de Parkinson/diagnóstico , Teorema de Bayes , Máquina de Vectores de Soporte , Benchmarking , Aprendizaje Automático
7.
J Digit Imaging ; 36(5): 2088-2099, 2023 10.
Artículo en Inglés | MEDLINE | ID: mdl-37340195

RESUMEN

Segmentation is a crucial step in extracting the medical image features for clinical diagnosis. Though multiple metrics have been proposed to evaluate the segmentation performance, there is no clear study on how or to what extent the segmentation errors will affect the diagnostic related features used in clinical practice. Therefore, we proposed a segmentation robustness plot (SRP) to build the link between segmentation errors and clinical acceptance, where relative area under the curve (R-AUC) was designed to help clinicians to identify the robust diagnostic related image features. In experiments, we first selected representative radiological series from time series (cardiac first-pass perfusion) and spatial series (T2 weighted images on brain tumors) of magnetic resonance images, respectively. Then, dice similarity coefficient (DSC) and Hausdorff distance (HD), as the widely used evaluation metrics, were used to systematically control the degree of the segmentation errors. Finally, the differences between diagnostic related image features extracted from the ground truth and the derived segmentation were analyzed, using the statistical method large sample size T-test to calculate the corresponding p values. The results are denoted in the SRP, where the x-axis indicates the segmentation performance using the aforementioned evaluation metric, and the y-axis shows the severity of the corresponding feature changes, which are expressed in either the p values for a single case or the proportion of patients without significant change. The experimental results in SRP show that when DSC is above 0.95 and HD is below 3 mm, the segmentation errors will not change the features significantly in most cases. However, when segmentation gets worse, additional metrics are required for further analysis. In this way, the proposed SRP indicates the impact of the segmentation errors on the severity of the corresponding feature changes. By using SRP, one could easily define the acceptable segmentation errors in a challenge. Additionally, the R-AUC calculated from SRP provides an objective reference to help the selection of reliable features in image analysis.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Imagen por Resonancia Magnética , Humanos , Imagen por Resonancia Magnética/métodos , Radiografía , Procesamiento de Imagen Asistido por Computador/métodos , Corazón
8.
Curr Hypertens Rep ; 24(11): 523-533, 2022 11.
Artículo en Inglés | MEDLINE | ID: mdl-35731335

RESUMEN

PURPOSE OF REVIEW: To provide an overview of the literature regarding the use of machine learning algorithms to predict hypertension. A systematic review was performed to select recent articles on the subject. RECENT FINDINGS: The screening of the articles was conducted using a machine learning algorithm (ASReview). A total of 21 articles published between January 2018 and May 2021 were identified and compared according to variable selection, train-test split, data balancing, outcome definition, final algorithm, and performance metrics. Overall, the articles achieved an area under the ROC curve (AUROC) between 0.766 and 1.00. The algorithms most frequently identified as having the best performance were support vector machines (SVM), extreme gradient boosting (XGBoost), and random forest. Machine learning algorithms are a promising tool to improve preventive clinical decisions and targeted public health policies for hypertension. However, technical factors such as outcome definition, availability of the final code, predictive performance, explainability, and data leakage need to be consistently and critically evaluated.


Asunto(s)
Hipertensión , Algoritmos , Área Bajo la Curva , Humanos , Hipertensión/diagnóstico , Aprendizaje Automático , Máquina de Vectores de Soporte
9.
Sensors (Basel) ; 22(1)2022 Jan 05.
Artículo en Inglés | MEDLINE | ID: mdl-35009933

RESUMEN

Low-cost sensors (LCS) are becoming popular for air quality monitoring (AQM). They promise high spatial and temporal resolutions at low-cost. In addition, citizen science applications such as personal exposure monitoring can be implemented effortlessly. However, the reliability of the data is questionable due to various error sources involved in the LCS measurement. Furthermore, sensor performance drift over time is another issue. Hence, the adoption of LCS by regulatory agencies is still evolving. Several studies have been conducted to improve the performance of low-cost sensors. This article summarizes the existing studies on the state-of-the-art of LCS for AQM. We conceptualize a step by step procedure to establish a sustainable AQM setup with LCS that can produce reliable data. The selection of sensors, calibration and evaluation, hardware setup, evaluation metrics and inferences, and end user-specific applications are various stages in the LCS-based AQM setup we propose. We present a critical analysis at every step of the AQM setup to obtain reliable data from the low-cost measurement. Finally, we conclude this study with future scope to improve the availability of air quality data.


Asunto(s)
Contaminantes Atmosféricos , Contaminación del Aire , Contaminantes Atmosféricos/análisis , Contaminación del Aire/análisis , Calibración , Monitoreo del Ambiente , Reproducibilidad de los Resultados
10.
Rob Auton Syst ; 148: 103917, 2022 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-34720413

RESUMEN

The coronavirus disease 2019 (COVID-19) outbreak has increased mortality and morbidity world-wide. Oropharyngeal swabbing is a well-known and commonly used sampling technique for COVID-19 diagnose around the world. We developed a robot to assist with COVID-19 oropharyngeal swabbing to prevent frontline clinical staff from being infected. The robot integrates a UR5 manipulator, rigid-flexible coupling (RFC) manipulator, force-sensing and control subsystem, visual subsystem and haptic device. The robot has strength in intrinsically safe and high repeat positioning accuracy. In addition, we also achieve one-dimensional constant force control in the automatic scheme (AS). Compared with the rigid sampling robot, the developed robot can perform the oropharyngeal swabbing procedure more safely and gently, reducing risk. Alternatively, a novel robot control schemes called collaborative manipulation scheme (CMS) which combines a automatic phase and teleoperation phase is proposed. At last, comparative experiments of three schemes were conducted, including CMS, AS, and teleoperation scheme (TS). The experimental results shows that CMS obtained the highest score according to the evaluation equation. CMS has the excellent performance in quality, experience and adaption. Therefore, the proposal of CMS is meaningful which is more suitable for robot-sampling.

11.
Int J Comput Vis ; 129(2): 548-578, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-33642696

RESUMEN

Multi-object tracking (MOT) has been notoriously difficult to evaluate. Previous metrics overemphasize the importance of either detection or association. To address this, we present a novel MOT evaluation metric, higher order tracking accuracy (HOTA), which explicitly balances the effect of performing accurate detection, association and localization into a single unified metric for comparing trackers. HOTA decomposes into a family of sub-metrics which are able to evaluate each of five basic error types separately, which enables clear analysis of tracking performance. We evaluate the effectiveness of HOTA on the MOTChallenge benchmark, and show that it is able to capture important aspects of MOT performance not previously taken into account by established metrics. Furthermore, we show HOTA scores better align with human visual evaluation of tracking performance.

12.
J Community Psychol ; 49(8): 3215-3231, 2021 11.
Artículo en Inglés | MEDLINE | ID: mdl-34551131

RESUMEN

Philanthropic foundations play a major role in supporting community organizing efforts around the country. Even as they provide millions of dollars each year to enable organizing, they have also come under significant criticism for investment approaches that sometimes hinder and even undermine the core goals of organizing itself. This paper focuses on reflections from a funder seeking to support community organizing and the challenges of doing so within the structures of the philanthropic community. In particular the paper examines the way notions of power, theories of change used to justify philanthropic investments, and entrenched assumptions about social change shape the ability of funders to leverage their perch to support organizing. The paper also contextualizes these reflections in a broader debate about the appropriate relationship between funders and grassroots organizing.


Asunto(s)
Obtención de Fondos , Humanos , Cambio Social , Estados Unidos
13.
Cytometry A ; 97(3): 241-252, 2020 03.
Artículo en Inglés | MEDLINE | ID: mdl-32100455

RESUMEN

Single-cell platforms provide statistically large samples of snapshot observations capable of resolving intrercellular heterogeneity. Currently, there is a growing literature on algorithms that exploit this attribute in order to infer the trajectory of biological mechanisms, such as cell proliferation and differentiation. Despite the efforts, the trajectory inference methodology has not yet been used for addressing the challenging problem of learning the dynamics of protein signaling systems. In this work, we assess this prospect by testing the performance of this class of algorithms on four proteomic temporal datasets. To evaluate the learning quality, we design new general-purpose evaluation metrics that are able to quantify performance on (i) the biological meaning of the output, (ii) the consistency of the inferred trajectory, (iii) the algorithm robustness, (iv) the correlation of the learning output with the initial dataset, and (v) the roughness of the cell parameter levels though the inferred trajectory. We show that experimental time alone is insufficient to provide knowledge about the order of proteins during signal transduction. Accordingly, we show that the inferred trajectories provide richer information about the underlying dynamics. We learn that established methods tested on high-dimensional data with small sample size, slow dynamics, and complex structures (e.g. bifurcations) cannot always work in the signaling setting. Among the methods we evaluate, Scorpius and a newly introduced approach that combines Diffusion Maps and Principal Curves were found to perform adequately in recovering the progression of signal transduction although their performance on some metrics varies from one dataset to another. The novel metrics we devise highlight that it is difficult to conclude, which one method is universally applicable for the task. Arguably, there are still many challenges and open problems to resolve. © 2020 The Authors. Cytometry Part A published by Wiley Periodicals, Inc. on behalf of International Society for Advancement of Cytometry.


Asunto(s)
Algoritmos , Proteómica , Humanos
14.
J Biomech Eng ; 142(11)2020 11 01.
Artículo en Inglés | MEDLINE | ID: mdl-32529240

RESUMEN

Faculty performance evaluation is an important element of assessment for departments and universities. A quantitative score is often needed for faculty annual evaluation, but its determination is often subjective, and it is hard to incorporate the versatile contributions of individual faculty members. Here, we propose a quantitative and objective faculty performance evaluation method. We established a well-structured quantitative evaluation system which scores faculty performance in key activities using expectation-based formula on key measures and then incorporates personalized flexible weights to integrate them into three area scores in teaching, research, and service as well as an overall score. It was implemented in a programed excel form, making it convenient to both faculty and evaluators and has generated very positive outcomes such as higher faculty satisfactory and improved productivity as indicated by associated increases in publications and new research grants etc. In conclusion, the quantitative faculty evaluation system provides more objective and transparent annual evaluation and a basis for making merit raise and award decisions. In addition, it can be readily adapted to evolving goals and needs of a department as well as different needs and cultures of different departments.


Asunto(s)
Motivación , Eficiencia , Docentes , Universidades
15.
Adv Exp Med Biol ; 1137: 45-60, 2019.
Artículo en Inglés | MEDLINE | ID: mdl-31183819

RESUMEN

In the previous chapter we were able to automatically process structured data to retrieve biomedical text about any chemical compound, such as caffeine. This chapter will provide a step-by-step introduction to how we can process that text using shell script commands, specifically extract information about diseases related to caffeine. The goal is to equip the reader with an essential set of skills to extract meaningful information from any text.


Asunto(s)
Minería de Datos/métodos , Procesamiento Automatizado de Datos , Cafeína , Programas Informáticos
16.
J Med Syst ; 43(8): 281, 2019 Jul 12.
Artículo en Inglés | MEDLINE | ID: mdl-31300894

RESUMEN

Online appointment scheduling systems have been designed in response to the problems of the traditional ones. In Iran, most outpatient clinics and our study population suffer from high patient' no-show rate and long waiting times because of not using online appointment scheduling system. In this study, the effect of an online appointment scheduling system was investigated by comparing the evaluation metrics of appointment scheduling before and after the intervention. This before-after pilot study was conducted on ten outpatient clinics with different specializations. Five clinics were selected as the intervention group and five clinics as the control group. A checklist was designed to evaluate appointment scheduling metrics. These checklists were completed from April to July 2017 in the pre-implementation phase and from September to December 2017 in the post-implementation phase. The evaluation metrics were evaluated before and after the implementation of the online appointment scheduling system. A total of 470 checklists containing 9034 patients' information and 460 checklists containing 9130 patients' information were completed at the pre- and post-implementation phases, respectively. There were significant positive effect on the improvement of the three metrics means, including Patient waiting time (38.2 min before vs. 23.8 min after the intervention), No-show rate (25% before vs. 11% after the intervention), and Physician punctuality (-30 min before vs. -14.2 min after the intervention). The use of an online appointment scheduling system was successful in improving several evaluation metrics in our target population and resulted in continued usage in intervention group clinics.


Asunto(s)
Instituciones de Atención Ambulatoria , Citas y Horarios , Internet , Lista de Verificación , Estudios de Evaluación como Asunto , Factores de Tiempo
17.
Neuroimage ; 152: 312-329, 2017 05 15.
Artículo en Inglés | MEDLINE | ID: mdl-28286318

RESUMEN

An important image processing step in spinal cord magnetic resonance imaging is the ability to reliably and accurately segment grey and white matter for tissue specific analysis. There are several semi- or fully-automated segmentation methods for cervical cord cross-sectional area measurement with an excellent performance close or equal to the manual segmentation. However, grey matter segmentation is still challenging due to small cross-sectional size and shape, and active research is being conducted by several groups around the world in this field. Therefore a grey matter spinal cord segmentation challenge was organised to test different capabilities of various methods using the same multi-centre and multi-vendor dataset acquired with distinct 3D gradient-echo sequences. This challenge aimed to characterize the state-of-the-art in the field as well as identifying new opportunities for future improvements. Six different spinal cord grey matter segmentation methods developed independently by various research groups across the world and their performance were compared to manual segmentation outcomes, the present gold-standard. All algorithms provided good overall results for detecting the grey matter butterfly, albeit with variable performance in certain quality-of-segmentation metrics. The data have been made publicly available and the challenge web site remains open to new submissions. No modifications were introduced to any of the presented methods as a result of this challenge for the purposes of this publication.


Asunto(s)
Mapeo Encefálico/métodos , Médula Cervical/anatomía & histología , Sustancia Gris/anatomía & histología , Procesamiento de Imagen Asistido por Computador/métodos , Adulto , Algoritmos , Femenino , Humanos , Imagen por Resonancia Magnética , Masculino , Persona de Mediana Edad , Reproducibilidad de los Resultados , Sustancia Blanca/anatomía & histología
18.
Sensors (Basel) ; 17(10)2017 Oct 13.
Artículo en Inglés | MEDLINE | ID: mdl-29027948

RESUMEN

In recent years, indoor localization systems have been the object of significant research activity and of growing interest for their great expected social impact and their impressive business potential. Application areas include tracking and navigation, activity monitoring, personalized advertising, Active and Assisted Living (AAL), traceability, Internet of Things (IoT) networks, and Home-land Security. In spite of the numerous research advances and the great industrial interest, no canned solutions have yet been defined. The diversity and heterogeneity of applications, scenarios, sensor and user requirements, make it difficult to create uniform solutions. From that diverse reality, a main problem is derived that consists in the lack of a consensus both in terms of the metrics and the procedures used to measure the performance of the different indoor localization and navigation proposals. This paper introduces the general lines of the EvAAL benchmarking framework, which is aimed at a fair comparison of indoor positioning systems through a challenging competition under complex, realistic conditions. To evaluate the framework capabilities, we show how it was used in the 2016 Indoor Positioning and Indoor Navigation (IPIN) Competition. The 2016 IPIN competition considered three different scenario dimensions, with a variety of use cases: (1) pedestrian versus robotic navigation, (2) smartphones versus custom hardware usage and (3) real-time positioning versus off-line post-processing. A total of four competition tracks were evaluated under the same EvAAL benchmark framework in order to validate its potential to become a standard for evaluating indoor localization solutions. The experience gained during the competition and feedback from track organizers and competitors showed that the EvAAL framework is flexible enough to successfully fit the very different tracks and appears adequate to compare indoor positioning systems.

19.
Sensors (Basel) ; 16(6)2016 Jun 10.
Artículo en Inglés | MEDLINE | ID: mdl-27294938

RESUMEN

This paper evaluates different wavelet-based cross-spectral image fusion strategies adopted to merge visible and infrared images. The objective is to find the best setup independently of the evaluation metric used to measure the performance. Quantitative performance results are obtained with state of the art approaches together with adaptations proposed in the current work. The options evaluated in the current work result from the combination of different setups in the wavelet image decomposition stage together with different fusion strategies for the final merging stage that generates the resulting representation. Most of the approaches evaluate results according to the application for which they are intended for. Sometimes a human observer is selected to judge the quality of the obtained results. In the current work, quantitative values are considered in order to find correlations between setups and performance of obtained results; these correlations can be used to define a criteria for selecting the best fusion strategy for a given pair of cross-spectral images. The whole procedure is evaluated with a large set of correctly registered visible and infrared image pairs, including both Near InfraRed (NIR) and Long Wave InfraRed (LWIR).

20.
Neural Netw ; 175: 106278, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-38581809

RESUMEN

In the field of deep learning, large quantities of data are typically required to effectively train models. This challenge has given rise to techniques like zero-shot learning (ZSL), which trains models on a set of "seen" classes and evaluates them on a set of "unseen" classes. Although ZSL has shown considerable potential, particularly with the employment of generative methods, its generalizability to real-world scenarios remains uncertain. The hypothesis of this work is that the performance of ZSL models is systematically influenced by the chosen "splits"; in particular, the statistical properties of the classes and attributes used in training. In this paper, we test this hypothesis by introducing the concepts of generalizability and robustness in attribute-based ZSL and carry out a variety of experiments to stress-test ZSL models against different splits. Our aim is to lay the groundwork for future research on ZSL models' generalizability, robustness, and practical applications. We evaluate the accuracy of state-of-the-art models on benchmark datasets and identify consistent trends in generalizability and robustness. We analyze how these properties vary based on the dataset type, differentiating between coarse- and fine-grained datasets, and our findings indicate significant room for improvement in both generalizability and robustness. Furthermore, our results demonstrate the effectiveness of dimensionality reduction techniques in improving the performance of state-of-the-art models in fine-grained datasets.


Asunto(s)
Aprendizaje Profundo , Redes Neurales de la Computación , Humanos , Algoritmos , Aprendizaje Automático
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA