Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 2.053
Filtrar
Más filtros

Tipo del documento
Intervalo de año de publicación
1.
Data Brief ; 56: 110775, 2024 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-39221011

RESUMEN

Bangladesh's agricultural landscape is significantly influenced by vegetable cultivation, which substantially enhances nutrition, the economy, and food security in the nation. Millions of people rely on vegetable production for their daily sustenance, generating considerable income for numerous farmers. However, leaf diseases frequently compromise the yield and quality of vegetable crops. Plant diseases are a common impediment to global agricultural productivity, adversely affecting crop quality and yield, leading to substantial economic losses for farmers. Early detection of plant leaf diseases is crucial for improving cultivation and vegetable production. Common diseases such as Bacterial Spot, Mosaic Virus, and Downy Mildew often reduce vegetable plant cultivation and severely impact vegetable production and the food economy. Consequently, many farmers in Bangladesh struggle to identify the specific diseases, incurring significant losses. This dataset contains 12,643 images of widely grown crops in Bangladesh, facilitating the identification of unhealthy leaves compared to healthy ones. The dataset includes images of vegetable leaves such as Bitter Gourd (2223 images), Bottle Gourd (1803 images), Eggplants (2944 images), Cauliflowers (1598 images), Cucumbers (1626 images), and Tomatoes (2449 images). Each vegetable class encompasses several common diseases that affect cultivation. By identifying early leaf diseases, this dataset will be invaluable for farmers and agricultural researchers alike.

2.
Front Pharmacol ; 15: 1424803, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39221152

RESUMEN

Background and aim: Pathological changes in the central nervous system (CNS) begin before the clinical symptoms of Alzheimer's Disease (AD) manifest, with the hippocampus being one of the first affected structures. Current treatments fail to alter AD progression. Traditional Chinese medicine (TCM) has shown potential in improving AD pathology through multi-target mechanisms. This study investigates pathological changes in AD hippocampal tissue and explores TCM active components that may alleviate these changes. Methods: GSE5281 and GSE173955 datasets were downloaded from GEO and normalized to identify differentially expressed genes (DEGs). Key functional modules and hub genes were analyzed using Cytoscape and R. Active TCM components were identified from literature and the Pharmacopoeia of the People's Republic of China. Enrichment analyses were performed on target genes overlapping with DEGs. Result: From the datasets, 76 upregulated and 363 downregulated genes were identified. Hub genes included SLAMF, CD34, ELN (upregulated) and ATP5F1B, VDAC1, VDAC2, HSPA8, ATP5F1C, PDHA1, UBB, SNCA, YWHAZ, PGK1 (downregulated). Literature review identified 33 active components from 23 herbal medicines. Target gene enrichment and analysis were performed for six components: dihydroartemisinin, berberine, naringenin, calycosin, echinacoside, and icariside II. Conclusion: Mitochondrial to synaptic vesicle dysfunction pathways were enriched in downregulated genes. Despite downregulation, UBB and SNCA proteins accumulate in AD brains. TCM studies suggest curcumin and echinacoside may improve hippocampal pathology and cognitive impairment in AD. Further investigation into their mechanisms is needed.

3.
Sci Rep ; 14(1): 20410, 2024 Sep 02.
Artículo en Inglés | MEDLINE | ID: mdl-39223219

RESUMEN

Accurate population data is crucial for assessing exposure in disaster risk assessments. In recent years, there has been a significant increase in the development of spatially gridded population datasets. Despite these datasets often using similar input data to derive population figures, notable differences arise when comparing them with direct ground-level observations. This study evaluates the precision and accuracy of flood exposure assessments using both known and generated gridded population datasets in Sweden. Specifically focusing on WorldPop and GHSPop, we compare these datasets against official national statistics at a 100 m grid cell resolution to assess their reliability in flood exposure analyses. Our objectives include quantifying the reliability of these datasets and examining the impact of data aggregation on estimated flood exposure across different administrative levels. The analysis reveals significant discrepancies in flood exposure estimates, underscoring the challenges associated with relying on generated gridded population data for precise flood risk assessments. Our findings emphasize the importance of careful dataset selection and highlight the potential for overestimation in flood risk analysis. This emphasises the critical need for validations against ground population data to ensure accurate flood risk management strategies.


Asunto(s)
Inundaciones , Suecia , Humanos , Medición de Riesgo , Desastres , Reproducibilidad de los Resultados
4.
Sensors (Basel) ; 24(16)2024 Aug 06.
Artículo en Inglés | MEDLINE | ID: mdl-39204791

RESUMEN

The rapid development of Internet of Things (IoT) technologies and the potential benefits of employing the vast datasets generated by IoT devices, including wearable sensors and camera systems, has ushered in a new era of opportunities for enhancing smart rehabilitation in various healthcare systems. Maintaining patient privacy is paramount in healthcare while providing smart insights and recommendations. This study proposed the adoption of federated learning to develop a scalable AI model for post-stroke assessment while protecting patients' privacy. This research compares the centralized (PSA-MNMF) model performance with the proposed scalable federated PSA-FL-CDM model for sensor- and camera-based datasets. The computational time indicates that the federated PSA-FL-CDM model significantly reduces the execution time and attains comparable performance while preserving the patient's privacy. Impact Statement-This research introduces groundbreaking contributions to stroke assessment by successfully implementing federated learning for the first time in this domain and applying consensus models in each node. It enables collaborative model training among multiple nodes or clients while ensuring the privacy of raw data. The study explores eight different clustering methods independently on each node, revolutionizing data organization based on similarities in stroke assessment. Additionally, the research applies the centralized PSA-MNMF consensus clustering technique to each client, resulting in more accurate and robust clustering solutions. By utilizing the FedAvg federated learning algorithm strategy, locally trained models are combined to create a global model that captures the collective knowledge of all participants. Comparative performance measurements and computational time analyses are conducted, facilitating a fair evaluation between centralized and federated learning models in stroke assessment. Moreover, the research extends beyond a single type of database by conducting experiments on two distinct datasets, wearable and camera-based, broadening the understanding of the proposed methods across different data modalities. These contributions develop stroke assessment methodologies, enabling efficient collaboration and accurate consensus clustering models and maintaining data privacy.


Asunto(s)
Accidente Cerebrovascular , Humanos , Algoritmos , Internet de las Cosas , Dispositivos Electrónicos Vestibles , Consenso , Análisis por Conglomerados , Aprendizaje Automático
5.
Sensors (Basel) ; 24(16)2024 Aug 11.
Artículo en Inglés | MEDLINE | ID: mdl-39204895

RESUMEN

Poplar (Populus) trees play a vital role in various industries and in environmental sustainability. They are widely used for paper production, timber, and as windbreaks, in addition to their significant contributions to carbon sequestration. Given their economic and ecological importance, effective disease management is essential. Convolutional Neural Networks (CNNs), particularly adept at processing visual information, are crucial for the accurate detection and classification of plant diseases. This study introduces a novel dataset of manually collected images of diseased poplar leaves from Uzbekistan and South Korea, enhancing the geographic diversity and application of the dataset. The disease classes consist of "Parsha (Scab)", "Brown-spotting", "White-Gray spotting", and "Rust", reflecting common afflictions in these regions. This dataset will be made publicly available to support ongoing research efforts. Employing the advanced YOLOv8 model, a state-of-the-art CNN architecture, we applied a Contrast Stretching technique prior to model training in order to enhance disease detection accuracy. This approach not only improves the model's diagnostic capabilities but also offers a scalable tool for monitoring and treating poplar diseases, thereby supporting the health and sustainability of these critical resources. This dataset, to our knowledge, will be the first of its kind to be publicly available, offering a valuable resource for researchers and practitioners worldwide.


Asunto(s)
Redes Neurales de la Computación , Enfermedades de las Plantas , Hojas de la Planta , Populus , Procesamiento de Imagen Asistido por Computador/métodos , República de Corea
6.
Sensors (Basel) ; 24(16)2024 Aug 11.
Artículo en Inglés | MEDLINE | ID: mdl-39204903

RESUMEN

Brain-computer interfaces (BCIs) are pivotal in translating neural activities into control commands for external assistive devices. Non-invasive techniques like electroencephalography (EEG) offer a balance of sensitivity and spatial-temporal resolution for capturing brain signals associated with motor activities. This work introduces MOVING, a Multi-Modal dataset of EEG signals and Virtual Glove Hand Tracking. This dataset comprises neural EEG signals and kinematic data associated with three hand movements-open/close, finger tapping, and wrist rotation-along with a rest period. The dataset, obtained from 11 subjects using a 32-channel dry wireless EEG system, also includes synchronized kinematic data captured by a Virtual Glove (VG) system equipped with two orthogonal Leap Motion Controllers. The use of these two devices allows for fast assembly (∼1 min), although introducing more noise than the gold standard devices for data acquisition. The study investigates which frequency bands in EEG signals are the most informative for motor task classification and the impact of baseline reduction on gesture recognition. Deep learning techniques, particularly EEGnetV4, are applied to analyze and classify movements based on the EEG data. This dataset aims to facilitate advances in BCI research and in the development of assistive devices for people with impaired hand mobility. This study contributes to the repository of EEG datasets, which is continuously increasing with data from other subjects, which is hoped to serve as benchmarks for new BCI approaches and applications.


Asunto(s)
Interfaces Cerebro-Computador , Electroencefalografía , Mano , Movimiento , Humanos , Electroencefalografía/métodos , Mano/fisiología , Movimiento/fisiología , Masculino , Adulto , Fenómenos Biomecánicos/fisiología , Femenino , Procesamiento de Señales Asistido por Computador
7.
Sensors (Basel) ; 24(16)2024 Aug 13.
Artículo en Inglés | MEDLINE | ID: mdl-39204933

RESUMEN

This paper presents the methodology and outcomes of creating the Rail Vista dataset, designed for detecting defects on railway tracks using machine and deep learning techniques. The dataset comprises 200,000 high-resolution images categorized into 19 distinct classes covering various railway infrastructure defects. The data collection involved a meticulous process including complex image capture methods, distortion techniques for data enrichment, and secure storage in a data warehouse using efficient binary file formats. This structured dataset facilitates effective training of machine/deep learning models, enhancing automated defect detection systems in railway safety and maintenance applications. The study underscores the critical role of high-quality datasets in advancing machine learning applications within the railway domain, highlighting future prospects for improving safety and reliability through automated recognition technologies.

8.
Br J Hosp Med (Lond) ; 85(8): 1-12, 2024 Aug 30.
Artículo en Inglés | MEDLINE | ID: mdl-39212573

RESUMEN

Adoption of electronic health record systems offers an opportunity to collate massive volumes of complex information about patient care. Healthcare data can inform performance management, enable predictive analytics and enhance strategic decision making. A data-driven approach to improving patient care is vital to address the growing burden of morbidity and mortality associated with major surgery. We describe our methodology for transforming and utilising process of care data in an electronic health record system to develop a registry for quality improvement purposes in patients undergoing major surgery at a single UK hospital. We highlight development of our data-driven vision, technical aspects of processing raw data into metrics relevant to clinical decision making, alongside challenges encountered. Finally, we outline how our data infrastructure supports clinical governance, quality improvement and research. In sharing our experiences, we hope to enable others to embed and access the transformative clinical insights that healthcare data can yield.


Asunto(s)
Registros Electrónicos de Salud , Mejoramiento de la Calidad , Centros de Atención Terciaria , Humanos , Centros de Atención Terciaria/organización & administración , Londres , Medicina Perioperatoria/métodos , Sistema de Registros
9.
Int J Med Inform ; 191: 105604, 2024 Aug 15.
Artículo en Inglés | MEDLINE | ID: mdl-39154600

RESUMEN

INTRODUCTION: Inherent variations between inter-center data can undermine the robustness of segmentation models when applied at a specific center (dataset shift). We investigated whether specialized center-specific models are more effective compared to generalist models based on multi-center data, and how center-specific data could enhance the performance of generalist models within a particular center using a fine-tuning transfer learning approach. For this purpose, we studied the dataset shift at center level and conducted a comparative analysis to assess the impact of data source on glioblastoma segmentation models. METHODS & MATERIALS: The three key components of dataset shift were studied: prior probability shift-variations in tumor size or tissue distribution among centers; covariate shift-inter-center MRI alterations; and concept shift-different criteria for tumor segmentation. BraTS 2021 dataset was used, which includes 1251 cases from 23 centers. Thereafter, 155 deep-learning models were developed and compared, including 1) generalist models trained with multi-center data, 2) specialized models using only center-specific data, and 3) fine-tuned generalist models using center-specific data. RESULTS: The three key components of dataset shift were characterized. The amount of covariate shift was substantial, indicating large variations in MR imaging between different centers. Glioblastoma segmentation models tend to perform best when using data from the application center. Generalist models, trained with over 700 samples, achieved a median Dice score of 88.98%. Specialized models surpassed this with 200 cases, while fine-tuned models outperformed with 50 cases. CONCLUSIONS: The influence of dataset shift on model performance is evident. Fine-tuned and specialized models, utilizing data from the evaluated center, outperform generalist models, which rely on data from other centers. These approaches could encourage medical centers to develop customized models for their local use, enhancing the accuracy and reliability of glioblastoma segmentation in a context where dataset shift is inevitable.

10.
Sci Rep ; 14(1): 18702, 2024 08 12.
Artículo en Inglés | MEDLINE | ID: mdl-39134549

RESUMEN

A new video based multi behavior dataset for cows, CBVD-5, is introduced in this paper. The dataset includes five cow behaviors: standing, lying down, foraging,rumination and drinking. The dataset comprises 107 cows from the entire barn, maintaining an 80% stocking density. Monitoring occurred over 96 h for these 20-month-old cows, considering varying light conditions and nighttime data to ensure standardization and inclusivity.The dataset consists of ranch monitoring footage collected by seven cameras, including 687 video segment samples and 206,100 image samples, covering five daily behaviors of cows. The data collection process entailed the deployment of cameras, hard drives, software, and servers for storage. Data annotation was conducted using the VIA web tool, leveraging the video expertise of pertinent professionals. The annotation coordinates and category labels of each individual cow in the image, as well as the generated configuration file, are also saved in the dataset. With this dataset,we propose a slowfast cow multi behavior recognition model based on video sequences as the baseline evaluation model. The experimental results show that the model can effectively learn corresponding category labels from the behavior type data of the dataset, with an error rate of 21.28% on the test set. In addition to cow behavior recognition, the dataset can also be used for cow target detection, and so on.The CBVD-5 dataset significantly influences dairy cow behavior recognition, advancing research, enriching data resources, standardizing datasets, enhancing dairy cow health and welfare monitoring, and fostering agricultural intelligence development. Additionally, it serves educational and training needs, supporting research and practical applications in related fields. The dataset will be made freely available to researchers world-wide.


Asunto(s)
Conducta Animal , Grabación en Video , Bovinos , Animales , Conducta Animal/fisiología , Femenino
11.
Front Med (Lausanne) ; 11: 1393123, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39139784

RESUMEN

Introduction: Transparency and traceability are essential for establishing trustworthy artificial intelligence (AI). The lack of transparency in the data preparation process is a significant obstacle in developing reliable AI systems which can lead to issues related to reproducibility, debugging AI models, bias and fairness, and compliance and regulation. We introduce a formal data preparation pipeline specification to improve upon the manual and error-prone data extraction processes used in AI and data analytics applications, with a focus on traceability. Methods: We propose a declarative language to define the extraction of AI-ready datasets from health data adhering to a common data model, particularly those conforming to HL7 Fast Healthcare Interoperability Resources (FHIR). We utilize the FHIR profiling to develop a common data model tailored to an AI use case to enable the explicit declaration of the needed information such as phenotype and AI feature definitions. In our pipeline model, we convert complex, high-dimensional electronic health records data represented with irregular time series sampling to a flat structure by defining a target population, feature groups and final datasets. Our design considers the requirements of various AI use cases from different projects which lead to implementation of many feature types exhibiting intricate temporal relations. Results: We implement a scalable and high-performant feature repository to execute the data preparation pipeline definitions. This software not only ensures reliable, fault-tolerant distributed processing to produce AI-ready datasets and their metadata including many statistics alongside, but also serve as a pluggable component of a decision support application based on a trained AI model during online prediction to automatically prepare feature values of individual entities. We deployed and tested the proposed methodology and the implementation in three different research projects. We present the developed FHIR profiles as a common data model, feature group definitions and feature definitions within a data preparation pipeline while training an AI model for "predicting complications after cardiac surgeries". Discussion: Through the implementation across various pilot use cases, it has been demonstrated that our framework possesses the necessary breadth and flexibility to define a diverse array of features, each tailored to specific temporal and contextual criteria.

12.
Plant Methods ; 20(1): 124, 2024 Aug 13.
Artículo en Inglés | MEDLINE | ID: mdl-39138512

RESUMEN

BACKGROUND: Chinese Cymbidium orchids, cherished for their deep-rooted cultural significance and significant economic value in China, have spawned a rich tapestry of cultivars. However, these orchid cultivars are facing challenges from insufficient cultivation practices and antiquated techniques, including cultivar misclassification, complex identification, and the proliferation of counterfeit products. Current commercial techniques and academic research primarily emphasize species identification of orchids, rather than delving into that of orchid cultivars within species. RESULTS: To bridge this gap, the authors dedicated over a year to collecting a cultivar image dataset for Chinese Cymbidium orchids named Orchid2024. This dataset contains over 150,000 images spanning 1,275 different categories, involving visits to 20 cities across 12 provincial administrative regions in China to gather pertinent data. Subsequently, we introduced various visual parameter-efficient fine-tuning (PEFT) methods to expedite model development, achieving the highest top-1 accuracy of 86.14% and top-5 accuracy of 95.44%. CONCLUSION: Experimental results demonstrate the complexity of the dataset while highlighting the considerable promise of PEFT methods within flower image classification. We believe that our work not only provides a practical tool for orchid researchers, growers and market participants, but also provides a unique and valuable resource for further exploring fine-grained image classification tasks. The dataset and code are available at https://github.com/pengyingshu/Orchid2024 .

13.
Heliyon ; 10(14): e34516, 2024 Jul 30.
Artículo en Inglés | MEDLINE | ID: mdl-39148969

RESUMEN

Objective: Ulcerative Colitis (UC) manifests as a chronic inflammatory condition of the intestines, marked by ongoing immune system dysregulation. Disulfidptosis, a newly identified cell death mechanism, is intimately linked to the onset and advancement of inflammation. However, the role of disulfidptosis in UC remains unclear. Methods: We screened differentially expressed genes (DEGs) associated with disulfidptosis in multiple UC datasets, narrowed down the target gene number using lasso regression, and conducted immune infiltration analysis and constructed a clinical diagnostic model. Additionally, we explored the association between disulfidptosis-related key genes and disease remission in UC patients receiving biologic therapy. Finally, we confirmed the expression of key genes in FHC cells and UC tissue samples. Results: In the differential analysis, we identified 20 DEGs associated with disulfidptosis. Immune infiltration results revealed that five genes (PDLIM1, SLC7A11, MYH10, NUBPL, OXSM) exhibited strong correlations with immune cells and pathways. Using GO, KEGG and WGCNA analyses, we discovered that gene modules highly correlated with disulfidptosis-related gene expression were significantly enriched in inflammation-related pathways. Additionally, we developed a nomogram based on these five immune-related disulfidptosis genes for UC diagnosis, showing robust diagnostic capability and clinical efficacy. Kaplan-Meier survival analysis revealed a significant link between changes in the expression levels of these cell genes and disease remission in UC patients receiving biologic therapy. In line with previous studies, similar expression changes of the target gene were seen in both UC cell models and tissue samples. Conclusions: This study utilized bioinformatic analysis and machine learning to identify and analyze features associated with disulfidptosis in multiple UC datasets. This enhances our comprehension of the role disulfidptosis plays in intestinal immunity and inflammation in UC, providing new perspectives for developing innovative treatments for UC.

14.
J Environ Manage ; 368: 122009, 2024 Aug 15.
Artículo en Inglés | MEDLINE | ID: mdl-39151335

RESUMEN

The analysis of risk awareness should be the initial stage in integrated natural hazard risk management to promote appropriate and effective measures for mitigating risks and strengthening social resilience inside the multi-risk framework. Nevertheless, earlier studies focused on cross-sectional data and overlooked the changes in risk awareness levels and associated independent variables with time. This study analyzes for the first time a balanced nationwide panel dataset of 1612 respondent-year observations from Switzerland (period 2015-2021, including the epidemic of COVID-19) to examine and compare the effects of potential independent variables on the four dimensions of natural hazard risk awareness (NHRA), ranging from the broadest dimension of Relevance to higher dimensions of Perceived Probability of an event, Perceived Threat to life and valuables, and Perceived Situational Threat. The analysis in this study incorporates multiple methods of Random-Effect Model (RE), Generalized Linear Model (GLM), and mediation analysis. Results show that NHRA increased in Switzerland to different extents (up to 23.24%) depending on the dimension. Event memory, perceived information impact and reported individual informed level appeared to be the most consistent independent variables positively influencing panel NHRA. Among these, perceived information impact as an important indicator of risk communication, was also found to serve as a mediator from risk preparedness to risk awareness. By encouraging residents to engage in "Begin Doing Before Thinking" (BDBT) programs to leverage subliminal effects and self-reflection, this study proposes that behavior-cognition feedback loops may facilitate a virtuous cycle. Our promising observations provide recommendations for an effective awareness-rising strategy design and suggest extensive insights from potential short-interval panel analysis in the future.

15.
Environ Sci Ecotechnol ; 22: 100450, 2024 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-39161573

RESUMEN

Rainfall data with high spatial and temporal resolutions are essential for urban hydrological modeling. Ubiquitous surveillance cameras can continuously record rainfall events through video and audio, so they have been recognized as potential rain gauges to supplement professional rainfall observation networks. Since video-based rainfall estimation methods can be affected by variable backgrounds and lighting conditions, audio-based approaches could be a supplement without suffering from these conditions. However, most audio-based approaches focus on rainfall-level classification rather than rainfall intensity estimation. Here, we introduce a dataset named Surveillance Audio Rainfall Intensity Dataset (SARID) and a deep learning model for estimating rainfall intensity. First, we created the dataset through audio of six real-world rainfall events. This dataset's audio recordings are segmented into 12,066 pieces and annotated with rainfall intensity and environmental information, such as underlying surfaces, temperature, humidity, and wind. Then, we developed a deep learning-based baseline using Mel-Frequency Cepstral Coefficients (MFCC) and Transformer architecture to estimate rainfall intensity from surveillance audio. Validated from ground truth data, our baseline achieves a root mean absolute error of 0.88 mm h-1 and a coefficient of correlation of 0.765. Our findings demonstrate the potential of surveillance audio-based models as practical and effective tools for rainfall observation systems, initiating a new chapter in rainfall intensity estimation. It offers a novel data source for high-resolution hydrological sensing and contributes to the broader landscape of urban sensing, emergency response, and resilience.

16.
Accid Anal Prev ; 207: 107748, 2024 Aug 18.
Artículo en Inglés | MEDLINE | ID: mdl-39159592

RESUMEN

Driving risk prediction emerges as a pivotal technology within the driving safety domain, facilitating the formulation of targeted driving intervention strategies to enhance driving safety. The driving safety undergoes continuous evolution in response to the complexities of the traffic environment, representing a dynamic and ongoing serialization process. The evolutionary trend of this sequence offers valuable information pertinent to driving safety research. However, existing research on driving risk prediction has primarily concentrated on forecasting a single index, such as the driving safety level or the extreme value within a specified future timeframe. This approach often neglects the intrinsic properties that characterize the temporal evolution of driving safety. Leveraging the high-D natural driving dataset, this study employs the multi-step time series forecasting methodology to predict the risk evolution sequence throughout the car-following process, elucidates the benefits of the multi-step time series forecasting approach, and contrasts the predictive efficacy on driving safety levels across various temporal windows. The empirical findings demonstrate that the time series prediction model proficiently captures essential dynamics such as risk evolution trends, amplitudes, and turning points. Consequently, it provides predictions that are significantly more robust and comprehensive than those obtained from a single risk index. The TsLeNet proposed in this study integrates a 2D convolutional network architecture with a dual attention mechanism, adeptly capturing and synthesizing multiple features across time steps. This integration significantly enhances the prediction precision at each temporal interval. Comparative analyses with other mainstream models reveal that TsLeNet achieves the best performance in terms of prediction accuracy and efficiency. Concurrently, this research undertakes a comprehensive analysis of the temporal distribution of errors, the impact pattern of features on risk sequence, and the applicability of interaction features among surrounding vehicles. The adoption of multi-step time series forecasting approach not only offers a novel perspective for analyzing and exploring driving safety, but also furnishes the design and development of targeted driving intervention systems.

17.
J Neural Eng ; 21(4)2024 Aug 20.
Artículo en Inglés | MEDLINE | ID: mdl-39094617

RESUMEN

Objective.This study aims to address the challenges associated with data-driven electroencephalography (EEG) data analysis by introducing a standardised library calledBIDSAlign. This library efficiently processes and merges heterogeneous EEG datasets from different sources into a common standard template. The goal of this work is to create an environment that allows to preprocess public datasets in order to provide data for the effective training of deep learning (DL) architectures.Approach.The library can handle both Brain Imaging Data Structure (BIDS) and non-BIDS datasets, allowing the user to easily preprocess multiple public datasets. It unifies the EEG recordings acquired with different settings by defining a common pipeline and a specified channel template. An array of visualisation functions is provided inside the library, together with a user-friendly graphical user interface to assist non-expert users throughout the workflow.Main results.BIDSAlign enables the effective use of public EEG datasets, providing valuable medical insights, even for non-experts in the field. Results from applying the library to datasets from OpenNeuro demonstrate its ability to extract significant medical knowledge through an end-to-end workflow, facilitating group analysis, visual comparison and statistical testing.Significance.BIDSAlign solves the lack of large EEG datasets by aligning multiple datasets to a standard template. This unlocks the potential of public EEG data for training DL models. It paves the way to promising contributions based on DL to clinical and non-clinical EEG research, offering insights that can inform neurological disease diagnosis and treatment strategies.


Asunto(s)
Electroencefalografía , Electroencefalografía/métodos , Humanos , Bases de Datos Factuales , Aprendizaje Profundo , Procesamiento de Señales Asistido por Computador
18.
Data Brief ; 55: 110723, 2024 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-39156666

RESUMEN

The underwater environment is characterized by complex light traversal, encompassing effects such as color loss, contrast loss, water distortion, backscatter, light attenuation, and color cast, which vary depending on water purity, depth, and other factors. The dataset presented in this paper is prepared with 100 ground-truth images and 1,50,000 synthetic underwater images. This dataset approximates the effects of underwater environment with implementable combinations of color cast, blurring, low-light, and contrast reduction. These effects and their combinations, with different severity levels are applied to each ground-truth image to generate as many as 150 synthetic underwater images. In addition to the dataset of 1,50,100 images, a comprehensive set of 21 focus metrics, including the average contrast measure operator, Brenner's gradient-based metric, and many others, are calculated for each image.

19.
Brain Inform ; 11(1): 21, 2024 Aug 21.
Artículo en Inglés | MEDLINE | ID: mdl-39167115

RESUMEN

Epileptic seizure (ES) detection is an active research area, that aims at patient-specific ES detection with high accuracy from electroencephalogram (EEG) signals. The early detection of seizure is crucial for timely medical intervention and prevention of further injuries of the patients. This work proposes a robust deep learning framework called HyEpiSeiD that extracts self-trained features from the pre-processed EEG signals using a hybrid combination of convolutional neural network followed by two gated recurrent unit layers and performs prediction based on those extracted features. The proposed HyEpiSeiD framework is evaluated on two public datasets, the UCI Epilepsy and Mendeley datasets. The proposed HyEpiSeiD model achieved 99.01% and 97.50% classification accuracy, respectively, outperforming most of the state-of-the-art methods in epilepsy detection domain.

20.
Data Brief ; 55: 110753, 2024 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-39149720

RESUMEN

Today, speech synthesis is a part of our daily lives in computers all around the world. Central Kurdish Speech Corpus Construction is a speech corpus that is a primary data source for developing a speech system. There are still two main issues that prevent them from achieving the best possible performance, the lack of efficiency in training and analysis, and the difficulty in modelling. The biggest obstacle against text-to-speech in the Kurdish language is that there is a lack of text and speech recognition tools compounded by the fact that around 30 million people speak the Kurdish language in different countries. To address this issue, this corpus introduced a large vocabulary of Kurdish Text-to-Speech Dataset (KTTS, Gigant), including a pronunciation lexicon and speech corpus for the Central Kurdish dialect. A variety of subjects is comprised to record these sentences. The sentences are recorded in a voice recording studio by a Kurdish man who is a dubber. The goal of the speech corpus is to create a collection of sentences that accurately reflect the real data about the Central Kurdish dialect. A combination of audio and visual sources is used to record the 6,078 sentences of 12 document topics. They were recorded in a controlled environment using microphones that were not noisy. The total record duration is 13.63 h. The recorded sentences are in the ".wav" format.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA