Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 25
Filtrar
Más filtros

Bases de datos
Tipo del documento
Intervalo de año de publicación
1.
PLoS Comput Biol ; 20(2): e1011774, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-38422112

RESUMEN

Dendritic spines are the seat of most excitatory synapses in the brain, and a cellular structure considered central to learning, memory, and activity-dependent plasticity. The quantification of dendritic spines from light microscopy data is usually performed by humans in a painstaking and error-prone process. We found that human-to-human variability is substantial (inter-rater reliability 82.2±6.4%), raising concerns about the reproducibility of experiments and the validity of using human-annotated 'ground truth' as an evaluation method for computational approaches of spine identification. To address this, we present DeepD3, an open deep learning-based framework to robustly quantify dendritic spines in microscopy data in a fully automated fashion. DeepD3's neural networks have been trained on data from different sources and experimental conditions, annotated and segmented by multiple experts and they offer precise quantification of dendrites and dendritic spines. Importantly, these networks were validated in a number of datasets on varying acquisition modalities, species, anatomical locations and fluorescent indicators. The entire DeepD3 open framework, including the fully segmented training data, a benchmark that multiple experts have annotated, and the DeepD3 model zoo is fully available, addressing the lack of openly available datasets of dendritic spines while offering a ready-to-use, flexible, transparent, and reproducible spine quantification method.


Asunto(s)
Benchmarking , Espinas Dendríticas , Humanos , Reproducibilidad de los Resultados , Encéfalo , Colorantes
2.
PLoS Comput Biol ; 15(4): e1006699, 2019 04.
Artículo en Inglés | MEDLINE | ID: mdl-30958870

RESUMEN

We present Stytra, a flexible, open-source software package, written in Python and designed to cover all the general requirements involved in larval zebrafish behavioral experiments. It provides timed stimulus presentation, interfacing with external devices and simultaneous real-time tracking of behavioral parameters such as position, orientation, tail and eye motion in both freely-swimming and head-restrained preparations. Stytra logs all recorded quantities, metadata, and code version in standardized formats to allow full provenance tracking, from data acquisition through analysis to publication. The package is modular and expandable for different experimental protocols and setups. Current releases can be found at https://github.com/portugueslab/stytra. We also provide complete documentation with examples for extending the package to new stimuli and hardware, as well as a schema and parts list for behavioral setups. We showcase Stytra by reproducing previously published behavioral protocols in both head-restrained and freely-swimming larvae. We also demonstrate the use of the software in the context of a calcium imaging experiment, where it interfaces with other acquisition devices. Our aims are to enable more laboratories to easily implement behavioral experiments, as well as to provide a platform for sharing stimulus protocols that permits easy reproduction of experiments and straightforward validation. Finally, we demonstrate how Stytra can serve as a platform to design behavioral experiments involving tracking or visual stimulation with other animals and provide an example integration with the DeepLabCut neural network-based tracking method.


Asunto(s)
Conducta Animal/fisiología , Programas Informáticos , Animales , Biología Computacional , Sistemas de Computación , Diseño de Equipo , Larva/fisiología , Metadatos , Modelos Neurológicos , Redes Neurales de la Computación , Estimulación Luminosa , Lenguajes de Programación , Restricción Física , Grabación en Video , Pez Cebra/fisiología
3.
J Biol Chem ; 289(4): 1971-80, 2014 Jan 24.
Artículo en Inglés | MEDLINE | ID: mdl-24311784

RESUMEN

Inherited erythromelalgia (IEM) causes debilitating episodic neuropathic pain characterized by burning in the extremities. Inherited "paroxysmal extreme pain disorder" (PEPD) differs in its clinical picture and affects proximal body areas like the rectal, ocular, or jaw regions. Both pain syndromes have been linked to mutations in the voltage-gated sodium channel Nav1.7. Electrophysiological characterization shows that IEM-causing mutations generally enhance activation, whereas mutations leading to PEPD alter fast inactivation. Previously, an A1632E mutation of a patient with overlapping symptoms of IEM and PEPD was reported (Estacion, M., Dib-Hajj, S. D., Benke, P. J., Te Morsche, R. H., Eastman, E. M., Macala, L. J., Drenth, J. P., and Waxman, S. G. (2008) NaV1.7 Gain-of-function mutations as a continuum. A1632E displays physiological changes associated with erythromelalgia and paroxysmal extreme pain disorder mutations and produces symptoms of both disorders. J. Neurosci. 28, 11079-11088), displaying a shift of both activation and fast inactivation. Here, we characterize a new mutation of Nav1.7, A1632T, found in a patient suffering from IEM. Although transfection of A1632T in sensory neurons resulted in hyperexcitability and spontaneous firing of dorsal root ganglia (DRG) neurons, whole-cell patch clamp of transfected HEK cells revealed that Nav1.7 activation was unaltered by the A1632T mutation but that steady-state fast inactivation was shifted to more depolarized potentials. This is a characteristic normally attributed to PEPD-causing mutations. In contrast to the IEM/PEPD crossover mutation A1632E, A1632T failed to slow current decay (i.e. open-state inactivation) and did not increase resurgent currents, which have been suggested to contribute to high-frequency firing in physiological and pathological conditions. Reduced fast inactivation without increased resurgent currents induces symptoms of IEM, not PEPD, in the new Nav1.7 mutation, A1632T. Therefore, persistent and resurgent currents are likely to determine whether a mutation in Nav1.7 leads to IEM or PEPD.


Asunto(s)
Sustitución de Aminoácidos , Eritromelalgia/metabolismo , Mutación Missense , Canal de Sodio Activado por Voltaje NAV1.7/metabolismo , Dolor/metabolismo , Recto/anomalías , Eritromelalgia/genética , Eritromelalgia/patología , Femenino , Ganglios Espinales/metabolismo , Ganglios Espinales/patología , Células HEK293 , Humanos , Transporte Iónico/genética , Masculino , Canal de Sodio Activado por Voltaje NAV1.7/genética , Dolor/genética , Dolor/patología , Recto/metabolismo , Recto/patología , Células Receptoras Sensoriales/metabolismo , Células Receptoras Sensoriales/patología
4.
J Voice ; 2024 Feb 22.
Artículo en Inglés | MEDLINE | ID: mdl-38395653

RESUMEN

The Glottal Area Waveform (GAW) is an important component in quantitative clinical voice assessment, providing valuable insights into vocal fold function. In this study, we introduce a novel method employing Variational Autoencoders (VAEs) to generate synthetic GAWs. Our approach enables the creation of synthetic GAWs that closely replicate real-world data, offering a versatile tool for researchers and clinicians. We elucidate the process of manipulating the VAE latent space using the Glottal Opening Vector (GlOVe). The GlOVe allows precise control over the synthetic closure and opening of the vocal folds. By utilizing the GlOVe, we generate synthetic laryngeal biosignals. These biosignals accurately reflect vocal fold behavior, allowing for the emulation of realistic glottal opening changes. This manipulation extends to the introduction of arbitrary oscillations in the vocal folds, closely resembling real vocal fold oscillations. The range of factor coefficient values enables the generation of diverse biosignals with varying frequencies and amplitudes. Our results demonstrate that this approach yields highly accurate laryngeal biosignals, with the Normalized Mean Absolute Error values for various frequencies ranging from 9.6 â‹… 10-3 to 1.20 â‹… 10-2 for different experimented frequencies, alongside a remarkable training effectiveness, reflected in reductions of up to approximately 89.52% in key loss components. This proposed method may have implications for downstream speech synthesis and phonetics research, offering the potential for advanced and natural-sounding speech technologies.

5.
Biomed Opt Express ; 15(4): 2175-2186, 2024 Apr 01.
Artículo en Inglés | MEDLINE | ID: mdl-38633078

RESUMEN

Three-dimensional stacks acquired with confocal or two-photon microscopy are crucial for studying neuroanatomy. However, high-resolution image stacks acquired at multiple depths are time-consuming and susceptible to photobleaching. In vivo microscopy is further prone to motion artifacts. In this work, we suggest that deep neural networks with sine activation functions encoding implicit neural representations (SIRENs) are suitable for predicting intermediate planes and correcting motion artifacts, addressing the aforementioned shortcomings. We show that we can accurately estimate intermediate planes across multiple micrometers and fully automatically and unsupervised estimate a motion-corrected denoised picture. We show that noise statistics can be affected by SIRENs, however, rescued by a downstream denoising neural network, shown exemplarily with the recovery of dendritic spines. We believe that the application of these technologies will facilitate more efficient acquisition and superior post-processing in the future.

6.
Neoplasia ; 49: 100953, 2024 03.
Artículo en Inglés | MEDLINE | ID: mdl-38232493

RESUMEN

PURPOSE: Individual prediction of treatment response is crucial for personalized treatment in multimodal approaches against head-and-neck squamous cell carcinoma (HNSCC). So far, no reliable predictive parameters for treatment schemes containing immunotherapy have been identified. This study aims to predict treatment response to induction chemo-immunotherapy based on the peripheral blood immune status in patients with locally advanced HNSCC. METHODS: The peripheral blood immune phenotype was assessed in whole blood samples in patients treated in the phase II CheckRad-CD8 trial as part of the pre-planned translational research program. Blood samples were analyzed by multicolor flow cytometry before (T1) and after (T2) induction chemo-immunotherapy with cisplatin/docetaxel/durvalumab/tremelimumab. Machine Learning techniques were used to predict pathological complete response (pCR) after induction therapy. RESULTS: The tested classifier methods (LDA, SVM, LR, RF, DT, and XGBoost) allowed a distinct prediction of pCR. Highest accuracy was achieved with a low number of features represented as principal components. Immune parameters obtained from the absolute difference (lT2-T1l) allowed the best prediction of pCR. In general, less than 30 parameters and at most 10 principal components were needed for highly accurate predictions. Across several datasets, cells of the innate immune system such as polymorphonuclear cells, monocytes, and plasmacytoid dendritic cells are most prominent. CONCLUSIONS: Our analyses imply that alterations of the innate immune cell distribution in the peripheral blood following induction chemo-immuno-therapy is highly predictive for pCR in HNSCC.


Asunto(s)
Carcinoma de Células Escamosas , Neoplasias de Cabeza y Cuello , Humanos , Carcinoma de Células Escamosas de Cabeza y Cuello/terapia , Neoplasias de Cabeza y Cuello/tratamiento farmacológico , Carcinoma de Células Escamosas/tratamiento farmacológico , Carcinoma de Células Escamosas/genética , Quimioterapia de Inducción/métodos , Inmunofenotipificación , Inmunoterapia , Linfocitos T CD8-positivos , Inmunidad Innata
7.
IEEE J Transl Eng Health Med ; 11: 137-144, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-36816097

RESUMEN

High-speed videoendoscopy is a major tool for quantitative laryngology. Glottis segmentation and glottal midline detection are crucial for computing vocal fold-specific, quantitative parameters. However, fully automated solutions show limited clinical applicability. Especially unbiased glottal midline detection remains a challenging problem. We developed a multitask deep neural network for glottis segmentation and glottal midline detection. We used techniques from pose estimation to estimate the anterior and posterior points in endoscopy images. Neural networks were set up in TensorFlow/Keras and trained and evaluated with the BAGLS dataset. We found that a dual decoder deep neural network termed GlottisNetV2 outperforms the previously proposed GlottisNet in terms of MAPE on the test dataset (1.85% to 6.3%) while converging faster. Using various hyperparameter tunings, we allow fast and directed training. Using temporal variant data on an additional data set designed for this task, we can improve the median prediction accuracy from 2.1% to 1.76% when using 12 consecutive frames and additional temporal filtering. We found that temporal glottal midline detection using a dual decoder architecture together with keypoint estimation allows accurate midline prediction. We show that our proposed architecture allows stable and reliable glottal midline predictions ready for clinical use and analysis of symmetry measures.


Asunto(s)
Glotis , Pliegues Vocales , Redes Neurales de la Computación , Endoscopía
8.
J Speech Lang Hear Res ; 66(2): 565-572, 2023 02 13.
Artículo en Inglés | MEDLINE | ID: mdl-36716396

RESUMEN

PURPOSE: This research note illustrates the effects of video data with nonsquare pixels on the pixel-based measures obtained from videofluoroscopic swallow studies (VFSS). METHOD: Six pixel-based distance and area measures were obtained from two different videoflouroscopic study units; both yielding videos with nonsquare pixels with different pixel aspect ratios (PARs). The swallowing measures were obtained from the original VFSS videos and from the videos after their pixels were squared. RESULTS: The results demonstrated significant multivariate effects both in video type (original vs. squared) and in the interaction between video type and sample (two video recordings of different patients, different PARs, and opposing tilt angles of the external reference). A wide range of variabilities was observed on the pixel-based measures between original and squared videos with the percent deviation ranging from 0.1% to 9.1% with the maximum effect size of 7.43. CONCLUSIONS: This research note demonstrates the effect of disregarding PAR to distance and area pixel-based parameters. In addition, we present a multilevel roadmap to prevent possible measurement errors that could occur. At the planning stage, the PAR of video source should be identified, and, at the analyses stage, video data should be prescaled prior to analysis with PAR-unaware software. No methodology in prior absolute or relative pixel-based studies reports adjustment to the PAR prior to measurements nor identify the PAR as a possible source of variation within the literature. Addressing PAR will improve the precision and stability of pixel-based VFSS findings and improve comparability within and across clinical and research settings. SUPPLEMENTAL MATERIAL: https://doi.org/10.23641/asha.21957134.


Asunto(s)
Trastornos de Deglución , Humanos , Trastornos de Deglución/diagnóstico por imagen , Deglución , Grabación en Video/métodos , Programas Informáticos , Fluoroscopía/métodos
9.
Diagn Pathol ; 18(1): 121, 2023 Nov 03.
Artículo en Inglés | MEDLINE | ID: mdl-37924082

RESUMEN

PURPOSE: Although neural networks have shown remarkable performance in medical image analysis, their translation into clinical practice remains difficult due to their lack of interpretability. An emerging field that addresses this problem is Explainable AI. METHODS: Here, we aimed to investigate the ability of Convolutional Neural Networks (CNNs) to classify head and neck cancer histopathology. To this end, we manually annotated 101 histopathological slides of locally advanced head and neck squamous cell carcinoma. We trained a CNN to classify tumor and non-tumor tissue, and another CNN to semantically segment four classes - tumor, non-tumor, non-specified tissue, and background. We applied Explainable AI techniques, namely Grad-CAM and HR-CAM, to both networks and explored important features that contributed to their decisions. RESULTS: The classification network achieved an accuracy of 89.9% on previously unseen data. Our segmentation network achieved a class-averaged Intersection over Union score of 0.690, and 0.782 for tumor tissue in particular. Explainable AI methods demonstrated that both networks rely on features agreeing with the pathologist's expert opinion. CONCLUSION: Our work suggests that CNNs can predict head and neck cancer with high accuracy. Especially if accompanied by visual explanations, CNNs seem promising for assisting pathologists in the assessment of cancer sections.


Asunto(s)
Neoplasias de Cabeza y Cuello , Procesamiento de Imagen Asistido por Computador , Humanos , Procesamiento de Imagen Asistido por Computador/métodos , Redes Neurales de la Computación , Carcinoma de Células Escamosas de Cabeza y Cuello
10.
Cancer Med ; 2023 Dec 22.
Artículo en Inglés | MEDLINE | ID: mdl-38132808

RESUMEN

BACKGROUND: The significance of different histological spreading patterns of tumor tissue in oral tongue squamous cell carcinoma (TSCC) is well known. Our aim was to construct a numeric parameter on a continuous scale, that is, the modified Polsby-Popper (MPP) score, to describe the aggressiveness of tumor growth and infiltration, with the potential to analyze hematoxylin and eosin-stained whole slide images (WSIs) in an automated manner. We investigated the application of the MPP score in predicting survival and cervical lymph node metastases as well as in determining patients at risk in the context of different surgical margin scenarios. METHODS: We developed a semiautomated image analysis pipeline to detect areas belonging to the tumor tissue compartment. Perimeter and area measurements of all detected tissue regions were derived, and a specific mathematical formula was applied to reflect the perimeter/area ratio in a comparable, observer-independent manner across digitized WSIs. We demonstrated the plausibility of the MPP score by correlating it with well-established clinicopathologic parameters. We then performed survival analysis to assess the relevance of the MPP score, with an emphasis on different surgical margin scenarios. Machine learning models were developed to assess the relevance of the MPP score in predicting survival and occult cervical nodal metastases. RESULTS: The MPP score was associated with unfavorable tumor growth and infiltration patterns, the presence of lymph node metastases, the extracapsular spread of tumor cells, and higher tumor thickness. Higher MPP scores were associated with worse overall survival (OS) and tongue carcinoma-specific survival (TCSS), both when assessing all pT-categories and pT1-pT2 categories only; moreover, higher MPP scores were associated with a significantly worse TCSS in cases where a cancer-free surgical margin of <5 mm could be achieved on the main surgical specimen. This discriminatory capacity remained constant when examining pT1-pT2 categories only. Importantly, the MPP score could successfully define cases at risk in terms of metastatic disease in pT1-pT2 cancer where tumor thickness failed to exhibit a significant predictive value. Machine learning (ML) models incorporating the MPP score could predict the 5-year TCSS efficiently. Furthermore, we demonstrated that machine learning models that predict occult cervical lymph node involvement can benefit from including the MPP score. CONCLUSIONS: We introduced an objective, quantifiable, and observer-independent parameter, the MPP score, representing the aggressiveness of tumor growth and infiltration in TSCC. We showed its prognostic relevance especially in pT1-pT2 category TSCC, and its possible use in ML models predicting TCSS and occult lymph node metastases.

11.
PLoS One ; 17(9): e0266989, 2022.
Artículo en Inglés | MEDLINE | ID: mdl-36129922

RESUMEN

Deep Learning has a large impact on medical image analysis and lately has been adopted for clinical use at the point of care. However, there is only a small number of reports of long-term studies that show the performance of deep neural networks (DNNs) in such an environment. In this study, we measured the long-term performance of a clinically optimized DNN for laryngeal glottis segmentation. We have collected the video footage for two years from an AI-powered laryngeal high-speed videoendoscopy imaging system and found that the footage image quality is stable across time. Next, we determined the DNN segmentation performance on lossy and lossless compressed data revealing that only 9% of recordings contain segmentation artifacts. We found that lossy and lossless compression is on par for glottis segmentation, however, lossless compression provides significantly superior image quality. Lastly, we employed continual learning strategies to continuously incorporate new data into the DNN to remove the aforementioned segmentation artifacts. With modest manual intervention, we were able to largely alleviate these segmentation artifacts by up to 81%. We believe that our suggested deep learning-enhanced laryngeal imaging platform consistently provides clinically sound results, and together with our proposed continual learning scheme will have a long-lasting impact on the future of laryngeal imaging.


Asunto(s)
Laringe , Sistemas de Atención de Punto , Artefactos , Glotis/diagnóstico por imagen , Procesamiento de Imagen Asistido por Computador/métodos , Laringe/diagnóstico por imagen , Redes Neurales de la Computación
12.
Sci Rep ; 12(1): 14292, 2022 08 22.
Artículo en Inglés | MEDLINE | ID: mdl-35995933

RESUMEN

Glottis segmentation is a crucial step to quantify endoscopic footage in laryngeal high-speed videoendoscopy. Recent advances in deep neural networks for glottis segmentation allow for a fully automatic workflow. However, exact knowledge of integral parts of these deep segmentation networks remains unknown, and understanding the inner workings is crucial for acceptance in clinical practice. Here, we show that a single latent channel as a bottleneck layer is sufficient for glottal area segmentation using systematic ablations. We further demonstrate that the latent space is an abstraction of the glottal area segmentation relying on three spatially defined pixel subtypes allowing for a transparent interpretation. We further provide evidence that the latent space is highly correlated with the glottal area waveform, can be encoded with four bits, and decoded using lean decoders while maintaining a high reconstruction accuracy. Our findings suggest that glottis segmentation is a task that can be highly optimized to gain very efficient and explainable deep neural networks, important for application in the clinic. In the future, we believe that online deep learning-assisted monitoring is a game-changer in laryngeal examinations.


Asunto(s)
Glotis , Laringe , Endoscopía , Glotis/diagnóstico por imagen , Procesamiento de Imagen Asistido por Computador , Redes Neurales de la Computación , Grabación en Video
13.
J Expo Sci Environ Epidemiol ; 32(5): 727-734, 2022 09.
Artículo en Inglés | MEDLINE | ID: mdl-34611302

RESUMEN

BACKGROUND: In the CoVID-19 pandemic, singing came into focus as a high-risk activity for the infection with airborne viruses and was therefore forbidden by many governmental administrations. OBJECTIVE: The aim of this study is to investigate the effectiveness of surgical masks regarding the spatial and temporal dispersion of aerosol and droplets during professional singing. METHODS: Ten professional singers performed a passage of the Ludwig van Beethoven's "Ode of Joy" in two experimental setups-each with and without surgical masks. First, they sang with previously inhaled vapor of e-cigarettes. The emitted cloud was recorded by three cameras to measure its dispersion dynamics. Secondly, the naturally expelled larger droplets were illuminated by a laser light sheet and recorded by a high-speed camera. RESULTS: The exhaled vapor aerosols were decelerated and deflected by the mask and stayed in the singer's near-field around and above their heads. In contrast, without mask, the aerosols spread widely reaching distances up to 1.3 m. The larger droplets were reduced by up to 86% with a surgical mask worn. SIGNIFICANCE: The study shows that surgical masks display an effective tool to reduce the range of aerosol dispersion during singing. In combination with an appropriate aeration strategy for aerosol removal, choir singers could be positioned in a more compact assembly without contaminating neighboring singers all singers.


Asunto(s)
COVID-19 , Sistemas Electrónicos de Liberación de Nicotina , Canto , Humanos , Máscaras , Pandemias , Aerosoles y Gotitas Respiratorias
14.
Sci Rep ; 11(1): 13760, 2021 07 02.
Artículo en Inglés | MEDLINE | ID: mdl-34215788

RESUMEN

High-speed videoendoscopy is an important tool to study laryngeal dynamics, to quantify vocal fold oscillations, to diagnose voice impairments at laryngeal level and to monitor treatment progress. However, there is a significant lack of an open source, expandable research tool that features latest hardware and data analysis. In this work, we propose an open research platform termed OpenHSV that is based on state-of-the-art, commercially available equipment and features a fully automatic data analysis pipeline. A publicly available, user-friendly graphical user interface implemented in Python is used to interface the hardware. Video and audio data are recorded in synchrony and are subsequently fully automatically analyzed. Video segmentation of the glottal area is performed using efficient deep neural networks to derive glottal area waveform and glottal midline. Established quantitative, clinically relevant video and audio parameters were implemented and computed. In a preliminary clinical study, we recorded video and audio data from 28 healthy subjects. Analyzing these data in terms of image quality and derived quantitative parameters, we show the applicability, performance and usefulness of OpenHSV. Therefore, OpenHSV provides a valid, standardized access to high-speed videoendoscopy data acquisition and analysis for voice scientists, highlighting its use as a valuable research tool in understanding voice physiology. We envision that OpenHSV serves as basis for the next generation of clinical HSV systems.


Asunto(s)
Glotis/cirugía , Enfermedades de la Laringe/cirugía , Laringoscopía/métodos , Laringe/cirugía , Adolescente , Adulto , Femenino , Glotis/diagnóstico por imagen , Glotis/fisiopatología , Humanos , Enfermedades de la Laringe/diagnóstico por imagen , Enfermedades de la Laringe/patología , Laringoscopía/instrumentación , Laringe/diagnóstico por imagen , Laringe/patología , Masculino , Persona de Mediana Edad , Redes Neurales de la Computación , Grabación en Video , Pliegues Vocales/diagnóstico por imagen , Pliegues Vocales/fisiopatología , Pliegues Vocales/cirugía , Voz/fisiología , Trastornos de la Voz/diagnóstico por imagen , Trastornos de la Voz/fisiopatología , Trastornos de la Voz/cirugía , Calidad de la Voz/fisiología , Adulto Joven
15.
Nat Commun ; 12(1): 6694, 2021 11 18.
Artículo en Inglés | MEDLINE | ID: mdl-34795244

RESUMEN

Animals must adapt their behavior to survive in a changing environment. Behavioral adaptations can be evoked by two mechanisms: feedback control and internal-model-based control. Feedback controllers can maintain the sensory state of the animal at a desired level under different environmental conditions. In contrast, internal models learn the relationship between the motor output and its sensory consequences and can be used to recalibrate behaviors. Here, we present multiple unpredictable perturbations in visual feedback to larval zebrafish performing the optomotor response and show that they react to these perturbations through a feedback control mechanism. In contrast, if a perturbation is long-lasting, fish adapt their behavior by updating a cerebellum-dependent internal model. We use modelling and functional imaging to show that the neuronal requirements for these mechanisms are met in the larval zebrafish brain. Our results illustrate the role of the cerebellum in encoding internal models and how these can calibrate neuronal circuits involved in reactive behaviors depending on the interactions between animal and environment.


Asunto(s)
Cerebelo/fisiología , Retroalimentación Fisiológica/fisiología , Retroalimentación Sensorial/fisiología , Pez Cebra/fisiología , Adaptación Fisiológica/fisiología , Animales , Animales Modificados Genéticamente , Encéfalo/citología , Encéfalo/fisiología , Cerebelo/citología , Humanos , Larva/genética , Larva/fisiología , Aprendizaje/fisiología , Neuronas/fisiología , Pez Cebra/genética
16.
PLoS One ; 16(2): e0246136, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-33529244

RESUMEN

In voice research, uncovering relations between the oscillating vocal folds, being the sound source of phonation, and the resulting perceived acoustic signal are of great interest. This is especially the case in the context of voice disorders, such as functional dysphonia (FD). We investigated 250 high-speed videoendoscopy (HSV) recordings with simultaneously recorded acoustic signals (124 healthy females, 60 FD females, 44 healthy males, 22 FD males). 35 glottal area waveform (GAW) parameters and 14 acoustic parameters were calculated for each recording. Linear and non-linear relations between GAW and acoustic parameters were investigated using Pearson correlation coefficients (PCC) and distance correlation coefficients (DCC). Further, norm values for parameters obtained from 250 ms long sustained phonation data (vowel /i/) were provided. 26 PCCs in females (5.3%) and 8 in males (1.6%) were found to be statistically significant (|corr.| ≥ 0.3). Only minor differences were found between PCCs and DCCs, indicating presence of weak non-linear dependencies between parameters. Fundamental frequency was involved in the majority of all relevant PCCs between GAW and acoustic parameters (19 in females and 7 in males). The most distinct difference between correlations in females and males was found for the parameter Period Variability Index. The study shows only weak relations between investigated acoustic and GAW-parameters. This indicates that the reduction of the complex 3D glottal dynamics to the 1D-GAW may erase laryngeal dynamic characteristics that are reflected within the acoustic signal. Hence, other GAW parameters, 2D-, 3D-laryngeal dynamics and vocal tract parameters should be further investigated towards potential correlations to the acoustic signal.


Asunto(s)
Disfonía/fisiopatología , Glotis/fisiopatología , Laringoscopía/métodos , Acústica , Adulto , Anciano , Estudios de Casos y Controles , Femenino , Humanos , Laringoscopía/instrumentación , Masculino , Persona de Mediana Edad , Grabación en Video , Calidad de la Voz , Adulto Joven
17.
J Speech Lang Hear Res ; 64(6): 1889-1903, 2021 06 04.
Artículo en Inglés | MEDLINE | ID: mdl-34000199

RESUMEN

Purpose High-speed videoendoscopy (HSV) is an emerging, but barely used, endoscopy technique in the clinic to assess and diagnose voice disorders because of the lack of dedicated software to analyze the data. HSV allows to quantify the vocal fold oscillations by segmenting the glottal area. This challenging task has been tackled by various studies; however, the proposed approaches are mostly limited and not suitable for daily clinical routine. Method We developed a user-friendly software in C# that allows the editing, motion correction, segmentation, and quantitative analysis of HSV data. We further provide pretrained deep neural networks for fully automatic glottis segmentation. Results We freely provide our software Glottis Analysis Tools (GAT). Using GAT, we provide a general threshold-based region growing platform that enables the user to analyze data from various sources, such as in vivo recordings, ex vivo recordings, and high-speed footage of artificial vocal folds. Additionally, especially for in vivo recordings, we provide three robust neural networks at various speed and quality settings to allow a fully automatic glottis segmentation needed for application by untrained personnel. GAT further evaluates video and audio data in parallel and is able to extract various features from the video data, among others the glottal area waveform, that is, the changing glottal area over time. In total, GAT provides 79 unique quantitative analysis parameters for video- and audio-based signals. Many of these parameters have already been shown to reflect voice disorders, highlighting the clinical importance and usefulness of the GAT software. Conclusion GAT is a unique tool to process HSV and audio data to determine quantitative, clinically relevant parameters for research, diagnosis, and treatment of laryngeal disorders. Supplemental Material https://doi.org/10.23641/asha.14575533.


Asunto(s)
Aprendizaje Profundo , Laringe , Glotis , Humanos , Laringoscopía , Fonación , Programas Informáticos , Vibración , Grabación en Video , Pliegues Vocales
18.
Sci Rep ; 10(1): 20723, 2020 11 26.
Artículo en Inglés | MEDLINE | ID: mdl-33244031

RESUMEN

A healthy voice is crucial for verbal communication and hence in daily as well as professional life. The basis for a healthy voice are the sound producing vocal folds in the larynx. A hallmark of healthy vocal fold oscillation is the symmetric motion of the left and right vocal fold. Clinically, videoendoscopy is applied to assess the symmetry of the oscillation and evaluated subjectively. High-speed videoendoscopy, an emerging method that allows quantification of the vocal fold oscillation, is more commonly employed in research due to the amount of data and the complex, semi-automatic analysis. In this study, we provide a comprehensive evaluation of methods that detect fully automatically the glottal midline. We used a biophysical model to simulate different vocal fold oscillations, extended the openly available BAGLS dataset using manual annotations, utilized both, simulations and annotated endoscopic images, to train deep neural networks at different stages of the analysis workflow, and compared these to established computer vision algorithms. We found that classical computer vision perform well on detecting the glottal midline in glottis segmentation data, but are outperformed by deep neural networks on this task. We further suggest GlottisNet, a multi-task neural architecture featuring the simultaneous prediction of both, the opening between the vocal folds and the symmetry axis, leading to a huge step forward towards clinical applicability of quantitative, deep learning-assisted laryngeal endoscopy, by fully automating segmentation and midline detection.

19.
IEEE J Transl Eng Health Med ; 8: 2100511, 2020.
Artículo en Inglés | MEDLINE | ID: mdl-32518739

RESUMEN

BACKGROUND: Various voice assessment tools, such as questionnaires and aerodynamic voice characteristics, can be used to assess vocal function of individuals. However, not much is known about the best combinations of these parameters in identification of functional dysphonia in clinical settings. METHODS: This study investigated six scores from clinically commonly used questionnaires and seven acoustic parameters. 514 females and 277 males were analyzed. The subjects were divided into three groups: one healthy group (N01) (49 females, 50 males) and two disordered groups with perceptually hoarse (FD23) (220 females, 96 males) and perceptually not hoarse (FD01) (245 females, 131 males) sounding voices. A tree stumps Adaboost approach was applied to find the subset of parameters that best separates the groups. Subsequently, it was determined if this parameter subset reflects treatment outcome for 120 female and 51 male patients by pairwise pre- and post-treatment comparisons of parameters. RESULTS: The questionnaire "Voice-related-quality-of-Life" and three objective parameters ("maximum fundamental frequency", "maximum Intensity" and "Jitter Percent") were sufficient to separate the groups (accuracy ranging from 0.690 (FD01 vs. FD23, females) to 0.961 (N01 vs. FD23, females)). Our study suggests that a reduced parameter subset (4 out of 13) is sufficient to separate these three groups. All parameters reflected treatment outcome for patients with hoarse voices, Voice-related-quality-of-Life showed improvement for the not hoarse group (FD01). CONCLUSION: Results show that single parameters are insufficient to separate voice disorders but a set of several well-chosen parameters is. These findings will help to optimize and reduce clinical assessment time.

20.
Sci Data ; 7(1): 186, 2020 06 19.
Artículo en Inglés | MEDLINE | ID: mdl-32561845

RESUMEN

Laryngeal videoendoscopy is one of the main tools in clinical examinations for voice disorders and voice research. Using high-speed videoendoscopy, it is possible to fully capture the vocal fold oscillations, however, processing the recordings typically involves a time-consuming segmentation of the glottal area by trained experts. Even though automatic methods have been proposed and the task is particularly suited for deep learning methods, there are no public datasets and benchmarks available to compare methods and to allow training of generalizing deep learning models. In an international collaboration of researchers from seven institutions from the EU and USA, we have created BAGLS, a large, multihospital dataset of 59,250 high-speed videoendoscopy frames with individually annotated segmentation masks. The frames are based on 640 recordings of healthy and disordered subjects that were recorded with varying technical equipment by numerous clinicians. The BAGLS dataset will allow an objective comparison of glottis segmentation methods and will enable interested researchers to train their own models and compare their methods.


Asunto(s)
Endoscopía , Glotis/fisiología , Grabación en Video , Pliegues Vocales/fisiología , Trastornos de la Voz/diagnóstico , Glotis/diagnóstico por imagen , Humanos , Pliegues Vocales/diagnóstico por imagen
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA