Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 61
Filtrar
1.
J Neurosci ; 43(10): 1830-1844, 2023 03 08.
Artigo em Inglês | MEDLINE | ID: mdl-36717226

RESUMO

The amyloid precursor protein (APP) is linked to the genetics and pathogenesis of Alzheimer's disease (AD). It is the parent protein of the ß-amyloid (Aß) peptide, the main constituent of the amyloid plaques found in an AD brain. The pathways from APP to Aß are intensively studied, yet the normal functions of APP itself have generated less interest. We report here that glutamate stimulation of neuronal activity leads to a rapid increase in App gene expression. In mouse and human neurons, elevated APP protein changes the structure of the axon initial segment (AIS) where action potentials are initiated. The AIS is shortened in length and shifts away from the cell body. The GCaMP8f Ca2+ reporter confirms the predicted decrease in neuronal activity. NMDA antagonists or knockdown of App block the glutamate effects. The actions of APP on the AIS are cell-autonomous; exogenous Aß, either fibrillar or oligomeric, has no effect. In culture, APPSwe (a familial AD mutation) induces larger AIS changes than wild type APP. Ankyrin G and ßIV-spectrin, scaffolding proteins of the AIS, both physically associate with APP, more so in AD brains. Finally, in humans with sporadic AD or in the R1.40 AD mouse model, both females and males, neurons have elevated levels of APP protein that invade the AIS. In vivo as in vitro, this increased APP is associated with a significant shortening of the AIS. The findings outline a new role for the APP and encourage a reconsideration of its relationship to AD.SIGNIFICANCE STATEMENT While the amyloid precursor protein (APP) has long been associated with Alzheimer's disease (AD), the normal functions of the full-length Type I membrane protein have been largely unexplored. We report here that the levels of APP protein increase with neuronal activity. In vivo and in vitro, modest amounts of excess APP alter the properties of the axon initial segment. The ß-amyloid peptide derived from APP is without effect. Consistent with the observed changes in the axon initial segment which would be expected to decrease action potential firing, we show that APP expression depresses neuronal activity. In mouse AD models and human sporadic AD, APP physically associates with the scaffolding proteins of the axon initial segment, suggesting a relationship with AD dementia.


Assuntos
Doença de Alzheimer , Segmento Inicial do Axônio , Masculino , Feminino , Camundongos , Humanos , Animais , Precursor de Proteína beta-Amiloide/genética , Precursor de Proteína beta-Amiloide/metabolismo , Doença de Alzheimer/metabolismo , Segmento Inicial do Axônio/metabolismo , Peptídeos beta-Amiloides/metabolismo , Proteínas de Membrana , Camundongos Transgênicos , Modelos Animais de Doenças
2.
IEEE Trans Pattern Anal Mach Intell ; 44(1): 361-372, 2022 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-32750822

RESUMO

Optical flow is a crucial component of the feature space for early visual processing of dynamic scenes especially in new applications such as self-driving vehicles, drones and autonomous robots. The dynamic vision sensors are well suited for such applications because of their asynchronous, sparse and temporally precise representation of the visual dynamics. Many algorithms proposed for computing visual flow for these sensors suffer from the aperture problem as the direction of the estimated flow is governed by the curvature of the object rather than the true motion direction. Some methods that do overcome this problem by temporal windowing under-utilize the true precise temporal nature of the dynamic sensors. In this paper, we propose a novel multi-scale plane fitting based visual flow algorithm that is robust to the aperture problem and also computationally fast and efficient. Our algorithm performs well in many scenarios ranging from fixed camera recording simple geometric shapes to real world scenarios such as camera mounted on a moving car and can successfully perform event-by-event motion estimation of objects in the scene to allow for predictions of upto 500 ms i.e., equivalent to 10 to 25 frames with traditional cameras.

3.
Sci Rep ; 11(1): 10783, 2021 05 24.
Artigo em Inglês | MEDLINE | ID: mdl-34031442

RESUMO

The neural encoding of visual features in primary visual cortex (V1) is well understood, with strong correlates to low-level perception, making V1 a strong candidate for vision restoration through neuroprosthetics. However, the functional relevance of neural dynamics evoked through external stimulation directly imposed at the cortical level is poorly understood. Furthermore, protocols for designing cortical stimulation patterns that would induce a naturalistic perception of the encoded stimuli have not yet been established. Here, we demonstrate a proof of concept by solving these issues through a computational model, combining (1) a large-scale spiking neural network model of cat V1 and (2) a virtual prosthetic system transcoding the visual input into tailored light-stimulation patterns which drive in situ the optogenetically modified cortical tissue. Using such virtual experiments, we design a protocol for translating simple Fourier contrasted stimuli (gratings) into activation patterns of the optogenetic matrix stimulator. We then quantify the relationship between spatial configuration of the imposed light pattern and the induced cortical activity. Our simulations in the absence of visual drive (simulated blindness) show that optogenetic stimulation with a spatial resolution as low as 100 [Formula: see text]m, and light intensity as weak as [Formula: see text] photons/s/cm[Formula: see text] is sufficient to evoke activity patterns in V1 close to those evoked by normal vision.


Assuntos
Optogenética/métodos , Estimulação Luminosa/métodos , Córtex Visual/fisiologia , Animais , Olho Artificial , Humanos , Modelos Teóricos , Estudo de Prova de Conceito , Vias Visuais , Percepção Visual
4.
Front Neurosci ; 15: 672161, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-34054420

RESUMO

Autonomous flight for large aircraft appears to be within our reach. However, launching autonomous systems for everyday missions still requires an immense interdisciplinary research effort supported by pointed policies and funding. We believe that concerted endeavors in the fields of neuroscience, mathematics, sensor physics, robotics, and computer science are needed to address remaining crucial scientific challenges. In this paper, we argue for a bio-inspired approach to solve autonomous flying challenges, outline the frontier of sensing, data processing, and flight control within a neuromorphic paradigm, and chart directions of research needed to achieve operational capabilities comparable to those we observe in nature. One central problem of neuromorphic computing is learning. In biological systems, learning is achieved by adaptive and relativistic information acquisition characterized by near-continuous information retrieval with variable rates and sparsity. This results in both energy and computational resource savings being an inspiration for autonomous systems. We consider pertinent features of insect, bat and bird flight behavior as examples to address various vital aspects of autonomous flight. Insects exhibit sophisticated flight dynamics with comparatively reduced complexity of the brain. They represent excellent objects for the study of navigation and flight control. Bats and birds enable more complex models of attention and point to the importance of active sensing for conducting more complex missions. The implementation of neuromorphic paradigms for autonomous flight will require fundamental changes in both traditional hardware and software. We provide recommendations for sensor hardware and processing algorithm development to enable energy efficient and computationally effective flight control.

5.
Commun Biol ; 4(1): 125, 2021 01 27.
Artigo em Inglês | MEDLINE | ID: mdl-33504896

RESUMO

Vision restoration is an ideal medical application for optogenetics, because the eye provides direct optical access to the retina for stimulation. Optogenetic therapy could be used for diseases involving photoreceptor degeneration, such as retinitis pigmentosa or age-related macular degeneration. We describe here the selection, in non-human primates, of a specific optogenetic construct currently tested in a clinical trial. We used the microbial opsin ChrimsonR, and showed that the AAV2.7m8 vector had a higher transfection efficiency than AAV2 in retinal ganglion cells (RGCs) and that ChrimsonR fused to tdTomato (ChR-tdT) was expressed more efficiently than ChrimsonR. Light at 600 nm activated RGCs transfected with AAV2.7m8 ChR-tdT, from an irradiance of 1015 photons.cm-2.s-1. Vector doses of 5 × 1010 and 5 × 1011 vg/eye transfected up to 7000 RGCs/mm2 in the perifovea, with no significant immune reaction. We recorded RGC responses from a stimulus duration of 1 ms upwards. When using the recorded activity to decode stimulus information, we obtained an estimated visual acuity of 20/249, above the level of legal blindness (20/400). These results lay the groundwork for the ongoing clinical trial with the AAV2.7m8 - ChR-tdT vector for vision restoration in patients with retinitis pigmentosa.


Assuntos
Optogenética , Estimulação Luminosa , Degeneração Retiniana/terapia , Visão Ocular/fisiologia , Animais , Equipamentos e Provisões , Feminino , Humanos , Macaca fascicularis , Masculino , Optogenética/instrumentação , Optogenética/métodos , Reconhecimento Visual de Modelos/fisiologia , Estimulação Luminosa/instrumentação , Estimulação Luminosa/métodos , Primatas , Degeneração Retiniana/fisiopatologia , Degeneração Retiniana/reabilitação , Terapias em Estudo/instrumentação , Terapias em Estudo/métodos
6.
J Clin Monit Comput ; 35(4): 835-847, 2021 08.
Artigo em Inglês | MEDLINE | ID: mdl-32533529

RESUMO

We evaluated the performance of a new device to control the administration of fluid alone or co-administration of fluid and norepinephrine in a pig model of haemorrhagic shock in two sets of experiments. In the first one, resuscitation was guided using continuous arterial pressure measurements (three groups: resuscitation with fluid by a physician, CL resuscitation with fluid, and CL resuscitation with fluid and norepinephrine). In the second one, resuscitation was guided using discontinuous arterial pressure measurements (three groups: CL resuscitation with fluid alone, CL resuscitation with fluid and moderate dose norepinephrine, and CL resuscitation with fluid and a high dose of norepinephrine). Pigs were resuscitated for 1 h. In the first set of experiments, proportion of time spent in the target area of 78-88 mmHg of systolic arterial pressure was not statistically different between the three groups: manual, 71.2% (39.1-80.1); CL with fluid, 87.8% (68.3-97.4); and CL with fluid and norepinephrine, 78.1% (59.2-83.6), p = 0.151. In the second set of experiments, performance of CL resuscitation with fluid or with combination of fluid and high or moderate dose of norepinephrine was not significantly different (p = 0.543 for time in target). Pigs resuscitated with norepinephrine required less fluid and had less haemodilution than pigs resuscitated with fluid alone. Performance of CL resuscitation using continuous arterial pressure measurement was not significantly different than optimised manual treatment by a dedicated physician. Performance of CL resuscitation was reduced with discontinuous arterial pressure measurements in comparison with continuous arterial pressure measurements.


Assuntos
Choque Hemorrágico , Animais , Hidratação , Norepinefrina/uso terapêutico , Projetos Piloto , Ressuscitação , Choque Hemorrágico/terapia , Suínos
7.
Front Neurosci ; 14: 587, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-32848527

RESUMO

We present the first purely event-based method for face detection using the high temporal resolution properties of an event-based camera to detect the presence of a face in a scene using eye blinks. Eye blinks are a unique and stable natural dynamic temporal signature of human faces across population that can be fully captured by event-based sensors. We show that eye blinks have a unique temporal signature over time that can be easily detected by correlating the acquired local activity with a generic temporal model of eye blinks that has been generated from a wide population of users. In a second stage once a face has been located it becomes possible to apply a probabilistic framework to track its spatial location for each incoming event while using eye blinks to correct for drift and tracking errors. Results are shown for several indoor and outdoor experiments. We also release an annotated data set that can be used for future work on the topic.

8.
Sensors (Basel) ; 20(12)2020 Jun 16.
Artigo em Inglês | MEDLINE | ID: mdl-32560238

RESUMO

Neuromorphic vision sensors detect changes in luminosity taking inspiration from mammalian retina and providing a stream of events with high temporal resolution, also known as Dynamic Vision Sensors (DVS). This continuous stream of events can be used to extract spatio-temporal patterns from a scene. A time-surface represents a spatio-temporal context for a given spatial radius around an incoming event from a sensor at a specific time history. Time-surfaces can be organized in a hierarchical way to extract features from input events using the Hierarchy Of Time-Surfaces algorithm, hereinafter HOTS. HOTS can be organized in consecutive layers to extract combination of features in a similar way as some deep-learning algorithms do. This work introduces a novel FPGA architecture for accelerating HOTS network. This architecture is mainly based on block-RAM memory and the non-restoring square root algorithm, requiring basic components and enabling it for low-power low-latency embedded applications. The presented architecture has been tested on a Zynq 7100 platform at 100 MHz. The results show that the latencies are in the range of 1 µ s to 6.7 µ s, requiring a maximum dynamic power consumption of 77 mW. This system was tested with a gesture recognition dataset, obtaining an accuracy loss for 16-bit precision of only 1.2% with respect to the original software HOTS.

9.
Front Neurosci ; 14: 420, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-32528239

RESUMO

Precise spike timing and temporal coding are used extensively within the nervous system of insects and in the sensory periphery of higher order animals. However, conventional Artificial Neural Networks (ANNs) and machine learning algorithms cannot take advantage of this coding strategy, due to their rate-based representation of signals. Even in the case of artificial Spiking Neural Networks (SNNs), identifying applications where temporal coding outperforms the rate coding strategies of ANNs is still an open challenge. Neuromorphic sensory-processing systems provide an ideal context for exploring the potential advantages of temporal coding, as they are able to efficiently extract the information required to cluster or classify spatio-temporal activity patterns from relative spike timing. Here we propose a neuromorphic model inspired by the sand scorpion to explore the benefits of temporal coding, and validate it in an event-based sensory-processing task. The task consists in localizing a target using only the relative spike timing of eight spatially-separated vibration sensors. We propose two different approaches in which the SNNs learns to cluster spatio-temporal patterns in an unsupervised manner and we demonstrate how the task can be solved both analytically and through numerical simulation of multiple SNN models. We argue that the models presented are optimal for spatio-temporal pattern classification using precise spike timing in a task that could be used as a standard benchmark for evaluating event-based sensory processing models based on temporal coding.

10.
Front Neurosci ; 14: 275, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-32327968

RESUMO

In this paper, we introduce a framework for dynamic gesture recognition with background suppression operating on the output of a moving event-based camera. The system is developed to operate in real-time using only the computational capabilities of a mobile phone. It introduces a new development around the concept of time-surfaces. It also presents a novel event-based methodology to dynamically remove backgrounds that uses the high temporal resolution properties of event-based cameras. To our knowledge, this is the first Android event-based framework for vision-based recognition of dynamic gestures running on a smartphone without off-board processing. We assess the performances by considering several scenarios in both indoors and outdoors, for static and dynamic conditions, in uncontrolled lighting conditions. We also introduce a new event-based dataset for gesture recognition with static and dynamic backgrounds (made publicly available). The set of gestures has been selected following a clinical trial to allow human-machine interaction for the visually impaired and older adults. We finally report comparisons with prior work that addressed event-based gesture recognition reporting comparable results, without the use of advanced classification techniques nor power greedy hardware.

11.
Nat Biomed Eng ; 4(2): 172-180, 2020 02.
Artigo em Inglês | MEDLINE | ID: mdl-31792423

RESUMO

Retinal dystrophies and age-related macular degeneration related to photoreceptor degeneration can cause blindness. In blind patients, although the electrical activation of the residual retinal circuit can provide useful artificial visual perception, the resolutions of current retinal prostheses have been limited either by large electrodes or small numbers of pixels. Here we report the evaluation, in three awake non-human primates, of a previously reported near-infrared-light-sensitive photovoltaic subretinal prosthesis. We show that multipixel stimulation of the prosthesis within radiation safety limits enabled eye tracking in the animals, that they responded to stimulations directed at the implant with repeated saccades and that the implant-induced responses were present two years after device implantation. Our findings pave the way for the clinical evaluation of the prosthesis in patients affected by dry atrophic age-related macular degeneration.


Assuntos
Degeneração Macular/reabilitação , Movimentos Sacádicos , Visão Ocular/fisiologia , Percepção Visual , Próteses Visuais , Animais , Modelos Animais de Doenças , Medições dos Movimentos Oculares , Macaca fascicularis , Degeneração Macular/fisiopatologia , Masculino , Estimulação Luminosa , Células Ganglionares da Retina/fisiologia
12.
Front Neurosci ; 13: 827, 2019.
Artigo em Inglês | MEDLINE | ID: mdl-31496927

RESUMO

Most dynamic systems are controlled by discrete time controllers. One of the main challenges faced during the design of a digital control law is the selection of the appropriate sampling time. A small sampling time will increase the accuracy of the controlled output at the expense of heavy computations. In contrast, a large sampling time will decrease the computational power needed to update the control law at the expense of a smaller stability region. In addition, once the setpoint is reached, the controlled input is still updated, making the overall controlled system not energetically efficient. To be more efficient, one can update the control law based on a significant fixed change of the controlled signal (send-on-delta or event-based controller). Like for time-based discretization, the amplitude of the significant change must be chosen carefully to avoid oscillations around the setpoint (e.g., if the setpoint is in between two samples) or an unnecessary increase of the samples number needed to reach the setpoint with a given accuracy. This paper proposes a novel non-linear event-based discretization method based on inter-events duration. We demonstrate that our new method reaches an arbitrary accuracy independently of the setpoint amplitude without increasing the network data transmission bandwidth. The method decreases the overall number of samples needed to estimate the states of a dynamical system and the update rate of an actuator, making it more energetically efficient.

13.
Neural Comput ; 31(6): 1114-1138, 2019 06.
Artigo em Inglês | MEDLINE | ID: mdl-30979350

RESUMO

In this work, we propose a two-layered descriptive model for motion processing from retina to the cortex, with an event-based input from the asynchronous time-based image sensor (ATIS) camera. Spatial and spatiotemporal filtering of visual scenes by motion energy detectors has been implemented in two steps in a simple layer of a lateral geniculate nucleus model and a set of three-dimensional Gabor kernels, eventually forming a probabilistic population response. The high temporal resolution of independent and asynchronous local sensory pixels from the ATIS provides a realistic stimulation to study biological motion processing, as well as developing bio-inspired motion processors for computer vision applications. Our study combines two significant theories in neuroscience: event-based stimulation and probabilistic sensory representation. We have modeled how this might be done at the vision level, as well as suggesting this framework as a generic computational principle among different sensory modalities.


Assuntos
Modelos Neurológicos , Percepção de Movimento , Estimulação Luminosa/métodos , Córtex Visual , Humanos , Percepção de Movimento/fisiologia , Probabilidade , Retina/fisiologia , Visão Ocular/fisiologia , Córtex Visual/fisiologia
14.
Sci Rep ; 9(1): 3744, 2019 03 06.
Artigo em Inglês | MEDLINE | ID: mdl-30842458

RESUMO

Depth from defocus is an important mechanism that enables vision systems to perceive depth. While machine vision has developed several algorithms to estimate depth from the amount of defocus present at the focal plane, existing techniques are slow, energy demanding and mainly relying on numerous acquisitions and massive amounts of filtering operations on the pixels' absolute luminance value. Recent advances in neuromorphic engineering allow an alternative to this problem, with the use of event-based silicon retinas and neural processing devices inspired by the organizing principles of the brain. In this paper, we present a low power, compact and computationally inexpensive setup to estimate depth in a 3D scene in real time at high rates that can be directly implemented with massively parallel, compact, low-latency and low-power neuromorphic engineering devices. Exploiting the high temporal resolution of the event-based silicon retina, we are able to extract depth at 100 Hz for a power budget lower than a 200 mW (10 mW for the camera, 90 mW for the liquid lens and ~100 mW for the computation). We validate the model with experimental results, highlighting features that are consistent with both computational neuroscience and recent findings in the retina physiology. We demonstrate its efficiency with a prototype of a neuromorphic hardware system and provide testable predictions on the role of spike-based representations and temporal dynamics in biological depth from defocus experiments reported in the literature.


Assuntos
Percepção de Profundidade/fisiologia , Processamento de Imagem Assistida por Computador/métodos , Visão Ocular/fisiologia , Potenciais de Ação/fisiologia , Algoritmos , Encéfalo/fisiologia , Computadores , Modelos Neurológicos , Redes Neurais de Computação , Neurônios/fisiologia , Retina/fisiologia
15.
Front Neurosci ; 13: 1338, 2019.
Artigo em Inglês | MEDLINE | ID: mdl-31969799

RESUMO

This paper introduces an new open-source, header-only and modular C++ framework to facilitate the implementation of event-driven algorithms. The framework relies on three independent components: sepia (file IO), tarsier (algorithms), and chameleon (display). Our benchmarks show that algorithms implemented with tarsier are faster and have a lower latency than identical implementations in other state-of-the-art frameworks, thanks to static polymorphism (compile-time pipeline assembly). The observer pattern used throughout the framework encourages implementations that better reflect the event-driven nature of the algorithms and the way they process events, easing future translation to neuromorphic hardware. The framework integrates drivers to communicate with the DVS, the DAVIS, the Opal Kelly ATIS, and the CCam ATIS.

16.
J Vis ; 18(11): 6, 2018 10 01.
Artigo em Inglês | MEDLINE | ID: mdl-30347093

RESUMO

The optic quality of the eyes is, at least in part, determined by pupil size. Large pupils let more light enter the eyes, but degrade the point spread function, and thus the spatial resolution that can be achieved (Campbell & Gregory, 1960). In natural conditions, the pupil is mainly driven by the luminance (and possibly the color and contrast) at the gazed location, but is also modulated by attention and cognitive factors. Whether changes in eyes' optics related to pupil size modulation by luminance and attention impacts visual processing was assessed in two experiments. In Experiment 1, we measured pupil size using a constantly visible display made of four disks with different luminance levels, with no other task than fixating the disks in succession. The results confirmed that pupil size depends on the luminance of the gazed stimulus. Experiment 2, using similar settings as Experiment 1, used a two-interval forced-choice design to test whether discriminating high spatial frequencies that requires covert attention to parafoveal stimuli is better during the fixation of bright disks that entails a small pupil size, and hence better eyes' optics, as compared to fixating dark disks that entails a large pupil size, and hence poorer eyes' optics. As in Experiment 1, we observed large modulations of pupil size depending on the luminance of the gazed stimulus, but pupil dynamics was more variable, with marked pupil dilation during stimulus encoding, presumably because the demanding spatial frequency discrimination task engaged attention. However, discrimination performance and mean pupil size were not correlated. Despite this lack of correlation, the slopes of pupil dilation during stimulus encoding were correlated to performance, while the slopes of pupil dilation during decision-making were not. We discuss these results regarding the possible functional roles of pupil size modulations.


Assuntos
Atenção/fisiologia , Luz , Pupila/fisiologia , Reflexo Pupilar/efeitos da radiação , Processamento Espacial/fisiologia , Adulto , Feminino , Humanos , Masculino , Óptica e Fotônica , Visão Ocular , Adulto Jovem
17.
IEEE Trans Biomed Circuits Syst ; 12(6): 1467-1474, 2018 12.
Artigo em Inglês | MEDLINE | ID: mdl-30334806

RESUMO

Johnson-Nyquist noise is the electronic noise generated by the thermal agitation of charge carriers, which increases when the sensor overheats. Current high-speed cameras used in low-light conditions are often cooled down to reduce thermal noise and increase their signal to noise ratio. These sensors, however, record hundreds of frames per second, which takes time, requires energy, and heavy computing power due to the substantial data load. Event-based sensors benefit from a high temporal resolution and record the information in a sparse manner. Based on an asynchronous time-based image sensor, we developed another version of this event-based camera whose pixels were designed for low-light applications and added a Peltier-effect-based cooling system at the back of the sensor in order to reduce thermal noise. We show the benefits from thermal noise reduction and study the improvement of the signal to noise ratio in the estimation of event-based normal flow norm and angle and particle tracking in microscopy.


Assuntos
Algoritmos , Temperatura Baixa , Processamento de Imagem Assistida por Computador/instrumentação , Processamento de Imagem Assistida por Computador/métodos , Razão Sinal-Ruído , Desenho de Equipamento , Microscopia
18.
Ann Intensive Care ; 8(1): 89, 2018 Sep 17.
Artigo em Inglês | MEDLINE | ID: mdl-30225814

RESUMO

BACKGROUND: Closed-loop resuscitation can improve personalization of care, decrease workload and bring expert knowledge in isolated areas. We have developed a new device to control the administration of fluid or simultaneous co-administration of fluid and norepinephrine using arterial pressure. METHOD: We evaluated the performance of our prototype in a rodent model of haemorrhagic shock. After haemorrhagic shock, rats were randomized to five experimental groups: three were resuscitated with fluid and two with co-administration of fluid and norepinephrine. Among groups resuscitated with fluid, one was resuscitated by a physician and two were resuscitated according to two different closed-loop algorithms. Among groups resuscitated with fluid and norepinephrine, one was resuscitated by a physician and the other one by the closed-loop device. The precision of arterial pressure during the resuscitation period was assessed using rising time, time passed in the target area and performance error calculations. RESULTS: Groups resuscitated with fluid had similar performances and passed as much time in the target area of 80-90 mmHg as the manual group [manual: 76.8% (67.9-78.2), closed-loop: 64.6% (45.7-72.9) and 80.9% (59.1-85.3)]. Rats resuscitated with fluid and norepinephrine using closed-loop passed similar time in target area than manual group [closed-loop: 74.4% (58.4-84.5) vs. manual: 60.1% (46.1-72.4)] but had shorter rising time to reach target area [160 s (106-187) vs. 434 s (254-1081)] than those resuscitated by a physician. Rats resuscitated with co-administration of fluid and norepinephrine required less fluid and had less hemodilution than rats resuscitated with fluid alone. Lactate decrease was similar between groups resuscitated with fluid alone and fluid with norepinephrine. CONCLUSIONS: This study assessed extensively the performances of several algorithms for closed-loop resuscitation of haemorrhagic shock with fluid alone and with co-administration of fluid and norepinephrine. The performance of the closed-loop algorithms tested was similar to physician-guided treatment with considerable saving of work for the caregiver. Arterial pressure closed-loop guided algorithms can be extended to combined administration of fluid and norepinephrine.

19.
Artigo em Inglês | MEDLINE | ID: mdl-30222585

RESUMO

This paper introduces an event-based luminance-free algorithm for line and segment detection from the output of asynchronous event-based neuromorphic retinas. These recent biomimetic vision sensors are composed of autonomous pixels, each of them asynchronously generating visual events that encode relative changes in pixels' illumination at high temporal resolutions. This frame-free approach results in an increased energy efficiency and in real-time operation, making these sensors especially suitable for applications such as autonomous robotics. The proposed algorithm is based on an iterative event-based weighted least squares fitting, and it is consequently well suited to the high temporal resolution and asynchronous acquisition of neuromorphic cameras: parameters of a current line are updated for each event attributed (i.e., spatio-temporally close) to it, while implicitly forgetting the contribution of older events according to a speed-tuned exponentially decaying function. A detection occurs if a measure of activity, i.e., implicit measure of the number of contributing events and using the same decay function, exceeds a given threshold. The speed-tuned decreasing function is based on a measure of the apparent motion, i.e., the optical flow computed around each event. This latter ensures that the algorithm behaves independently of the edges' dynamics. Line segments are then extracted from the lines, allowing for the tracking of the corresponding endpoints. We provide experiments showing the accuracy of our algorithm and study the influence of the apparent velocity and relative orientation of the observed edges. Finally, evaluations of its computational efficiency show that this algorithm can be envisioned for high-speed applications, such as vision-based robotic navigation.

20.
Front Neurosci ; 12: 442, 2018.
Artigo em Inglês | MEDLINE | ID: mdl-30013461

RESUMO

3D reconstruction from multiple viewpoints is an important problem in machine vision that allows recovering tridimensional structures from multiple two-dimensional views of a given scene. Reconstructions from multiple views are conventionally achieved through a process of pixel luminance-based matching between different views. Unlike conventional machine vision methods that solve matching ambiguities by operating only on spatial constraints and luminance, this paper introduces a fully time-based solution to stereovision using the high temporal resolution of neuromorphic asynchronous event-based cameras. These cameras output dynamic visual information in the form of what is known as "change events" that encode the time, the location and the sign of the luminance changes. A more advanced event-based camera, the Asynchronous Time-based Image Sensor (ATIS), in addition of change events, encodes absolute luminance as time differences. The stereovision problem can then be formulated solely in the time domain as a problem of events coincidences detection problem. This work is improving existing event-based stereovision techniques by adding luminance information that increases the matching reliability. It also introduces a formulation that does not require to build local frames (though it is still possible) from the luminances which can be costly to implement. Finally, this work also introduces a methodology for time based stereovision in the context of binocular and trinocular configurations using time based event matching criterion combining for the first time all together: space, time, luminance, and motion.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...