Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 167
Filtrar
1.
Artigo em Inglês | MEDLINE | ID: mdl-39321002

RESUMO

Smoothing filters are widely used in EEG signal processing for noise removal while preserving signals' features. Inspired by our recent work on Upscale and Downscale Representation (UDR), this paper proposes a cascade arrangement of some effective image-processing techniques for signal filtering in the image domain. The UDR concept is to visualize EEG signals at an appropriate line width and convert it to a binary image. The smoothing process is then conducted by skeletonizing the signal object to a unit width and projecting it back to the time domain. Two successive UDRs could result in a better-smoothing performance, but their binary image conversion should be restricted. The process is computationally ineffective, especially at higher line width values. Cascaded Thinning UDR (CTUDR) is proposed, exploiting morphological operations to perform a two-stage upscale and downscale within one binary image representation. CTUDR is verified on a signal smoothing and classification task and compared with conventional techniques, such as the Moving Average, the Binomial, the Median, and the Savitzky Golay filters. Simulated EEG data with added white Gaussian noise is employed in the former, while cognitive conflict data obtained from a 3D object selection task is utilized in the latter. CTUDR outperforms its counterparts, scoring the best fitting error and correlation coefficient in signal smoothing while achieving the highest gain in Accuracy (0.7640%) and F-measure (0.7607%) when used as a smoothing filter for training data of EEGNet.


Assuntos
Algoritmos , Eletroencefalografia , Processamento de Sinais Assistido por Computador , Eletroencefalografia/métodos , Humanos , Razão Sinal-Ruído , Distribuição Normal , Simulação por Computador , Processamento de Imagem Assistida por Computador/métodos
2.
Spinal Cord ; 2024 Sep 12.
Artigo em Inglês | MEDLINE | ID: mdl-39266672

RESUMO

STUDY DESIGN: Randomised controlled trial. OBJECTIVES: The objective is to describe an electroencephalography (EEG) neurofeedback intervention that will be provided in a randomised controlled trial for people with neuropathic pain following spinal cord injury (SCI): the StoPain Trial. In this trial, participants in the treatment group will implement an EEG neurofeedback system as an analgesic intervention at home, while participants in the control group will continue with the treatments available to them in the community. SETTING: University-based study in Sydney, Australia. METHODS/RESULTS: This manuscript describes the rationale and components of the EEG neurofeedback intervention designed for individuals with SCI neuropathic pain and intended for home-based implementation. Our report is based on the criteria of the Template for Intervention Description and Replication (TIDieR) checklist, and includes why the efficacy of EEG neurofeedback will be investigated, what will be provided, who will administer it, and how, where, when, and how much the EEG neurofeedback intervention will be administered. CONCLUSIONS: This manuscript provides a detailed description of a complex intervention used in a randomised controlled trial. This description will facilitate the subsequent interpretation of the trial results and allow for the replication of the intervention in clinical practice and future trials. SPONSORSHIP: Australian Government Medical Research Future Fund (2020 Rare Cancers Rare Diseases and Unmet Needs Scheme: 2006020).

3.
Artigo em Inglês | MEDLINE | ID: mdl-39190511

RESUMO

Decoding natural language from noninvasive brain signals has been an exciting topic with the potential to expand the applications of brain-computer interface (BCI) systems. However, current methods face limitations in decoding sentences from electroencephalography (EEG) signals. Improving decoding performance requires the development of a more effective encoder for the EEG modality. Nonetheless, learning generalizable EEG representations remains a challenge due to the relatively small scale of existing EEG datasets. In this paper, we propose enhancing the EEG encoder to improve subsequent decoding performance. Specifically, we introduce the discrete Conformer encoder (D-Conformer) to transform EEG signals into discrete representations and bootstrap the learning process by imposing EEG-language alignment from the early training stage. The D-Conformer captures both local and global patterns from EEG signals and discretizes the EEG representation, making the representation more resilient to variations, while early-stage EEG-language alignment mitigates the limitations of small EEG datasets and facilitates the learning of the semantic representations from EEG signals. These enhancements result in improved EEG representations and decoding performance. We conducted extensive experiments and ablation studies to thoroughly evaluate the proposed method. Utilizing the D-Conformer encoder and bootstrapping training strategy, our approach demonstrates superior decoding performance across various tasks, including word-level, sentence-level, and sentiment-level decoding from EEG signals. Specifically, in word-level classification, we show that our encoding method produces more distinctive representations and higher classification performance compared to the EEG encoders from existing methods. At the sentence level, our model outperformed the baseline by 5.45%, achieving a BLEU-1 score of 42.31%. Furthermore, in sentiment classification, our model exceeded the baseline by 14%, achieving a sentiment classification accuracy of 69.3%.


Assuntos
Algoritmos , Interfaces Cérebro-Computador , Eletroencefalografia , Processamento de Linguagem Natural , Humanos , Eletroencefalografia/métodos , Idioma , Semântica , Aprendizado de Máquina Supervisionado , Masculino , Feminino , Adulto , Reprodutibilidade dos Testes
4.
Sci Rep ; 14(1): 13217, 2024 06 08.
Artigo em Inglês | MEDLINE | ID: mdl-38851836

RESUMO

Anticipating human decisions while performing complex tasks remains a formidable challenge. This study proposes a multimodal machine-learning approach that leverages image features and electroencephalography (EEG) data to predict human response correctness in a demanding visual searching task. Notably, we extract a novel set of image features pertaining to object relationships using the Segment Anything Model (SAM), which enhances prediction accuracy compared to traditional features. Additionally, our approach effectively utilizes a combination of EEG signals and image features to streamline the feature set required for the Random Forest Classifier (RFC) while maintaining high accuracy. The findings of this research hold substantial potential for developing advanced fault alert systems, particularly in critical decision-making environments such as the medical and defence sectors.


Assuntos
Tomada de Decisões , Eletroencefalografia , Aprendizado de Máquina , Humanos , Masculino , Feminino , Adulto , Adulto Jovem , Algoritmos
5.
IEEE Open J Eng Med Biol ; 5: 180-190, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38606398

RESUMO

A significant issue for traffic safety has been drowsy driving for decades. A number of studies have investigated the effects of acute fatigue on spectral power; and recent research has revealed that drowsy driving is associated with a variety of brain connections in a specific cortico-cortical pathway. In spite of this, it is still unclear how different brain regions are connected in drowsy driving at different levels of daily fatigue. This study identified the brain connectivity-behavior relationship among three different daily fatigue levels (low-, median- and high-fatigue) with the EEG data transfer entropy. According to the results, only low- and medium-fatigue groups demonstrated an inverted U-shaped change in connectivity from high performance to poor behavioral performance. In addition, from low- to high-fatigue groups, connectivity magnitude decreased in the frontal region and increased in the occipital region. These study results suggest that brain connectivity and driving behavior would be affected by different levels of daily fatigue.

6.
PLoS One ; 19(4): e0301052, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-38630669

RESUMO

Stress is a prevalent bodily response universally experienced and significantly affects a person's mental and cognitive state. The P300 response is a commonly observed brain behaviour that provides insight into a person's cognitive state. Previous works have documented the effects of stress on the P300 behaviour; however, only a few have explored the performance in a mobile and naturalistic experimental setup. Our study examined the effects of stress on the human brain's P300 behaviour through a height exposure experiment that incorporates complex visual, vestibular, and proprioceptive stimuli. A more complex sensory environment could produce translatable findings toward real-world behaviour and benefit emerging technologies such as brain-computer interfaces. Seventeen participants experienced our experiment that elicited the stress response through physical and virtual height exposure. We found two unique groups within our participants that exhibited contrasting behavioural performance and P300 target reaction response when exposed to stressors (from walking at heights). One group performed worse when exposed to heights and exhibited a significant decrease in parietal P300 peak amplitude and increased beta and gamma power. On the other hand, the group less affected by stress exhibited a change in their N170 peak amplitude and alpha/mu rhythm desynchronisation. The findings of our study suggest that a more individualised approach to assessing a person's behaviour performance under stress can aid in understanding P300 performance when experiencing stress.


Assuntos
Encéfalo , Potenciais Evocados P300 , Humanos , Potenciais Evocados P300/fisiologia , Encéfalo/fisiologia , Simulação por Computador , Ritmo alfa , Cabeça , Eletroencefalografia
7.
J Neural Eng ; 21(2)2024 Mar 19.
Artigo em Inglês | MEDLINE | ID: mdl-38295415

RESUMO

Objective. Brain-computer interface (BCI) technology is poised to play a prominent role in modern work environments, especially a collaborative environment where humans and machines work in close proximity, often with physical contact. In a physical human robot collaboration (pHRC), the robot performs complex motion sequences. Any unexpected robot behavior or faulty interaction might raise safety concerns. Error-related potentials, naturally generated by the brain when a human partner perceives an error, have been extensively employed in BCI as implicit human feedback to adapt robot behavior to facilitate a safe and intuitive interaction. However, the integration of BCI technology with error-related potential for robot control demands failure-free integration of highly uncertain electroencephalography (EEG) signals, particularly influenced by the physical and cognitive state of the user. As a higher workload on the user compromises their access to cognitive resources needed for error awareness, it is crucial to study how mental workload variations impact the error awareness as it might raise safety concerns in pHRC. In this study, we aim to study how cognitive workload affects the error awareness of a human user engaged in a pHRC.Approach. We designed a blasting task with an abrasive industrial robot and manipulated the mental workload with a secondary arithmetic task of varying difficulty. EEG data, perceived workload, task and physical performance were recorded from 24 participants moving the robot arm. The error condition was achieved by the unexpected stopping of the robot in 33% of trials.Main results. We observed a diminished amplitude for the prediction error negativity (PEN) and error positivity (Pe), indicating reduced error awareness with increasing mental workload. We further observed an increased frontal theta power and increasing trend in the central alpha and central beta power after the unexpected robot stopping compared to when the robot stopped correctly at the target. We also demonstrate that a popular convolution neural network model, EEGNet, could predict the amplitudes of PEN and Pe from the EEG data prior to the error.Significance. This prediction model could be instrumental in developing an online prediction model that could forewarn the system and operators of the diminished error awareness of the user, alluding to a potential safety breach in error-related potential-based BCI system for pHRC. Therefore, our work paves the way for embracing BCI technology in pHRC to optimally adapt the robot behavior for personalized user experience using real-time brain activity, enriching the quality of the interaction.


Assuntos
Interfaces Cérebro-Computador , Robótica , Humanos , Carga de Trabalho/psicologia , Eletroencefalografia/métodos , Cognição
8.
Artigo em Inglês | MEDLINE | ID: mdl-38010936

RESUMO

Drowsy driving is one of the primary causes of driving fatalities. Electroencephalography (EEG), a method for detecting drowsiness directly from brain activity, has been widely used for detecting driver drowsiness in real-time. Recent studies have revealed the great potential of using brain connectivity graphs constructed based on EEG data for drowsy state predictions. However, traditional brain connectivity networks are irrelevant to the downstream prediction tasks. This article proposes a connectivity-aware graph neural network (CAGNN) using a self-attention mechanism that can generate task-relevant connectivity networks via end-to-end training. Our method achieved an accuracy of 72.6% and outperformed other convolutional neural networks (CNNs) and graph generation methods based on a drowsy driving dataset. In addition, we introduced a squeeze-and-excitation (SE) block to capture important features and demonstrated that the SE attention score can reveal the most important feature band. We compared our generated connectivity graphs in the drowsy and alert states and found drowsiness connectivity patterns, including significantly reduced occipital connectivity and interregional connectivity. Additionally, we performed a post hoc interpretability analysis and found that our method could identify drowsiness features such as alpha spindles. Our code is available online at https://github.com/ALEX95GOGO/CAGNN.


Assuntos
Condução de Veículo , Humanos , Eletroencefalografia/métodos , Encéfalo , Redes Neurais de Computação , Vigília
9.
IEEE Trans Cybern ; 54(5): 3275-3285, 2024 May.
Artigo em Inglês | MEDLINE | ID: mdl-37027534

RESUMO

Fuzzy neural networks (FNNs) have been very successful at handling uncertainty in data using fuzzy mappings and if-then rules. However, they suffer from generalization and dimensionality issues. Although deep neural networks (DNNs) represent a step toward processing high-dimensional data, their capacity to address data uncertainty is limited. Furthermore, deep learning algorithms designed to improve robustness are either time consuming or yield unsatisfactory performance. In this article, we propose a robust fuzzy neural network (RFNN) to overcome these problems. The network contains an adaptive inference engine that is capable of handling samples with high-level uncertainty and high dimensions. Unlike traditional FNNs that use a fuzzy AND operation to calculate the firing strength for each rule, our inference engine is able to learn the firing strength adaptively. It also further processes the uncertainty in membership function values. Taking advantage of the learning ability of neural networks, the acquired fuzzy sets can be learned from training inputs automatically to cover the input space well. Furthermore, the consequent layer uses neural network structures to enhance the reasoning ability of the fuzzy rules when dealing with complex inputs. Experiments on a range of datasets show that RFNN delivers state-of-the-art accuracy even at very high levels of uncertainty. Our code is available online. https://github.com/leijiezhang/RFNN.

10.
Artigo em Inglês | MEDLINE | ID: mdl-38082585

RESUMO

Detecting concealed objects presents a significant challenge for human and artificial intelligent systems. Detecting concealed objects task necessitates a high level of human attention and cognitive effort to complete the task successfully. Thus, in this study, we use concealed objects as stimuli for our decision-making experimental paradigms to quantify participants' decision-making performance. We applied a deep learning model, Bi-directional Long Short Term Memory (BiLSTM), to predict the participant's decision accuracy by using their electroencephalogram (EEG) signals as input. The classifier model demonstrated high accuracy, reaching 96.1% with an epoching time range of 500 ms following the stimulus event onset. The results revealed that the parietal-occipital brain region provides highly informative information for the classifier in the concealed visual searching tasks. Furthermore, the neural mechanism underlying the concealed visual-searching and decision-making process was explained by analyzing serial EEG components. The findings of this study could contribute to the development of a fault alert system, which has the potential to improve human decision-making performance.


Assuntos
Encéfalo , Eletroencefalografia , Humanos , Inteligência Artificial , Mapeamento Encefálico , Atenção
11.
Artigo em Inglês | MEDLINE | ID: mdl-38082701

RESUMO

Situational awareness (SA) is vital for understanding our surroundings. Multiple variables, including inattentive blindness (IB), contribute to the deterioration of SA, which may have detrimental effects on individuals' cognitive performance. IB occurs due to attentional limitations, ignoring critical information and resulting in a loss of SA and a decline in general performance, particularly in complicated situations requiring substantial cognitive resources. To the best of our knowledge, however, past research has not fully uncovered the neurological characteristics of IB nor classified these characteristics in life-alike virtual situations. Therefore, the purpose of this study is to determine whether ERP dynamics in the brain may be utilised as a neural feature to predict the occurrence of IB using machine learning (ML) algorithms. In a virtual reality simulation of an IB experiment, 30 participants' behaviour and Electroencephalography (EEG) measurements were obtained. Participants were given a target detection task in the IB experiment without knowing the unattended shapes displayed on the background building. The targets were presented in three different sensory modalities (auditory, visual, and visual-auditory). On the post-experiment questionnaire, participants who claimed not to have noticed the unattended shapes were assigned to the IB group. Subsequently, the Aware group was formed from individuals who reported seeing the unattended shapes. Using EEGNet to classify IB and Aware groups demonstrated a high classification performance. According to the research, ERP brain dynamics are associated with the awareness of unattended shapes and have the potential to serve as a reliable indication for predicting the visual consciousness of unexpected objects.(p/)(p)Clinical relevance- This research offers a potential brain marker for the mixed-reality and BCI systems that will be used in the future to identify cognitive deterioration, maintain attentional capacity, and prevent disasters.


Assuntos
Atenção , Encéfalo , Humanos , Cognição , Potenciais Evocados , Cegueira
12.
Artigo em Inglês | MEDLINE | ID: mdl-38083669

RESUMO

Object recognition is a complex cognitive process in which information is integrated and processed by various brain regions. Previous studies have shown that both the visual and temporal cortices are active during object recognition and identification. However, although object recognition and object identification are similar, these processes are considered distinct functions in the brain. Despite this, the differentiation between object recognition and identification has yet to be clearly defined for use in brain-computer interface (BCI) applications. This research aims to utilize neural features related to object recognition and identification and classify these features to differentiate between the two processes. The results demonstrate that several classifiers achieved high levels of accuracy, with the XGBoost classifier using a Linear Booster achieving the highest accuracy at 96% and a F1 score of 0.97. This ability to distinguish between object recognition and identification can be a beneficial aspect of a BCI object recognition system as it could help determine the intended target object for a user.


Assuntos
Interfaces Cérebro-Computador , Eletroencefalografia , Encéfalo , Percepção Visual
13.
Artigo em Inglês | MEDLINE | ID: mdl-38051624

RESUMO

Object recognition and object identification are multifaceted cognitive operations that require various brain regions to synthesize and process information. Prior research has evidenced the activity of both visual and temporal cortices during these tasks. Notwithstanding their similarities, object recognition and identification are recognized as separate brain functions. Drawing from the two-stream hypothesis, our investigation aims to understand whether the channels within the ventral and dorsal streams contain pertinent information for effective model learning regarding object recognition and identification tasks. By utilizing the data we collected during the object recognition and identification experiment, we scrutinized EEGNet models, trained using channels that replicate the two-stream hypothesis pathways, against a model trained using all available channels. The outcomes reveal that the model trained solely using the temporal region delivered a high accuracy level in classifying four distinct object categories. Specifically, the object recognition and object identification models achieved an accuracy of 89% and 85%, respectively. By incorporating the channels that mimic the ventral stream, the model's accuracy was further improved, with the object recognition model and object identification model achieving an accuracy of 95% and 94%, respectively. Furthermore, the Grad-CAM result of the trained models revealed a significant contribution from the ventral and dorsal stream channels toward the training of the EEGNet model. The aim of our study is to pinpoint the optimal channel configuration that provides a swift and accurate brain-computer interface system for object recognition and identification.


Assuntos
Reconhecimento Visual de Modelos , Percepção Visual , Humanos , Encéfalo , Lobo Temporal , Imageamento por Ressonância Magnética , Eletroencefalografia , Mapeamento Encefálico
14.
Artigo em Inglês | MEDLINE | ID: mdl-37938954

RESUMO

Deep-learning models have been widely used in image recognition tasks due to their strong feature-learning ability. However, most of the current deep-learning models are "black box" systems that lack a semantic explanation of how they reached their conclusions. This makes it difficult to apply these methods to complex medical image recognition tasks. The vision transformer (ViT) model is the most commonly used deep-learning model with a self-attention mechanism that shows the region of influence as compared to traditional convolutional networks. Thus, ViT offers greater interpretability. However, medical images often contain lesions of variable size in different locations, which makes it difficult for a deep-learning model with a self-attention module to reach correct and explainable conclusions. We propose a multigranularity random walk transformer (MGRW-Transformer) model guided by an attention mechanism to find the regions that influence the recognition task. Our method divides the image into multiple subimage blocks and transfers them to the ViT module for classification. Simultaneously, the attention matrix output from the multiattention layer is fused with the multigranularity random walk module. Within the multigranularity random walk module, the segmented image blocks are used as nodes to construct an undirected graph using the attention node as a starting node and guiding the coarse-grained random walk. We appropriately divide the coarse blocks into finer ones to manage the computational cost and combine the results based on the importance of the discovered features. The result is that the model offers a semantic interpretation of the input image, a visualization of the interpretation, and insight into how the decision was reached. Experimental results show that our method improves classification performance with medical images while presenting an understandable interpretation for use by medical professionals.

15.
PLoS One ; 18(10): e0290431, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37878584

RESUMO

Wearable smart glasses are an emerging technology gaining popularity in the assistive technologies industry. Smart glasses aids typically leverage computer vision and other sensory information to translate the wearer's surrounding into computer-synthesized speech. In this work, we explored the potential of a new technique known as "acoustic touch" to provide a wearable spatial audio solution for assisting people who are blind in finding objects. In contrast to traditional systems, this technique uses smart glasses to sonify objects into distinct sound auditory icons when the object enters the device's field of view. We developed a wearable Foveated Audio Device to study the efficacy and usability of using acoustic touch to search, memorize, and reach items. Our evaluation study involved 14 participants, 7 blind or low-visioned and 7 blindfolded sighted (as a control group) participants. We compared the wearable device to two idealized conditions, a verbal clock face description and a sequential audio presentation through external speakers. We found that the wearable device can effectively aid the recognition and reaching of an object. We also observed that the device does not significantly increase the user's cognitive workload. These promising results suggest that acoustic touch can provide a wearable and effective method of sensory augmentation.


Assuntos
Acústica , Percepção do Tato , Humanos , Cegueira , Fala , Visão Ocular
16.
IEEE Trans Pattern Anal Mach Intell ; 45(12): 14709-14726, 2023 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-37651495

RESUMO

Information can be quantified and expressed by uncertainty, and improving the decision level of uncertain information is vital in modeling and processing uncertain information. Dempster-Shafer evidence theory can model and process uncertain information effectively. However, the Dempster combination rule may provide counter-intuitive results when dealing with highly conflicting information, leading to a decline in decision level. Thus, measuring conflict is significant in the improvement of decision level. Motivated by this issue, this paper proposes a novel method to measure the discrepancy between bodies of evidence. First, the model of dynamic fractal probability transformation is proposed to effectively obtain more information about the non-specificity of basic belief assignments (BBAs). Then, we propose the higher-order fractal belief Rényi divergence (HOFBReD). HOFBReD can effectively measure the discrepancy between BBAs. Moreover, it is the first belief Rényi divergence that can measure the discrepancy between BBAs with dynamic fractal probability transformation. HoFBReD has several properties in terms of probability transformation as well as measurement. When the dynamic fractal probability transformation ends, HoFBReD is equivalent to measuring the Rényi divergence between the pignistic probability transformations of BBAs. When the BBAs degenerate to the probability distributions, HoFBReD will also degenerate to or be related to several well-known divergences. In addition, based on HoFBReD, a novel multisource information fusion algorithm is proposed. A pattern classification experiment with real-world datasets is presented to compare the proposed algorithm with other methods. The experiment results indicate that the proposed algorithm has a higher average pattern recognition accuracy with all datasets than other methods. The proposed discrepancy measurement method and multisource information algorithm contribute to the improvement of decision level.

17.
Med Biol Eng Comput ; 61(11): 3003-3019, 2023 Nov.
Artigo em Inglês | MEDLINE | ID: mdl-37563528

RESUMO

Brain-computer interfaces (BCIs) allow communication between the brain and the external world. This type of technology has been extensively studied. However, BCI instruments with high signal quality are typically heavy and large. Thus, recording electroencephalography (EEG) signals is an inconvenient task. In recent years, system-on-chip (SoC) approaches have been integrated into BCI research, and sensors for wireless portable devices have been developed; however, there is still considerable work to be done. As neuroscience research has advanced, EEG signal analyses have come to require more accurate data. Due to the limited bandwidth of Bluetooth wireless transmission technology, EEG measurement systems with more than 16 channels must be used to reduce the sampling rate and prevent data loss. Therefore, the goal of this study was to develop a multichannel, high-resolution (24-bit), high-sampling-rate EEG BCI device that transmits signals via Wi-Fi. We believe that this system can be used in neuroscience research. The EEG acquisition system proposed in this work is based on a Cortex-M4 microcontroller with a Wi-Fi subsystem, providing a multichannel design and improved signal quality. This system is compatible with wet sensors, Ag/AgCl electrodes, and dry sensors. A LabVIEW-based user interface receives EEG data via Wi-Fi transmission and saves the raw data for offline analysis. In previous cognitive experiments, event tags have been communicated using Recommended Standard 232 (RS-232). The developed system was validated through event-related potential (ERP) and steady-state visually evoked potential (SSVEP) experiments. Our experimental results demonstrate that this system is suitable for recording EEG measurements and has potential in practical applications. The advantages of the developed system include its high sampling rate and high amplification. Additionally, in the future, Internet of Things (IoT) technology can be integrated into the system for remote real-time analysis or edge computing.


Assuntos
Interfaces Cérebro-Computador , Dispositivos Eletrônicos Vestíveis , Eletroencefalografia/métodos , Potenciais Evocados , Córtex Cerebral , Potenciais Evocados Visuais
18.
Artigo em Inglês | MEDLINE | ID: mdl-37220054

RESUMO

Federated learning is an emerging learning paradigm where multiple clients collaboratively train a machine learning model in a privacy-preserving manner. Personalized federated learning extends this paradigm to overcome heterogeneity across clients by learning personalized models. Recently, there have been some initial attempts to apply transformers to federated learning. However, the impacts of federated learning algorithms on self-attention have not yet been studied. In this article, we investigate this relationship and reveal that federated averaging (FedAvg) algorithms actually have a negative impact on self-attention in cases of data heterogeneity, which limits the capabilities of the transformer model in federated learning settings. To address this issue, we propose FedTP, a novel transformer-based federated learning framework that learns personalized self-attention for each client while aggregating the other parameters among the clients. Instead of using a vanilla personalization mechanism that maintains personalized self-attention layers of each client locally, we develop a learn-to-personalize mechanism to further encourage the cooperation among clients and to increase the scalability and generalization of FedTP. Specifically, we achieve this by learning a hypernetwork on the server that outputs the personalized projection matrices of self-attention layers to generate clientwise queries, keys, and values. Furthermore, we present the generalization bound for FedTP with the learn-to-personalize mechanism. Extensive experiments verify that FedTP with the learn-to-personalize mechanism yields state-of-the-art performance in the non-IID scenarios. Our code is available online https://github.com/zhyczy/FedTP.

19.
Artigo em Inglês | MEDLINE | ID: mdl-37145943

RESUMO

Research and development of electroencephalogram (EEG) based brain-computer interfaces (BCIs) have advanced rapidly, partly due to deeper understanding of the brain and wide adoption of sophisticated machine learning approaches for decoding the EEG signals. However, recent studies have shown that machine learning algorithms are vulnerable to adversarial attacks. This paper proposes to use narrow period pulse for poisoning attack of EEG-based BCIs, which makes adversarial attacks much easier to implement. One can create dangerous backdoors in the machine learning model by injecting poisoning samples into the training set. Test samples with the backdoor key will then be classified into the target class specified by the attacker. What most distinguishes our approach from previous ones is that the backdoor key does not need to be synchronized with the EEG trials, making it very easy to implement. The effectiveness and robustness of the backdoor attack approach is demonstrated, highlighting a critical security concern for EEG-based BCIs and calling for urgent attention to address it.


Assuntos
Interfaces Cérebro-Computador , Humanos , Eletroencefalografia , Algoritmos , Aprendizado de Máquina , Encéfalo
20.
Artigo em Inglês | MEDLINE | ID: mdl-37079422

RESUMO

Identifying meaningful brain activities is critical in brain-computer interface (BCI) applications. Recently, an increasing number of neural network approaches have been proposed to recognize EEG signals. However, these approaches depend heavily on using complex network structures to improve the performance of EEG recognition and suffer from the deficit of training data. Inspired by the waveform characteristics and processing methods shared between EEG and speech signals, we propose Speech2EEG, a novel EEG recognition method that leverages pretrained speech features to improve the accuracy of EEG recognition. Specifically, a pretrained speech processing model is adapted to the EEG domain to extract multichannel temporal embeddings. Then, several aggregation methods, including the weighted average, channelwise aggregation, and channel-and-depthwise aggregation, are implemented to exploit and integrate the multichannel temporal embeddings. Finally, a classification network is used to predict EEG categories based on the integrated features. Our work is the first to explore the use of pretrained speech models for EEG signal analysis as well as the effective ways to integrate the multichannel temporal embeddings from the EEG signal. Extensive experimental results suggest that the proposed Speech2EEG method achieves state-of-the-art performance on two challenging motor imagery (MI) datasets, the BCI IV-2a and BCI IV-2b datasets, with accuracies of 89.5% and 84.07% , respectively. Visualization analysis of the multichannel temporal embeddings show that the Speech2EEG architecture can capture useful patterns related to MI categories, which can provide a novel solution for subsequent research under the constraints of a limited dataset scale.


Assuntos
Interfaces Cérebro-Computador , Fala , Humanos , Imaginação , Redes Neurais de Computação , Eletroencefalografia/métodos , Algoritmos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA