Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 2 de 2
Filtrar
Mais filtros

Base de dados
Assunto principal
Ano de publicação
Tipo de documento
Intervalo de ano de publicação
1.
Sensors (Basel) ; 24(13)2024 Jun 26.
Artigo em Inglês | MEDLINE | ID: mdl-39000940

RESUMO

With the increasing frequency and severity of disasters and accidents, there is a growing need for efficient emergency alert systems. The ultra-high definition (UHD) broadcasting service based on Advanced Television Systems Committee (ATSC) 3.0, a leading terrestrial digital broadcasting system, offers such capabilities, including a wake-up function for minimizing damage through early alerts. In case of a disaster situation, the emergency alert wake-up signal is transmitted, allowing UHD TVs to be activated, enabling individuals to receive emergency alerts and access emergency broadcasting content. However, conventional methods for detecting the bootstrap signal, essential for this function, typically require an ATSC 3.0 demodulator. In this paper, we propose a novel deep learning-based method capable of detecting an emergency wake-up signal without the need for an ATSC 3.0. The proposed method leverages deep learning techniques, specifically a deep neural network (DNN) structure for bootstrap detection and a convolutional neural network (CNN) structure for wake-up signal demodulation and to detect the bootstrap and 2 bit emergency alert wake-up signal. Specifically, our method eliminates the need for Fast Fourier Transform (FFT), frequency synchronization, and interleaving processes typically required by a demodulator. By applying a deep learning in the time domain, we simplify the detection process, allowing for the detection of an emergency alert signal without the full suite of demodulator components required for ATSC 3.0. Furthermore, we have verified the performance of the deep learning-based method using ATSC 3.0-based RF signals and a commercial Software-Defined Radio (SDR) platform in a real environment.

2.
Sensors (Basel) ; 22(15)2022 Aug 01.
Artigo em Inglês | MEDLINE | ID: mdl-35957313

RESUMO

Despite advanced machine learning methods, the implementation of emotion recognition systems based on real-world video content remains challenging. Videos may contain data such as images, audio, and text. However, the application of multimodal models using two or more types of data to real-world video media (CCTV, illegally filmed content, etc.) lacking sound or subtitles is difficult. Although facial expressions in image sequences can be utilized in emotion recognition, the diverse identities of individuals in real-world content limits computational models of relationships between facial expressions. This study proposed a transformation model which employed a video vision transformer to focus on facial expression sequences in videos. It effectively understood and extracted facial expression information from the identities of individuals, instead of fusing multimodal models. The design entailed capture of higher-quality facial expression information through mixed-token embedding facial expression sequences augmented via various methods into a single data representation, and comprised two modules: spatial and temporal encoders. Further, temporal position embedding, focusing on relationships between video frames, was proposed and subsequently applied to the temporal encoder module. The performance of the proposed algorithm was compared with that of conventional methods on two emotion recognition datasets of video content, with results demonstrating its superiority.


Assuntos
Reconhecimento Facial , Algoritmos , Face , Expressão Facial , Humanos , Aprendizado de Máquina
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA