Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 42
Filter
1.
Sci Rep ; 14(1): 12763, 2024 06 04.
Article in English | MEDLINE | ID: mdl-38834661

ABSTRACT

With the continuous progress of technology, the subject of life science plays an increasingly important role, among which the application of artificial intelligence in the medical field has attracted more and more attention. Bell facial palsy, a neurological ailment characterized by facial muscle weakness or paralysis, exerts a profound impact on patients' facial expressions and masticatory abilities, thereby inflicting considerable distress upon their overall quality of life and mental well-being. In this study, we designed a facial attribute recognition model specifically for individuals with Bell's facial palsy. The model utilizes an enhanced SSD network and scientific computing to perform a graded assessment of the patients' condition. By replacing the VGG network with a more efficient backbone, we improved the model's accuracy and significantly reduced its computational burden. The results show that the improved SSD network has an average precision of 87.9% in the classification of light, middle and severe facial palsy, and effectively performs the classification of patients with facial palsy, where scientific calculations also increase the precision of the classification. This is also one of the most significant contributions of this article, which provides intelligent means and objective data for future research on intelligent diagnosis and treatment as well as progressive rehabilitation.


Subject(s)
Bell Palsy , Humans , Bell Palsy/diagnosis , Bell Palsy/physiopathology , Neural Networks, Computer , Female , Male , Facial Expression , Adult , Artificial Intelligence , Middle Aged , Facial Paralysis/diagnosis , Facial Paralysis/physiopathology , Facial Paralysis/psychology , Facial Recognition , Automated Facial Recognition/methods
2.
BMC Pediatr ; 24(1): 361, 2024 May 24.
Article in English | MEDLINE | ID: mdl-38783283

ABSTRACT

BACKGROUND: Noonan syndrome (NS) is a rare genetic disease, and patients who suffer from it exhibit a facial morphology that is characterized by a high forehead, hypertelorism, ptosis, inner epicanthal folds, down-slanting palpebral fissures, a highly arched palate, a round nasal tip, and posteriorly rotated ears. Facial analysis technology has recently been applied to identify many genetic syndromes (GSs). However, few studies have investigated the identification of NS based on the facial features of the subjects. OBJECTIVES: This study develops advanced models to enhance the accuracy of diagnosis of NS. METHODS: A total of 1,892 people were enrolled in this study, including 233 patients with NS, 863 patients with other GSs, and 796 healthy children. We took one to 10 frontal photos of each subject to build a dataset, and then applied the multi-task convolutional neural network (MTCNN) for data pre-processing to generate standardized outputs with five crucial facial landmarks. The ImageNet dataset was used to pre-train the network so that it could capture generalizable features and minimize data wastage. We subsequently constructed seven models for facial identification based on the VGG16, VGG19, VGG16-BN, VGG19-BN, ResNet50, MobileNet-V2, and squeeze-and-excitation network (SENet) architectures. The identification performance of seven models was evaluated and compared with that of six physicians. RESULTS: All models exhibited a high accuracy, precision, and specificity in recognizing NS patients. The VGG19-BN model delivered the best overall performance, with an accuracy of 93.76%, precision of 91.40%, specificity of 98.73%, and F1 score of 78.34%. The VGG16-BN model achieved the highest AUC value of 0.9787, while all models based on VGG architectures were superior to the others on the whole. The highest scores of six physicians in terms of accuracy, precision, specificity, and the F1 score were 74.00%, 75.00%, 88.33%, and 61.76%, respectively. The performance of each model of facial recognition was superior to that of the best physician on all metrics. CONCLUSION: Models of computer-assisted facial recognition can improve the rate of diagnosis of NS. The models based on VGG19-BN and VGG16-BN can play an important role in diagnosing NS in clinical practice.


Subject(s)
Noonan Syndrome , Humans , Noonan Syndrome/diagnosis , Child , Female , Male , Child, Preschool , Neural Networks, Computer , Infant , Adolescent , Automated Facial Recognition/methods , Diagnosis, Computer-Assisted/methods , Sensitivity and Specificity , Case-Control Studies
3.
PLoS One ; 19(5): e0304610, 2024.
Article in English | MEDLINE | ID: mdl-38820451

ABSTRACT

Face Morphing Attacks pose a threat to the security of identity documents, especially with respect to a subsequent access control process, because they allow both involved individuals to use the same document. Several algorithms are currently being developed to detect Morphing Attacks, often requiring large data sets of morphed face images for training. In the present study, face embeddings are used for two different purposes: first, to pre-select images for the subsequent large-scale generation of Morphing Attacks, and second, to detect potential Morphing Attacks. Previous studies have demonstrated the power of embeddings in both use cases. However, we aim to build on these studies by adding the more powerful MagFace model to both use cases, and by performing comprehensive analyses of the role of embeddings in pre-selection and attack detection in terms of the vulnerability of face recognition systems and attack detection algorithms. In particular, we use recent developments to assess the attack potential, but also investigate the influence of morphing algorithms. For the first objective, an algorithm is developed that pairs individuals based on the similarity of their face embeddings. Different state-of-the-art face recognition systems are used to extract embeddings in order to pre-select the face images and different morphing algorithms are used to fuse the face images. The attack potential of the differently generated morphed face images will be quantified to compare the usability of the embeddings for automatically generating a large number of successful Morphing Attacks. For the second objective, we compare the performance of the embeddings of two state-of-the-art face recognition systems with respect to their ability to detect morphed face images. Our results demonstrate that ArcFace and MagFace provide valuable face embeddings for image pre-selection. Various open-source and commercial-off-the-shelf face recognition systems are vulnerable to the generated Morphing Attacks, and their vulnerability increases when image pre-selection is based on embeddings compared to random pairing. In particular, landmark-based closed-source morphing algorithms generate attacks that pose a high risk to any tested face recognition system. Remarkably, more accurate face recognition systems show a higher vulnerability to Morphing Attacks. Among the systems tested, commercial-off-the-shelf systems were the most vulnerable to Morphing Attacks. In addition, MagFace embeddings stand out as a robust alternative for detecting morphed face images compared to the previously used ArcFace embeddings. The results endorse the benefits of face embeddings for more effective image pre-selection for face morphing and for more accurate detection of morphed face images, as demonstrated by extensive analysis of various designed attacks. The MagFace model is a powerful alternative to the often-used ArcFace model in detecting attacks and can increase performance depending on the use case. It also highlights the usability of embeddings to generate large-scale morphed face databases for various purposes, such as training Morphing Attack Detection algorithms as a countermeasure against attacks.


Subject(s)
Algorithms , Computer Security , Humans , Face , Image Processing, Computer-Assisted/methods , Automated Facial Recognition/methods , Facial Recognition
4.
Sensors (Basel) ; 24(10)2024 May 18.
Article in English | MEDLINE | ID: mdl-38794068

ABSTRACT

Most facial analysis methods perform well in standardized testing but not in real-world testing. The main reason is that training models cannot easily learn various human features and background noise, especially for facial landmark detection and head pose estimation tasks with limited and noisy training datasets. To alleviate the gap between standardized and real-world testing, we propose a pseudo-labeling technique using a face recognition dataset consisting of various people and background noise. The use of our pseudo-labeled training dataset can help to overcome the lack of diversity among the people in the dataset. Our integrated framework is constructed using complementary multitask learning methods to extract robust features for each task. Furthermore, introducing pseudo-labeling and multitask learning improves the face recognition performance by enabling the learning of pose-invariant features. Our method achieves state-of-the-art (SOTA) or near-SOTA performance on the AFLW2000-3D and BIWI datasets for facial landmark detection and head pose estimation, with competitive face verification performance on the IJB-C test dataset for face recognition. We demonstrate this through a novel testing methodology that categorizes cases as soft, medium, and hard based on the pose values of IJB-C. The proposed method achieves stable performance even when the dataset lacks diverse face identifications.


Subject(s)
Automated Facial Recognition , Face , Head , Humans , Face/anatomy & histology , Face/diagnostic imaging , Head/diagnostic imaging , Automated Facial Recognition/methods , Algorithms , Machine Learning , Facial Recognition , Databases, Factual , Image Processing, Computer-Assisted/methods
5.
Neural Netw ; 175: 106275, 2024 Jul.
Article in English | MEDLINE | ID: mdl-38653078

ABSTRACT

Face Anti-Spoofing (FAS) seeks to protect face recognition systems from spoofing attacks, which is applied extensively in scenarios such as access control, electronic payment, and security surveillance systems. Face anti-spoofing requires the integration of local details and global semantic information. Existing CNN-based methods rely on small stride or image patch-based feature extraction structures, which struggle to capture spatial and cross-layer feature correlations effectively. Meanwhile, Transformer-based methods have limitations in extracting discriminative detailed features. To address the aforementioned issues, we introduce a multi-stage CNN-Transformer-based framework, which extracts local features through the convolutional layer and long-distance feature relationships via self-attention. Based on this, we proposed a cross-attention multi-stage feature fusion, employing semantically high-stage features to query task-relevant features in low-stage features for further cross-stage feature fusion. To enhance the discrimination of local features for subtle differences, we design pixel-wise material classification supervision and add a auxiliary branch in the intermediate layers of the model. Moreover, to address the limitations of a single acquisition environment and scarcity of acquisition devices in the existing Near-Infrared dataset, we create a large-scale Near-Infrared Face Anti-Spoofing dataset with 380k pictures of 1040 identities. The proposed method could achieve the state-of-the-art in OULU-NPU and our proposed Near-Infrared dataset at just 1.3GFlops and 3.2M parameter numbers, which demonstrate the effective of the proposed method.


Subject(s)
Neural Networks, Computer , Humans , Automated Facial Recognition/methods , Image Processing, Computer-Assisted/methods , Face , Computer Security , Algorithms
6.
Comput Math Methods Med ; 2021: 7748350, 2021.
Article in English | MEDLINE | ID: mdl-34824599

ABSTRACT

The application of face detection and recognition technology in security monitoring systems has made a huge contribution to public security. Face detection is an essential first step in many face analysis systems. In complex scenes, the accuracy of face detection would be limited because of the missing and false detection of small faces, due to image quality, face scale, light, and other factors. In this paper, a two-level face detection model called SR-YOLOv5 is proposed to address some problems of dense small faces in actual scenarios. The research first optimized the backbone and loss function of YOLOv5, which is aimed at achieving better performance in terms of mean average precision (mAP) and speed. Then, to improve face detection in blurred scenes or low-resolution situations, we integrated image superresolution technology on the detection head. In addition, some representative deep-learning algorithm based on face detection is discussed by grouping them into a few major categories, and the popular face detection benchmarks are enumerated in detail. Finally, the wider face dataset is used to train and test the SR-YOLOv5 model. Compared with multitask convolutional neural network (MTCNN), Contextual Multi-Scale Region-based CNN (CMS-RCNN), Finding Tiny Faces (HR), Single Shot Scale-invariant Face Detector (S3FD), and TinaFace algorithms, it is verified that the proposed model has higher detection precision, which is 0.7%, 0.6%, and 2.9% higher than the top one. SR-YOLOv5 can effectively use face information to accurately detect hard-to-detect face targets in complex scenes.


Subject(s)
Algorithms , Automated Facial Recognition/methods , Face/anatomy & histology , Neural Networks, Computer , Automated Facial Recognition/statistics & numerical data , Computational Biology , Deep Learning , Humans , Security Measures/statistics & numerical data
7.
PLoS One ; 16(10): e0257923, 2021.
Article in English | MEDLINE | ID: mdl-34648520

ABSTRACT

Facial imaging and facial recognition technologies, now common in our daily lives, also are increasingly incorporated into health care processes, enabling touch-free appointment check-in, matching patients accurately, and assisting with the diagnosis of certain medical conditions. The use, sharing, and storage of facial data is expected to expand in coming years, yet little is documented about the perspectives of patients and participants regarding these uses. We developed a pair of surveys to gather public perspectives on uses of facial images and facial recognition technologies in healthcare and in health-related research in the United States. We used Qualtrics Panels to collect responses from general public respondents using two complementary and overlapping survey instruments; one focused on six types of biometrics (including facial images and DNA) and their uses in a wide range of societal contexts (including healthcare and research) and the other focused on facial imaging, facial recognition technology, and related data practices in health and research contexts specifically. We collected responses from a diverse group of 4,048 adults in the United States (2,038 and 2,010, from each survey respectively). A majority of respondents (55.5%) indicated they were equally worried about the privacy of medical records, DNA, and facial images collected for precision health research. A vignette was used to gauge willingness to participate in a hypothetical precision health study, with respondents split as willing to (39.6%), unwilling to (30.1%), and unsure about (30.3%) participating. Nearly one-quarter of respondents (24.8%) reported they would prefer to opt out of the DNA component of a study, and 22.0% reported they would prefer to opt out of both the DNA and facial imaging component of the study. Few indicated willingness to pay a fee to opt-out of the collection of their research data. Finally, respondents were offered options for ideal governance design of their data, as "open science"; "gated science"; and "closed science." No option elicited a majority response. Our findings indicate that while a majority of research participants might be comfortable with facial images and facial recognition technologies in healthcare and health-related research, a significant fraction expressed concern for the privacy of their own face-based data, similar to the privacy concerns of DNA data and medical records. A nuanced approach to uses of face-based data in healthcare and health-related research is needed, taking into consideration storage protection plans and the contexts of use.


Subject(s)
Automated Facial Recognition/methods , Biomedical Research/methods , Data Management/methods , Delivery of Health Care/methods , Facial Recognition , Information Dissemination/methods , Public Opinion , Adolescent , Adult , Aged , Female , Humans , Male , Medical Records , Middle Aged , Privacy , Surveys and Questionnaires , United States , Young Adult
8.
PLoS One ; 16(10): e0258672, 2021.
Article in English | MEDLINE | ID: mdl-34665834

ABSTRACT

The aim of this study was to develop and evaluate a machine vision algorithm to assess the pain level in horses, using an automatic computational classifier based on the Horse Grimace Scale (HGS) and trained by machine learning method. The use of the Horse Grimace Scale is dependent on a human observer, who most of the time does not have availability to evaluate the animal for long periods and must also be well trained in order to apply the evaluation system correctly. In addition, even with adequate training, the presence of an unknown person near an animal in pain can result in behavioral changes, making the evaluation more complex. As a possible solution, the automatic video-imaging system will be able to monitor pain responses in horses more accurately and in real-time, and thus allow an earlier diagnosis and more efficient treatment for the affected animals. This study is based on assessment of facial expressions of 7 horses that underwent castration, collected through a video system positioned on the top of the feeder station, capturing images at 4 distinct timepoints daily for two days before and four days after surgical castration. A labeling process was applied to build a pain facial image database and machine learning methods were used to train the computational pain classifier. The machine vision algorithm was developed through the training of a Convolutional Neural Network (CNN) that resulted in an overall accuracy of 75.8% while classifying pain on three levels: not present, moderately present, and obviously present. While classifying between two categories (pain not present and pain present) the overall accuracy reached 88.3%. Although there are some improvements to be made in order to use the system in a daily routine, the model appears promising and capable of measuring pain on images of horses automatically through facial expressions, collected from video images.


Subject(s)
Automated Facial Recognition/methods , Orchiectomy/adverse effects , Pain Measurement/veterinary , Algorithms , Animals , Databases, Factual , Deep Learning , Facial Recognition , Horses , Humans , Image Interpretation, Computer-Assisted/methods , Neural Networks, Computer , Orchiectomy/veterinary , Video Recording
9.
IEEE Trans Image Process ; 30: 7636-7648, 2021.
Article in English | MEDLINE | ID: mdl-34469297

ABSTRACT

Convolutional neural networks are capable of extracting powerful representations for face recognition. However, they tend to suffer from poor generalization due to imbalanced data distributions where a small number of classes are over-represented (e.g. frontal or non-occluded faces) and some of the remaining rarely appear (e.g. profile or heavily occluded faces). This is the reason why the performance is dramatically degraded in minority classes. For example, this issue is serious for recognizing masked faces in the scenario of ongoing pandemic of the COVID-19. In this work, we propose an Attention Augmented Network, called AAN-Face, to handle this issue. First, an attention erasing (AE) scheme is proposed to randomly erase units in attention maps. This well prepares models towards occlusions or pose variations. Second, an attention center loss (ACL) is proposed to learn a center for each attention map, so that the same attention map focuses on the same facial part. Consequently, discriminative facial regions are emphasized, while useless or noisy ones are suppressed. Third, the AE and the ACL are incorporated to form the AAN-Face. Since the discriminative parts are randomly removed by the AE, the ACL is encouraged to learn different attention centers, leading to the localization of diverse and complementary facial parts. Comprehensive experiments on various test datasets, especially on masked faces, demonstrate that our AAN-Face models outperform the state-of-the-art methods, showing the importance and effectiveness.


Subject(s)
Automated Facial Recognition/methods , Face/anatomy & histology , Image Processing, Computer-Assisted/methods , Neural Networks, Computer , COVID-19 , Humans , Masks
10.
Plast Reconstr Surg ; 148(1): 45-54, 2021 Jul 01.
Article in English | MEDLINE | ID: mdl-34181603

ABSTRACT

BACKGROUND: Patients desire face-lifting procedures primarily to appear younger, more refreshed, and attractive. Because there are few objective studies assessing the success of face-lift surgery, the authors used artificial intelligence, in the form of convolutional neural network algorithms alongside FACE-Q patient-reported outcomes, to evaluate perceived age reduction and patient satisfaction following face-lift surgery. METHODS: Standardized preoperative and postoperative (1 year) images of 50 consecutive patients who underwent face-lift procedures (platysmaplasty, superficial musculoaponeurotic system-ectomy, cheek minimal access cranial suspension malar lift, or fat grafting) were used by four neural networks (trained to identify age based on facial features) to estimate age reduction after surgery. In addition, FACE-Q surveys were used to measure patient-reported facial aesthetic outcome. Patient satisfaction was compared to age reduction. RESULTS: The neural network preoperative age accuracy score demonstrated that all four neural networks were accurate in identifying ages (mean score, 100.8). Patient self-appraisal age reduction reported a greater age reduction than neural network age reduction after a face lift (-6.7 years versus -4.3 years). FACE-Q scores demonstrated a high level of patient satisfaction for facial appearance (75.1 ± 8.1), quality of life (82.4 ± 8.3), and satisfaction with outcome (79.0 ± 6.3). Finally, there was a positive correlation between neural network age reduction and patient satisfaction. CONCLUSION: Artificial intelligence algorithms can reliably estimate the reduction in apparent age after face-lift surgery; this estimated age reduction correlates with patient satisfaction. CLINICAL QUESTION/LEVEL OF EVIDENCE: Diagnostic, IV.


Subject(s)
Automated Facial Recognition/statistics & numerical data , Deep Learning/statistics & numerical data , Patient Satisfaction/statistics & numerical data , Rejuvenation , Rhytidoplasty/statistics & numerical data , Aged , Automated Facial Recognition/methods , Face/diagnostic imaging , Face/surgery , Feasibility Studies , Female , Follow-Up Studies , Humans , Image Processing, Computer-Assisted/methods , Image Processing, Computer-Assisted/statistics & numerical data , Middle Aged , Patient Reported Outcome Measures , Postoperative Period , Preoperative Period , Quality of Life , Reproducibility of Results , Treatment Outcome
11.
Plast Reconstr Surg ; 148(1): 162-169, 2021 Jul 01.
Article in English | MEDLINE | ID: mdl-34181613

ABSTRACT

BACKGROUND: Despite the wide range of cleft lip morphology, consistent scales to categorize preoperative severity do not exist. Machine learning has been used to increase accuracy and efficiency in detection and rating of multiple conditions, yet it has not been applied to cleft disease. The authors tested a machine learning approach to automatically detect and measure facial landmarks and assign severity grades using preoperative photographs. METHODS: Preoperative images were collected from 800 unilateral cleft lip patients, manually annotated for cleft-specific landmarks, and rated using a previously validated severity scale by eight expert reviewers. Five convolutional neural network models were trained for landmark detection and severity grade assignment. Mean squared error loss and Pearson correlation coefficient for cleft width ratio, nostril width ratio, and severity grade assignment were calculated. RESULTS: All five models performed well in landmark detection and severity grade assignment, with the largest and most complex model, Residual Network, performing best (mean squared error, 24.41; cleft width ratio correlation, 0.943; nostril width ratio correlation, 0.879; severity correlation, 0.892). The mobile device-compatible network, MobileNet, also showed a high degree of accuracy (mean squared error, 36.66; cleft width ratio correlation, 0.901; nostril width ratio correlation, 0.705; severity correlation, 0.860). CONCLUSIONS: Machine learning models demonstrate the ability to accurately measure facial features and assign severity grades according to validated scales. Such models hold promise for the creation of a simple, automated approach to classifying cleft lip morphology. Further potential exists for a mobile telephone-based application to provide real-time feedback to improve clinical decision making and patient counseling.


Subject(s)
Cleft Lip/diagnosis , Deep Learning , Image Processing, Computer-Assisted/methods , Nose/abnormalities , Severity of Illness Index , Anatomic Landmarks , Automated Facial Recognition/methods , Cleft Lip/complications , Cleft Lip/surgery , Clinical Decision-Making , Counseling , Datasets as Topic , Feasibility Studies , Humans , Mobile Applications , Nose/diagnostic imaging , Nose/surgery , Photography , Preoperative Period , Remote Consultation , Rhinoplasty
12.
Eur J Med Genet ; 64(9): 104267, 2021 Sep.
Article in English | MEDLINE | ID: mdl-34161860

ABSTRACT

Down syndrome is one of the most common chromosomal anomalies affecting the world's population, with an estimated frequency of 1 in 700 live births. Despite its relatively high prevalence, diagnostic rates based on clinical features have remained under 70% for most of the developed world and even lower in countries with limited resources. While genetic and cytogenetic confirmation greatly increases the diagnostic rate, such resources are often non-existent in many low- and middle-income countries, particularly in Sub-Saharan Africa. To address the needs of countries with limited resources, the implementation of mobile, user-friendly and affordable technologies that aid in diagnosis would greatly increase the odds of success for a child born with a genetic condition. Given that the Democratic Republic of the Congo is estimated to have one of the highest rates of birth defects in the world, our team sought to determine if smartphone-based facial analysis technology could accurately detect Down syndrome in individuals of Congolese descent. Prior to technology training, we confirmed the presence of trisomy 21 using low-cost genomic applications that do not need advanced expertise to utilize and are available in many low-resourced countries. Our software technology trained on 132 Congolese subjects had a significantly improved performance (91.67% accuracy, 95.45% sensitivity, 87.88% specificity) when compared to previous technology trained on individuals who are not of Congolese origin (p < 5%). In addition, we provide the list of most discriminative facial features of Down syndrome and their ranges in the Congolese population. Collectively, our technology provides low-cost and accurate diagnosis of Down syndrome in the local population.


Subject(s)
Automated Facial Recognition/methods , Down Syndrome/pathology , Facies , Image Processing, Computer-Assisted/methods , Automated Facial Recognition/economics , Automated Facial Recognition/standards , Democratic Republic of the Congo , Developing Countries , Down Syndrome/genetics , Genetic Testing , Humans , Image Processing, Computer-Assisted/economics , Image Processing, Computer-Assisted/standards , Machine Learning , Sensitivity and Specificity
13.
Plast Surg Nurs ; 41(2): 112-116, 2021.
Article in English | MEDLINE | ID: mdl-34033638

ABSTRACT

The number of applications for facial recognition technology is increasing due to the improvement in image quality, artificial intelligence, and computer processing power that has occurred during the last decades. Algorithms can be used to convert facial anthropometric landmarks into a computer representation, which can be used to help identify nonverbal information about an individual's health status. This article discusses the potential ways a facial recognition tool can perform a health assessment. Because facial attributes may be considered biometric data, clinicians should be informed about the clinical, ethical, and legal issues associated with its use.


Subject(s)
Automated Facial Recognition/instrumentation , Health Status , Nursing Assessment/methods , Artificial Intelligence/trends , Automated Facial Recognition/methods , Humans , Nursing Assessment/standards
14.
IEEE Trans Image Process ; 30: 5313-5326, 2021.
Article in English | MEDLINE | ID: mdl-34038362

ABSTRACT

In this paper, we propose a structure-coherent deep feature learning method for face alignment. Unlike most existing face alignment methods which overlook the facial structure cues, we explicitly exploit the relation among facial landmarks to make the detector robust to hard cases such as occlusion and large pose. Specifically, we leverage a landmark-graph relational network to enforce the structural relationships among landmarks. We consider the facial landmarks as structural graph nodes and carefully design the neighborhood to passing features among the most related nodes. Our method dynamically adapts the weights of node neighborhood to eliminate distracted information from noisy nodes, such as occluded landmark point. Moreover, different from most previous works which only tend to penalize the landmarks absolute position during the training, we propose a relative location loss to enhance the information of relative location of landmarks. This relative location supervision further regularizes the facial structure. Our approach considers the interactions among facial landmarks and can be easily implemented on top of any convolutional backbone to boost the performance. Extensive experiments on three popular benchmarks, including WFLW, COFW and 300W, demonstrate the effectiveness of the proposed method. In particular, due to explicit structure modeling, our approach is especially robust to challenging cases resulting in impressive low failure rate on COFW and WFLW datasets. The model and code are publicly available at https://github.com/BeierZhu/Sturcture-Coherency-Face-Alignment.


Subject(s)
Automated Facial Recognition/methods , Deep Learning , Face/anatomy & histology , Anatomic Landmarks/anatomy & histology , Databases, Factual , Humans
15.
Mol Genet Genomic Med ; 9(5): e1636, 2021 05.
Article in English | MEDLINE | ID: mdl-33773094

ABSTRACT

INTRODUCTION: Patients with Noonan and Williams-Beuren syndrome present similar facial phenotypes modulated by their ethnic background. Although distinctive facial features have been reported, studies show a variable incidence of those characteristics in populations with diverse ancestry. Hence, a differential diagnosis based on reported facial features can be challenging. Although accurate diagnoses are possible with genetic testing, they are not available in developing and remote regions. METHODS: We used a facial analysis technology to identify the most discriminative facial metrics between 286 patients with Noonan and 161 with Williams-Beuren syndrome with diverse ethnic background. We quantified the most discriminative metrics, and their ranges both globally and in different ethnic groups. We also created population-based appearance images that are useful not only as clinical references but also for training purposes. Finally, we trained both global and ethnic-specific machine learning models with previous metrics to distinguish between patients with Noonan and Williams-Beuren syndromes. RESULTS: We obtained a classification accuracy of 85.68% in the global population evaluated using cross-validation, which improved to 90.38% when we adapted the facial metrics to the ethnicity of the patients (p = 0.024). CONCLUSION: Our facial analysis provided for the first time quantitative reference facial metrics for the differential diagnosis Noonan and Williams-Beuren syndromes in diverse populations.


Subject(s)
Automated Facial Recognition/methods , Diagnosis, Computer-Assisted/methods , Face/pathology , Noonan Syndrome/diagnosis , Phenotype , Williams Syndrome/diagnosis , Adolescent , Adult , Automated Facial Recognition/standards , Child , Child, Preschool , Diagnosis, Computer-Assisted/standards , Diagnosis, Differential , Female , Humans , Infant , Machine Learning , Male , Sensitivity and Specificity
16.
Biomed Res Int ; 2021: 6696357, 2021.
Article in English | MEDLINE | ID: mdl-33778081

ABSTRACT

BACKGROUND: Sedentary lifestyle and work from home schedules due to the ongoing COVID-19 pandemic in 2020 have caused a significant rise in obesity across adults. With limited visits to the doctors during this period to avoid possible infections, there is currently no way to measure or track obesity. METHODS: We reviewed the literature on relationships between obesity and facial features, in white, black, hispanic-latino, and Korean populations and validated them against a cohort of Indian participants (n = 106). The body mass index (BMI) and waist-to-hip ratio (WHR) were obtained using anthropometric measurements, and body fat mass (BFM), percentage body fat (PBF), and visceral fat area (VFA) were measured using body composition analysis. Facial pictures were also collected and processed to characterize facial geometry. Regression analysis was conducted to determine correlations between body fat parameters and facial model parameters. RESULTS: Lower facial geometry was highly correlated with BMI (R 2 = 0.77) followed by PBF (R 2 = 0.72), VFA (R 2 = 0.65), WHR (R 2 = 0.60), BFM (R 2 = 0.59), and weight (R 2 = 0.54). CONCLUSIONS: The ability to predict obesity using facial images through mobile application or telemedicine can help with early diagnosis and timely medical intervention for people with obesity during the pandemic.


Subject(s)
Anthropometry/methods , Automated Facial Recognition/methods , COVID-19/epidemiology , Obesity/diagnosis , Adult , Body Composition , Body Mass Index , Body Weight , Facial Recognition/physiology , Female , Humans , Male , Middle Aged , Obesity/epidemiology , Obesity/metabolism , Pandemics , Predictive Value of Tests , Prognosis , Risk Factors , SARS-CoV-2/isolation & purification , Waist Circumference , Waist-Hip Ratio
17.
IEEE Trans Image Process ; 30: 2627-2642, 2021.
Article in English | MEDLINE | ID: mdl-33523811

ABSTRACT

Light field (LF) cameras provide rich spatio-angular visual representations by sensing the visual scene from multiple perspectives and have recently emerged as a promising technology to boost the performance of human-machine systems such as biometrics and affective computing. Despite the significant success of LF representation for constrained facial image analysis, this technology has never been used for face and expression recognition in the wild. In this context, this paper proposes a new deep face and expression recognition solution, called CapsField, based on a convolutional neural network and an additional capsule network that utilizes dynamic routing to learn hierarchical relations between capsules. CapsField extracts the spatial features from facial images and learns the angular part-whole relations for a selected set of 2D sub-aperture images rendered from each LF image. To analyze the performance of the proposed solution in the wild, the first in the wild LF face dataset, along with a new complementary constrained face dataset captured from the same subjects recorded earlier have been captured and are made available. A subset of the in the wild dataset contains facial images with different expressions, annotated for usage in the context of face expression recognition tests. An extensive performance assessment study using the new datasets has been conducted for the proposed and relevant prior solutions, showing that the CapsField proposed solution achieves superior performance for both face and expression recognition tasks when compared to the state-of-the-art.


Subject(s)
Automated Facial Recognition/methods , Deep Learning , Face , Image Processing, Computer-Assisted/methods , Face/anatomy & histology , Face/diagnostic imaging , Facial Expression , Female , Humans , Information Storage and Retrieval , Male
18.
Neuroimage ; 231: 117845, 2021 05 01.
Article in English | MEDLINE | ID: mdl-33582276

ABSTRACT

Recent advances in automated face recognition algorithms have increased the risk that de-identified research MRI scans may be re-identifiable by matching them to identified photographs using face recognition. A variety of software exist to de-face (remove faces from) MRI, but their ability to prevent face recognition has never been measured and their image modifications can alter automated brain measurements. In this study, we compared three popular de-facing techniques and introduce our mri_reface technique designed to minimize effects on brain measurements by replacing the face with a population average, rather than removing it. For each technique, we measured 1) how well it prevented automated face recognition (i.e. effects on exceptionally-motivated individuals) and 2) how it altered brain measurements from SPM12, FreeSurfer, and FSL (i.e. effects on the average user of de-identified data). Before de-facing, 97% of scans from a sample of 157 volunteers were correctly matched to photographs using automated face recognition. After de-facing with popular software, 28-38% of scans still retained enough data for successful automated face matching. Our proposed mri_reface had similar performance with the best existing method (fsl_deface) at preventing face recognition (28-30%) and it had the smallest effects on brain measurements in more pipelines than any other, but these differences were modest.


Subject(s)
Automated Facial Recognition/methods , Biomedical Research/methods , Brain/diagnostic imaging , Image Processing, Computer-Assisted/methods , Magnetic Resonance Imaging/methods , Neuroimaging/methods , Adult , Aged , Aged, 80 and over , Algorithms , Automated Facial Recognition/trends , Brain/physiology , Female , Humans , Image Processing, Computer-Assisted/trends , Magnetic Resonance Imaging/trends , Male , Middle Aged , Neuroimaging/trends , Software/trends
19.
Neural Netw ; 136: 233-243, 2021 Apr.
Article in English | MEDLINE | ID: mdl-33257223

ABSTRACT

Recently, convolutional neural networks (CNNs)-based facial landmark detection methods have achieved great success. However, most of existing CNN-based facial landmark detection methods have not attempted to activate multiple correlated facial parts and learn different semantic features from them that they can not accurately model the relationships among the local details and can not fully explore more discriminative and fine semantic features, thus they suffer from partial occlusions and large pose variations. To address these problems, we propose a cross-order cross-semantic deep network (CCDN) to boost the semantic features learning for robust facial landmark detection. Specifically, a cross-order two-squeeze multi-excitation (CTM) module is proposed to introduce the cross-order channel correlations for more discriminative representations learning and multiple attention-specific part activation. Moreover, a novel cross-order cross-semantic (COCS) regularizer is designed to drive the network to learn cross-order cross-semantic features from different activation for facial landmark detection. It is interesting to show that by integrating the CTM module and COCS regularizer, the proposed CCDN can effectively activate and learn more fine and complementary cross-order cross-semantic features to improve the accuracy of facial landmark detection under extremely challenging scenarios. Experimental results on challenging benchmark datasets demonstrate the superiority of our CCDN over state-of-the-art facial landmark detection methods.


Subject(s)
Automated Facial Recognition/methods , Neural Networks, Computer , Semantic Web , Face , Humans , Semantics
20.
IEEE Trans Neural Netw Learn Syst ; 32(4): 1560-1574, 2021 04.
Article in English | MEDLINE | ID: mdl-32310806

ABSTRACT

This article focuses on a new and practical problem in single-sample per person face recognition (SSPP FR), i.e., SSPP FR with a contaminated biometric enrolment database (SSPP-ce FR), where the SSPP-based enrolment database is contaminated by nuisance facial variations in the wild, such as poor lightings, expression change, and disguises (e.g., wearing sunglasses, hat, and scarf). In SSPP-ce FR, the most popular generic learning methods will suffer serious performance degradation because the prototype plus variation (P+V) model used in these methods is no longer suitable in such scenarios. The reasons are twofold. First, the contaminated enrolment samples could yield bad prototypes to represent the persons. Second, the generated variation dictionary is simply based on the subtraction of the average face from generic samples of the same person and cannot well depict the intrapersonal variations. To address the SSPP-ce FR problem, we propose a novel iterative dynamic generic learning (IDGL) method, where the labeled enrolment database and the unlabeled query set are fed into a dynamic label feedback network for learning. Specifically, IDGL first recovers the prototypes for the contaminated enrolment samples via a semisupervised low-rank representation (SSLRR) framework and learns a representative variation dictionary by extracting the "sample-specific" corruptions from an auxiliary generic set. Then, it puts them into the P+V model to estimate labels for query samples. Subsequently, the estimated labels will be used as feedback to modify the SSLRR, thus updating new prototypes for the next round of P+V-based label estimation. With the dynamic learning network, the accuracy of the estimated labels is improved iteratively by virtue of the steadily enhanced prototypes. Experiments on various benchmark face data sets have demonstrated the superiority of IDGL over state-of-the-art counterparts.


Subject(s)
Automated Facial Recognition/methods , Biometric Identification/methods , Machine Learning , Algorithms , Artificial Intelligence , Biometry , Databases, Factual , Deep Learning , Feedback , Humans , Neural Networks, Computer
SELECTION OF CITATIONS
SEARCH DETAIL
...