Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 52
Filtrar
1.
Sensors (Basel) ; 22(13)2022 Jun 25.
Artigo em Inglês | MEDLINE | ID: mdl-35808322

RESUMO

For 5G and future Internet, in this paper, we propose a task allocation method for future Internet application to reduce the total latency in a mobile edge computing (MEC) platform with three types of servers: a dedicated MEC server, a shared MEC server, and a cloud server. For this platform, we first calculate the delay between sending a task and receiving a response for the dedicated MEC server, shared MEC server, and cloud server by considering the processing time and transmission delay. Here, the transmission delay for the shared MEC server is derived using queueing theory. Then, we formulate an optimization problem for task allocation to minimize the total latency for all tasks. By solving this optimization problem, tasks can be allocated to the MEC servers and cloud server appropriately. In addition, we propose a heuristic algorithm to obtain the approximate optimal solution in a shorter time. This heuristic algorithm consists of four algorithms: a main algorithm and three additional algorithms. In this algorithm, tasks are divided into two groups, and task allocation is executed for each group. We compare the performance of our proposed heuristic algorithm with the solution obtained by three other methods and investigate the effectiveness of our algorithm. Numerical examples are used to demonstrate the effectiveness of our proposed heuristic algorithm. From some results, we observe that our proposed heuristic algorithm can perform task allocation in a short time and can effectively reduce the total latency in a short time. We conclude that our proposed heuristic algorithm is effective for task allocation in a MEC platform with multiple types of MEC servers.


Assuntos
Algoritmos , Computação em Nuvem , Heurística Computacional , Previsões , Internet/tendências
2.
Food Res Int ; 150(Pt A): 110753, 2021 12.
Artigo em Inglês | MEDLINE | ID: mdl-34865771

RESUMO

Bioactive peptides are short peptides (3-20 amino acid residues in length) endowed of specific biological activities. The identification and characterization of bioactive peptides of food origin are crucial to better understand the physiological consequences of food, as well as to design novel foods, ingredients, supplements, and diets to counteract mild metabolic disorders. For this reason, the identification of bioactive peptides is also relevant from a pharmaceutical standpoint. Nevertheless, the systematic identification of bioactive sequences of food origin is still challenging and relies mainly on the so defined "bottom-up" approaches, which rarely results in the total identification of most active sequences. Conversely, "top-down" approaches aim at identifying bioactive sequences with certain features and may be more suitable for the precise identification of very potent bioactive peptides. In this context, this work presents a top-down, computer-assisted and hypothesis-driven identification of potent angiotensin I converting enzyme inhibitory tripeptides, as a proof of principle. A virtual library of 6840 tripeptides was screened in silico to identify potential highly potent inhibitory peptides. Then, computational results were confirmed experimentally and a very potent novel sequence, LMP was identified. LMP showed an IC50 of 15.8 and 6.8 µM in cell-free and cell-based assays, respectively. In addition, a bioinformatics approach was used to search potential food sources of LMP. Yolk proteins were identified as a possible relevant source to analyze in further experiments. Overall, the method presented may represent a powerful and versatile framework for a systematic, high-throughput and top-down identification of bioactive peptides.


Assuntos
Heurística Computacional , Peptidil Dipeptidase A , Computadores , Suplementos Nutricionais , Peptídeos
3.
PLoS One ; 16(12): e0259786, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-34855771

RESUMO

Team formation (TF) in social networks exploits graphs (i.e., vertices = experts and edges = skills) to represent a possible collaboration between the experts. These networks lead us towards building cost-effective research teams irrespective of the geolocation of the experts and the size of the dataset. Previously, large datasets were not closely inspected for the large-scale distributions & relationships among the researchers, resulting in the algorithms failing to scale well on the data. Therefore, this paper presents a novel TF algorithm for expert team formation called SSR-TF based on two metrics; communication cost and graph reduction, that will become a basis for future TF's. In SSR-TF, communication cost finds the possibility of collaboration between researchers. The graph reduction scales the large data to only appropriate skills and the experts, resulting in real-time extraction of experts for collaboration. This approach is tested on five organic and benchmark datasets, i.e., UMP, DBLP, ACM, IMDB, and Bibsonomy. The SSR-TF algorithm is able to build cost-effective teams with the most appropriate experts-resulting in the formation of more communicative teams with high expertise levels.


Assuntos
Algoritmos , Comportamento Cooperativo , Rede Social , Gráficos por Computador , Heurística Computacional , Bases de Dados Factuais , Humanos , Filmes Cinematográficos
4.
Comput Math Methods Med ; 2021: 5595180, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-34790252

RESUMO

A common gynecological disease in the world is breast cancer that early diagnosis of this disease can be very effective in its treatment. The use of image processing methods and pattern recognition techniques in automatic breast detection from mammographic images decreases human errors and increments the rapidity of diagnosis. In this paper, mammographic images are analyzed using image processing techniques and a pipeline structure for the diagnosis of the cancerous masses. In the first stage, the quality of mammogram images and the contrast of abnormal areas in the image are improved by using image contrast improvement and a noise decline. A method based on color space is then used for image segmentation that is followed by mathematical morphology. Then, for feature image extraction, a combined gray-level cooccurrence matrix (GLCM) and discrete wavelet transform (DWT) method is used. At last, a new optimized version of convolutional neural network (CNN) and a new improved metaheuristic, called Advanced Thermal Exchange Optimizer, are used for the classification of the features. A comparison of the simulations of the proposed technique with three different techniques from the literature applied on the MIAS mammogram database is performed to show its superiority. Results show that the accuracy of diagnosing cancer cases for the proposed method and applied on the MIAS database is 93.79%, and sensitivity and specificity are obtained 96.89% and 67.7%, respectively.


Assuntos
Algoritmos , Neoplasias da Mama/diagnóstico por imagem , Neoplasias da Mama/diagnóstico , Mamografia/métodos , Redes Neurais de Computação , Biologia Computacional , Heurística Computacional , Simulação por Computador , Bases de Dados Factuais , Feminino , Humanos , Interpretação de Imagem Assistida por Computador/estatística & dados numéricos , Mamografia/estatística & dados numéricos , Intensificação de Imagem Radiográfica/métodos , Análise de Ondaletas
5.
Proc Natl Acad Sci U S A ; 118(43)2021 10 26.
Artigo em Inglês | MEDLINE | ID: mdl-34675075

RESUMO

In this paper, we introduce the Layer-Peeled Model, a nonconvex, yet analytically tractable, optimization program, in a quest to better understand deep neural networks that are trained for a sufficiently long time. As the name suggests, this model is derived by isolating the topmost layer from the remainder of the neural network, followed by imposing certain constraints separately on the two parts of the network. We demonstrate that the Layer-Peeled Model, albeit simple, inherits many characteristics of well-trained neural networks, thereby offering an effective tool for explaining and predicting common empirical patterns of deep-learning training. First, when working on class-balanced datasets, we prove that any solution to this model forms a simplex equiangular tight frame, which, in part, explains the recently discovered phenomenon of neural collapse [V. Papyan, X. Y. Han, D. L. Donoho, Proc. Natl. Acad. Sci. U.S.A. 117, 24652-24663 (2020)]. More importantly, when moving to the imbalanced case, our analysis of the Layer-Peeled Model reveals a hitherto-unknown phenomenon that we term Minority Collapse, which fundamentally limits the performance of deep-learning models on the minority classes. In addition, we use the Layer-Peeled Model to gain insights into how to mitigate Minority Collapse. Interestingly, this phenomenon is first predicted by the Layer-Peeled Model before being confirmed by our computational experiments.


Assuntos
Aprendizado Profundo , Redes Neurais de Computação , Heurística Computacional , Bases de Dados Factuais/estatística & dados numéricos , Humanos , Dinâmica não Linear , Processos Estocásticos
6.
Comput Math Methods Med ; 2021: 2213194, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-34545291

RESUMO

Predicting postoperative survival of lung cancer patients (LCPs) is an important problem of medical decision-making. However, the imbalanced distribution of patient survival in the dataset increases the difficulty of prediction. Although the synthetic minority oversampling technique (SMOTE) can be used to deal with imbalanced data, it cannot identify data noise. On the other hand, many studies use a support vector machine (SVM) combined with resampling technology to deal with imbalanced data. However, most studies require manual setting of SVM parameters, which makes it difficult to obtain the best performance. In this paper, a hybrid improved SMOTE and adaptive SVM method is proposed for imbalance data to predict the postoperative survival of LCPs. The proposed method is divided into two stages: in the first stage, the cross-validated committees filter (CVCF) is used to remove noise samples to improve the performance of SMOTE. In the second stage, we propose an adaptive SVM, which uses fuzzy self-tuning particle swarm optimization (FPSO) to optimize the parameters of SVM. Compared with other advanced algorithms, our proposed method obtains the best performance with 95.11% accuracy, 95.10% G-mean, 95.02% F1, and 95.10% area under the curve (AUC) for predicting postoperative survival of LCPs.


Assuntos
Algoritmos , Neoplasias Pulmonares/mortalidade , Neoplasias Pulmonares/cirurgia , Máquina de Vetores de Suporte , Biologia Computacional , Heurística Computacional , Bases de Dados Factuais , Lógica Fuzzy , Humanos , Análise de Sobrevida
7.
Comput Math Methods Med ; 2021: 5527698, 2021.
Artigo em Inglês | MEDLINE | ID: mdl-34239598

RESUMO

Skin cancer is the most common cancer of the body. It is estimated that more than one million people worldwide develop skin cancer each year. Early detection of this cancer has a high effect on the disease treatment. In this paper, a new optimal and automatic pipeline approach has been proposed for the diagnosis of this disease from dermoscopy images. The proposed method includes a noise reduction process before processing for eliminating the noises. Then, the Otsu method as one of the widely used thresholding method is used to characterize the region of interest. Afterward, 20 different features are extracted from the image. To reduce the method complexity, a new modified version of the Thermal Exchange Optimization Algorithm is performed to the features. This improves the method precision and consistency. To validate the proposed method's efficiency, it is implemented to the American Cancer Society database, its results are compared with some state-of-the-art methods, and the final results showed the superiority of the proposed method against the others.


Assuntos
Algoritmos , Dermoscopia/métodos , Melanoma/diagnóstico por imagem , Neoplasias Cutâneas/diagnóstico por imagem , Biologia Computacional , Heurística Computacional , Simulação por Computador , Bases de Dados Factuais , Aprendizado Profundo , Dermoscopia/estatística & dados numéricos , Detecção Precoce de Câncer/métodos , Detecção Precoce de Câncer/estatística & dados numéricos , Humanos , Interpretação de Imagem Assistida por Computador/métodos , Interpretação de Imagem Assistida por Computador/estatística & dados numéricos , Melanoma/classificação , Reconhecimento Automatizado de Padrão/métodos , Reconhecimento Automatizado de Padrão/estatística & dados numéricos , Razão Sinal-Ruído , Neoplasias Cutâneas/classificação , Máquina de Vetores de Suporte , Termografia/métodos , Termografia/estatística & dados numéricos
8.
J Comput Biol ; 28(8): 789-803, 2021 08.
Artigo em Inglês | MEDLINE | ID: mdl-34161175

RESUMO

For DNA sequence analysis, we are facing challenging tasks such as the identification of structural variants, sequencing repetitive regions, and phasing of alleles. Those challenging tasks suffer from the short length of sequencing reads, where each read may cover less than 2 single nucleotide polymorphism (SNP), or less than two occurrences of a repeated region. It is believed that long reads can help to solve those challenging tasks. In this study, we have designed new algorithms for mapping long reads to reference genomes. We have also designed efficient and effective heuristic algorithms for local alignments of long reads against the corresponding segments of the reference genome. To design the new mapping algorithm, we formulate the problem as the longest common subsequence with distance constraints. The local alignment heuristic algorithm is based on the idea of recursive alignment of k-mers, where the size of k differs in each round. We have implemented all the algorithms in C++ and produce a software package named mapAlign. Experiments on real data sets showed that the newly proposed approach can generate better alignments in terms of both identity and alignment scores for both Nanopore and single molecule real time sequencing (SMRT) data sets. For human individuals of both Nanopore and SMRT data sets, the new method can successfully math/align 91.53% and 85.36% of letters from reads to identical letters on reference genomes, respectively. In comparison, the best known method can only align 88.44% and 79.08% letters of reads for Nanopore and SMRT data sets, respectively. Our method is also faster than the best known method.


Assuntos
Mapeamento Cromossômico/métodos , Alinhamento de Sequência/métodos , Algoritmos , Heurística Computacional , Variação Genética , Humanos , Análise de Sequência de DNA , Imagem Individual de Molécula
9.
Food Chem ; 335: 127681, 2021 Jan 15.
Artigo em Inglês | MEDLINE | ID: mdl-32739803

RESUMO

In this study the Lagrange interpolation optimization algorithm based on two variables with respect to all experimental replicates (POA), was compared with two other heuristics methods (WOA and GOA). Modification of the apple surface by an edible nano coating solution in food packaging was used as case study. The experiment was performed as a factorial test based on completely randomized design by 100 permutations data sets. Results showed a significant difference between the three optimization methods (POA, WOA and GOA) which indicates the necessity of optimization and also efficiency of the present POA. The optimum result by POA, similar to a rose petal property, could rise 72% in surface contact angle (CA). The scanning electron microscopy (SEM) images of the derived surfaces showed almost a uniform spherical nanoparticles morphology. Remarkable advantages of this new approach are no additional material requirement, healthful, easy, inexpensive, fast and affordable technique for surface improvement.


Assuntos
Algoritmos , Quitosana , Embalagem de Alimentos , Nanopartículas/química , Quitosana/química , Heurística Computacional , Interações Hidrofóbicas e Hidrofílicas , Microscopia Eletrônica de Varredura
10.
Biomed Res Int ; 2020: 8506365, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-32908920

RESUMO

Finding an optimal set of discriminative features is still a crucial but challenging task in biomedical science. The complexity of the task is intensified when any of the two scenarios arise: a highly dimensioned dataset and a small sample-sized dataset. The first scenario poses a big challenge to existing machine learning approaches since the search space for identifying the most relevant feature subset is so diverse to be explored quickly while utilizing minimal computational resources. On the other hand, the second aspect poses a challenge of too few samples to learn from. Though many hybrid metaheuristic approaches (i.e., combining multiple search algorithms) have been proposed in the literature to address these challenges with very attractive performance compared to their counterpart standard standalone metaheuristics, more superior hybrid approaches can be achieved if the individual metaheuristics within the proposed hybrid algorithms are improved prior to the hybridization. Motivated by this, we propose a new hybrid Excited- (E-) Adaptive Cuckoo Search- (ACS-) Intensification Dedicated Grey Wolf Optimization (IDGWO), i.e., EACSIDGWO. EACSIDGWO is an algorithm where the step size of ACS and the nonlinear control strategy of parameter a → of the IDGWO are innovatively made adaptive via the concept of the complete voltage and current responses of a direct current (DC) excited resistor-capacitor (RC) circuit. Since the population has a higher diversity at early stages of the proposed EACSIDGWO algorithm, both the ACS and IDGWO are jointly involved in local exploitation. On the other hand, to enhance mature convergence at latter stages of the proposed algorithm, the role of ACS is switched to global exploration while the IDGWO is still left conducting the local exploitation. To prove that the proposed algorithm is superior in providing a good learning from fewer instances and an optimal feature selection from information-rich biomedical data, all these while maintaining a high classification accuracy of the data, the EACSIDGWO is employed to solve the feature selection problem. The EACSIDGWO as a feature selector is tested on six standard biomedical datasets from the University of California at Irvine (UCI) repository. The experimental results are compared with the state-of-the-art feature selection techniques, including binary ant-colony optimization (BACO), binary genetic algorithm (BGA), binary particle swarm optimization (BPSO), and extended binary cuckoo search algorithm (EBCSA). These results reveal that the EACSIDGWO has comprehensive superiority in tackling the feature selection problem, which proves the capability of the proposed algorithm in solving real-world complex problems. Furthermore, the superiority of the proposed algorithm is proved via various numerical techniques like ranking methods and statistical analysis.


Assuntos
Algoritmos , Biologia Computacional , Heurística Computacional , Simulação por Computador , Interpretação Estatística de Dados , Bases de Dados Factuais , Humanos
11.
PLoS One ; 15(8): e0232384, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-32750052

RESUMO

We propose a process graph (P-graph) approach to develop ecosystem networks from knowledge of the properties of the component species. Originally developed as a process engineering tool for designing industrial plants, the P-graph framework has key advantages over conventional ecological network analysis techniques based on input-output models. A P-graph is a bipartite graph consisting of two types of nodes, which we propose to represent components of an ecosystem. Compartments within ecosystems (e.g., organism species) are represented by one class of nodes, while the roles or functions that they play relative to other compartments are represented by a second class of nodes. This bipartite graph representation enables a powerful, unambiguous representation of relationships among ecosystem compartments, which can come in tangible (e.g., mass flow in predation) or intangible form (e.g., symbiosis). For example, within a P-graph, the distinct roles of bees as pollinators for some plants and as prey for some animals can be explicitly represented, which would not otherwise be possible using conventional ecological network analysis. After a discussion of the mapping of ecosystems into P-graph, we also discuss how this framework can be used to guide understanding of complex networks that exist in nature. Two component algorithms of P-graph, namely maximal structure generation (MSG) and solution structure generation (SSG), are shown to be particularly useful for ecological network analysis. These algorithms enable candidate ecosystem networks to be deduced based on current scientific knowledge on the individual ecosystem components. This method can be used to determine the (a) effects of loss of specific ecosystem compartments due to extinction, (b) potential efficacy of ecosystem reconstruction efforts, and (c) maximum sustainable exploitation of human ecosystem services by humans. We illustrate the use of P-graph for the analysis of ecosystem compartment loss using a small-scale stylized case study, and further propose a new criticality index that can be easily derived from SSG results.


Assuntos
Ecossistema , Algoritmos , Animais , Gráficos por Computador , Heurística Computacional , Cadeia Alimentar , Heurística , Humanos , Conceitos Matemáticos , Modelos Biológicos , Biologia de Sistemas , Teoria de Sistemas
12.
J Am Med Inform Assoc ; 27(10): 1625-1638, 2020 10 01.
Artigo em Inglês | MEDLINE | ID: mdl-32766692

RESUMO

OBJECTIVE: The study sought to describe the literature related to the development of methods for auditing the Unified Medical Language System (UMLS), with particular attention to identifying errors and inconsistencies of attributes of the concepts in the UMLS Metathesaurus. MATERIALS AND METHODS: We applied the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) approach by searching the MEDLINE database and Google Scholar for studies referencing the UMLS and any of several terms related to auditing, error detection, and quality assurance. A qualitative analysis and summarization of articles that met inclusion criteria were performed. RESULTS: Eighty-three studies were reviewed in detail. We first categorized techniques based on various aspects including concepts, concept names, and synonymy (n = 37), semantic type assignments (n = 36), hierarchical relationships (n = 24), lateral relationships (n = 12), ontology enrichment (n = 8), and ontology alignment (n = 18). We also categorized the methods according to their level of automation (ie, automated systematic, automated heuristic, or manual) and the type of knowledge used (ie, intrinsic or extrinsic knowledge). CONCLUSIONS: This study is a comprehensive review of the published methods for auditing the various conceptual aspects of the UMLS. Categorizing the auditing techniques according to the various aspects will enable the curators of the UMLS as well as researchers comprehensive easy access to this wealth of knowledge (eg, for auditing lateral relationships in the UMLS). We also reviewed ontology enrichment and alignment techniques due to their critical use of and impact on the UMLS.


Assuntos
Controle de Qualidade , Unified Medical Language System , Heurística Computacional , Web Semântica
13.
PLoS One ; 15(6): e0234625, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-32525939

RESUMO

In many objective optimization problems (MaOPs), more than three distinct objectives are optimized. The challenging part in MaOPs is to get the Pareto approximation (PA) with high diversity and good convergence. In Literature, in order to solve the issue of diversity and convergence in MaOPs, many approaches are proposed using different multi objective evolutionary algorithms (MOEAs). Moreover, to get better results, the researchers use the sets of reference points to differentiate the solutions and to model the search process, it further evaluates and selects the non-dominating solutions by using the reference set of solutions. Furthermore, this technique is used in some of the swarm-based evolutionary algorithms. In this paper, we have used some effective adaptations of bat algorithm with the previous mentioned approach to effectively handle the many objective problems. Moreover, we have called this algorithm as many objective bat algorithm (MaOBAT). This algorithm is a biologically inspired algorithm, which uses echolocation power of micro bats. Each bat represents a complete solution, which can be evaluated based on the problem specific fitness function and then based on the dominance relationship, non-dominated solutions are selected. In proposed MaOBAT, dominance rank is used as dominance relationship (dominance rank of a solution means by how many other solutions a solution dominated). In our proposed strategy, dynamically allocated set of reference points are used, allowing the algorithm to have good convergence and high diversity pareto fronts (PF). The experimental results show that the proposed algorithm has significant advantages over several state-of-the-art algorithms in terms of the quality of the solution.


Assuntos
Heurística Computacional , Redes Neurais de Computação
14.
Genomics ; 112(5): 3207-3217, 2020 09.
Artigo em Inglês | MEDLINE | ID: mdl-32526247

RESUMO

Cancer subtype stratification, which may help to make a better decision in treating cancerous patients, is one of the most crucial and challenging problems in cancer studies. To this end, various computational methods such as Feature selection, which enhances the accuracy of the classification and is an NP-Hard problem, have been proposed. However, the performance of the applied methods is still low and can be increased by the state-of-the-art and efficient methods. We used 11 efficient and popular meta-heuristic algorithms including WCC, LCA, GA, PSO, ACO, ICA, LA, HTS, FOA, DSOS and CUK along with SVM classifier to stratify human breast cancer molecular subtypes using mRNA and micro-RNA expression data. The applied algorithms select 186 mRNAs and 116 miRNAs out of 9692 mRNAs and 489 miRNAs, respectively. Although some of the selected mRNAs and miRNAs are common in different algorithms results, six miRNAs including miR-190b, miR-18a, miR-301a, miR-34c-5p, miR-18b, and miR-129-5p were selected by equal or more than three different algorithms. Further, six mRNAs, including HAUS6, LAMA2, TSPAN33, PLEKHM3, GFRA3, and DCBLD2, were chosen through two different algorithms. We have reported these miRNAs and mRNAs as important diagnostic biomarkers to the stratification of breast cancer subtypes. By investigating the literature, it is also observed that most of our reported mRNAs and miRNAs have been proposed and introduced as biomarkers in cancer subtypes stratification.


Assuntos
Algoritmos , Neoplasias da Mama/classificação , MicroRNAs/metabolismo , RNA Mensageiro/metabolismo , Biomarcadores Tumorais/genética , Neoplasias da Mama/genética , Neoplasias da Mama/metabolismo , Heurística Computacional , Feminino , Humanos , Máquina de Vetores de Suporte
15.
Evol Comput ; 28(3): 379-404, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-31295020

RESUMO

This article presents a method to generate diverse and challenging new test instances for continuous black-box optimization. Each instance is represented as a feature vector of exploratory landscape analysis measures. By projecting the features into a two-dimensional instance space, the location of existing test instances can be visualized, and their similarities and differences revealed. New instances are generated through genetic programming which evolves functions with controllable characteristics. Convergence to selected target points in the instance space is used to drive the evolutionary process, such that the new instances span the entire space more comprehensively. We demonstrate the method by generating two-dimensional functions to visualize its success, and ten-dimensional functions to test its scalability. We show that the method can recreate existing test functions when target points are co-located with existing functions, and can generate new functions with entirely different characteristics when target points are located in empty regions of the instance space. Moreover, we test the effectiveness of three state-of-the-art algorithms on the new set of instances. The results demonstrate that the new set is not only more diverse than a well-known benchmark set, but also more challenging for the tested algorithms. Hence, the method opens up a new avenue for developing test instances with controllable characteristics, necessary to expose the strengths and weaknesses of algorithms, and drive algorithm development.


Assuntos
Algoritmos , Benchmarking/métodos , Benchmarking/estatística & dados numéricos , Evolução Biológica , Biologia Computacional , Heurística Computacional , Análise de Componente Principal , Probabilidade , Software
16.
IEEE Trans Cybern ; 50(9): 4053-4065, 2020 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-31295135

RESUMO

The rapid development of online social networks not only enables prompt and convenient dissemination of desirable information but also incurs fast and wide propagation of undesirable information. A common way to control the spread of pollutants is to block some nodes, but such a strategy may affect the service quality of a social network and leads to a high control cost if too many nodes are blocked. This paper considers the node selection problem as a biobjective optimization problem to find a subset of nodes to be blocked so that the effect of the control is maximized while the cost of the control is minimized. To solve this problem, we design an ant colony optimization algorithm with an adaptive dimension size selection under the multiobjective evolutionary algorithm framework based on decomposition (MOEA/D-ADACO). The proposed algorithm divides the biobjective problem into a set of single-objective subproblems and each ant takes charge of optimizing one subproblem. Moreover, two types of pheromone and heuristic information are incorporated into MOEA/D-ADACO, that is, pheromone and heuristic information of dimension size selection and that of node selection. While constructing solutions, the ants first determine the dimension size according to the former type of pheromone and heuristic information. Then, the ants select a specific number of nodes to build solutions according to the latter type of pheromone and heuristic information. Experiments conducted on a set of real-world online social networks confirm that the proposed biobjective optimization model and the developed MOEA/D-ADACO are promising for the pollutant spreading control.


Assuntos
Disseminação de Informação , Modelos Biológicos , Rede Social , Algoritmos , Heurística Computacional , Poluentes Ambientais , Internet , Modelos Estatísticos , Feromônios
17.
Br J Math Stat Psychol ; 73(3): 375-396, 2020 11.
Artigo em Inglês | MEDLINE | ID: mdl-31512759

RESUMO

Most partitioning methods used in psychological research seek to produce homogeneous groups (i.e., groups with low intra-group dissimilarity). However, there are also applications where the goal is to provide heterogeneous groups (i.e., groups with high intra-group dissimilarity). Examples of these anticlustering contexts include construction of stimulus sets, formation of student groups, assignment of employees to project work teams, and assembly of test forms from a bank of items. Unfortunately, most commercial software packages are not equipped to accommodate the objective criteria and constraints that commonly arise for anticlustering problems. Two important objective criteria for anticlustering based on information in a dissimilarity matrix are: a diversity measure based on within-cluster sums of dissimilarities; and a dispersion measure based on the within-cluster minimum dissimilarities. In many instances, it is possible to find a partition that provides a large improvement in one of these two criteria with little (or no) sacrifice in the other criterion. For this reason, it is of significant value to explore the trade-offs that arise between these two criteria. Accordingly, the key contribution of this paper is the formulation of a bicriterion optimization problem for anticlustering based on the diversity and dispersion criteria, along with heuristics to approximate the Pareto efficient set of partitions. A motivating example and computational study are provided within the framework of test assembly.


Assuntos
Análise por Conglomerados , Modelos Estatísticos , Psicologia/estatística & dados numéricos , Algoritmos , Heurística Computacional , Simulação por Computador , Avaliação Educacional/estatística & dados numéricos , Humanos , Testes Neuropsicológicos/estatística & dados numéricos , Psicometria/estatística & dados numéricos
18.
Evol Comput ; 28(3): 437-461, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-31120773

RESUMO

Selection hyper-heuristics (HHs) are randomised search methodologies which choose and execute heuristics during the optimisation process from a set of low-level heuristics. A machine learning mechanism is generally used to decide which low-level heuristic should be applied in each decision step. In this article, we analyse whether sophisticated learning mechanisms are always necessary for HHs to perform well. To this end we consider the most simple HHs from the literature and rigorously analyse their performance for the LeadingOnes benchmark function. Our analysis shows that the standard Simple Random, Permutation, Greedy, and Random Gradient HHs show no signs of learning. While the former HHs do not attempt to learn from the past performance of low-level heuristics, the idea behind the Random Gradient HH is to continue to exploit the currently selected heuristic as long as it is successful. Hence, it is embedded with a reinforcement learning mechanism with the shortest possible memory. However, the probability that a promising heuristic is successful in the next step is relatively low when perturbing a reasonable solution to a combinatorial optimisation problem. We generalise the "simple" Random Gradient HH so success can be measured over a fixed period of time τ, instead of a single iteration. For LeadingOnes we prove that the Generalised Random Gradient (GRG) HH can learn to adapt the neighbourhood size of Randomised Local Search to optimality during the run. As a result, we prove it has the best possible performance achievable with the low-level heuristics (Randomised Local Search with different neighbourhood sizes), up to lower-order terms. We also prove that the performance of the HH improves as the number of low-level local search heuristics to choose from increases. In particular, with access to k low-level local search heuristics, it outperforms the best-possible algorithm using any subset of the k heuristics. Finally, we show that the advantages of GRG over Randomised Local Search and Evolutionary Algorithms using standard bit mutation increase if the anytime performance is considered (i.e., the performance gap is larger if approximate solutions are sought rather than exact ones). Experimental analyses confirm these results for different problem sizes (up to n=108) and shed some light on the best choices for the parameter τ in various situations.


Assuntos
Algoritmos , Heurística Computacional , Benchmarking/estatística & dados numéricos , Evolução Biológica , Simulação por Computador , Aprendizado de Máquina , Ferramenta de Busca
19.
IEEE/ACM Trans Comput Biol Bioinform ; 17(4): 1430-1439, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-30418914

RESUMO

Protein tertiary structure prediction is an important open challenge in bioinformatics and requires effective methods to accurately evaluate the quality of protein 3-D models generated computationally. Many quality assessment (QA) methods have been proposed over the past three decades. However, the accuracy or robustness is unsatisfactory for practical applications. In this paper, two new heuristic QA methods are proposed: MUfoldQA_S and MUfoldQA_C. The MUfoldQA_S is a quasi-single-model QA method that assesses the model quality based on the known protein structures with similar sequences. This algorithm can be directly applied to protein fragments without the necessity of building a full structural model. A BLOSUM-based heuristic is also introduced to help differentiate accurate templates from poor ones. In MUfoldQA_C, the ideas from MUfoldQA_S were combined with the consensus approach to create a multi-model QA method that could also utilize information from existing reference models and have demonstrated improved performance. Extensive experimental results of these two methods have shown significant improvement over existing methods. In addition, both methods have been blindly tested in the CASP12 world-wide competition in the protein structure prediction field and ranked as top performers in their respective categories.


Assuntos
Biologia Computacional , Heurística Computacional , Modelos Moleculares , Estrutura Terciária de Proteína , Software , Algoritmos , Biologia Computacional/métodos , Biologia Computacional/normas , Proteínas/química
20.
IEEE/ACM Trans Comput Biol Bioinform ; 17(4): 1253-1261, 2020.
Artigo em Inglês | MEDLINE | ID: mdl-30403637

RESUMO

Epistasis learning, which is aimed at detecting associations between multiple Single Nucleotide Polymorphisms (SNPs) and complex diseases, has gained increasing attention in genome wide association studies. Although much work has been done on mapping the SNPs underlying complex diseases, there is still difficulty in detecting epistatic interactions due to the lack of heuristic information to expedite the search process. In this study, a method EACO is proposed to detect epistatic interactions based on the ant colony optimization (ACO) algorithm, the highlights of which are the introduced heuristic information, fitness function, and a candidate solutions filtration strategy. The heuristic information multi-SURF* is introduced into EACO for identifying epistasis, which is incorporated into ant-decision rules to guide the search with linear time. Two functionally complementary fitness functions, mutual information and the Gini index, are combined to effectively evaluate the associations between SNP combinations and the phenotype. Furthermore, a strategy for candidate solutions filtration is provided to adaptively retain all optimal solutions which yields a more accurate way for epistasis searching. Experiments of EACO, as well as three ACO based methods (AntEpiSeeker, MACOED, and epiACO) and four commonly used methods (BOOST, SNPRuler, TEAM, and epiMODE) are performed on both simulation data sets and a real data set of age-related macular degeneration. Results indicate that EACO is promising in identifying epistasis.


Assuntos
Algoritmos , Biologia Computacional/métodos , Heurística Computacional , Epistasia Genética/genética , Estudo de Associação Genômica Ampla , Humanos , Degeneração Macular/genética
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...