Your browser doesn't support javascript.
loading
: 20 | 50 | 100
1 - 20 de 29
2.
Trends Hear ; 27: 23312165231211437, 2023.
Article En | MEDLINE | ID: mdl-37990543

Preference for noise reduction (NR) strength differs between individuals. The purpose of this study was (1) to investigate whether hearing loss influences this preference, (2) to find the number of distinct settings required to classify participants in similar groups based on their preference for NR strength, and (3) to estimate the number of paired comparisons needed to predict to which preference group a participant belongs. A paired comparison paradigm was used in which participants listened to pairs of speech-in-noise stimuli processed by NR with 10 different strength settings. Participants indicated their preferred sound sample. The 30 participants were divided into three groups according to hearing status (normal hearing, mild hearing loss, and moderate hearing loss). The results showed that (1) participants with moderate hearing loss preferred stronger NR than participants with normal hearing; (2) cluster analysis based solely on the preference for NR strength showed that the data could be described well by dividing the participants into three preference clusters; (3) the appropriate cluster membership could be found with 15 paired comparisons. We conclude that on average, a higher hearing loss is related to a preference for stronger NR, at least for our NR algorithm and our participants. The results show that it might be possible to use a limited set of pre-set NR strengths that can be chosen clinically. For our NR one might use three settings: no NR, intermediate NR, and strong NR. Paired comparisons might be used to find the optimal one of the three settings.


Deafness , Hearing Aids , Hearing Loss, Sensorineural , Hearing Loss , Speech Perception , Humans , Hearing Loss, Sensorineural/diagnosis , Hearing Loss/diagnosis , Hearing
3.
Cell Rep Methods ; 3(8): 100560, 2023 08 28.
Article En | MEDLINE | ID: mdl-37671023

In protein design, the energy associated with a huge number of sequence-conformer perturbations has to be routinely estimated. Hence, enhancing the throughput and accuracy of these energy calculations can profoundly improve design success rates and enable tackling more complex design problems. In this work, we explore the possibility of tensorizing the energy calculations and apply them in a protein design framework. We use this framework to design enhanced proteins with anti-cancer and radio-tracing functions. Particularly, we designed multispecific binders against ligands of the epidermal growth factor receptor (EGFR), where the tested design could inhibit EGFR activity in vitro and in vivo. We also used this method to design high-affinity Cu2+ binders that were stable in serum and could be readily loaded with copper-64 radionuclide. The resulting molecules show superior functional properties for their respective applications and demonstrate the generalizable potential of the described protein design approach.


Copper Radioisotopes , ErbB Receptors , Eye, Artificial , Orthotic Devices , Phosphorylation
4.
Sleep Med ; 98: 9-12, 2022 10.
Article En | MEDLINE | ID: mdl-35764010

OBJECTIVE: We have used an obstructive apnea index of ≥3 as treatment indication for infants with Robin sequence (RS), while the obstructive apnea-hypopnea index (OAHI) and a threshold of ≥5 is often used internationally. We wanted to know whether these two result in similar indications, and what the interobserver variability is with either asessement. METHODS: Twenty lab-based overnight sleep recordings from infants with isolated RS (median age: 7 days, range 2-38) were scored based on the 2020 American Academy of Sleep Medicine guidelines, including or excluding obstructive hypopneas. RESULTS: Median obstructive apnea index (OAI) was 18 (interquartile range: 7.6-38) including only apneas, and 35 (18-54) if obstructive hypopneas were also considered as respiratory events (OAHI). Obstructive sleep apnea (OSA) severity was re-classified from moderate to severe for two infants when obstructive hypopneas were also considered, but this did not lead to a change in clinical treatment decisions for either infant. Median interobserver agreement was 0.86 (95% CI 0.70-0.94) for the OAI, and 0.60 (0.05-0.84) for the OAHI. CONCLUSION: Inclusion of obstructive hypopneas when assessing OSA severity in RS infants doubled the obstructive event rate, but impaired interobserver agreement and would not have changed clinical management.


Physicians , Pierre Robin Syndrome , Sleep Apnea, Obstructive , Child , Humans , Infant , Pierre Robin Syndrome/complications , Polysomnography , Sleep
5.
Mol Cell Proteomics ; 19(12): 2157-2168, 2020 12.
Article En | MEDLINE | ID: mdl-33067342

Cross-linking MS (XL-MS) has been recognized as an effective source of information about protein structures and interactions. In contrast to regular peptide identification, XL-MS has to deal with a quadratic search space, where peptides from every protein could potentially be cross-linked to any other protein. To cope with this search space, most tools apply different heuristics for search space reduction. We introduce a new open-source XL-MS database search algorithm, OpenPepXL, which offers increased sensitivity compared with other tools. OpenPepXL searches the full search space of an XL-MS experiment without using heuristics to reduce it. Because of efficient data structures and built-in parallelization OpenPepXL achieves excellent runtimes and can also be deployed on large compute clusters and cloud services while maintaining a slim memory footprint. We compared OpenPepXL to several other commonly used tools for identification of noncleavable labeled and label-free cross-linkers on a diverse set of XL-MS experiments. In our first comparison, we used a data set from a fraction of a cell lysate with a protein database of 128 targets and 128 decoys. At 5% FDR, OpenPepXL finds from 7% to over 50% more unique residue pairs (URPs) than other tools. On data sets with available high-resolution structures for cross-link validation OpenPepXL reports from 7% to over 40% more structurally validated URPs than other tools. Additionally, we used a synthetic peptide data set that allows objective validation of cross-links without relying on structural information and found that OpenPepXL reports at least 12% more validated URPs than other tools. It has been built as part of the OpenMS suite of tools and supports Windows, macOS, and Linux operating systems. OpenPepXL also supports the MzIdentML 1.2 format for XL-MS identification results. It is freely available under a three-clause BSD license at https://openms.org/openpepxl.


Cross-Linking Reagents/chemistry , Peptides/analysis , Software , Algorithms , Amino Acid Sequence , Databases, Protein , HEK293 Cells , Humans , Mass Spectrometry , Models, Molecular , Peptides/chemistry , Ribosomes/metabolism
6.
Diagn Pathol ; 15(1): 130, 2020 Oct 23.
Article En | MEDLINE | ID: mdl-33097073

BACKGROUND: The conventional method for the diagnosis of malaria parasites is the microscopic examination of stained blood films, which is time consuming and requires expertise. We introduce computer-based image segmentation and life stage classification with a random forest classifier. Segmentation and stage classification are performed on a large dataset of malaria parasites with ground truth labels provided by experts. METHODS: We made use of Giemsa stained images obtained from the blood of 16 patients infected with Plasmodium falciparum. Experts labeled the parasite types from each of the images. We applied a two-step approach: image segmentation followed by life stage classification. In segmentation, we classified each pixel as a parasite or non-parasite pixel using a random forest classifier. Performance was evaluated with classification accuracy, Dice coefficient and free-response receiver operating characteristic (FROC) analysis. In life stage classification, we classified each of the segmented objects into one of 8 classes: 6 parasite life stages, early ring, late ring or early trophozoite, mid trophozoite, early schizont, late schizont or segmented, and two other classes, white blood cell or debris. RESULTS: Our segmentation method gives an average cross-validated Dice coefficient of 0.82 which is a 13% improvement compared to the Otsu method. The Otsu method achieved a True Positive Fraction (TPF) of 0.925 at the expense of a False Positive Rate (FPR) of 2.45. At the same TPF of 0.925, our method achieved an FPR of 0.92, an improvement by more than a factor two. We find that inclusion of average intensity of the whole image as feature for the random forest considerably improves segmentation performance. We obtain an overall accuracy of 58.8% when classifying all life stages. Stages are mostly confused with their neighboring stages. When we reduce the life stages to ring, trophozoite and schizont only, we obtain an accuracy of 82.7%. CONCLUSION: Pixel classification gives better segmentation performance than the conventional Otsu method. Effects of staining and background variations can be reduced with the inclusion of average intensity features. The proposed method and data set can be used in the development of automatic tools for the detection and stage classification of malaria parasites. The data set is publicly available as a benchmark for future studies.


Image Processing, Computer-Assisted/methods , Malaria, Falciparum/diagnosis , Plasmodium falciparum , Algorithms , Humans , Life Cycle Stages , Malaria, Falciparum/blood , Plasmodium falciparum/growth & development
7.
J Proteome Res ; 19(3): 1060-1072, 2020 03 06.
Article En | MEDLINE | ID: mdl-31975601

Accurate protein inference in the presence of shared peptides is still one of the key problems in bottom-up proteomics. Most protein inference tools employing simple heuristic inference strategies are efficient but exhibit reduced accuracy. More advanced probabilistic methods often exhibit better inference quality but tend to be too slow for large data sets. Here, we present a novel protein inference method, EPIFANY, combining a loopy belief propagation algorithm with convolution trees for efficient processing of Bayesian networks. We demonstrate that EPIFANY combines the reliable protein inference of Bayesian methods with significantly shorter runtimes. On the 2016 iPRG protein inference benchmark data, EPIFANY is the only tested method that finds all true-positive proteins at a 5% protein false discovery rate (FDR) without strict prefiltering on the peptide-spectrum match (PSM) level, yielding an increase in identification performance (+10% in the number of true positives and +14% in partial AUC) compared to previous approaches. Even very large data sets with hundreds of thousands of spectra (which are intractable with other Bayesian and some non-Bayesian tools) can be processed with EPIFANY within minutes. The increased inference quality including shared peptides results in better protein inference results and thus increased robustness of the biological hypotheses generated. EPIFANY is available as open-source software for all major platforms at https://OpenMS.de/epifany.


Algorithms , Proteomics , Bayes Theorem , Databases, Protein , Proteins , Software
8.
Front Pharmacol ; 10: 1384, 2019.
Article En | MEDLINE | ID: mdl-31849651

In synergy studies, one focuses on compound combinations that promise a synergistic or antagonistic effect. With the help of high-throughput techniques, a huge amount of compound combinations can be screened and filtered for suitable candidates for a more detailed analysis. Those promising candidates are chosen based on the deviance between a measured response and an expected non-interactive response. A non-interactive response is based on a principle of no interaction, such as Loewe Additivity or Bliss Independence. In a previous study, we introduced, an explicit formulation of the hitherto implicitly defined Loewe Additivity, the so-called Explicit Mean Equation. In the current study we show that this Explicit Mean Equation outperforms the original implicit formulation of Loewe Additivity and Bliss Independence when measuring synergy in terms of the deviance between measured and expected response, called the lack-of-fit. Further, we show that computing synergy as lack-of-fit outperforms a parametric approach. We show this on two datasets of compound combinations that are categorized into synergistic, non-interactive, and antagonistic.

9.
PLoS One ; 14(5): e0216559, 2019.
Article En | MEDLINE | ID: mdl-31071186

RATIONALE & OBJECTIVE: Early prediction of chronic kidney disease (CKD) progression to end-stage kidney disease (ESKD) currently use Cox models including baseline estimated glomerular filtration rate (eGFR) only. Alternative approaches include a Cox model that includes eGFR slope determined over a baseline period of time, a Cox model with time varying GFR, or a joint modeling approach. We studied if these more complex approaches may further improve ESKD prediction. STUDY DESIGN: Prospective cohort. SETTING & PARTICIPANTS: We re-used data from two CKD cohorts including patients with baseline eGFR >30ml/min per 1.73m2. MASTERPLAN (N = 505; 55 ESKD events) was used as development dataset, and NephroTest (N = 1385; 72 events) for validation. PREDICTORS: All models included age, sex, eGFR, and albuminuria, known prognostic markers for ESKD. ANALYTICAL APPROACH: We trained the models on the MASTERPLAN data and determined discrimination and calibration for each model at 2 years follow-up for a prediction horizon of 2 years in the NephroTest cohort. We benchmarked the predictive performance against the Kidney Failure Risk Equation (KFRE). RESULTS: The C-statistics for the KFRE was 0.94 (95%CI 0.86 to 1.01). Performance was similar for the Cox model with time-varying eGFR (0.92 [0.84 to 0.97]), eGFR (0.95 [0.90 to 1.00]), and the joint model 0.91 [0.87 to 0.96]). The Cox model with eGFR slope showed the best calibration. CONCLUSION: In the present studies, where the outcome was rare and follow-up data was highly complete, the joint models did not offer improvement in predictive performance over more traditional approaches such as a survival model with time-varying eGFR, or a model with eGFR slope.


Kidney Failure, Chronic/diagnosis , Models, Statistical , Renal Insufficiency, Chronic/complications , Risk Assessment/methods , Disease Progression , Female , Glomerular Filtration Rate , Humans , Kidney Failure, Chronic/etiology , Kidney Function Tests , Male , Middle Aged , Prognosis , Prospective Studies
10.
Front Pharmacol ; 9: 31, 2018.
Article En | MEDLINE | ID: mdl-29467650

High-throughput techniques allow for massive screening of drug combinations. To find combinations that exhibit an interaction effect, one filters for promising compound combinations by comparing to a response without interaction. A common principle for no interaction is Loewe Additivity which is based on the assumption that no compound interacts with itself and that two doses from different compounds having the same effect are equivalent. It then should not matter whether a component is replaced by the other or vice versa. We call this assumption the Loewe Additivity Consistency Condition (LACC). We derive explicit and implicit null reference models from the Loewe Additivity principle that are equivalent when the LACC holds. Of these two formulations, the implicit formulation is the known General Isobole Equation (Loewe, 1928), whereas the explicit one is the novel contribution. The LACC is violated in a significant number of cases. In this scenario the models make different predictions. We analyze two data sets of drug screening that are non-interactive (Cokol et al., 2011; Yadav et al., 2015) and show that the LACC is mostly violated and Loewe Additivity not defined. Further, we compare the measurements of the non-interactive cases of both data sets to the theoretical null reference models in terms of bias and mean squared error. We demonstrate that the explicit formulation of the null reference model leads to smaller mean squared errors than the implicit one and is much faster to compute.

11.
Sci Rep ; 8(1): 1507, 2018 01 24.
Article En | MEDLINE | ID: mdl-29367629

The visual system is able to recognize body motion from impoverished stimuli. This requires combining stimulus information with visual priors. We present a new visual illusion showing that one of these priors is the assumption that bodies are typically illuminated from above. A change of illumination direction from above to below flips the perceived locomotion direction of a biological motion stimulus. Control experiments show that the underlying mechanism is different from shape-from-shading and directly combines information about body motion with a lighting-from-above prior. We further show that the illusion is critically dependent on the intrinsic luminance gradients of the most mobile parts of the moving body. We present a neural model with physiologically plausible mechanisms that accounts for the illusion and shows how the illumination prior might be encoded within the visual pathway. Our experiments demonstrate, for the first time, a direct influence of illumination priors in high-level motion vision.


Illusions , Lighting/methods , Motion Perception , Visual Pathways/physiology , Humans , Models, Neurological
12.
Bioinformatics ; 34(5): 803-811, 2018 03 01.
Article En | MEDLINE | ID: mdl-29069283

Motivation: Computational models in biology are frequently underdetermined, due to limits in our capacity to measure biological systems. In particular, mechanistic models often contain parameters whose values are not constrained by a single type of measurement. It may be possible to achieve better model determination by combining the information contained in different types of measurements. Bayesian statistics provides a convenient framework for this, allowing a quantification of the reduction in uncertainty with each additional measurement type. We wished to explore whether such integration is feasible and whether it can allow computational models to be more accurately determined. Results: We created an ordinary differential equation model of cell cycle regulation in budding yeast and integrated data from 13 different studies covering different experimental techniques. We found that for some parameters, a single type of measurement, relative time course mRNA expression, is sufficient to constrain them. Other parameters, however, were only constrained when two types of measurements were combined, namely relative time course and absolute transcript concentration. Comparing the estimates to measurements from three additional, independent studies, we found that the degradation and transcription rates indeed matched the model predictions in order of magnitude. The predicted translation rate was incorrect however, thus revealing a deficiency in the model. Since this parameter was not constrained by any of the measurement types separately, it was only possible to falsify the model when integrating multiple types of measurements. In conclusion, this study shows that integrating multiple measurement types can allow models to be more accurately determined. Availability and implementation: The models and files required for running the inference are included in the Supplementary information. Contact: l.wessels@nki.nl. Supplementary information: Supplementary data are available at Bioinformatics online.


Computational Biology/methods , Models, Biological , Bayes Theorem , Saccharomycetales/genetics , Saccharomycetales/metabolism
13.
BMC Syst Biol ; 10(1): 100, 2016 10 21.
Article En | MEDLINE | ID: mdl-27769238

BACKGROUND: Computational models in biology are characterized by a large degree of uncertainty. This uncertainty can be analyzed with Bayesian statistics, however, the sampling algorithms that are frequently used for calculating Bayesian statistical estimates are computationally demanding, and each algorithm has unique advantages and disadvantages. It is typically unclear, before starting an analysis, which algorithm will perform well on a given computational model. RESULTS: We present BCM, a toolkit for the Bayesian analysis of Computational Models using samplers. It provides efficient, multithreaded implementations of eleven algorithms for sampling from posterior probability distributions and for calculating marginal likelihoods. BCM includes tools to simplify the process of model specification and scripts for visualizing the results. The flexible architecture allows it to be used on diverse types of biological computational models. In an example inference task using a model of the cell cycle based on ordinary differential equations, BCM is significantly more efficient than existing software packages, allowing more challenging inference problems to be solved. CONCLUSIONS: BCM represents an efficient one-stop-shop for computational modelers wishing to use sampler-based Bayesian statistics.


Computational Biology/methods , Computer Simulation , Software , Algorithms , Bayes Theorem , Kinetics , Models, Biological
14.
BMC Bioinformatics ; 15: 342, 2014 Oct 21.
Article En | MEDLINE | ID: mdl-25336059

BACKGROUND: Millions of cells are present in thousands of images created in high-throughput screening (HTS). Biologists could classify each of these cells into a phenotype by visual inspection. But in the presence of millions of cells this visual classification task becomes infeasible. Biologists train classification models on a few thousand visually classified example cells and iteratively improve the training data by visual inspection of the important misclassified phenotypes. Classification methods differ in performance and performance evaluation time. We present a comparative study of computational performance of gentle boosting, joint boosting CellProfiler Analyst (CPA), support vector machines (linear and radial basis function) and linear discriminant analysis (LDA) on two data sets of HT29 and HeLa cancer cells. RESULTS: For the HT29 data set we find that gentle boosting, SVM (linear) and SVM (RBF) are close in performance but SVM (linear) is faster than gentle boosting and SVM (RBF). For the HT29 data set the average performance difference between SVM (RBF) and SVM (linear) is 0.42 %. For the HeLa data set we find that SVM (RBF) outperforms other classification methods and is on average 1.41 % better in performance than SVM (linear). CONCLUSIONS: Our study proposes SVM (linear) for iterative improvement of the training data and SVM (RBF) for the final classifier to classify all unlabeled cells in the whole data set.


Computational Biology/methods , High-Throughput Screening Assays/methods , Molecular Imaging , Discriminant Analysis , HT29 Cells , HeLa Cells , Humans , Linear Models , Support Vector Machine
15.
J Proteome Res ; 13(9): 3871-80, 2014 Sep 05.
Article En | MEDLINE | ID: mdl-25102230

A challenge in proteomics is that many observations are missing with the probability of missingness increasing as abundance decreases. Adjusting for this informative missingness is required to assess accurately which proteins are differentially abundant. We propose an empirical Bayesian random censoring threshold (EBRCT) model that takes the pattern of missingness in account in the identification of differential abundance. We compare our model with four alternatives, one that considers the missing values as missing completely at random (MCAR model), one with a fixed censoring threshold for each protein species (fixed censoring model) and two imputation models, k-nearest neighbors (IKNN) and singular value thresholding (SVTI). We demonstrate that the EBRCT model bests all alternative models when applied to the CPTAC study 6 benchmark data set. The model is applicable to any label-free peptide or protein quantification pipeline and is provided as an R script.


Bayes Theorem , Models, Statistical , Proteomics/methods , Mass Spectrometry , Proteins/analysis , Proteins/chemistry , ROC Curve
16.
Front Comput Neurosci ; 7: 111, 2013.
Article En | MEDLINE | ID: mdl-24032015

Despite anecdotal reports that humans retain acquired motor skills for many years, if not a lifetime, long-term memory of motor skills has received little attention. While numerous neuroimaging studies showed practice-induced cortical plasticity, the behavioral correlates, what is retained and also what is forgotten, are little understood. This longitudinal case study on four subjects presents detailed kinematic analyses of humans practicing a bimanual polyrhythmic task over 2 months with retention tests after 6 months and, for two subjects, after 8 years. Results showed that individuals not only retained the task, but also reproduced their individual "style" of performance, even after 8 years. During practice, variables such as the two hands' frequency ratio and relative phase, changed at different rates, indicative of multiple time scales of neural processes. Frequency leakage across hands, reflecting intermanual crosstalk, attenuated at a significantly slower rate and was the only variable not maintained after 8 years. Complementing recent findings on neuroplasticity in gray and white matter, our study presents new behavioral evidence that highlights the multi-scale process of practice-induced changes and its remarkable persistence. Results suggest that motor memory may comprise not only higher-level task variables but also individual kinematic signatures.

17.
Vision Res ; 91: 84-92, 2013 Oct 18.
Article En | MEDLINE | ID: mdl-23942288

Participants viewed pairs of ellipses differing in size and aspect ratio (short axis divided by long axis length). In separate experiments with identical stimuli participants were asked to indicate the larger or the more circular ellipse of the pair. First, the size discrimination thresholds decreased with an increase in the circularity of the ellipses. Second, size discrimination thresholds were lower than aspect ratio thresholds, except for the circle and more elongated ellipses where both were similar. Third, there was also an effect of size on aspect ratio discrimination such that larger stimuli appeared more circular.


Discrimination, Psychological/physiology , Form Perception/physiology , Size Perception/physiology , Adult , Analysis of Variance , Female , Humans , Male , Photic Stimulation/methods , Psychometrics , Sensory Thresholds/physiology
18.
PLoS Comput Biol ; 8(4): e1002450, 2012.
Article En | MEDLINE | ID: mdl-22496630

Activity regulated neurotransmission shapes the computational properties of a neuron and involves the concerted action of many proteins. Classical, intuitive working models often assign specific proteins to specific steps in such complex cellular processes, whereas modern systems theories emphasize more integrated functions of proteins. To test how often synaptic proteins participate in multiple steps in neurotransmission we present a novel probabilistic method to analyze complex functional data from genetic perturbation studies on neuronal secretion. Our method uses a mixture of probabilistic principal component analyzers to cluster genetic perturbations on two distinct steps in synaptic secretion, vesicle priming and fusion, and accounts for the poor standardization between different studies. Clustering data from 121 perturbations revealed that different perturbations of a given protein are often assigned to different steps in the release process. Furthermore, vesicle priming and fusion are inversely correlated for most of those perturbations where a specific protein domain was mutated to create a gain-of-function variant. Finally, two different modes of vesicle release, spontaneous and action potential evoked release, were affected similarly by most perturbations. This data suggests that the presynaptic protein network has evolved as a highly integrated supramolecular machine, which is responsible for both spontaneous and activity induced release, with a group of core proteins using different domains to act on multiple steps in the release process.


Models, Neurological , Nerve Tissue Proteins/metabolism , Neurons/metabolism , Neurotransmitter Agents/metabolism , Synapses/metabolism , Synaptic Vesicles/metabolism , Animals , Computer Simulation , Humans
19.
Mol Ecol ; 20(6): 1144-54, 2011 Mar.
Article En | MEDLINE | ID: mdl-21255171

Ecological functional genomics, dealing with the responses of organisms to their natural environment is confronted with a complex pattern of variation and a large number of confounding environmental factors. For gene expression studies to provide meaningful information on conditions deviating from normal, a baseline or normal operating range (NOR) response needs to be established which indicates how an organism's transcriptome reacts to naturally varying ecological factors. Here we determine the transcriptional plasticity of a soil arthropod, Folsomia candida, exposed to various natural environments, as part of a first attempt in establishing such a NOR. Animals were exposed to 26 different field soils after which gene expression levels were measured. The main factor found to regulate gene expression was soil-type (sand or clay). Cell homeostasis and DNA replication were affected in collembolans exposed to sandy soil, indicating general stress. Multivariate analysis identified soil fertility as the main factor influencing gene expression. Regarding land-use, only forest soils showed an expression pattern deviating from the others. No significant effect of land-use, agricultural practice or soil type on fitness was observed, but arsenic concentration was negatively correlated with reproductive output. In conclusion, transcriptional responses remained within a limited range across the different land-uses but were significantly affected by soil-type. This may be caused by the contrasting soil physicochemical properties to which F. candida strongly responds. The broad range of conditions over which this soil-living detritivore is able to survive and reproduce, indicates a strategy of high plasticity, which comes with extensive gene expression regulation.


Arthropods/genetics , Ecology/methods , Animals , DNA Replication/genetics , Multivariate Analysis
20.
Bioinformatics ; 25(12): 1484-91, 2009 Jun 15.
Article En | MEDLINE | ID: mdl-19336444

MOTIVATION: To date, there is little knowledge about one of the processes fundamental to the biology of Plasmodium falciparum, gene regulation including transcriptional control. We use noisy threshold models to identify regulatory sequence elements explaining membership to a gene expression cluster where each cluster consists of genes active during the part of the developmental cycle inside a red blood cell. Our approach is both able to capture the combinatorial nature of gene regulation and to incorporate uncertainty about the functionality of putative regulatory sequence elements. RESULTS: We find a characteristic pattern where the most common motifs tend to be absent upstream of genes active in the first half of the cycle and present upstream of genes active in the second half. We find no evidence that motif's score, orientation, location and multiplicity improves prediction of gene expression. Through comparative genome analysis, we find a list of potential transcription factors and their associated motifs. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.


Computational Biology/methods , Erythrocytes/parasitology , Gene Expression Regulation , Plasmodium falciparum/genetics , Animals , DNA, Protozoan/chemistry , Gene Expression Profiling/methods , Genome, Protozoan , Transcription Factors/genetics
...