RESUMO
Historical data from control groups in animal toxicity studies is currently mainly used for comparative purposes to assess validity and robustness of study results. Due to the highly controlled environment in which the studies are performed and the homogeneity of the animal collectives it has been proposed to use the historical data for building so-called virtual control groups, which could replace partly or entirely the concurrent control. This would constitute a substantial contribution to the reduction of animal use in safety studies. Before the concept can be implemented, the prerequisites regarding data collection, curation and statistical evaluation together with a validation strategy need to be identified to avoid any impairment of the study outcome and subsequent consequences for human risk assessment. To further assess and develop the concept of virtual control groups the transatlantic think tank for toxicology (t4) sponsored a workshop with stakeholders from the pharmaceutical and chemical industry, academia, FDA, pharmaceutical, contract research organizations (CROs), and non-governmental organizations in Washington, which took place in March 2023. This report summarizes the current efforts of a European initiative to share, collect and curate animal control data in a centralized database and the first approaches to identify optimal matching criteria between virtual controls and the treatment arms of a study as well as first reflections about strategies for a qualification procedure and potential pitfalls of the concept.
Animal safety studies are usually performed with three groups of animals where increasing amounts of the test chemical are given to the animals and one control group where the animals do not receive the test chemical. The design of such studies, the characteristics of the animals, and the measured parameters are often very similar from study to study. Therefore, it has been suggested that measurement data from the control groups could be reused from study to study to lower the total number of animals per study. This could reduce animal use by up to 25% for such standardized studies. A workshop was held to discuss the pros and cons of such a concept and what would have to be done to implement it without threatening the reliability of the study outcome or the resulting human risk assessment.
Assuntos
Pesquisa , Animais , Grupos Controle , Preparações FarmacêuticasRESUMO
In silico toxicology protocols are meant to support computationally-based assessments using principles that ensure that results can be generated, recorded, communicated, archived, and then evaluated in a uniform, consistent, and reproducible manner. We investigated the availability of in silico models to predict the carcinogenic potential of pregabalin using the ten key characteristics of carcinogens as a framework for organizing mechanistic studies. Pregabalin is a single-species carcinogen producing only one type of tumor, hemangiosarcomas in mice via a nongenotoxic mechanism. The overall goal of this exercise is to test the ability of in silico models to predict nongenotoxic carcinogenicity with pregabalin as a case study. The established mode of action (MOA) of pregabalin is triggered by tissue hypoxia, leading to oxidative stress (KC5), chronic inflammation (KC6), and increased cell proliferation (KC10) of endothelial cells. Of these KCs, in silico models are available only for selected endpoints in KC5, limiting the usefulness of computational tools in prediction of pregabalin carcinogenicity. KC1 (electrophilicity), KC2 (genotoxicity), and KC8 (receptor-mediated effects), for which predictive in silico models exist, do not play a role in this mode of action. Confidence in the overall assessments is considered to be medium to high for KCs 1, 2, 5, 6, 7 (immune system effects), 8, and 10 (cell proliferation), largely due to the high-quality experimental data. In order to move away from dependence on animal data, development of reliable in silico models for prediction of oxidative stress, chronic inflammation, immunosuppression, and cell proliferation will be critical for the ability to predict nongenotoxic compound carcinogenicity.
RESUMO
N-Nitrosamines (NAs) are a class of reactive organic chemicals that humans may be exposed to from environmental sources, food but also impurities in pharmaceutical preparations. Some NAs were identified as DNA-reactive mutagens and many of those have been classified as probable human carcinogens. Beyond high-potency mutagenic carcinogens that need to be strictly controlled, NAs of low potency need to be considered for risk assessment as well. NA impurities and nitrosylated products of active pharmaceutical ingredients (APIs) often arise from production processes or degradation. Most NAs require metabolic activation to ultimately become carcinogens, and their activation can be appropriately described by first-principles computational chemistry approaches. To this end, we treat NA-induced DNA alkylation as a series of subsequent association and dissociation reaction steps that can be calculated stringently by density functional theory (DFT), including α-hydroxylation, proton transfer, hydroxyl elimination, direct SN2/SNAr DNA alkylation, competing hydrolysis and SN1 reactions. Both toxification and detoxification reactions are considered. The activation reactions are modeled by DFT at a high level of theory with an appropriate solvent model to compute Gibbs free energies of the reactions (thermodynamical effects) and activation barriers (kinetic effects). We study congeneric series of aliphatic and cyclic NAs to identify trends. Overall, this work reveals detailed insight into mechanisms of activation for NAs, suggesting that individual steric and electronic factors have directing and rate-determining influence on the formation of carbenium ions as the ultimate pro-mutagens and thus carcinogens. Therefore, an individual risk assessment of NAs is suggested, as exemplified for the complex API-like 4-(N-nitroso-N-methyl)aminoantipyrine which is considered as low-potency NA by in silico prediction.
Assuntos
Nitrosaminas , Humanos , Nitrosaminas/metabolismo , Carcinógenos/metabolismo , Mutagênicos , DNA , Preparações FarmacêuticasRESUMO
mRNA vaccines hold tremendous potential in disease control and prevention for their flexibility with respect to production, application, and design. Recent breakthroughs in mRNA vaccination would have not been possible without major advances in lipid nanoparticles (LNPs) technologies. We developed an LNP containing a novel ionizable cationic lipid, Lipid-1, and three well known excipients. An in silico toxicity hazard assessment for genotoxicity, a genotoxicity assessment, and a dose range finding toxicity study were performed to characterize the safety profile of Lipid-1. The in silico toxicity hazard assessment, utilizing two prediction systems DEREK and Leadscope, did not find any structural alert for mutagenicity and clastogenicity, and prediction in the statistical models were all negative. In addition, applying a read-across approach a structurally very similar compound was tested negative in two in vitro assays confirming the low genotoxicity potential of Lipid-1. A dose range finding toxicity study in rabbits, receiving a single intramuscular injection of either different doses of an mRNA encoding Influenza Hemagglutinin H3 antigen encapsulated in the LNP containing Lipid-1 or the empty LNP, evaluated local tolerance and systemic toxicity during a 2-week observation period. Only rabbits exposed to the vaccine were able to develop a specific IgG response, indicating an appropriate vaccine take. The vaccine was well tolerated up to 250 µg mRNA/injection, which was defined as the No Observed Adverse Effect Level (NOAEL). These results support the use of the LNP containing Lipid-1 as an mRNA delivery system for different vaccine formulations and its deployment into clinical trials.
Assuntos
Lipídeos , Nanopartículas , Animais , Lipídeos/química , Lipídeos/toxicidade , Lipossomos , Nanopartículas/química , Nanopartículas/toxicidade , RNA Mensageiro/genética , CoelhosRESUMO
Acute toxicity in silico models are being used to support an increasing number of application areas including (1) product research and development, (2) product approval and registration as well as (3) the transport, storage and handling of chemicals. The adoption of such models is being hindered, in part, because of a lack of guidance describing how to perform and document an in silico analysis. To address this issue, a framework for an acute toxicity hazard assessment is proposed. This framework combines results from different sources including in silico methods and in vitro or in vivo experiments. In silico methods that can assist the prediction of in vivo outcomes (i.e., LD50) are analyzed concluding that predictions obtained using in silico approaches are now well-suited for reliably supporting assessment of LD50-based acute toxicity for the purpose of GHS classification. A general overview is provided of the endpoints from in vitro studies commonly evaluated for predicting acute toxicity (e.g., cytotoxicity/cytolethality as well as assays targeting specific mechanisms). The increased understanding of pathways and key triggering mechanisms underlying toxicity and the increased availability of in vitro data allow for a shift away from assessments solely based on endpoints such as LD50, to mechanism-based endpoints that can be accurately assessed in vitro or by using in silico prediction models. This paper also highlights the importance of an expert review of all available information using weight-of-evidence considerations and illustrates, using a series of diverse practical use cases, how in silico approaches support the assessment of acute toxicity.
RESUMO
Historically, identifying carcinogens has relied primarily on tumor studies in rodents, which require enormous resources in both money and time. In silico models have been developed for predicting rodent carcinogens but have not yet found general regulatory acceptance, in part due to the lack of a generally accepted protocol for performing such an assessment as well as limitations in predictive performance and scope. There remains a need for additional, improved in silico carcinogenicity models, especially ones that are more human-relevant, for use in research and regulatory decision-making. As part of an international effort to develop in silico toxicological protocols, a consortium of toxicologists, computational scientists, and regulatory scientists across several industries and governmental agencies evaluated the extent to which in silico models exist for each of the recently defined 10 key characteristics (KCs) of carcinogens. This position paper summarizes the current status of in silico tools for the assessment of each KC and identifies the data gaps that need to be addressed before a comprehensive in silico carcinogenicity protocol can be developed for regulatory use.
RESUMO
Hepatotoxicity is one of the most frequently observed adverse effects resulting from exposure to a xenobiotic. For example, in pharmaceutical research and development it is one of the major reasons for drug withdrawals, clinical failures, and discontinuation of drug candidates. The development of faster and cheaper methods to assess hepatotoxicity that are both more sustainable and more informative is critically needed. The biological mechanisms and processes underpinning hepatotoxicity are summarized and experimental approaches to support the prediction of hepatotoxicity are described, including toxicokinetic considerations. The paper describes the increasingly important role of in silico approaches and highlights challenges to the adoption of these methods including the lack of a commonly agreed upon protocol for performing such an assessment and the need for in silico solutions that take dose into consideration. A proposed framework for the integration of in silico and experimental information is provided along with a case study describing how computational methods have been used to successfully respond to a regulatory question concerning non-genotoxic impurities in chemically synthesized pharmaceuticals.
RESUMO
The kidneys, heart and lungs are vital organ systems evaluated as part of acute or chronic toxicity assessments. New methodologies are being developed to predict these adverse effects based on in vitro and in silico approaches. This paper reviews the current state of the art in predicting these organ toxicities. It outlines the biological basis, processes and endpoints for kidney toxicity, pulmonary toxicity, respiratory irritation and sensitization as well as functional and structural cardiac toxicities. The review also covers current experimental approaches, including off-target panels from secondary pharmacology batteries. Current in silico approaches for prediction of these effects and mechanisms are described as well as obstacles to the use of in silico methods. Ultimately, a commonly accepted protocol for performing such assessment would be a valuable resource to expand the use of such approaches across different regulatory and industrial applications. However, a number of factors impede their widespread deployment including a lack of a comprehensive mechanistic understanding, limited in vitro testing approaches and limited in vivo databases suitable for modeling, a limited understanding of how to incorporate absorption, distribution, metabolism, and excretion (ADME) considerations into the overall process, a lack of in silico models designed to predict a safe dose and an accepted framework for organizing the key characteristics of these organ toxicants.
RESUMO
Sharing legacy data from in vivo toxicity studies offers the opportunity to analyze the variability of control groups stratified for strain, age, duration of study, vehicle and other experimental conditions. Historical animal control group data may lead to a repository, which could be used to construct virtual control groups (VCGs) for toxicity studies. VCGs are an established concept in clinical trials, but the idea of replacing living beings with virtual data sets has so far not been introduced into the design of regulatory animal studies. The use of VCGs has the potential of a 25% reduction in animal use by replacing the control group animals with existing randomized data sets. Prerequisites for such an approach are the availability of large and well-structured control data sets as well as thorough statistical evaluations. the foundation of data sharing has been laid within the Innovative Medicines Initiatives projects eTOX and eTRANSAFE. For a proof of principle participating companies have started to collect control group data for subacute (4-week) GLP studies with Wistar rats (the strain preferentially used in Europe) and are characterizing these data for its variability. In a second step, the control group data will be shared among the companies and cross-company variability will be investigated. In a third step, a set of studies will be analyzed to assess whether the use of VCG data would have influenced the outcome of the study compared to the real control group.
Assuntos
Bases de Dados Factuais , Avaliação Pré-Clínica de Medicamentos/métodos , Disseminação de Informação , Projetos de Pesquisa , Testes de Toxicidade/métodos , Bases de ConhecimentoRESUMO
In silico toxicology (IST) approaches to rapidly assess chemical hazard, and usage of such methods is increasing in all applications but especially for regulatory submissions, such as for assessing chemicals under REACH as well as the ICH M7 guideline for drug impurities. There are a number of obstacles to performing an IST assessment, including uncertainty in how such an assessment and associated expert review should be performed or what is fit for purpose, as well as a lack of confidence that the results will be accepted by colleagues, collaborators and regulatory authorities. To address this, a project to develop a series of IST protocols for different hazard endpoints has been initiated and this paper describes the genetic toxicity in silico (GIST) protocol. The protocol outlines a hazard assessment framework including key effects/mechanisms and their relationships to endpoints such as gene mutation and clastogenicity. IST models and data are reviewed that support the assessment of these effects/mechanisms along with defined approaches for combining the information and evaluating the confidence in the assessment. This protocol has been developed through a consortium of toxicologists, computational scientists, and regulatory scientists across several industries to support the implementation and acceptance of in silico approaches.
Assuntos
Modelos Teóricos , Mutagênicos/toxicidade , Projetos de Pesquisa , Toxicologia/métodos , Animais , Simulação por Computador , Humanos , Testes de Mutagenicidade , Medição de RiscoRESUMO
For fasiglifam (TAK875) and its metabolites the substance-specific mechanisms of liver toxicity were studied. Metabolism studies were run to identify a putatively reactive acyl glucuronide metabolite. In vitro cytotoxicity and caspase 3/7 activation were assessed in primary human and dog hepatocytes in 2D and 3D cell culture. Involvement of glutathione (GSH) detoxication system in mediating cytotoxicity was determined by assessing potentiation of cytotoxicity in a GSH depleted in vitro system. In addition, potential mitochondrial liabilities of the compounds were assessed in a whole-cell mitochondrial functional assay. Fasiglifam showed moderate cytotoxicity in human primary hepatocytes in the classical 2D cytotoxicity assays and also in the complex 3D human liver microtissue (hLiMT) after short-term treatment (24 hours or 48 hours) with TC50 values of 56 to 68 µM (adenosine triphosphate endpoint). The long-term treatment for 14 days in the hLiMT resulted in a slight TC50 shift over time of 2.7/3.6 fold lower vs 24-hour treatment indicating possibly a higher risk for cytotoxicity during long-term treatment. Cellular GSH depletion and impairment of mitochondrial function by TAK875 and its metabolites evaluated by Seahorse assay could not be found being involved in DILI reported for TAK875. The acyl glucuronide metabolites of TAK875 have been finally identified to be the dominant reason for liver toxicity.
Assuntos
Benzofuranos/toxicidade , Ácidos Graxos não Esterificados/metabolismo , Fígado/efeitos dos fármacos , Receptores Acoplados a Proteínas G/agonistas , Sulfonas/toxicidade , Animais , Benzofuranos/metabolismo , Células Cultivadas , Cães , Glutationa/metabolismo , Hepatócitos/efeitos dos fármacos , Hepatócitos/metabolismo , Humanos , Microssomos Hepáticos/efeitos dos fármacos , Microssomos Hepáticos/metabolismo , Mitocôndrias Hepáticas/efeitos dos fármacos , Mitocôndrias Hepáticas/metabolismo , Ratos , Receptores Acoplados a Proteínas G/metabolismo , Sulfonas/metabolismoRESUMO
(Quantitative) structure-activity relationship or (Q)SAR predictions of DNA-reactive mutagenicity are important to support both the design of new chemicals and the assessment of impurities, degradants, metabolites, extractables and leachables, as well as existing chemicals. Aromatic N-oxides represent a class of compounds that are often considered alerting for mutagenicity yet the scientific rationale of this structural alert is not clear and has been questioned. Because aromatic N-oxide-containing compounds may be encountered as impurities, degradants and metabolites, it is important to accurately predict mutagenicity of this chemical class. This article analysed a series of publicly available aromatic N-oxide data in search of supporting information. The article also used a previously developed structure-activity relationship (SAR) fingerprint methodology where a series of aromatic N-oxide substructures was generated and matched against public and proprietary databases, including pharmaceutical data. An assessment of the number of mutagenic and non-mutagenic compounds matching each substructure across all sources was used to understand whether the general class or any specific subclasses appear to lead to mutagenicity. This analysis resulted in a downgrade of the general aromatic N-oxide alert. However, it was determined there were enough public and proprietary data to assign the quindioxin and related chemicals as well as benzo[c][1,2,5]oxadiazole 1-oxide subclasses as alerts. The overall results of this analysis were incorporated into Leadscope's expert-rule-based model to enhance its predictive accuracy.
Assuntos
Óxidos N-Cíclicos/química , Dano ao DNA/efeitos dos fármacos , Mutagênicos/química , Relação Quantitativa Estrutura-Atividade , Óxidos N-Cíclicos/toxicidade , Mutagênese/efeitos dos fármacos , Testes de Mutagenicidade , Mutagênicos/toxicidadeRESUMO
The International Council for Harmonization (ICH) M7 guideline describes a hazard assessment process for impurities that have the potential to be present in a drug substance or drug product. In the absence of adequate experimental bacterial mutagenicity data, (Q)SAR analysis may be used as a test to predict impurities' DNA reactive (mutagenic) potential. However, in certain situations, (Q)SAR software is unable to generate a positive or negative prediction either because of conflicting information or because the impurity is outside the applicability domain of the model. Such results present challenges in generating an overall mutagenicity prediction and highlight the importance of performing a thorough expert review. The following paper reviews pharmaceutical and regulatory experiences handling such situations. The paper also presents an analysis of proprietary data to help understand the likelihood of misclassifying a mutagenic impurity as non-mutagenic based on different combinations of (Q)SAR results. This information may be taken into consideration when supporting the (Q)SAR results with an expert review, especially when out-of-domain results are generated during a (Q)SAR evaluation.
Assuntos
Contaminação de Medicamentos , Guias como Assunto , Mutagênicos/classificação , Relação Quantitativa Estrutura-Atividade , Indústria Farmacêutica , Órgãos Governamentais , Mutagênicos/toxicidade , Medição de RiscoRESUMO
The present publication surveys several applications of in silico (i.e., computational) toxicology approaches across different industries and institutions. It highlights the need to develop standardized protocols when conducting toxicity-related predictions. This contribution articulates the information needed for protocols to support in silico predictions for major toxicological endpoints of concern (e.g., genetic toxicity, carcinogenicity, acute toxicity, reproductive toxicity, developmental toxicity) across several industries and regulatory bodies. Such novel in silico toxicology (IST) protocols, when fully developed and implemented, will ensure in silico toxicological assessments are performed and evaluated in a consistent, reproducible, and well-documented manner across industries and regulatory bodies to support wider uptake and acceptance of the approaches. The development of IST protocols is an initiative developed through a collaboration among an international consortium to reflect the state-of-the-art in in silico toxicology for hazard identification and characterization. A general outline for describing the development of such protocols is included and it is based on in silico predictions and/or available experimental data for a defined series of relevant toxicological effects or mechanisms. The publication presents a novel approach for determining the reliability of in silico predictions alongside experimental data. In addition, we discuss how to determine the level of confidence in the assessment based on the relevance and reliability of the information.
Assuntos
Simulação por Computador , Testes de Toxicidade/métodos , Toxicologia/métodos , Animais , HumanosRESUMO
Over the past decades, pharmaceutical companies have conducted a large number of high-quality in vivo repeat-dose toxicity (RDT) studies for regulatory purposes. As part of the eTOX project, a high number of these studies have been compiled and integrated into a database. This valuable resource can be queried directly, but it can be further exploited to build predictive models. As the studies were originally conducted to investigate the properties of individual compounds, the experimental conditions across the studies are highly heterogeneous. Consequently, the original data required normalization/standardization, filtering, categorization and integration to make possible any data analysis (such as building predictive models). Additionally, the primary objectives of the RDT studies were to identify toxicological findings, most of which do not directly translate to in vivo endpoints. This article describes a method to extract datasets containing comparable toxicological properties for a series of compounds amenable for building predictive models. The proposed strategy starts with the normalization of the terms used within the original reports. Then, comparable datasets are extracted from the database by applying filters based on the experimental conditions. Finally, carefully selected profiles of toxicological findings are mapped to endpoints of interest, generating QSAR-like tables. In this work, we describe in detail the strategy and tools used for carrying out these transformations and illustrate its application in a data sample extracted from the eTOX database. The suitability of the resulting tables for developing hazard-predicting models was investigated by building proof-of-concept models for in vivo liver endpoints.
Assuntos
Bases de Dados Factuais , Avaliação Pré-Clínica de Medicamentos/métodos , Efeitos Colaterais e Reações Adversas Relacionados a Medicamentos , Determinação de Ponto Final , Modelos Teóricos , Testes de Toxicidade/métodos , Mineração de Dados , Avaliação Pré-Clínica de Medicamentos/normas , Avaliação Pré-Clínica de Medicamentos/estatística & dados numéricos , Previsões , Disseminação de Informação , Medição de Risco , Testes de Toxicidade/normas , Testes de Toxicidade/estatística & dados numéricosRESUMO
The sharing of legacy preclinical safety data among pharmaceutical companies and its integration with other information sources offers unprecedented opportunities to improve the early assessment of drug safety. Here, we discuss the experience of the eTOX project, which was established through the Innovative Medicines Initiative to explore this possibility.
Assuntos
Avaliação Pré-Clínica de Medicamentos/métodos , Indústria Farmacêutica/métodos , Efeitos Colaterais e Reações Adversas Relacionados a Medicamentos , Disseminação de Informação , Humanos , Medição de Risco/métodosRESUMO
Metabolomics, also often referred as "metabolic profiling," is the systematic profiling of metabolites in biofluids or tissues of organisms and their temporal changes. In the last decade, metabolomics has become more and more popular in drug development, molecular medicine, and other biotechnology fields, since it profiles directly the phenotype and changes thereof in contrast to other "-omics" technologies. The increasing popularity of metabolomics has been possible only due to the enormous development in the technology and bioinformatics fields. In particular, the analytical technologies supporting metabolomics, i.e., NMR, UPLC-MS, and GC-MS, have evolved into sensitive and highly reproducible platforms allowing the determination of hundreds of metabolites in parallel. This chapter describes the best practices of metabolomics as seen today. All important steps of metabolic profiling in drug development and molecular medicine are described in great detail, starting from sample preparation to determining the measurement details of all analytical platforms, and finally to discussing the corresponding specific steps of data analysis.
Assuntos
Espectroscopia de Ressonância Magnética/métodos , Metabolômica/métodos , Cromatografia Gasosa-Espectrometria de Massas , MetabolomaRESUMO
The present study applies a systems biology approach for the in silico predictive modeling of drug toxicity on the basis of high-quality preclinical drug toxicity data with the aim of increasing the mechanistic understanding of toxic effects of compounds at different levels (pathway, cell, tissue, organ). The model development was carried out using 77 compounds for which gene expression data for treated primary human hepatocytes is available in the LINCS database and for which rodent in vivo hepatotoxicity information is available in the eTOX database. The data from LINCS were used to determine the type and number of pathways disturbed by each compound and to estimate the extent of disturbance (network perturbation elasticity), and were used to analyze the correspondence with the in vivo information from eTOX. Predictive models were developed through this integrative analysis, and their specificity and sensitivity were assessed. The quality of the predictions was determined on the basis of the area under the curve (AUC) of plots of true positive vs. false positive rates (ROC curves). The ROC AUC reached values of up to 0.9 (out of 1.0) for some hepatotoxicity endpoints. Moreover, the most frequently disturbed metabolic pathways were determined across the studied toxicants. They included, e.g., mitochondrial beta-oxidation of fatty acids and amino acid metabolism. The process was exemplified by successful predictions on various statins. In conclusion, an entirely new approach linking gene expression alterations to the prediction of complex organ toxicity was developed and evaluated.
Assuntos
Regulação da Expressão Gênica/genética , Hepatócitos/efeitos dos fármacos , Redes e Vias Metabólicas/efeitos dos fármacos , Alternativas aos Testes com Animais , Animais , Bases de Dados Factuais , Avaliação Pré-Clínica de Medicamentos/métodos , Efeitos Colaterais e Reações Adversas Relacionados a Medicamentos/genética , Humanos , Técnicas In Vitro , Fígado/efeitos dos fármacos , Redes e Vias Metabólicas/genética , Modelos Estatísticos , Ratos , Sensibilidade e EspecificidadeRESUMO
Statistical-based and expert rule-based models built using public domain mutagenicity knowledge and data are routinely used for computational (Q)SAR assessments of pharmaceutical impurities in line with the approach recommended in the ICH M7 guideline. Knowledge from proprietary corporate mutagenicity databases could be used to increase the predictive performance for selected chemical classes as well as expand the applicability domain of these (Q)SAR models. This paper outlines a mechanism for sharing knowledge without the release of proprietary data. Primary aromatic amine mutagenicity was selected as a case study because this chemical class is often encountered in pharmaceutical impurity analysis and mutagenicity of aromatic amines is currently difficult to predict. As part of this analysis, a series of aromatic amine substructures were defined and the number of mutagenic and non-mutagenic examples for each chemical substructure calculated across a series of public and proprietary mutagenicity databases. This information was pooled across all sources to identify structural classes that activate or deactivate aromatic amine mutagenicity. This structure activity knowledge, in combination with newly released primary aromatic amine data, was incorporated into Leadscope's expert rule-based and statistical-based (Q)SAR models where increased predictive performance was demonstrated.