Publications Search
Explore how scientists all over the world use DrugBank in their research.
Published in 2016
READ PUBLICATION →

Identify potential drugs for cardiovascular diseases caused by stress-induced genes in vascular smooth muscle cells.

Authors: Huang CH, Ciou JS, Chen ST, Kok VC, Chung Y, Tsai JJ, Kurubanjerdjit N, Huang CF, Ng KL

Abstract: BACKGROUND: Abnormal proliferation of vascular smooth muscle cells (VSMC) is a major cause of cardiovascular diseases (CVDs). Many studies suggest that vascular injury triggers VSMC dedifferentiation, which results in VSMC changes from a contractile to a synthetic phenotype; however, the underlying molecular mechanisms are still unclear. METHODS: In this study, we examined how VSMC responds under mechanical stress by using time-course microarray data. A three-phase study was proposed to investigate the stress-induced differentially expressed genes (DEGs) in VSMC. First, DEGs were identified by using the moderated t-statistics test. Second, more DEGs were inferred by using the Gaussian Graphical Model (GGM). Finally, the topological parameters-based method and cluster analysis approach were employed to predict the last batch of DEGs. To identify the potential drugs for vascular diseases involve VSMC proliferation, the drug-gene interaction database, Connectivity Map (cMap) was employed. Success of the predictions were determined using in-vitro data, i.e. MTT and clonogenic assay. RESULTS: Based on the differential expression calculation, at least 23 DEGs were found, and the findings were qualified by previous studies on VSMC. The results of gene set enrichment analysis indicated that the most often found enriched biological processes are cell-cycle-related processes. Furthermore, more stress-induced genes, well supported by literature, were found by applying graph theory to the gene association network (GAN). Finally, we showed that by processing the cMap input queries with a cluster algorithm, we achieved a substantial increase in the number of potential drugs with experimental IC50 measurements. With this novel approach, we have not only successfully identified the DEGs, but also improved the DEGs prediction by performing the topological and cluster analysis. Moreover, the findings are remarkably validated and in line with the literature. Furthermore, the cMap and DrugBank resources were used to identify potential drugs and targeted genes for vascular diseases involve VSMC proliferation. Our findings are supported by in-vitro experimental IC50, binding activity data and clinical trials. CONCLUSION: This study provides a systematic strategy to discover potential drugs and target genes, by which we hope to shed light on the treatments of VSMC proliferation associated diseases.
Published in 2016
READ PUBLICATION →

Enhancing Extraction of Drug-Drug Interaction from Literature Using Neutral Candidates, Negation, and Clause Dependency.

Authors: Bokharaeian B, Diaz A, Chitsaz H

Abstract: MOTIVATION: Supervised biomedical relation extraction plays an important role in biomedical natural language processing, endeavoring to obtain the relations between biomedical entities. Drug-drug interactions, which are investigated in the present paper, are notably among the critical biomedical relations. Thus far many methods have been developed with the aim of extracting DDI relations. However, unfortunately there has been a scarcity of comprehensive studies on the effects of negation, complex sentences, clause dependency, and neutral candidates in the course of DDI extraction from biomedical articles. RESULTS: Our study proposes clause dependency features and a number of features for identifying neutral candidates as well as negation cues and scopes. Furthermore, our experiments indicate that the proposed features significantly improve the performance of the relation extraction task combined with other kernel methods. We characterize the contribution of each category of features and finally conclude that neutral candidate features have the most prominent role among all of the three categories.
Published in 2016
READ PUBLICATION →

Computerized techniques pave the way for drug-drug interaction prediction and interpretation.

Authors: Safdari R, Ferdousi R, Aziziheris K, Niakan-Kalhori SR, Omidi Y

Abstract: INTRODUCTION: Health care industry also patients penalized by medical errors that are inevitable but highly preventable. Vast majority of medical errors are related to adverse drug reactions, while drug-drug interactions (DDIs) are the main cause of adverse drug reactions (ADRs). DDIs and ADRs have mainly been reported by haphazard case studies. Experimental in vivo and in vitro researches also reveals DDI pairs. Laboratory and experimental researches are valuable but also expensive and in some cases researchers may suffer from limitations. METHODS: In the current investigation, the latest published works were studied to analyze the trend and pattern of the DDI modelling and the impacts of machine learning methods. Applications of computerized techniques were also investigated for the prediction and interpretation of DDIs. RESULTS: Computerized data-mining in pharmaceutical sciences and related databases provide new key transformative paradigms that can revolutionize the treatment of diseases and hence medical care. Given that various aspects of drug discovery and pharmacotherapy are closely related to the clinical and molecular/biological information, the scientifically sound databases (e.g., DDIs, ADRs) can be of importance for the success of pharmacotherapy modalities. CONCLUSION: A better understanding of DDIs not only provides a robust means for designing more effective medicines but also grantees patient safety.
Published in 2016
READ PUBLICATION →

A graph-based approach to construct target-focused libraries for virtual screening.

Authors: Naderi M, Alvin C, Ding Y, Mukhopadhyay S, Brylinski M

Abstract: BACKGROUND: Due to exorbitant costs of high-throughput screening, many drug discovery projects commonly employ inexpensive virtual screening to support experimental efforts. However, the vast majority of compounds in widely used screening libraries, such as the ZINC database, will have a very low probability to exhibit the desired bioactivity for a given protein. Although combinatorial chemistry methods can be used to augment existing compound libraries with novel drug-like compounds, the broad chemical space is often too large to be explored. Consequently, the trend in library design has shifted to produce screening collections specifically tailored to modulate the function of a particular target or a protein family. METHODS: Assuming that organic compounds are composed of sets of rigid fragments connected by flexible linkers, a molecule can be decomposed into its building blocks tracking their atomic connectivity. On this account, we developed eSynth, an exhaustive graph-based search algorithm to computationally synthesize new compounds by reconnecting these building blocks following their connectivity patterns. RESULTS: We conducted a series of benchmarking calculations against the Directory of Useful Decoys, Enhanced database. First, in a self-benchmarking test, the correctness of the algorithm is validated with the objective to recover a molecule from its building blocks. Encouragingly, eSynth can efficiently rebuild more than 80 % of active molecules from their fragment components. Next, the capability to discover novel scaffolds is assessed in a cross-benchmarking test, where eSynth successfully reconstructed 40 % of the target molecules using fragments extracted from chemically distinct compounds. Despite an enormous chemical space to be explored, eSynth is computationally efficient; half of the molecules are rebuilt in less than a second, whereas 90 % take only about a minute to be generated. CONCLUSIONS: eSynth can successfully reconstruct chemically feasible molecules from molecular fragments. Furthermore, in a procedure mimicking the real application, where one expects to discover novel compounds based on a small set of already developed bioactives, eSynth is capable of generating diverse collections of molecules with the desired activity profiles. Thus, we are very optimistic that our effort will contribute to targeted drug discovery. eSynth is freely available to the academic community at www.brylinski.org/content/molecular-synthesis.Graphical abstractAssuming that organic compounds are composed of sets of rigid fragments connected by flexible linkers, a molecule can be decomposed into its building blocks tracking their atomic connectivity. Here, we developed eSynth, an automated method to synthesize new compounds by reconnecting these building blocks following the connectivity patterns via an exhaustive graph-based search algorithm. eSynth opens up a possibility to rapidly construct virtual screening libraries for targeted drug discovery.
Published in 2016
READ PUBLICATION →

Integrated Approaches to Drug Discovery for Oxidative Stress-Related Retinal Diseases.

Authors: Nishimura Y, Hara H

Abstract: Excessive oxidative stress induces dysregulation of functional networks in the retina, resulting in retinal diseases such as glaucoma, age-related macular degeneration, and diabetic retinopathy. Although various therapies have been developed to reduce oxidative stress in retinal diseases, most have failed to show efficacy in clinical trials. This may be due to oversimplification of target selection for such a complex network as oxidative stress. Recent advances in high-throughput technologies have facilitated the collection of multilevel omics data, which has driven growth in public databases and in the development of bioinformatics tools. Integration of the knowledge gained from omics databases can be used to generate disease-related biological networks and to identify potential therapeutic targets within the networks. Here, we provide an overview of integrative approaches in the drug discovery process and provide simple examples of how the approaches can be exploited to identify oxidative stress-related targets for retinal diseases.
Published in 2016
READ PUBLICATION →

ClassyFire: automated chemical classification with a comprehensive, computable taxonomy.

Authors: Djoumbou Feunang Y, Eisner R, Knox C, Chepelev L, Hastings J, Owen G, Fahy E, Steinbeck C, Subramanian S, Bolton E, Greiner R, Wishart DS

Abstract: BACKGROUND: Scientists have long been driven by the desire to describe, organize, classify, and compare objects using taxonomies and/or ontologies. In contrast to biology, geology, and many other scientific disciplines, the world of chemistry still lacks a standardized chemical ontology or taxonomy. Several attempts at chemical classification have been made; but they have mostly been limited to either manual, or semi-automated proof-of-principle applications. This is regrettable as comprehensive chemical classification and description tools could not only improve our understanding of chemistry but also improve the linkage between chemistry and many other fields. For instance, the chemical classification of a compound could help predict its metabolic fate in humans, its druggability or potential hazards associated with it, among others. However, the sheer number (tens of millions of compounds) and complexity of chemical structures is such that any manual classification effort would prove to be near impossible. RESULTS: We have developed a comprehensive, flexible, and computable, purely structure-based chemical taxonomy (ChemOnt), along with a computer program (ClassyFire) that uses only chemical structures and structural features to automatically assign all known chemical compounds to a taxonomy consisting of >4800 different categories. This new chemical taxonomy consists of up to 11 different levels (Kingdom, SuperClass, Class, SubClass, etc.) with each of the categories defined by unambiguous, computable structural rules. Furthermore each category is named using a consensus-based nomenclature and described (in English) based on the characteristic common structural properties of the compounds it contains. The ClassyFire webserver is freely accessible at http://classyfire.wishartlab.com/. Moreover, a Ruby API version is available at https://bitbucket.org/wishartlab/classyfire_api, which provides programmatic access to the ClassyFire server and database. ClassyFire has been used to annotate over 77 million compounds and has already been integrated into other software packages to automatically generate textual descriptions for, and/or infer biological properties of over 100,000 compounds. Additional examples and applications are provided in this paper. CONCLUSION: ClassyFire, in combination with ChemOnt (ClassyFire's comprehensive chemical taxonomy), now allows chemists and cheminformaticians to perform large-scale, rapid and automated chemical classification. Moreover, a freely accessible API allows easy access to more than 77 million "ClassyFire" classified compounds. The results can be used to help annotate well studied, as well as lesser-known compounds. In addition, these chemical classifications can be used as input for data integration, and many other cheminformatics-related tasks.
Published in 2016
READ PUBLICATION →

Determining the Degree of Promiscuity of Extensively Assayed Compounds.

Authors: Jasial S, Hu Y, Bajorath J

Abstract: In the context of polypharmacology, an emerging concept in drug discovery, promiscuity is rationalized as the ability of compounds to specifically interact with multiple targets. Promiscuity of drugs and bioactive compounds has thus far been analyzed computationally on the basis of activity annotations, without taking assay frequencies or inactivity records into account. Most recent estimates have indicated that bioactive compounds interact on average with only one to two targets, whereas drugs interact with six or more. In this study, we have further extended promiscuity analysis by identifying the most extensively assayed public domain compounds and systematically determining their promiscuity. These compounds were tested in hundreds of assays against hundreds of targets. In our analysis, assay promiscuity was distinguished from target promiscuity and separately analyzed for primary and confirmatory assays. Differences between the degree of assay and target promiscuity were surprisingly small and average and median degrees of target promiscuity of 2.6 to 3.4 and 2.0 were determined, respectively. Thus, target promiscuity remained at a low level even for most extensively tested active compounds. These findings provide further evidence that bioactive compounds are less promiscuous than drugs and have implications for pharmaceutical research. In addition to a possible explanation that drugs are more extensively tested for additional targets, the results would also support a "promiscuity enrichment model" according to which promiscuous compounds might be preferentially selected for therapeutic efficacy during clinical evaluation to ultimately become drugs.
Published in 2016
READ PUBLICATION →

Centralizing content and distributing labor: a community model for curating the very long tail of microbial genomes.

Authors: Putman TE, Burgstaller-Muehlbacher S, Waagmeester A, Wu C, Su AI, Good BM

Abstract: The last 20 years of advancement in sequencing technologies have led to sequencing thousands of microbial genomes, creating mountains of genetic data. While efficiency in generating the data improves almost daily, applying meaningful relationships between taxonomic and genetic entities on this scale requires a structured and integrative approach. Currently, knowledge is distributed across a fragmented landscape of resources from government-funded institutions such as National Center for Biotechnology Information (NCBI) and UniProt to topic-focused databases like the ODB3 database of prokaryotic operons, to the supplemental table of a primary publication. A major drawback to large scale, expert-curated databases is the expense of maintaining and extending them over time. No entity apart from a major institution with stable long-term funding can consider this, and their scope is limited considering the magnitude of microbial data being generated daily. Wikidata is an openly editable, semantic web compatible framework for knowledge representation. It is a project of the Wikimedia Foundation and offers knowledge integration capabilities ideally suited to the challenge of representing the exploding body of information about microbial genomics. We are developing a microbial specific data model, based on Wikidata's semantic web compatibility, which represents bacterial species, strains and the gene and gene products that define them. Currently, we have loaded 43,694 gene and 37,966 protein items for 21 species of bacteria, including the human pathogenic bacteriaChlamydia trachomatis.Using this pathogen as an example, we explore complex interactions between the pathogen, its host, associated genes, other microbes, disease and drugs using the Wikidata SPARQL endpoint. In our next phase of development, we will add another 99 bacterial genomes and their gene and gene products, totaling approximately 900,000 additional entities. This aggregation of knowledge will be a platform for community-driven collaboration, allowing the networking of microbial genetic data through the sharing of knowledge by both the data and domain expert.
Published in 2016
READ PUBLICATION →

An integrative data analysis platform for gene set analysis and knowledge discovery in a data warehouse framework.

Authors: Chen YA, Tripathi LP, Mizuguchi K

Abstract: Data analysis is one of the most critical and challenging steps in drug discovery and disease biology. A user-friendly resource to visualize and analyse high-throughput data provides a powerful medium for both experimental and computational biologists to understand vastly different biological data types and obtain a concise, simplified and meaningful output for better knowledge discovery. We have previously developed TargetMine, an integrated data warehouse optimized for target prioritization. Here we describe how upgraded and newly modelled data types in TargetMine can now survey the wider biological and chemical data space, relevant to drug discovery and development. To enhance the scope of TargetMine from target prioritization to broad-based knowledge discovery, we have also developed a new auxiliary toolkit to assist with data analysis and visualization in TargetMine. This toolkit features interactive data analysis tools to query and analyse the biological data compiled within the TargetMine data warehouse. The enhanced system enables users to discover new hypotheses interactively by performing complicated searches with no programming and obtaining the results in an easy to comprehend output format. Database URL: http://targetmine.mizuguchilab.org.
Published in 2016
READ PUBLICATION →

COMPUTING THERAPY FOR PRECISION MEDICINE: COLLABORATIVE FILTERING INTEGRATES AND PREDICTS MULTI-ENTITY INTERACTIONS.

Authors: Regenbogen S, Wilkins AD, Lichtarge O

Abstract: Biomedicine produces copious information it cannot fully exploit. Specifically, there is considerable need to integrate knowledge from disparate studies to discover connections across domains. Here, we used a Collaborative Filtering approach, inspired by online recommendation algorithms, in which non-negative matrix factorization (NMF) predicts interactions among chemicals, genes, and diseases only from pairwise information about their interactions. Our approach, applied to matrices derived from the Comparative Toxicogenomics Database, successfully recovered Chemical-Disease, Chemical-Gene, and Disease-Gene networks in 10-fold cross-validation experiments. Additionally, we could predict each of these interaction matrices from the other two. Integrating all three CTD interaction matrices with NMF led to good predictions of STRING, an independent, external network of protein-protein interactions. Finally, this approach could integrate the CTD and STRING interaction data to improve Chemical-Gene cross-validation performance significantly, and, in a time-stamped study, it predicted information added to CTD after a given date, using only data prior to that date. We conclude that collaborative filtering can integrate information across multiple types of biological entities, and that as a first step towards precision medicine it can compute drug repurposing hypotheses.