Feeds:
Posts
Comments

Archive for the ‘Advanced Drug Manufacturing Technology’ Category

From High-Throughput Assay to Systems Biology: New Tools for Drug Discovery

Curator: Stephen J. Williams, PhD

Marc W. Kirschner*

Department of Systems Biology
Harvard Medical School

Boston, Massachusetts 02115

With the new excitement about systems biology, there is understandable interest in a definition. This has proven somewhat difficult. Scientific fields, like spe­cies, arise by descent with modification, so in their ear­liest forms even the founders of great dynasties are only marginally different than their sister fields and spe­cies. It is only in retrospect that we can recognize the significant founding events. Before embarking on a def­inition of systems biology, it may be worth remember­ing that confusion and controversy surrounded the in­troduction of the term “molecular biology,” with claims that it hardly differed from biochemistry. Yet in retro­spect molecular biology was new and different. It intro­duced both new subject matter and new technological approaches, in addition to a new style.

As a point of departure for systems biology, consider the quintessential experiment in the founding of molec­ular biology, the one gene one enzyme hypothesis of Beadle and Tatum. This experiment first connected the genotype directly to the phenotype on a molecular level, although efforts in that direction can certainly be found in the work of Archibald Garrod, Sewell Wright, and others. Here a protein (in this case an enzyme) is seen to be a product of a single gene, and a single function; the completion of a specific step in amino acid biosynthesis is the direct result. It took the next 30 years to fill in the gaps in this process. Yet the one gene one enzyme hypothesis looks very different to us today. What is the function of tubulin, of PI-3 kinase or of rac? Could we accurately predict the phenotype of a nonle­thal mutation in these genes in a multicellular organ­ism? Although we can connect structure to the gene, we can no longer infer its larger purpose in the cell or in the organism. There are too many purposes; what the protein does is defined by context. The context also includes a history, either developmental or physiologi­cal. Thus the behavior of the Wnt signaling pathway depends on the previous lineage, the “where and when” questions of embryonic development. Similarly the behavior of the immune system depends on previ­ous experience in a variable environment. All of these features stress how inadequate an explanation for function we can achieve solely by trying to identify genes (by annotating them!) and characterizing their transcriptional control circuits.

That we are at a crossroads in how to explore biology is not at all clear to many. Biology is hardly in its dotage; the process of discovery seems to have been per­fected, accelerated, and made universally applicable to all fields of biology. With the completion of the human genome and the genomes of other species, we have a glimpse of many more genes than we ever had before to study. We are like naturalists discovering a new con­tinent, enthralled with the diversity itself. But we have also at the same time glimpsed the finiteness of this list of genes, a disturbingly small list. We have seen that the diversity of genes cannot approximate the diversity of functions within an organism. In response, we have argued that combinatorial use of small numbers of components can generate all the diversity that is needed. This has had its recent incarnation in the sim­plistic view that the rules of cis-regulatory control on DNA can directly lead to an understanding of organ­isms and their evolution. Yet this assumes that the gene products can be linked together in arbitrary combina­tions, something that is not assured in chemistry. It also downplays the significant regulatory features that in­volve interactions between gene products, their local­ization, binding, posttranslational modification, degra­dation, etc. The big question to understand in biology is not regulatory linkage but the nature of biological systems that allows them to be linked together in many nonlethal and even useful combinations. More and more we come to realize that understanding the con­served genes and their conserved circuits will require an understanding of their special properties that allow them to function together to generate different pheno­types in different tissues of metazoan organisms. These circuits may have certain robustness, but more impor­tant they have adaptability and versatility. The ease of putting conserved processes under regulatory control is an inherent design feature of the processes them­selves. Among other things it loads the deck in evolu­tionary variation and makes it more feasible to generate useful phenotypes upon which selection can act.

Systems biology offers an opportunity to study how the phenotype is generated from the genotype and with it a glimpse of how evolution has crafted the pheno­type. One aspect of systems biology is the develop­ment of techniques to examine broadly the level of pro­tein, RNA, and DNA on a gene by gene basis and even the posttranslational modification and localization of proteins. In a very short time we have witnessed the development of high-throughput biology, forcing us to consider cellular processes in toto. Even though much of the data is noisy and today partially inconsistent and incomplete, this has been a radical shift in the way we tear apart problems one interaction at a time. When coupled with gene deletions by RNAi and classical methods, and with the use of chemical tools tailored to proteins and protein domains, these high-throughput techniques become still more powerful.

High-throughput biology has opened up another im­portant area of systems biology: it has brought us out into the field again or at least made us aware that there is a world outside our laboratories. Our model systems have been chosen intentionally to be of limited genetic diversity and examined in a highly controlled and repro­ducible environment. The real world of ecology, evolu­tion, and human disease is a very different place. When genetics separated from the rest of biology in the early part of the 20th century, most geneticists sought to understand heredity and chose to study traits in the organism that could be easily scored and could be used to reveal genetic mechanisms. This was later ex­tended to powerful effect to use genetics to study cell biological and developmental mechanisms. Some ge­neticists, including a large school in Russia in the early 20th century, continued to study the genetics of natural populations, focusing on traits important for survival. That branch of genetics is coming back strongly with the power of phenotypic assays on the RNA and pro­tein level. As human beings we are most concerned not with using our genetic misfortunes to unravel biology’s complexity (important as that is) but with the role of our genetics in our individual survival. The context for understanding this is still not available, even though the data are now coming in torrents, for many of the genes that will contribute to our survival will have small quan­titative effects, partially masked or accentuated by other genetic and environmental conditions. To under­stand the genetic basis of disease will require not just mapping these genes but an understanding of how the phenotype is created in the first place and the messy interactions between genetic variation and environ­mental variation.

Extracts and explants are relatively accessible to syn­thetic manipulation. Next there is the explicit recon­struction of circuits within cells or the deliberate modifi­cation of those circuits. This has occurred for a while in biology, but the difference is that now we wish to construct or intervene with the explicit purpose of de­scribing the dynamical features of these synthetic or partially synthetic systems. There are more and more tools to intervene and more and more tools to measure. Although these fall short of total descriptions of cells and organisms, the detailed information will give us a sense of the special life-like processes of circuits, pro­teins, cells in tissues, and whole organisms in their en­vironment. This meso-scale systems biology will help establish the correspondence between molecules and large-scale physiology.

You are probably running out of patience for some definition of systems biology. In any case, I do not think the explicit definition of systems biology should come from me but should await the words of the first great modern systems biologist. She or he is probably among us now. However, if forced to provide some kind of label for systems biology, I would simply say that systems biology is the study of the behavior of complex biologi­cal organization and processes in terms of the molecu­lar constituents. It is built on molecular biology in its special concern for information transfer, on physiology for its special concern with adaptive states of the cell and organism, on developmental biology for the impor­tance of defining a succession of physiological states in that process, and on evolutionary biology and ecol­ogy for the appreciation that all aspects of the organ­ism are products of selection, a selection we rarely understand on a molecular level. Systems biology attempts all of this through quantitative measurement, modeling, reconstruction, and theory. Systems biology is not a branch of physics but differs from physics in that the primary task is to understand how biology gen­erates variation. No such imperative to create variation exists in the physical world. It is a new principle that Darwin understood and upon which all of life hinges. That sounds different enough for me to justify a new field and a new name. Furthermore, the success of sys­tems biology is essential if we are to understand life; its success is far from assured—a good field for those seeking risk and adventure.

Source: “Meaning of Systems Biology” Cell, Vol. 121, 503–504, May 20, 2005, DOI 10.1016/j.cell.2005.05.005

Old High-throughput Screening, Once the Gold Standard in Drug Development, Gets a Systems Biology Facelift

From Phenotypic Hit to Chemical Probe: Chemical Biology Approaches to Elucidate Small Molecule Action in Complex Biological Systems

Quentin T. L. Pasquer, Ioannis A. Tsakoumagkos and Sascha Hoogendoorn 

Molecules 202025(23), 5702; https://doi.org/10.3390/molecules25235702

Abstract

Biologically active small molecules have a central role in drug development, and as chemical probes and tool compounds to perturb and elucidate biological processes. Small molecules can be rationally designed for a given target, or a library of molecules can be screened against a target or phenotype of interest. Especially in the case of phenotypic screening approaches, a major challenge is to translate the compound-induced phenotype into a well-defined cellular target and mode of action of the hit compound. There is no “one size fits all” approach, and recent years have seen an increase in available target deconvolution strategies, rooted in organic chemistry, proteomics, and genetics. This review provides an overview of advances in target identification and mechanism of action studies, describes the strengths and weaknesses of the different approaches, and illustrates the need for chemical biologists to integrate and expand the existing tools to increase the probability of evolving screen hits to robust chemical probes.

5.1.5. Large-Scale Proteomics

While FITExP is based on protein expression regulation during apoptosis, a study of Ruprecht et al. showed that proteomic changes are induced both by cytotoxic and non-cytotoxic compounds, which can be detected by mass spectrometry to give information on a compound’s mechanism of action. They developed a large-scale proteome-wide mass spectrometry analysis platform for MOA studies, profiling five lung cancer cell lines with over 50 drugs. Aggregation analysis over the different cell lines and the different compounds showed that one-quarter of the drugs changed the abundance of their protein target. This approach allowed target confirmation of molecular degraders such as PROTACs or molecular glues. Finally, this method yielded unexpected off-target mechanisms for the MAP2K1/2 inhibitor PD184352 and the ALK inhibitor ceritinib [97]. While such a mapping approach clearly provides a wealth of information, it might not be easily attainable for groups that are not equipped for high-throughput endeavors.

All-in-all, mass spectrometry methods have gained a lot of traction in recent years and have been successfully applied for target deconvolution and MOA studies of small molecules. As with all high-throughput methods, challenges lie in the accessibility of the instruments (both from a time and cost perspective) and data analysis of complex and extensive data sets.

5.2. Genetic Approaches

Both label-based and mass spectrometry proteomic approaches are based on the physical interaction between a small molecule and a protein target, and focus on the proteome for target deconvolution. It has been long realized that genetics provides an alternative avenue to understand a compound’s action, either through precise modification of protein levels, or by inducing protein mutations. First realized in yeast as a genetically tractable organism over 20 years ago, recent advances in genetic manipulation of mammalian cells have opened up important opportunities for target identification and MOA studies through genetic screening in relevant cell types [98]. Genetic approaches can be roughly divided into two main areas, with the first centering on the identification of mutations that confer compound resistance (Figure 3a), and the second on genome-wide perturbation of gene function and the concomitant changes in sensitivity to the compound (Figure 3b). While both methods can be used to identify or confirm drug targets, the latter category often provides many additional insights in the compound’s mode of action.

Figure 3. Genetic methods for target identification and mode of action studies. Schematic representations of (a) resistance cloning, and (b) chemogenetic interaction screens.

5.2.1. Resistance Cloning

The “gold standard” in drug target confirmation is to identify mutations in the presumed target protein that render it insensitive to drug treatment. Conversely, different groups have sought to use this principle as a target identification method based on the concept that cells grown in the presence of a cytotoxic drug will either die or develop mutations that will make them resistant to the compound. With recent advances in deep sequencing it is now possible to then scan the transcriptome [99] or genome [100] of the cells for resistance-inducing mutations. Genes that are mutated are then hypothesized to encode the protein target. For this approach to be successful, there are two initial requirements: (1) the compound needs to be cytotoxic for resistant clones to arise, and (2) the cell line needs to be genetically unstable for mutations to occur in a reasonable timeframe.

In 2012, the Kapoor group demonstrated in a proof-of-concept study that resistance cloning in mammalian cells, coupled to transcriptome sequencing (RNA-seq), yields the known polo-like kinase 1 (PLK1) target of the small molecule BI 2536. For this, they used the cancer cell line HCT-116, which is deficient in mismatch repair and consequently prone to mutations. They generated and sequenced multiple resistant clones, and clustered the clones based on similarity. PLK1 was the only gene that was mutated in multiple groups. Of note, one of the groups did not contain PLK1 mutations, but rather developed resistance through upregulation of ABCBA1, a drug efflux transporter, which is a general and non-specific resistance mechanism [101]. In a following study, they optimized their pipeline “DrugTargetSeqR”, by counter-screening for these types of multidrug resistance mechanisms so that these clones were excluded from further analysis (Figure 3a). Furthermore, they used CRISPR/Cas9-mediated gene editing to determine which mutations were sufficient to confer drug resistance, and as independent validation of the biochemical relevance of the obtained hits [102].

While HCT-116 cells are a useful model cell line for resistance cloning because of their genomic instability, they may not always be the cell line of choice, depending on the compound and process that is studied. Povedana et al. used CRISPR/Cas9 to engineer mismatch repair deficiencies in Ewing sarcoma cells and small cell lung cancer cells. They found that deletion of MSH2 results in hypermutations in these normally mutationally silent cells, resulting in the formation of resistant clones in the presence of bortezomib, MLN4924, and CD437, which are all cytotoxic compounds [103]. Recently, Neggers et al. reasoned that CRISPR/Cas9-induced non-homologous end-joining repair could be a viable strategy to create a wide variety of functional mutants of essential genes through in-frame mutations. Using a tiled sgRNA library targeting 75 target genes of investigational neoplastic drugs in HAP1 and K562 cells, they generated several KPT-9274 (an anticancer agent with unknown target)-resistant clones, and subsequent deep sequencing showed that the resistant clones were enriched in NAMPT sgRNAs. Direct target engagement was confirmed by co-crystallizing the compound with NAMPT [104]. In addition to these genetic mutation strategies, an alternative method is to grow the cells in the presence of a mutagenic chemical to induce higher mutagenesis rates [105,106].

When there is already a hypothesis on the pathway involved in compound action, the resistance cloning methodology can be extended to non-cytotoxic compounds. Sekine et al. developed a fluorescent reporter model for the integrated stress response, and used this cell line for target deconvolution of a small molecule inhibitor towards this pathway (ISRIB). Reporter cells were chemically mutagenized, and ISRIB-resistant clones were isolated by flow cytometry, yielding clones with various mutations in the delta subunit of guanine nucleotide exchange factor eIF2B [107].

While there are certainly successful examples of resistance cloning yielding a compound’s direct target as discussed above, resistance could also be caused by mutations or copy number alterations in downstream components of a signaling pathway. This is illustrated by clinical examples of acquired resistance to small molecules, nature’s way of “resistance cloning”. For example, resistance mechanisms in Hedgehog pathway-driven cancers towards the Smoothened inhibitor vismodegib include compound-resistant mutations in Smoothened, but also copy number changes in downstream activators SUFU and GLI2 [108]. It is, therefore, essential to conduct follow-up studies to confirm a direct interaction between a compound and the hit protein, as well as a lack of interaction with the mutated protein.

5.2.3. “Chemogenomics”: Examples of Gene-Drug Interaction Screens

When genetic perturbations are combined with small molecule drugs in a chemogenetic interaction screen, the effect of a gene’s perturbation on compound action is studied. Gene perturbation can render the cells resistant to the compound (suppressor interaction), or conversely, result in hypersensitivity and enhanced compound potency (synergistic interaction) [5,117,121]. Typically, cells are treated with the compound at a sublethal dose, to ascertain that both types of interactions can be found in the final dataset, and often it is necessary to use a variety of compound doses (i.e., LD20, LD30, LD50) and timepoints to obtain reliable insights (Figure 3b).

An early example of successful coupling of a phenotypic screen and downstream genetic screening for target identification is the study of Matheny et al. They identified STF-118804 as a compound with antileukemic properties. Treatment of MV411 cells, stably transduced with a high complexity, genome-wide shRNA library, with STF-118804 (4 rounds of increasing concentration) or DMSO control resulted in a marked depletion of cells containing shRNAs against nicotinamide phosphoribosyl transferase (NAMPT) [122].

The Bassik lab subsequently directly compared the performance of shRNA-mediated knockdown versus CRISPR/Cas9-knockout screens for the target elucidation of the antiviral drug GSK983. The data coming out of both screens were complementary, with the shRNA screen resulting in hits leading to the direct compound target and the CRISPR screen giving information on cellular mechanisms of action of the compound. A reason for this is likely the level of protein depletion that is reached by these methods: shRNAs lead to decreased protein levels, which is advantageous when studying essential genes. However, knockdown may not result in a phenotype for non-essential genes, in which case a full CRISPR-mediated knockout is necessary to observe effects [123].

Another NAMPT inhibitor was identified in a CRISPR/Cas9 “haplo-insufficiency (HIP)”-like approach [124]. Haploinsuffiency profiling is a well-established system in yeast which is performed in a ~50% protein background by heterozygous deletions [125]. As there is no control over CRISPR-mediated loss of alleles, compound treatment was performed at several timepoints after addition of the sgRNA library to HCT116 cells stably expressing Cas9, in the hope that editing would be incomplete at early timepoints, resulting in residual protein levels. Indeed, NAMPT was found to be the target of phenotypic hit LB-60-OF61, especially at earlier timepoints, confirming the hypothesis that some level of protein needs to be present to identify a compound’s direct target [124]. This approach was confirmed in another study, thereby showing that direct target identification through CRISPR-knockout screens is indeed possible [126].

An alternative strategy was employed by the Weissman lab, where they combined genome-wide CRISPR-interference and -activation screens to identify the target of the phase 3 drug rigosertib. They focused on hits that had opposite action in both screens, as in sensitizing in one but protective in the other, which were related to microtubule stability. In a next step, they created chemical-genetic profiles of a variety of microtubule destabilizing agents, rationalizing that compounds with the same target will have similar drug-gene interactions. For this, they made a focused library of sgRNAs, based on the most high-ranking hits in the rigosertib genome-wide CRISPRi screen, and compared the focused screen results of the different compounds. The profile for rigosertib clustered well with that of ABT-571, and rigorous target validation studies confirmed rigosertib binding to the colchicine binding site of tubulin—the same site as occupied by ABT-571 [127].

From the above examples, it is clear that genetic screens hold a lot of promise for target identification and MOA studies for small molecules. The CRISPR screening field is rapidly evolving, sgRNA libraries are continuously improving and increasingly commercially available, and new tools for data analysis are being developed [128]. The challenge lies in applying these screens to study compounds that are not cytotoxic, where finding the right dosage regimen will not be trivial.

SYSTEMS BIOLOGY AND CANCER RESEARCH & DRUG DISCOVERY

Integrative Analysis of Next-Generation Sequencing for Next-Generation Cancer Research toward Artificial Intelligence

Youngjun Park, Dominik Heider and Anne-Christin Hauschild. Cancers 202113(13), 3148; https://doi.org/10.3390/cancers13133148

Abstract

The rapid improvement of next-generation sequencing (NGS) technologies and their application in large-scale cohorts in cancer research led to common challenges of big data. It opened a new research area incorporating systems biology and machine learning. As large-scale NGS data accumulated, sophisticated data analysis methods became indispensable. In addition, NGS data have been integrated with systems biology to build better predictive models to determine the characteristics of tumors and tumor subtypes. Therefore, various machine learning algorithms were introduced to identify underlying biological mechanisms. In this work, we review novel technologies developed for NGS data analysis, and we describe how these computational methodologies integrate systems biology and omics data. Subsequently, we discuss how deep neural networks outperform other approaches, the potential of graph neural networks (GNN) in systems biology, and the limitations in NGS biomedical research. To reflect on the various challenges and corresponding computational solutions, we will discuss the following three topics: (i) molecular characteristics, (ii) tumor heterogeneity, and (iii) drug discovery. We conclude that machine learning and network-based approaches can add valuable insights and build highly accurate models. However, a well-informed choice of learning algorithm and biological network information is crucial for the success of each specific research question

1. Introduction

The development and widespread use of high-throughput technologies founded the era of big data in biology and medicine. In particular, it led to an accumulation of large-scale data sets that opened a vast amount of possible applications for data-driven methodologies. In cancer, these applications range from fundamental research to clinical applications: molecular characteristics of tumors, tumor heterogeneity, drug discovery and potential treatments strategy. Therefore, data-driven bioinformatics research areas have tailored data mining technologies such as systems biology, machine learning, and deep learning, elaborated in this review paper (see Figure 1 and Figure 2). For example, in systems biology, data-driven approaches are applied to identify vital signaling pathways [1]. This pathway-centric analysis is particularly crucial in cancer research to understand the characteristics and heterogeneity of the tumor and tumor subtypes. Consequently, this high-throughput data-based analysis enables us to explore characteristics of cancers with a systems biology and a systems medicine point of view [2].Combining high-throughput techniques, especially next-generation sequencing (NGS), with appropriate analytical tools has allowed researchers to gain a deeper systematic understanding of cancer at various biological levels, most importantly genomics, transcriptomics, and epigenetics [3,4]. Furthermore, more sophisticated analysis tools based on computational modeling are introduced to decipher underlying molecular mechanisms in various cancer types. The increasing size and complexity of the data required the adaptation of bioinformatics processing pipelines for higher efficiency and sophisticated data mining methodologies, particularly for large-scale, NGS datasets [5]. Nowadays, more and more NGS studies integrate a systems biology approach and combine sequencing data with other types of information, for instance, protein family information, pathway, or protein–protein interaction (PPI) networks, in an integrative analysis. Experimentally validated knowledge in systems biology may enhance analysis models and guides them to uncover novel findings. Such integrated analyses have been useful to extract essential information from high-dimensional NGS data [6,7]. In order to deal with the increasing size and complexity, the application of machine learning, and specifically deep learning methodologies, have become state-of-the-art in NGS data analysis.

Figure 1. Next-generation sequencing data can originate from various experimental and technological conditions. Depending on the purpose of the experiment, one or more of the depicted omics types (Genomics, Transcriptomics, Epigenomics, or Single-Cell Omics) are analyzed. These approaches led to an accumulation of large-scale NGS datasets to solve various challenges of cancer research, molecular characterization, tumor heterogeneity, and drug target discovery. For instance, The Cancer Genome Atlas (TCGA) dataset contains multi-omics data from ten-thousands of patients. This dataset facilitates a variety of cancer researches for decades. Additionally, there are also independent tumor datasets, and, frequently, they are analyzed and compared with the TCGA dataset. As the large scale of omics data accumulated, various machine learning techniques are applied, e.g., graph algorithms and deep neural networks, for dimensionality reduction, clustering, or classification. (Created with BioRender.com.)

Figure 2. (a) A multitude of different types of data is produced by next-generation sequencing, for instance, in the fields of genomics, transcriptomics, and epigenomics. (b) Biological networks for biomarker validation: The in vivo or in vitro experiment results are considered ground truth. Statistical analysis on next-generation sequencing data produces candidate genes. Biological networks can validate these candidate genes and highlight the underlying biological mechanisms (Section 2.1). (c) De novo construction of Biological Networks: Machine learning models that aim to reconstruct biological networks can incorporate prior knowledge from different omics data. Subsequently, the model will predict new unknown interactions based on new omics information (Section 2.2). (d) Network-based machine learning: Machine learning models integrating biological networks as prior knowledge to improve predictive performance when applied to different NGS data (Section 2.3). (Created with BioRender.com).

Therefore, a large number of studies integrate NGS data with machine learning and propose a novel data-driven methodology in systems biology [8]. In particular, many network-based machine learning models have been developed to analyze cancer data and help to understand novel mechanisms in cancer development [9,10]. Moreover, deep neural networks (DNN) applied for large-scale data analysis improved the accuracy of computational models for mutation prediction [11,12], molecular subtyping [13,14], and drug repurposing [15,16]. 

2. Systems Biology in Cancer Research

Genes and their functions have been classified into gene sets based on experimental data. Our understandings of cancer concentrated into cancer hallmarks that define the characteristics of a tumor. This collective knowledge is used for the functional analysis of unseen data.. Furthermore, the regulatory relationships among genes were investigated, and, based on that, a pathway can be composed. In this manner, the accumulation of public high-throughput sequencing data raised many big-data challenges and opened new opportunities and areas of application for computer science. Two of the most vibrantly evolving areas are systems biology and machine learning which tackle different tasks such as understanding the cancer pathways [9], finding crucial genes in pathways [22,53], or predicting functions of unidentified or understudied genes [54]. Essentially, those models include prior knowledge to develop an analysis and enhance interpretability for high-dimensional data [2]. In addition to understanding cancer pathways with in silico analysis, pathway activity analysis incorporating two different types of data, pathways and omics data, is developed to understand heterogeneous characteristics of the tumor and cancer molecular subtyping. Due to its advantage in interpretability, various pathway-oriented methods are introduced and become a useful tool to understand a complex diseases such as cancer [55,56,57].

In this section, we will discuss how two related research fields, namely, systems biology and machine learning, can be integrated with three different approaches (see Figure 2), namely, biological network analysis for biomarker validation, the use of machine learning with systems biology, and network-based models.

2.1. Biological Network Analysis for Biomarker Validation

The detection of potential biomarkers indicative of specific cancer types or subtypes is a frequent goal of NGS data analysis in cancer research. For instance, a variety of bioinformatics tools and machine learning models aim at identify lists of genes that are significantly altered on a genomic, transcriptomic, or epigenomic level in cancer cells. Typically, statistical and machine learning methods are employed to find an optimal set of biomarkers, such as single nucleotide polymorphisms (SNPs), mutations, or differentially expressed genes crucial in cancer progression. Traditionally, resource-intensive in vitro analysis was required to discover or validate those markers. Therefore, systems biology offers in silico solutions to validate such findings using biological pathways or gene ontology information (Figure 2b) [58]. Subsequently, gene set enrichment analysis (GSEA) [50] or gene set analysis (GSA) [59] can be used to evaluate whether these lists of genes are significantly associated with cancer types and their specific characteristics. GSA, for instance, is available via web services like DAVID [60] and g:Profiler [61]. Moreover, other applications use gene ontology directly [62,63]. In addition to gene-set-based analysis, there are other methods that focuse on the topology of biological networks. These approaches evaluate various network structure parameters and analyze the connectivity of two genes or the size and interconnection of their neighbors [64,65]. According to the underlying idea, the mutated gene will show dysfunction and can affect its neighboring genes. Thus, the goal is to find abnormalities in a specific set of genes linked with an edge in a biological network. For instance, KeyPathwayMiner can extract informative network modules in various omics data [66]. In summary, these approaches aim at predicting the effect of dysfunctional genes among neighbors according to their connectivity or distances from specific genes such as hubs [67,68]. During the past few decades, the focus of cancer systems biology extended towards the analysis of cancer-related pathways since those pathways tend to carry more information than a gene set. Such analysis is called Pathway Enrichment Analysis (PEA) [69,70]. The use of PEA incorporates the topology of biological networks. However, simultaneously, the lack of coverage issue in pathway data needs to be considered. Because pathway data does not cover all known genes yet, an integration analysis on omics data can significantly drop in genes when incorporated with pathways. Genes that can not be mapped to any pathway are called ‘pathway orphan.’ In this manner, Rahmati et al. introduced a possible solution to overcome the ‘pathway orphan’ issue [71]. At the bottom line, regardless of whether researchers consider gene-set or pathway-based enrichment analysis, the performance and accuracy of both methods are highly dependent on the quality of the external gene-set and pathway data [72].

2.2. De Novo Construction of Biological Networks

While the known fraction of existing biological networks barely scratches the surface of the whole system of mechanisms occurring in each organism, machine learning models can improve on known network structures and can guide potential new findings [73,74]. This area of research is called de novo network construction (Figure 2c), and its predictive models can accelerate experimental validation by lowering time costs [75,76]. This interplay between in silico biological networks building and mining contributes to expanding our knowledge in a biological system. For instance, a gene co-expression network helps discover gene modules having similar functions [77]. Because gene co-expression networks are based on expressional changes under specific conditions, commonly, inferring a co-expression network requires many samples. The WGCNA package implements a representative model using weighted correlation for network construction that leads the development of the network biology field [78]. Due to NGS developments, the analysis of gene co-expression networks subsequently moved from microarray-based to RNA-seq based experimental data [79]. However, integration of these two types of data remains tricky. Ballouz et al. compared microarray and NGS-based co-expression networks and found the existence of a bias originating from batch effects between the two technologies [80]. Nevertheless, such approaches are suited to find disease-specific co-expressional gene modules. Thus, various studies based on the TCGA cancer co-expression network discovered characteristics of prognostic genes in the network [81]. Accordingly, a gene co-expression network is a condition-specific network rather than a general network for an organism. Gene regulatory networks can be inferred from the gene co-expression network when various data from different conditions in the same organism are available. Additionally, with various NGS applications, we can obtain multi-modal datasets about regulatory elements and their effects, such as epigenomic mechanisms on transcription and chromatin structure. Consequently, a gene regulatory network can consist of solely protein-coding genes or different regulatory node types such as transcription factors, inhibitors, promoter interactions, DNA methylations, and histone modifications affecting the gene expression system [82,83]. More recently, researchers were able to build networks based on a particular experimental setup. For instance, functional genomics or CRISPR technology enables the high-resolution regulatory networks in an organism [84]. Other than gene co-expression or regulatory networks, drug target, and drug repurposing studies are active research areas focusing on the de novo construction of drug-to-target networks to allow the potential repurposing of drugs [76,85].

2.3. Network Based Machine Learning

A network-based machine learning model directly integrates the insights of biological networks within the algorithm (Figure 2d) to ultimately improve predictive performance concerning cancer subtyping or susceptibility to therapy. Following the establishment of high-quality biological networks based on NGS technologies, these biological networks were suited to be integrated into advanced predictive models. In this manner, Zhang et al., categorized network-based machine learning approaches upon their usage into three groups: (i) model-based integration, (ii) pre-processing integration, and (iii) post-analysis integration [7]. Network-based models map the omics data onto a biological network, and proper algorithms travel the network while considering both values of nodes and edges and network topology. In the pre-processing integration, pathway or other network information is commonly processed based on its topological importance. Meanwhile, in the post-analysis integration, omics data is processed solely before integration with a network. Subsequently, omics data and networks are merged and interpreted. The network-based model has advantages in multi-omics integrative analysis. Due to the different sensitivity and coverage of various omics data types, a multi-omics integrative analysis is challenging. However, focusing on gene-level or protein-level information enables a straightforward integration [86,87]. Consequently, when different machine learning approaches tried to integrate two or more different data types to find novel biological insights, one of the solutions is reducing the search space to gene or protein level and integrated heterogeneous datatypes [25,88].

In summary, using network information opens new possibilities for interpretation. However, as mentioned earlier, several challenges remain, such as the coverage issue. Current databases for biological networks do not cover the entire set of genes, transcripts, and interactions. Therefore, the use of networks can lead to loss of information for gene or transcript orphans. The following section will focus on network-based machine learning models and their application in cancer genomics. We will put network-based machine learning into the perspective of the three main areas of application, namely, molecular characterization, tumor heterogeneity analysis, and cancer drug discovery.

3. Network-Based Learning in Cancer Research

As introduced previously, the integration of machine learning with the insights of biological networks (Figure 2d) ultimately aims at improving predictive performance and interpretability concerning cancer subtyping or treatment susceptibility.

3.1. Molecular Characterization with Network Information

Various network-based algorithms are used in genomics and focus on quantifying the impact of genomic alteration. By employing prior knowledge in biological network algorithms, performance compared to non-network models can be improved. A prominent example is HotNet. The algorithm uses a thermodynamics model on a biological network and identifies driver genes, or prognostic genes, in pan-cancer data [89]. Another study introduced a network-based stratification method to integrate somatic alterations and expression signatures with network information [90]. These approaches use network topology and network-propagation-like algorithms. Network propagation presumes that genomic alterations can affect the function of neighboring genes. Two genes will show an exclusive pattern if two genes complement each other, and the function carried by those two genes is essential to an organism [91]. This unique exclusive pattern among genomic alteration is further investigated in cancer-related pathways. Recently, Ku et al. developed network-centric approaches and tackled robustness issues while studying synthetic lethality [92]. Although synthetic lethality was initially discovered in model organisms of genetics, it helps us to understand cancer-specific mutations and their functions in tumor characteristics [91].

Furthermore, in transcriptome research, network information is used to measure pathway activity and its application in cancer subtyping. For instance, when comparing the data of two or more conditions such as cancer types, GSEA as introduced in Section 2 is a useful approach to get an overview of systematic changes [50]. It is typically used at the beginning of a data evaluation [93]. An experimentally validated gene set can provide information about how different conditions affect molecular systems in an organism. In addition to the gene sets, different approaches integrate complex interaction information into GSEA and build network-based models [70]. In contrast to GSEA, pathway activity analysis considers transcriptome data and other omics data and structural information of a biological network. For example, PARADIGM uses pathway topology and integrates various omics in the analysis to infer a patient-specific status of pathways [94]. A benchmark study with pan-cancer data recently reveals that using network structure can show better performance [57]. In conclusion, while the loss of data is due to the incompleteness of biological networks, their integration improved performance and increased interpretability in many cases.

3.2. Tumor Heterogeneity Study with Network Information

The tumor heterogeneity can originate from two directions, clonal heterogeneity and tumor impurity. Clonal heterogeneity covers genomic alterations within the tumor [95]. While de novo mutations accumulate, the tumor obtains genomic alterations with an exclusive pattern. When these genomic alterations are projected on the pathway, it is possible to observe exclusive relationships among disease-related genes. For instance, the CoMEt and MEMo algorithms examine mutual exclusivity on protein–protein interaction networks [96,97]. Moreover, the relationship between genes can be essential for an organism. Therefore, models analyzing such alterations integrate network-based analysis [98].

In contrast, tumor purity is dependent on the tumor microenvironment, including immune-cell infiltration and stromal cells [99]. In tumor microenvironment studies, network-based models are applied, for instance, to find immune-related gene modules. Although the importance of the interaction between tumors and immune cells is well known, detailed mechanisms are still unclear. Thus, many recent NGS studies employ network-based models to investigate the underlying mechanism in tumor and immune reactions. For example, McGrail et al. identified a relationship between the DNA damage response protein and immune cell infiltration in cancer. The analysis is based on curated interaction pairs in a protein–protein interaction network [100]. Most recently, Darzi et al. discovered a prognostic gene module related to immune cell infiltration by using network-centric approaches [101]. Tu et al. presented a network-centric model for mining subnetworks of genes other than immune cell infiltration by considering tumor purity [102].

3.3. Drug Target Identification with Network Information

In drug target studies, network biology is integrated into pharmacology [103]. For instance, Yamanishi et al. developed novel computational methods to investigate the pharmacological space by integrating a drug-target protein network with genomics and chemical information. The proposed approaches investigated such drug-target network information to identify potential novel drug targets [104]. Since then, the field has continued to develop methods to study drug target and drug response integrating networks with chemical and multi-omic datasets. In a recent survey study by Chen et al., the authors compared 13 computational methods for drug response prediction. It turned out that gene expression profiles are crucial information for drug response prediction [105].

Moreover, drug-target studies are often extended to drug-repurposing studies. In cancer research, drug-repurposing studies aim to find novel interactions between non-cancer drugs and molecular features in cancer. Drug-repurposing (or repositioning) studies apply computational approaches and pathway-based models and aim at discovering potential new cancer drugs with a higher probability than de novo drug design [16,106]. Specifically, drug-repurposing studies can consider various areas of cancer research, such as tumor heterogeneity and synthetic lethality. As an example, Lee et al. found clinically relevant synthetic lethality interactions by integrating multiple screening NGS datasets [107]. This synthetic lethality and related-drug datasets can be integrated for an effective combination of anticancer therapeutic strategy with non-cancer drug repurposing.

4. Deep Learning in Cancer Research

DNN models develop rapidly and become more sophisticated. They have been frequently used in all areas of biomedical research. Initially, its development was facilitated by large-scale imaging and video data. While most data sets in the biomedical field would not typically be considered big data, the rapid data accumulation enabled by NGS made it suitable for the application of DNN models requiring a large amount of training data [108]. For instance, in 2019, Samiei et al. used TCGA-based large-scale cancer data as benchmark datasets for bioinformatics machine learning research such as Image-Net in the computer vision field [109]. Subsequently, large-scale public cancer data sets such as TCGA encouraged the wide usage of DNNs in the cancer domain [110]. Over the last decade, these state-of-the-art machine learning methods have been incorporated in many different biological questions [111].

In addition to public cancer databases such as TCGA, the genetic information of normal tissues is stored in well-curated databases such as GTEx [112] and 1000Genomes [113]. These databases are frequently used as control or baseline training data for deep learning [114]. Moreover, other non-curated large-scale data sources such as GEO (https://www.ncbi.nlm.nih.gov/geo/, accessed on 20 May 2021) can be leveraged to tackle critical aspects in cancer research. They store a large-scale of biological data produced under various experimental setups (Figure 1). Therefore, an integration of GEO data and other data requires careful preprocessing. Overall, an increasing amount of datasets facilitate the development of current deep learning in bioinformatics research [115].

4.1. Challenges for Deep Learning in Cancer Research

Many studies in biology and medicine used NGS and produced large amounts of data during the past few decades, moving the field to the big data era. Nevertheless, researchers still face a lack of data in particular when investigating rare diseases or disease states. Researchers have developed a manifold of potential solutions to overcome this lack of data challenges, such as imputation, augmentation, and transfer learning (Figure 3b). Data imputation aims at handling data sets with missing values [116]. It has been studied on various NGS omics data types to recover missing information [117]. It is known that gene expression levels can be altered by different regulatory elements, such as DNA-binding proteins, epigenomic modifications, and post-transcriptional modifications. Therefore, various models integrating such regulatory schemes have been introduced to impute missing omics data [118,119]. Some DNN-based models aim to predict gene expression changes based on genomics or epigenomics alteration. For instance, TDimpute aims at generating missing RNA-seq data by training a DNN on methylation data. They used TCGA and TARGET (https://ocg.cancer.gov/programs/target/data-matrix, accessed on 20 May 2021) data as proof of concept of the applicability of DNN for data imputation in a multi-omics integration study [120]. Because this integrative model can exploit information in different levels of regulatory mechanisms, it can build a more detailed model and achieve better performance than a model build on a single-omics dataset [117,121]. The generative adversarial network (GAN) is a DNN structure for generating simulated data that is different from the original data but shows the same characteristics [122]. GANs can impute missing omics data from other multi-omics sources. Recently, the GAN algorithm is getting more attention in single-cell transcriptomics because it has been recognized as a complementary technique to overcome the limitation of scRNA-seq [123]. In contrast to data imputation and generation, other machine learning approaches aim to cope with a limited dataset in different ways. Transfer learning or few-shot learning, for instance, aims to reduce the search space with similar but unrelated datasets and guide the model to solve a specific set of problems [124]. These approaches train models with data of similar characteristics and types but different data to the problem set. After pre-training the model, it can be fine-tuned with the dataset of interest [125,126]. Thus, researchers are trying to introduce few-shot learning models and meta-learning approaches to omics and translational medicine. For example, Select-ProtoNet applied the ProtoTypical Network [127] model to TCGA transcriptome data and classified patients into two groups according to their clinical status [128]. AffinityNet predicts kidney and uterus cancer subtypes with gene expression profiles [129].

Figure 3. (a) In various studies, NGS data transformed into different forms. The 2-D transformed form is for the convolution layer. Omics data is transformed into pathway level, GO enrichment score, or Functional spectra. (b) DNN application on different ways to handle lack of data. Imputation for missing data in multi-omics datasets. GAN for data imputation and in silico data simulation. Transfer learning pre-trained the model with other datasets and fine-tune. (c) Various types of information in biology. (d) Graph neural network examples. GCN is applied to aggregate neighbor information. (Created with BioRender.com).

4.2. Molecular Charactization with Network and DNN Model

DNNs have been applied in multiple areas of cancer research. For instance, a DNN model trained on TCGA cancer data can aid molecular characterization by identifying cancer driver genes. At the very early stage, Yuan et al. build DeepGene, a cancer-type classifier. They implemented data sparsity reduction methods and trained the DNN model with somatic point mutations [130]. Lyu et al. [131] and DeepGx [132] embedded a 1-D gene expression profile to a 2-D array by chromosome order to implement the convolution layer (Figure 3a). Other algorithms, such as the deepDriver, use k-nearest neighbors for the convolution layer. A predefined number of neighboring gene mutation profiles was the input for the convolution layer. It employed this convolution layer in a DNN by aggregating mutation information of the k-nearest neighboring genes [11]. Instead of embedding to a 2-D image, DeepCC transformed gene expression data into functional spectra. The resulting model was able to capture molecular characteristics by training cancer subtypes [14].

Another DNN model was trained to infer the origin of tissue from single-nucleotide variant (SNV) information of metastatic tumor. The authors built a model by using the TCGA/ICGC data and analyzed SNV patterns and corresponding pathways to predict the origin of cancer. They discovered that metastatic tumors retained their original cancer’s signature mutation pattern. In this context, their DNN model obtained even better accuracy than a random forest model [133] and, even more important, better accuracy than human pathologists [12].

4.3. Tumor Heterogeneity with Network and DNN Model

As described in Section 4.1, there are several issues because of cancer heterogeneity, e.g., tumor microenvironment. Thus, there are only a few applications of DNN in intratumoral heterogeneity research. For instance, Menden et al. developed ’Scaden’ to deconvolve cell types in bulk-cell sequencing data. ’Scaden’ is a DNN model for the investigation of intratumor heterogeneity. To overcome the lack of training datasets, researchers need to generate in silico simulated bulk-cell sequencing data based on single-cell sequencing data [134]. It is presumed that deconvolving cell types can be achieved by knowing all possible expressional profiles of the cell [36]. However, this information is typically not available. Recently, to tackle this problem, single-cell sequencing-based studies were conducted. Because of technical limitations, we need to handle lots of missing data, noises, and batch effects in single-cell sequencing data [135]. Thus, various machine learning methods were developed to process single-cell sequencing data. They aim at mapping single-cell data onto the latent space. For example, scDeepCluster implemented an autoencoder and trained it on gene-expression levels from single-cell sequencing. During the training phase, the encoder and decoder work as denoiser. At the same time, they can embed high-dimensional gene-expression profiles to lower-dimensional vectors [136]. This autoencoder-based method can produce biologically meaningful feature vectors in various contexts, from tissue cell types [137] to different cancer types [138,139].

4.4. Drug Target Identification with Networks and DNN Models

In addition to NGS datasets, large-scale anticancer drug assays enabled the training train of DNNs. Moreover, non-cancer drug response assay datasets can also be incorporated with cancer genomic data. In cancer research, a multidisciplinary approach was widely applied for repurposing non-oncology drugs to cancer treatment. This drug repurposing is faster than de novo drug discovery. Furthermore, combination therapy with a non-oncology drug can be beneficial to overcome the heterogeneous properties of tumors [85]. The deepDR algorithm integrated ten drug-related networks and trained deep autoencoders. It used a random-walk-based algorithm to represent graph information into feature vectors. This approach integrated network analysis with a DNN model validated with an independent drug-disease dataset [15].

The authors of CDRscan did an integrative analysis of cell-line-based assay datasets and other drug and genomics datasets. It shows that DNN models can enhance the computational model for improved drug sensitivity predictions [140]. Additionally, similar to previous network-based models, the multi-omics application of drug-targeted DNN studies can show higher prediction accuracy than the single-omics method. MOLI integrated genomic data and transcriptomic data to predict the drug responses of TCGA patients [141].

4.5. Graph Neural Network Model

In general, the advantage of using a biological network is that it can produce more comprehensive and interpretable results from high-dimensional omics data. Furthermore, in an integrative multi-omics data analysis, network-based integration can improve interpretability over traditional approaches. Instead of pre-/post-integration of a network, recently developed graph neural networks use biological networks as the base structure for the learning network itself. For instance, various pathways or interactome information can be integrated as a learning structure of a DNN and can be aggregated as heterogeneous information. In a GNN study, a convolution process can be done on the provided network structure of data. Therefore, the convolution on a biological network made it possible for the GNN to focus on the relationship among neighbor genes. In the graph convolution layer, the convolution process integrates information of neighbor genes and learns topological information (Figure 3d). Consequently, this model can aggregate information from far-distant neighbors, and thus can outperform other machine learning models [142].

In the context of the inference problem of gene expression, the main question is whether the gene expression level can be explained by aggregating the neighboring genes. A single gene inference study by Dutil et al. showed that the GNN model outperformed other DNN models [143]. Moreover, in cancer research, such GNN models can identify cancer-related genes with better performance than other network-based models, such as HotNet2 and MutSigCV [144]. A recent GNN study with a multi-omics integrative analysis identified 165 new cancer genes as an interactive partner for known cancer genes [145]. Additionally, in the synthetic lethality area, dual-dropout GNN outperformed previous bioinformatics tools for predicting synthetic lethality in tumors [146]. GNNs were also able to classify cancer subtypes based on pathway activity measures with RNA-seq data. Lee et al. implemented a GNN for cancer subtyping and tested five cancer types. Thus, the informative pathway was selected and used for subtype classification [147]. Furthermore, GNNs are also getting more attention in drug repositioning studies. As described in Section 3.3, drug discovery requires integrating various networks in both chemical and genomic spaces (Figure 3d). Chemical structures, protein structures, pathways, and other multi-omics data were used in drug-target identification and repurposing studies (Figure 3c). Each of the proposed applications has a specialty in the different purposes of drug-related tasks. Sun et al. summarized GNN-based drug discovery studies and categorized them into four classes: molecular property and activity prediction, interaction prediction, synthesis prediction, and de novo drug design. The authors also point out four challenges in the GNN-mediated drug discovery. At first, as we described before, there is a lack of drug-related datasets. Secondly, the current GNN models can not fully represent 3-D structures of chemical molecules and protein structures. The third challenge is integrating heterogeneous network information. Drug discovery usually requires a multi-modal integrative analysis with various networks, and GNNs can improve this integrative analysis. Lastly, although GNNs use graphs, stacked layers still make it hard to interpret the model [148].

4.6. Shortcomings in AI and Revisiting Validity of Biological Networks as Prior Knowledge

The previous sections reviewed a variety of DNN-based approaches that present a good performance on numerous applications. However, it is hardly a panacea for all research questions. In the following, we will discuss potential limitations of the DNN models. In general, DNN models with NGS data have two significant issues: (i) data requirements and (ii) interpretability. Usually, deep learning needs a large proportion of training data for reasonable performance which is more difficult to achieve in biomedical omics data compared to, for instance, image data. Today, there are not many NGS datasets that are well-curated and -annotated for deep learning. This can be an answer to the question of why most DNN studies are in cancer research [110,149]. Moreover, the deep learning models are hard to interpret and are typically considered as black-boxes. Highly stacked layers in the deep learning model make it hard to interpret its decision-making rationale. Although the methodology to understand and interpret deep learning models has been improved, the ambiguity in the DNN models’ decision-making hindered the transition between the deep learning model and translational medicine [149,150].

As described before, biological networks are employed in various computational analyses for cancer research. The studies applying DNNs demonstrated many different approaches to use prior knowledge for systematic analyses. Before discussing GNN application, the validity of biological networks in a DNN model needs to be shown. The LINCS program analyzed data of ’The Connectivity Map (CMap) project’ to understand the regulatory mechanism in gene expression by inferring the whole gene expression profiles from a small set of genes (https://lincsproject.org/, accessed on 20 May 2021) [151,152]. This LINCS program found that the gene expression level is inferrable with only nearly 1000 genes. They called this gene list ’landmark genes’. Subsequently, Chen et al. started with these 978 landmark genes and tried to predict other gene expression levels with DNN models. Integrating public large-scale NGS data showed better performance than the linear regression model. The authors conclude that the performance advantage originates from the DNN’s ability to model non-linear relationships between genes [153].

Following this study, Beltin et al. extensively investigated various biological networks in the same context of the inference of gene expression level. They set up a simplified representation of gene expression status and tried to solve a binary classification task. To show the relevance of a biological network, they compared various gene expression levels inferred from a different set of genes, neighboring genes in PPI, random genes, and all genes. However, in the study incorporating TCGA and GTEx datasets, the random network model outperformed the model build on a known biological network, such as StringDB [154]. While network-based approaches can add valuable insights to analysis, this study shows that it cannot be seen as the panacea, and a careful evaluation is required for each data set and task. In particular, this result may not represent biological complexity because of the oversimplified problem setup, which did not consider the relative gene-expressional changes. Additionally, the incorporated biological networks may not be suitable for inferring gene expression profiles because they consist of expression-regulating interactions, non-expression-regulating interactions, and various in vivo and in vitro interactions.

“ However, although recently sophisticated applications of deep learning showed improved accuracy, it does not reflect a general advancement. Depending on the type of NGS data, the experimental design, and the question to be answered, a proper approach and specific deep learning algorithms need to be considered. Deep learning is not a panacea. In general, to employ machine learning and systems biology methodology for a specific type of NGS data, a certain experimental design, a particular research question, the technology, and network data have to be chosen carefully.”

References

  1. Janes, K.A.; Yaffe, M.B. Data-driven modelling of signal-transduction networks. Nat. Rev. Mol. Cell Biol. 20067, 820–828. [Google Scholar] [CrossRef] [PubMed]
  2. Kreeger, P.K.; Lauffenburger, D.A. Cancer systems biology: A network modeling perspective. Carcinogenesis 201031, 2–8. [Google Scholar] [CrossRef] [PubMed]
  3. Vucic, E.A.; Thu, K.L.; Robison, K.; Rybaczyk, L.A.; Chari, R.; Alvarez, C.E.; Lam, W.L. Translating cancer ‘omics’ to improved outcomes. Genome Res. 201222, 188–195. [Google Scholar] [CrossRef]
  4. Hoadley, K.A.; Yau, C.; Wolf, D.M.; Cherniack, A.D.; Tamborero, D.; Ng, S.; Leiserson, M.D.; Niu, B.; McLellan, M.D.; Uzunangelov, V.; et al. Multiplatform analysis of 12 cancer types reveals molecular classification within and across tissues of origin. Cell 2014158, 929–944. [Google Scholar] [CrossRef] [PubMed]
  5. Hutter, C.; Zenklusen, J.C. The cancer genome atlas: Creating lasting value beyond its data. Cell 2018173, 283–285. [Google Scholar] [CrossRef]
  6. Chuang, H.Y.; Lee, E.; Liu, Y.T.; Lee, D.; Ideker, T. Network-based classification of breast cancer metastasis. Mol. Syst. Biol. 20073, 140. [Google Scholar] [CrossRef]
  7. Zhang, W.; Chien, J.; Yong, J.; Kuang, R. Network-based machine learning and graph theory algorithms for precision oncology. NPJ Precis. Oncol. 20171, 25. [Google Scholar] [CrossRef] [PubMed]
  8. Ngiam, K.Y.; Khor, W. Big data and machine learning algorithms for health-care delivery. Lancet Oncol. 201920, e262–e273. [Google Scholar] [CrossRef]
  9. Creixell, P.; Reimand, J.; Haider, S.; Wu, G.; Shibata, T.; Vazquez, M.; Mustonen, V.; Gonzalez-Perez, A.; Pearson, J.; Sander, C.; et al. Pathway and network analysis of cancer genomes. Nat. Methods 201512, 615. [Google Scholar]
  10. Reyna, M.A.; Haan, D.; Paczkowska, M.; Verbeke, L.P.; Vazquez, M.; Kahraman, A.; Pulido-Tamayo, S.; Barenboim, J.; Wadi, L.; Dhingra, P.; et al. Pathway and network analysis of more than 2500 whole cancer genomes. Nat. Commun. 202011, 729. [Google Scholar] [CrossRef]
  11. Luo, P.; Ding, Y.; Lei, X.; Wu, F.X. deepDriver: Predicting cancer driver genes based on somatic mutations using deep convolutional neural networks. Front. Genet. 201910, 13. [Google Scholar] [CrossRef]
  12. Jiao, W.; Atwal, G.; Polak, P.; Karlic, R.; Cuppen, E.; Danyi, A.; De Ridder, J.; van Herpen, C.; Lolkema, M.P.; Steeghs, N.; et al. A deep learning system accurately classifies primary and metastatic cancers using passenger mutation patterns. Nat. Commun. 202011, 728. [Google Scholar] [CrossRef]
  13. Chaudhary, K.; Poirion, O.B.; Lu, L.; Garmire, L.X. Deep learning–based multi-omics integration robustly predicts survival in liver cancer. Clin. Cancer Res. 201824, 1248–1259. [Google Scholar] [CrossRef]
  14. Gao, F.; Wang, W.; Tan, M.; Zhu, L.; Zhang, Y.; Fessler, E.; Vermeulen, L.; Wang, X. DeepCC: A novel deep learning-based framework for cancer molecular subtype classification. Oncogenesis 20198, 44. [Google Scholar] [CrossRef]
  15. Zeng, X.; Zhu, S.; Liu, X.; Zhou, Y.; Nussinov, R.; Cheng, F. deepDR: A network-based deep learning approach to in silico drug repositioning. Bioinformatics 201935, 5191–5198. [Google Scholar] [CrossRef]
  16. Issa, N.T.; Stathias, V.; Schürer, S.; Dakshanamurthy, S. Machine and deep learning approaches for cancer drug repurposing. In Seminars in Cancer Biology; Elsevier: Amsterdam, The Netherlands, 2020. [Google Scholar]
  17. Weinstein, J.N.; Collisson, E.A.; Mills, G.B.; Shaw, K.R.M.; Ozenberger, B.A.; Ellrott, K.; Shmulevich, I.; Sander, C.; Stuart, J.M.; Network, C.G.A.R.; et al. The cancer genome atlas pan-cancer analysis project. Nat. Genet. 201345, 1113. [Google Scholar] [CrossRef]
  18. The ICGC/TCGA Pan-Cancer Analysis of Whole Genomes Consortium. Pan-cancer analysis of whole genomes. Nature 2020578, 82. [Google Scholar] [CrossRef] [PubMed]
  19. King, M.C.; Marks, J.H.; Mandell, J.B. Breast and ovarian cancer risks due to inherited mutations in BRCA1 and BRCA2. Science 2003302, 643–646. [Google Scholar] [CrossRef] [PubMed]
  20. Courtney, K.D.; Corcoran, R.B.; Engelman, J.A. The PI3K pathway as drug target in human cancer. J. Clin. Oncol. 201028, 1075. [Google Scholar] [CrossRef] [PubMed]
  21. Parker, J.S.; Mullins, M.; Cheang, M.C.; Leung, S.; Voduc, D.; Vickery, T.; Davies, S.; Fauron, C.; He, X.; Hu, Z.; et al. Supervised risk predictor of breast cancer based on intrinsic subtypes. J. Clin. Oncol. 200927, 1160. [Google Scholar] [CrossRef]
  22. Yersal, O.; Barutca, S. Biological subtypes of breast cancer: Prognostic and therapeutic implications. World J. Clin. Oncol. 20145, 412. [Google Scholar] [CrossRef] [PubMed]
  23. Zhao, L.; Lee, V.H.; Ng, M.K.; Yan, H.; Bijlsma, M.F. Molecular subtyping of cancer: Current status and moving toward clinical applications. Brief. Bioinform. 201920, 572–584. [Google Scholar] [CrossRef] [PubMed]
  24. Jones, P.A.; Issa, J.P.J.; Baylin, S. Targeting the cancer epigenome for therapy. Nat. Rev. Genet. 201617, 630. [Google Scholar] [CrossRef] [PubMed]
  25. Huang, S.; Chaudhary, K.; Garmire, L.X. More is better: Recent progress in multi-omics data integration methods. Front. Genet. 20178, 84. [Google Scholar] [CrossRef]
  26. Chin, L.; Andersen, J.N.; Futreal, P.A. Cancer genomics: From discovery science to personalized medicine. Nat. Med. 201117, 297. [Google Scholar] [CrossRef] [PubMed]

Use of Systems Biology in Anti-Microbial Drug Development

Genomics, Computational Biology and Drug Discovery for Mycobacterial Infections: Fighting the Emergence of Resistance. Asma Munir, Sundeep Chaitanya Vedithi, Amanda K. Chaplin and Tom L. Blundell. Front. Genet., 04 September 2020 | https://doi.org/10.3389/fgene.2020.00965

In an earlier review article (Waman et al., 2019), we discussed various computational approaches and experimental strategies for drug target identification and structure-guided drug discovery. In this review we discuss the impact of the era of precision medicine, where the genome sequences of pathogens can give clues about the choice of existing drugs, and repurposing of others. Our focus is directed toward combatting antimicrobial drug resistance with emphasis on tuberculosis and leprosy. We describe structure-guided approaches to understanding the impacts of mutations that give rise to antimycobacterial resistance and the use of this information in the design of new medicines.

Genome Sequences and Proteomic Structural Databases

In recent years, there have been many focused efforts to define the amino-acid sequences of the M. tuberculosis pan-genome and then to define the three-dimensional structures and functional interactions of these gene products. This work has led to essential genes of the bacteria being revealed and to a better understanding of the genetic diversity in different strains that might lead to a selective advantage (Coll et al., 2018). This will help with our understanding of the mode of antibiotic resistance within these strains and aid structure-guided drug discovery. However, only ∼10% of the ∼4128 proteins have structures determined experimentally.

Several databases have been developed to integrate the genomic and/or structural information linked to drug resistance in Mycobacteria (Table 1). These invaluable resources can contribute to better understanding of molecular mechanisms involved in drug resistance and improvement in the selection of potential drug targets.

There is a dearth of information related to structural aspects of proteins from M. leprae and their oligomeric and hetero-oligomeric organization, which has limited the understanding of physiological processes of the bacillus. The structures of only 12 proteins have been solved and deposited in the protein data bank (PDB). However, the high sequence similarity in protein coding genes between M. leprae and M. tuberculosis allows computational methods to be used for comparative modeling of the proteins of M. leprae. Mainly monomeric models using single template modeling have been defined and deposited in the Swiss Model repository (Bienert et al., 2017), in Modbase (Pieper et al., 2014), and in a collection with other infectious disease agents (Sosa et al., 2018). There is a need for multi-template modeling and building homo- and hetero-oligomeric complexes to better understand the interfaces, druggability and impacts of mutations.

We are now exploiting Vivace, a multi-template modeling pipeline developed in our lab for modeling the proteomes of M. tuberculosis (CHOPIN, see above) and M. abscessus [Mabellini Database (Skwark et al., 2019)], to model the proteome of M. leprae. We emphasize the need for understanding the protein interfaces that are critical to function. An example of this is that of the RNA-polymerase holoenzyme complex from M. leprae. We first modeled the structure of this hetero-hexamer complex and later deciphered the binding patterns of rifampin (Vedithi et al., 2018Figures 1A,B). Rifampin is a known drug to treat tuberculosis and leprosy. Owing to high rifampin resistance in tuberculosis and emerging resistance in leprosy, we used an approach known as “Computational Saturation Mutagenesis”, to identify sites on the protein that are less impacted by mutations. In this study, we were able to understand the association between predicted impacts of mutations on the structure and phenotypic rifampin-resistance outcomes in leprosy.

FIGURE 2

Figure 2. (A) Stability changes predicted by mCSM for systematic mutations in the ß-subunit of RNA polymerase in M. leprae. The maximum destabilizing effect from among all 19 possible mutations at each residue position is considered as a weighting factor for the color map that gradients from red (high destabilizing effects) to white (neutral to stabilizing effects) (Vedithi et al., 2020). (B) One of the known mutations in the ß-subunit of RNA polymerase, the S437H substitution which resulted in a maximum destabilizing effect [-1.701 kcal/mol (mCSM)] among all 19 possibilities this position. In the mutant, histidine (residue in green) forms hydrogen bonds with S434 and Q438, aromatic interactions with F431, and other ring-ring and π interactions with the surrounding residues which can impact the shape of the rifampin binding pocket and rifampin affinity to the ß-subunit [-0.826 log(affinity fold change) (mCSM-lig)]. Orange dotted lines represent weak hydrogen bond interactions. Ring-ring and intergroup interactions are depicted in cyan. Aromatic interactions are represented in sky-blue and carbonyl interactions in pink dotted lines. Green dotted lines represent hydrophobic interactions (Vedithi et al., 2020).

Examples of Understanding and Combatting Resistance

The availability of whole genome sequences in the present era has greatly enhanced the understanding of emergence of drug resistance in infectious diseases like tuberculosis. The data generated by the whole genome sequencing of clinical isolates can be screened for the presence of drug-resistant mutations. A preliminary in silico analysis of mutations can then be used to prioritize experimental work to identify the nature of these mutations.

FIGURE 3

Figure 3. (A) Mechanism of isoniazid activation and INH-NAD adduct formation. (B) Mutations mapped (Munir et al., 2019) on the structure of KatG (PDB ID:1SJ2; Bertrand et al., 2004).

Other articles related to Computational Biology, Systems Biology, and Bioinformatics on this online journal include:

20th Anniversary and the Evolution of Computational Biology – International Society for Computational Biology

Featuring Computational and Systems Biology Program at Memorial Sloan Kettering Cancer Center, Sloan Kettering Institute (SKI), The Dana Pe’er Lab

Quantum Biology And Computational Medicine

Systems Biology Analysis of Transcription Networks, Artificial Intelligence, and High-End Computing Coming to Fruition in Personalized Oncology

Read Full Post »

Thriving Vaccines and Research: Weizmann Institute Coronavirus Research Development

Reporter: Amandeep Kaur, B.Sc., M.Sc.

In early February, Prof. Eran Segal updated in one of his tweets and mentioned that “We say with caution, the magic has started.”

The article reported that this statement by Prof. Segal was due to decreasing cases of COVID-19, severe infection cases and hospitalization of patients by rapid vaccination process throughout Israel. Prof. Segal emphasizes in another tweet to remain cautious over the country and informed that there is a long way to cover and searching for scientific solutions.

A daylong webinar entitled “COVID-19: The epidemic that rattles the world” was a great initiative by Weizmann Institute to share their scientific knowledge about the infection among the Israeli institutions and scientists. Prof. Gideon Schreiber and Dr. Ron Diskin organized the event with the support of the Weizmann Coronavirus Response Fund and Israel Society for Biochemistry and Molecular Biology. The speakers were invited from the Hebrew University of Jerusalem, Tel-Aviv University, the Israel Institute for Biological Research (IIBR), and Kaplan Medical Center who addressed the molecular structure and infection biology of the virus, treatments and medications for COVID-19, and the positive and negative effect of the pandemic.

The article reported that with the emergence of pandemic, the scientists at Weizmann started more than 60 projects to explore the virus from different range of perspectives. With the help of funds raised by communities worldwide for the Weizmann Coronavirus Response Fund supported scientists and investigators to elucidate the chemistry, physics and biology behind SARS-CoV-2 infection.

Prof. Avi Levy, the coordinator of the Weizmann Institute’s coronavirus research efforts, mentioned “The vaccines are here, and they will drastically reduce infection rates. But the coronavirus can mutate, and there are many similar infectious diseases out there to be dealt with. All of this research is critical to understanding all sorts of viruses and to preempting any future pandemics.”

The following are few important projects with recent updates reported in the article.

Mapping a hijacker’s methods

Dr. Noam Stern-Ginossar studied the virus invading strategies into the healthy cells and hijack the cell’s systems to divide and reproduce. The article reported that viruses take over the genetic translation system and mainly the ribosomes to produce viral proteins. Dr. Noam used a novel approach known as ‘ribosome profiling’ as her research objective and create a map to locate the translational events taking place inside the viral genome, which further maps the full repertoire of viral proteins produced inside the host.

She and her team members grouped together with the Weizmann’s de Botton Institute and researchers at IIBR for Protein Profiling and understanding the hijacking instructions of coronavirus and developing tools for treatment and therapies. Scientists generated a high-resolution map of the coding regions in the SARS-CoV-2 genome using ribosome-profiling techniques, which allowed researchers to quantify the expression of vital zones along the virus genome that regulates the translation of viral proteins. The study published in Nature in January, explains the hijacking process and reported that virus produces more instruction in the form of viral mRNA than the host and thus dominates the translation process of the host cell. Researchers also clarified that it is the misconception that virus forced the host cell to translate its viral mRNA more efficiently than the host’s own translation, rather high level of viral translation instructions causes hijacking. This study provides valuable insights for the development of effective vaccines and drugs against the COVID-19 infection.

Like chutzpah, some things don’t translate

Prof. Igor Ulitsky and his team worked on untranslated region of viral genome. The article reported that “Not all the parts of viral transcript is translated into protein- rather play some important role in protein production and infection which is unknown.” This region may affect the molecular environment of the translated zones. The Ulitsky group researched to characterize that how the genetic sequence of regions that do not translate into proteins directly or indirectly affect the stability and efficiency of the translating sequences.

Initially, scientists created the library of about 6,000 regions of untranslated sequences to further study their functions. In collaboration with Dr. Noam Stern-Ginossar’s lab, the researchers of Ulitsky’s team worked on Nsp1 protein and focused on the mechanism that how such regions affect the Nsp1 protein production which in turn enhances the virulence. The researchers generated a new alternative and more authentic protocol after solving some technical difficulties which included infecting cells with variants from initial library. Within few months, the researchers are expecting to obtain a more detailed map of how the stability of Nsp1 protein production is getting affected by specific sequences of the untranslated regions.

The landscape of elimination

The article reported that the body’s immune system consists of two main factors- HLA (Human Leukocyte antigen) molecules and T cells for identifying and fighting infections. HLA molecules are protein molecules present on the cell surface and bring fragments of peptide to the surface from inside the infected cell. These peptide fragments are recognized and destroyed by the T cells of the immune system. Samuels’ group tried to find out the answer to the question that how does the body’s surveillance system recognizes the appropriate peptide derived from virus and destroy it. They isolated and analyzed the ‘HLA peptidome’- the complete set of peptides bound to the HLA proteins from inside the SARS-CoV-2 infected cells.

After the analysis of infected cells, they found 26 class-I and 36 class-II HLA peptides, which are present in 99% of the population around the world. Two peptides from HLA class-I were commonly present on the cell surface and two other peptides were derived from coronavirus rare proteins- which mean that these specific coronavirus peptides were marked for easy detection. Among the identified peptides, two peptides were novel discoveries and seven others were shown to induce an immune response earlier. These results from the study will help to develop new vaccines against new coronavirus mutation variants.

Gearing up ‘chain terminators’ to battle the coronavirus

Prof. Rotem Sorek and his lab discovered a family of enzymes within bacteria that produce novel antiviral molecules. These small molecules manufactured by bacteria act as ‘chain terminators’ to fight against the virus invading the bacteria. The study published in Nature in January which reported that these molecules cause a chemical reaction that halts the virus’s replication ability. These new molecules are modified derivates of nucleotide which integrates at the molecular level in the virus and obstruct the works.

Prof. Sorek and his group hypothesize that these new particles could serve as a potential antiviral drug based on the mechanism of chain termination utilized in antiviral drugs used recently in the clinical treatments. Yeda Research and Development has certified these small novel molecules to a company for testing its antiviral mechanism against SARS-CoV-2 infection. Such novel discoveries provide evidences that bacterial immune system is a potential repository of many natural antiviral particles.

Resolving borderline diagnoses

Currently, Real-time Polymerase chain reaction (RT-PCR) is the only choice and extensively used for diagnosis of COVID-19 patients around the globe. Beside its benefits, there are problems associated with RT-PCR, false negative and false positive results and its limitation in detecting new mutations in the virus and emerging variants in the population worldwide. Prof. Eran Elinavs’ lab and Prof. Ido Amits’ lab are working collaboratively to develop a massively parallel, next-generation sequencing technique that tests more effectively and precisely as compared to RT-PCR. This technique can characterize the emerging mutations in SARS-CoV-2, co-occurring viral, bacterial and fungal infections and response patterns in human.

The scientists identified viral variants and distinctive host signatures that help to differentiate infected individuals from non-infected individuals and patients with mild symptoms and severe symptoms.

In Hadassah-Hebrew University Medical Center, Profs. Elinav and Amit are performing trails of the pipeline to test the accuracy in borderline cases, where RT-PCR shows ambiguous or incorrect results. For proper diagnosis and patient stratification, researchers calibrated their severity-prediction matrix. Collectively, scientists are putting efforts to develop a reliable system that resolves borderline cases of RT-PCR and identify new virus variants with known and new mutations, and uses data from human host to classify patients who are needed of close observation and extensive treatment from those who have mild complications and can be managed conservatively.

Moon shot consortium refining drug options

The ‘Moon shot’ consortium was launched almost a year ago with an initiative to develop a novel antiviral drug against SARS-CoV-2 and was led by Dr. Nir London of the Department of Chemical and Structural Biology at Weizmann, Prof. Frank von Delft of Oxford University and the UK’s Diamond Light Source synchroton facility.

To advance the series of novel molecules from conception to evidence of antiviral activity, the scientists have gathered support, guidance, expertise and resources from researchers around the world within a year. The article reported that researchers have built an alternative template for drug-discovery, full transparency process, which avoids the hindrance of intellectual property and red tape.

The new molecules discovered by scientists inhibit a protease, a SARS-CoV-2 protein playing important role in virus replication. The team collaborated with the Israel Institute of Biological Research and other several labs across the globe to demonstrate the efficacy of molecules not only in-vitro as well as in analysis against live virus.

Further research is performed including assaying of safety and efficacy of these potential drugs in living models. The first trial on mice has been started in March. Beside this, additional drugs are optimized and nominated for preclinical testing as candidate drug.

Source: https://www.weizmann.ac.il/WeizmannCompass/sections/features/the-vaccines-are-here-and-research-abounds

Other related articles were published in this Open Access Online Scientific Journal, including the following:

Identification of Novel genes in human that fight COVID-19 infection

Reporter: Amandeep Kaur, B.Sc., M.Sc. (ept. 5/2021)

https://pharmaceuticalintelligence.com/2021/04/19/identification-of-novel-genes-in-human-that-fight-covid-19-infection/

Fighting Chaos with Care, community trust, engagement must be cornerstones of pandemic response

Reporter: Amandeep Kaur, B.Sc., M.Sc. (ept. 5/2021)

https://pharmaceuticalintelligence.com/2021/04/13/fighting-chaos-with-care/

T cells recognize recent SARS-CoV-2 variants

Reporter: Aviva Lev-Ari, PhD, RN

https://pharmaceuticalintelligence.com/2021/03/30/t-cells-recognize-recent-sars-cov-2-variants/

Need for Global Response to SARS-CoV-2 Viral Variants

Reporter: Aviva Lev-Ari, PhD, RN

https://pharmaceuticalintelligence.com/2021/02/12/need-for-global-response-to-sars-cov-2-viral-variants/

Mechanistic link between SARS-CoV-2 infection and increased risk of stroke using 3D printed models and human endothelial cells

Reporter: Adina Hazan, PhD

https://pharmaceuticalintelligence.com/2020/12/28/mechanistic-link-between-sars-cov-2-infection-and-increased-risk-of-stroke-using-3d-printed-models-and-human-endothelial-cells/

Read Full Post »

National Resilience, Inc. is a first-of-its-kind manufacturing and technology company dedicated to broadening access to complex medicines and protecting biopharmaceutical supply chains against disruption – the Acquisition of Two Premier Biologics Manufacturing Facilities: Boston and in Ontario, Canada

 

Reporter: Aviva Lev-Ari, PhD, RN

Resilience’s new facility, located at 500 Soldiers Field Rd., Boston, MA. (Photo: Business Wire) – The Genzyme-Sanofi Building

 

SAN DIEGO & BOSTON–(BUSINESS WIRE)–Resilience (National Resilience, Inc.), a new company building the world’s most advanced biopharmaceutical manufacturing ecosystem, announced it has acquired two premier commercial manufacturing facilities in North America, joining other facilities already in Resilience’s network to boost total capacity under management to more than 750,000 square feet.

“These locations will serve as hubs for the future of biopharma manufacturing, leading the way and shaping the future of Resilience.”

  • The acquired facilities include a 310,000-square-foot plant in Boston, MA, purchased from Sanofi; and in a separate transaction,
  • a 136,000-square-foot plant in Mississauga, Ontario, Canada.

Both facilities, which currently produce commercial, marketed products, will see significant investments as Resilience adds capacity and capabilities to produce new therapies at these locations. In addition, the company has offered employment to the existing plant staff and intends to add more jobs at each facility.

“We have big plans for these facilities including investing in new capacity, applying new manufacturing technologies, creating jobs and bringing in new customers,” said Rahul Singhvi, Sc.D, Chief Executive Officer of Resilience. “These locations will serve as hubs for the future of biopharma manufacturing, leading the way and shaping the future of Resilience.”

As part of its agreement with Sanofi, Resilience will continue to manufacture a marketed product at the Boston location. The facility plan includes a build out to facilitate multi-modality manufacturing and state-of-the-art quality laboratories to ensure safe, reliable supply to patients. The facility itself is certified ISO 14001 (Environmental management system), OSHAS 18001 (Health & safety management system) and ISO 50001 (Energy management system).​

This is currently the largest of several facilities in Resilience’s growing biologics and advanced therapeutics manufacturing network, with plans to acquire and develop other sites in the U.S. this year. The facility offers 24/7/365 production, multiple 2000L bioreactors capacity and multiple downstream processing trains, with investment in additional capabilities to come.

Our state-of-the-art flexible facility in Mississauga, Ontario, provides upstream, downstream and aseptic fill finish, and is designed to comply with cGMP. The plant has been inspected and approved by multiple regulatory bodies, and handles development and commercialized products.

About Resilience

Resilience (National Resilience, Inc.) is a first-of-its-kind manufacturing and technology company dedicated to broadening access to complex medicines and protecting biopharmaceutical supply chains against disruption. Founded in 2020, the company is building a sustainable network of high-tech, end-to-end manufacturing solutions to ensure the medicines of today and tomorrow can be made quickly, safely, and at scale. Resilience offers the highest quality and regulatory capabilities, and flexible and adaptive facilities to serve partners of all sizes. By continuously advancing the science of biopharmaceutical manufacturing and development, Resilience frees partners to focus on the discoveries that improve patients’ lives.

For more information, visit www.Resilience.com.

Contacts

Ryan Flinn
Head of Communications
Ryan.flinn@Resilience.com
510-207-7616

Read Full Post »

Reporter and Curator: Dr. Sudipta Saha, Ph.D.

 

Parkinson’s Disease (PD), characterized by both motor and non-motor system pathology, is a common neurodegenerative disorder affecting about 1% of the population over age 60. Its prevalence presents an increasing social burden as the population ages. Since its introduction in the 1960’s, dopamine (DA)-replacement therapy (e.g., L-DOPA) has remained the gold standard treatment. While improving PD patients’ quality of life, the effects of treatment fade with disease progression and prolonged usage of these medications often (>80%) results in side effects including dyskinesias and motor fluctuations. Since the selective degeneration of A9 mDA neurons (mDANs) in the substantia nigra (SN) is a key pathological feature of the disease and is directly associated with the cardinal motor symptoms, dopaminergic cell transplantation has been proposed as a therapeutic strategy.

 

Researchers showed that mammalian fibroblasts can be converted into embryonic stem cell (ESC)-like induced pluripotent stem cells (iPSCs) by introducing four transcription factors i.e., Oct4, Sox2, Klf4, and c-Myc. This was then accomplished with human somatic cells, reprogramming them into human iPSCs (hiPSCs), offering the possibility of generating patient-specific stem cells. There are several major barriers to implementation of hiPSC-based cell therapy for PD. First, probably due to the limited understanding of the reprogramming process, wide variability exists between the differentiation potential of individual hiPSC lines. Second, the safety of hiPSC-based cell therapy has yet to be fully established. In particular, since any hiPSCs that remain undifferentiated or bear sub-clonal tumorigenic mutations have neoplastic potential, it is critical to eliminate completely such cells from a therapeutic product.

 

In the present study the researchers established human induced pluripotent stem cell (hiPSC)-based autologous cell therapy. Researchers reported a platform of core techniques for the production of mDA progenitors as a safe and effective therapeutic product. First, by combining metabolism-regulating microRNAs with reprogramming factors, a method was developed to more efficiently generate clinical grade iPSCs, as evidenced by genomic integrity and unbiased pluripotent potential. Second, a “spotting”-based in vitro differentiation methodology was established to generate functional and healthy mDA cells in a scalable manner. Third, a chemical method was developed that safely eliminates undifferentiated cells from the final product. Dopaminergic cells thus produced can express high levels of characteristic mDA markers, produce and secrete dopamine, and exhibit electrophysiological features typical of mDA cells. Transplantation of these cells into rodent models of PD robustly restored motor dysfunction and reinnervated host brain, while showing no evidence of tumor formation or redistribution of the implanted cells.

 

Together these results supported the promise of these techniques to provide clinically applicable personalized autologous cell therapy for PD. It was recognized by researchers that this methodology is likely to be more costly in dollars and manpower than techniques using off-the-shelf methods and allogenic cell lines. Nevertheless, the cost for autologous cell therapy may be expected to decrease steadily with technological refinement and automation. Given the significant advantages inherent in a cell source free of ethical concerns and with the potential to obviate the need for immunosuppression, with its attendant costs and dangers, it was proposed that this platform is suitable for the successful implementation of human personalized autologous cell therapy for PD.

 

References:

 

https://www.jci.org/articles/view/130767/pdf?elqTrackId=2fd7d0edee744f9cb6d70a686d7b273b

 

https://www.ncbi.nlm.nih.gov/pubmed/31714896

 

https://www.ncbi.nlm.nih.gov/pubmed/23666606

 

https://www.ncbi.nlm.nih.gov/pubmed/27343168

 

https://www.ncbi.nlm.nih.gov/pubmed/21495962

 

https://www.ncbi.nlm.nih.gov/pubmed/28083784

 

https://www.ncbi.nlm.nih.gov/pubmed/20336395

 

https://www.ncbi.nlm.nih.gov/pubmed/28585381

 

Read Full Post »

Extracellular RNA and their carriers in disease diagnosis and therapy, Volume 2 (Volume Two: Latest in Genomics Methodologies for Therapeutics: Gene Editing, NGS and BioInformatics, Simulations and the Genome Ontology), Part 1: Next Generation Sequencing (NGS)

Reporter and Curator: Dr. Sudipta Saha, Ph.D.

 

RNA plays various roles in determining how the information in our genes drives cell behavior. One of its roles is to carry information encoded by our genes from the cell nucleus to the rest of the cell where it can be acted on by other cell components. Rresearchers have now defined how RNA also participates in transmitting information outside cells, known as extracellular RNA or exRNA. This new role of RNA in cell-to-cell communication has led to new discoveries of potential disease biomarkers and therapeutic targets. Cells using RNA to talk to each other is a significant shift in the general thought process about RNA biology.

 

Researchers explored basic exRNA biology, including how exRNA molecules and their transport packages (or carriers) were made, how they were expelled by producer cells and taken up by target cells, and what the exRNA molecules did when they got to their destination. They encountered surprising complexity both in the types of carriers that transport exRNA molecules between cells and in the different types of exRNA molecules associated with the carriers. The researchers had to be exceptionally creative in developing molecular and data-centric tools to begin making sense of the complexity, and found that the type of carrier affected how exRNA messages were sent and received.

 

As couriers of information between cells, exRNA molecules and their carriers give researchers an opportunity to intercept exRNA messages to see if they are associated with disease. If scientists could change or engineer designer exRNA messages, it may be a new way to treat disease. The researchers identified potential exRNA biomarkers for nearly 30 diseases including cardiovascular disease, diseases of the brain and central nervous system, pregnancy complications, glaucoma, diabetes, autoimmune diseases and multiple types of cancer.

 

As for example some researchers found that exRNA in urine showed promise as a biomarker of muscular dystrophy where current studies rely on markers obtained through painful muscle biopsies. Some other researchers laid the groundwork for exRNA as therapeutics with preliminary studies demonstrating how researchers might load exRNA molecules into suitable carriers and target carriers to intended recipient cells, and determining whether engineered carriers could have adverse side effects. Scientists engineered carriers with designer RNA messages to target lab-grown breast cancer cells displaying a certain protein on their surface. In an animal model of breast cancer with the cell surface protein, the researchers showed a reduction in tumor growth after engineered carriers deposited their RNA cargo.

 

Other than the above research work the scientists also created a catalog of exRNA molecules found in human biofluids like plasma, saliva and urine. They analyzed over 50,000 samples from over 2000 donors, generating exRNA profiles for 13 biofluids. This included over 1000 exRNA profiles from healthy volunteers. The researchers found that exRNA profiles varied greatly among healthy individuals depending on characteristics like age and environmental factors like exercise. This means that exRNA profiles can give important and detailed information about health and disease, but careful comparisons need to be made with exRNA data generated from people with similar characteristics.

 

Next the researchers will develop tools to efficiently and reproducibly isolate, identify and analyze different carrier types and their exRNA cargos and allow analysis of one carrier and its cargo at a time. These tools will be shared with the research community to fill gaps in knowledge generated till now and to continue to move this field forward.

 

References:

 

https://www.nih.gov/news-events/news-releases/scientists-explore-new-roles-rna

 

https://www.cell.com/consortium/exRNA

 

https://www.sciencedaily.com/releases/2016/06/160606120230.htm

 

https://www.pasteur.fr/en/multiple-roles-rnas

 

https://www.nature.com/scitable/topicpage/rna-functions-352

 

https://www.umassmed.edu/rti/biology/role-of-rna-in-biology/

 

Read Full Post »

UPDATED on 3/17/2019

https://www.medpagetoday.com/cardiology/prevention/78202?xid=nl_mpt_SRCardiology_2019-02-25&eun=g99985d0r&utm_source=Sailthru&utm_medium=email&utm_campaign=CardioUpdate_022519&utm_term=NL_Spec_Cardiology_Update_Active

Transgender hormone therapy appears to increase cardiovascular risk. (Circulation)

A mobile app with a step-by-step guide to prepping vasoactive drugs for CPR of children in the emergency room substantially cut medication errors, drug preparation time, and delivery time compared with using infusion-rate tables in a study using manikins. (The Lancet Child & Adolescent Health)

 

Artificial ovary instead of conventional hormone replacement

Reporter and Curator: Dr. Sudipta Saha, Ph.D.

During menopause a woman’s ovaries stop working—leading to hot flashes, sleep problems, weight gain, and worse, bone deterioration. Now scientists are exploring whether transplanting lab-made ovaries might stop those symptoms. In one of the first efforts to explore the potential of such a technique, researchers say they used tissue engineering to construct artificial rat ovaries able to supply female hormones like estrogen and progesterone. A research carried out at Wake Forest Baptist Medical Center, suggests a potential alternative to the synthetic hormones millions of women take after reaching middle age. A paper describing the findings was published in Nature Communications.

 

Women going through menopause, as well as those who have undergone cancer treatment or had their ovaries removed for medical purposes, lose the ability to produce important hormones, including estrogen and progesterone. Lower levels of these hormones can affect a number of different body functions. To counteract unpleasant symptoms, many women turn to combinations of hormone replacement medications—synthetic estrogen and progestin. Pharmacologic hormone replacement therapy (pHRT) with estrogen alone or estrogen and progestogens is known to effectively ameliorate the unpleasant symptoms. But hormone replacement carries an increased risk of heart disease and breast cancer, so it’s not recommended for long-term use. In these circumstances artificial ovaries could be safer and more effective.

 

Regenerative medicine approaches that use cell-based hormone replacement therapy (cHRT) offer a potential solution to temporal control of hormone delivery and the ability to restore the HPO (Hypothalamo-Pituitary-Ovarian) axis in a way not possible with pHRT. Scientists have previously described an approach to achieve microencapsulation of ovarian cells that results in bioengineered constructs that replicate key structure-function relationships of ovarian follicles as an approach to cHRT. In the present study the scientists have adapted an isogeneic cell-based construct to provide a proof-of-concept for the potential benefits of cHRT.

 

Tissue or cell encapsulation may offer effective strategies to fabricate ovarian constructs for the purpose of fertility and/or hormone replacement. Approaches using segmental ovarian tissue or whole-follicle implantation (typically with a focus on cryopreservation of the tissue for reproductive purposes) have resulted in detectable hormone levels in the blood after transplantation. Previous studies have also shown that autotransplantation of frozen-thawed ovarian tissue can lead to hormone secretion for over 5 years in humans.

 

Although these approaches can be used to achieve the dual purpose of fertility and hormone replacement in premenopausal women undergoing premature ovarian failure, they would have limited application in postmenopausal women who only need hormone replacement to manage menopausal symptoms and in whom fertility is not desirable. In full development, the technology described in this research is focused on hormone replacement, would meet the needs of the latter group of women that is the postmenopausal women.

 

The cell-based system of hormone replacement described in this report offers an attractive alternative to traditional pharmacological approaches and is consistent with current guidelines in the U.S. and Europe recommending the lowest possible doses of hormone for replacement therapy. In the present research sustained stable hormone release over the course of 90 days of study was demonstrated. The study also demonstrated the effective end-organ outcomes in body fat composition, uterine health, and bone health. However, additional studies will be required to determine the sustainability of the hormone secretion of the constructs by measuring hormone levels from implanted constructs for periods longer than 3 months in the rat model.

 

This study highlights the potential utility of cHRT for the treatment and study of conditions associated with functional loss of the ovaries. Although longer-term studies would be of future interest, the 90-day duration of this rodent model study is consistent with others investigating osteoporosis in an ovariectomy model. However, this study provides a proof-of-concept for cHRT, it suffers the limitation that it is only an isogeneic-based construct implantation. Scientists think that further studies in either allogeneic or xenogeneic settings would be required with the construct design described in this report in the path towards clinical translation given that patients who would receive this type of treatment are unlikely to have sufficient autologous ovarian cells for transplantation.

 

Researchers from Copenhagen, Denmark, were recently able to isolate viable, early stage follicles in ovarian tissue. They have successfully stripped ovarian tissue from its cancerous cells and used the remaining scaffold to support the growth and survival of human follicles. This “artificial ovary” may help y to help women who have become infertile due to cancer and chemotherapy. But, the research is presently at a very preliminary stage and much research is still required to ensure that cancer cells are not reintroduced during the grafting process.

 

References:

 

https://www.technologyreview.com/the-download/609677/will-artificial-ovaries-mean-no-more-menopause/

 

https://www.nature.com/articles/s41467-017-01851-3

 

https://www.ncbi.nlm.nih.gov/pubmed/23274068

 

https://www.ncbi.nlm.nih.gov/pubmed/26210678

 

https://www.ncbi.nlm.nih.gov/pubmed/21954279

 

http://www.frontlinegenomics.com/news/24423/artificial-ovaries-hope-to-help-infertile-women-conceive-following-chemotherapy/

 

Read Full Post »

Cellular Guillotine Created for Studying Single-Cell Wound Repair

Reporter: Irina Robu, PhD

Using the century-old cutting method, it would take a researcher five hours to cut 100 cells, and by the time they were done, the cells they cut first would be well on their way to healing.

In an effort to comprehend how a single cell heal, mechanical engineer Sing Tand developed a microscopic guillotine that proficiently cuts cells into two.

Tang, who is an assistant professor of mechanical engineering at Stanford University knew that finding a way to competently slice the cell in two could lead to engineering self-healing materials and machines. In order, to efficiently slice a cell in two he developed a tool that could cut 150 cells in just over 2 minutes, and the cuts were much more standardized and synchronized in the stage of their repair process. They attained this rate by creating a scaled-up version of their tool with eight identical parallel channels that run simultaneously. Being able to efficiently study cell healing could eventually help scientists study and treat a variety of human diseases such as cancer and neurodegenerative diseases. Prior to Tang’s cellular guillotine, scientists used to slice cells by hand under a microscope using a glass needle which is a method that can lead to errors.

Tang’s method can be the Holy Grail of engineering self-healing materials and machines.

SOURCE

http://news.stanford.edu/2017/06/26/stanford-scientists-create-cellular-guillotine-studying-single-cell-wound-repair/

Read Full Post »

Ultra-Pure Melatonin Product Helps Maintain Sleep for Up to 7 Hours

Curator: Gail S. Thornton, M.A.

Co-Editor: The VOICES of Patients, Hospital CEOs, HealthCare Providers, Caregivers and Families: Personal Experience with Critical Care and Invasive Medical Procedures

 

The role of melatonin is important in regulating natural sleep and wake cycles. Typically, melatonin levels decline with age, significantly decreasing after age 40. An estimated 50 to 70 million Americans are affected by sleep difficulties – a process regulated by melatonin — and long-term sleep deprivation has been associated with negative health consequences, including an increased risk of diabetes, hypertension, heart attack, stroke, obesity, and depression.

Clinical data from a new pharmacokinetic study suggests that REMfresh®, the first and only continuous release and absorption melatonin (CRA-melatonin), helps maintain sleep for up to 7 hours. REMfresh® contains 99 percent ultra-pure melatonin and is sourced in Western Europe, a factor that is significant and important to many sleep specialists.

Three research abstracts on the REMfresh® data were published in an online supplement in the journal, Sleep, and were presented recently at the 31st Annual Meeting of the Associated Professional Sleep Societies LLC (APSS).

REMfresh Photo

Image SOURCE: Photograph courtesy of Physician’s Seal®

How REMfresh® Works

REMfresh® (CRA-melatonin) mimics the body’s own 7-hour Mesa Wave™, a natural pattern of melatonin blood levels during a normal night’s sleep cycle.

The study demonstrated the continuous release and absorption of 99 percent ultra-pure melatonin in REMfresh® (CRA-melatonin) was designed to induce sleep onset and provide continuous, lasting restorative sleep over 7 hours.

The scientifically advanced, patented formulation, called Ion Powered Pump (IPP™) technology, replicates the way in which the body naturally releases and absorbs melatonin, unlike conventional melatonin sleep products.

Since REMfresh® (CRA-melatonin) is not a drug, there is no drug hangover.

REMfresh MesaCurveNew-1

Image SOURCE: Diagram courtesy of Physician’s Seal®

 

Data Based on Scientifically Advanced Delivery Technology

According to the primary study author, David C. Brodner, M.D., “These study results represent an unparalleled breakthrough in drug-free, sleep maintenance that physicians and patients have been waiting for in a sleep product.” Dr. Brodner is a sleep specialist who is double board-certified in Otolaryngology – Head and Neck Surgery and Sleep Medicine and is the founder and principle physician at the Center for Sinus, Allergy, and Sleep Wellness in Palm Beach County, Florida.

Dr. Brodner said, “Melatonin products have been used primarily as a chronobiotic to address sleep disorders, such as jet lag and shift work. The patented delivery system in REMfresh mimics the body’s own natural sleep pattern, so individuals may experience consistent, restorative sleep and have an improved quality of life with this drug-free product.”

Study Findings With REMAKT

The study findings are based on REMAKT™ (REM Absorption Kinetics Trial), a U.S.-based randomized, crossover pharmacokinetic (PK) evaluation study in healthy, non-smoking adults that compared REMfresh® (CRA-melatonin) with a market-leading, immediate-release melatonin (IR-melatonin).

The study found that melatonin levels with REMfresh® exceeded the targeted sleep maintenance threshold for a median of 6.7 hours, compared with 3.7 hours with the leading IR-melatonin. Conversely, the levels of the market-leading IR-melatonin formulation dramatically increased 23 times greater than the targeted levels of exogenous melatonin for sleep maintenance and had a rapid decline in serum levels that did not allow melatonin levels to be maintained beyond 4 hours.

Additional analysis presented showed that REMfresh® (CRA-melatonin) builds upon the body of evidence from prolonged-release melatonin (PR-M), which demonstrated in well-conducted, placebo-controlled studies, statistically significant improvement in sleep quality, morning alertness, sleep latency and quality of life in patients aged 55 years and older compared with placebo.

REMfresh® (CRA-melatonin) was designed to overcome the challenges of absorption in the intestines, thereby extending the continual and gradual release pattern of melatonin through the night (known as the Mesa Wave™, a flat-topped hill with steep sides). There was a faster time to Cmax, which is anticipated to result in improved sleep onset, while the extended median plateau time to 6.7 hours and rapid fall-off in plasma levels at the end of the Mesa Wave™ may help to improve sleep maintenance and morning alertness.

REFERENCE/SOURCE

Physician’s Seal® and REMfresh® (www.remfresh.com)

REMfresh® press release, June 5, 2017 (http://www.prnewswire.com/news-releases/scientifically-advanced-delivery-technology-in-sleep-management-debuts-at-sleep-2017-with-clinical-data-showing-remfresh-the-first-and-only-continuous-release-and-absorption-melatonin-helps-maintain-sleep-for-up-to-7-hours-300468218.html)

Dr. David C. Brodner, Center for Sinus, Allergy, and Sleep Wellness  (http://www.brodnermd.com/sleep-hygiene.html)

Other related articles published in this Open Access Online Scientific Journal include the following:

2017

Sleep Research Society announces 2017 award recipients including Thomas S. Kilduff, PhD, Director, Center for Neuroscience at SRI International in Menlo Park, California

https://pharmaceuticalintelligence.com/2017/04/28/sleep-research-society-announces-2017-award-recipients-including-thomas-s-kilduff-phd-director-center-for-neuroscience-at-sri-international-in-menlo-park-california/

2016

Sleep Science

Genetic link to sleep and mood disorders

https://pharmaceuticalintelligence.com/2016/02/27/genetic-link-to-sleep-and-mood-disorders/

2015

Sleep quality, amyloid and cognitive decline

https://pharmaceuticalintelligence.com/2015/10/31/sleep-quality-amyloid-and-cognitive-decline/

2013

Day and Night Variation in Melatonin Level affects Plasma Membrane Redox System in Red Blood Cells

https://pharmaceuticalintelligence.com/2013/02/23/httpwww-ncbi-nlm-nih-govpubmed22561555/

Read Full Post »

Topical Solution for Combination Oncology Drug Therapy: Patch that delivers Drug, Gene, and Light-based Therapy to Tumor, Volume 2 (Volume Two: Latest in Genomics Methodologies for Therapeutics: Gene Editing, NGS and BioInformatics, Simulations and the Genome Ontology), Part 1: Next Generation Sequencing (NGS)

Topical Solution for Combination Oncology Drug Therapy: Patch that delivers Drug, Gene, and Light-based Therapy to Tumor

Reporter: Aviva Lev-Ari, PhD, RN

 

Self-assembled RNA-triple-helix hydrogel scaffold for microRNA modulation in the tumour microenvironment

Affiliations

  1. Massachusetts Institute of Technology, Institute for Medical Engineering and Science, Harvard-MIT Division for Health Sciences and Technology, Cambridge, Massachusetts 02139, USA
    • João Conde,
    • Nuria Oliva,
    • Mariana Atilano,
    • Hyun Seok Song &
    • Natalie Artzi
  2. School of Engineering and Materials Science, Queen Mary University of London, London E1 4NS, UK
    • João Conde
  3. Grup dEnginyeria de Materials, Institut Químic de Sarrià-Universitat Ramon Llull, Barcelona 08017, Spain
    • Mariana Atilano
  4. Division of Bioconvergence Analysis, Korea Basic Science Institute, Yuseong, Daejeon 169-148, Republic of Korea
    • Hyun Seok Song
  5. Broad Institute of MIT and Harvard, Cambridge, Massachusetts 02142, USA
    • Natalie Artzi
  6. Department of Medicine, Biomedical Engineering Division, Brigham and Womens Hospital, Harvard Medical School, Boston, Massachusetts 02115, USA
    • Natalie Artzi

Contributions

J.C. and N.A. conceived the project and designed the experiments. J.C., N.O., H.S.S. and M.A. performed the experiments, collected and analysed the data. J.C. and N.A. co-wrote the manuscript. All authors discussed the results and reviewed the manuscript.

Nature Materials
15,
353–363
(2016)
doi:10.1038/nmat4497
Received
22 April 2015
Accepted
26 October 2015
Published online
07 December 2015

The therapeutic potential of miRNA (miR) in cancer is limited by the lack of efficient delivery vehicles. Here, we show that a self-assembled dual-colour RNA-triple-helix structure comprising two miRNAs—a miR mimic (tumour suppressor miRNA) and an antagomiR (oncomiR inhibitor)—provides outstanding capability to synergistically abrogate tumours. Conjugation of RNA triple helices to dendrimers allows the formation of stable triplex nanoparticles, which form an RNA-triple-helix adhesive scaffold upon interaction with dextran aldehyde, the latter able to chemically interact and adhere to natural tissue amines in the tumour. We also show that the self-assembled RNA-triple-helix conjugates remain functional in vitro and in vivo, and that they lead to nearly 90% levels of tumour shrinkage two weeks post-gel implantation in a triple-negative breast cancer mouse model. Our findings suggest that the RNA-triple-helix hydrogels can be used as an efficient anticancer platform to locally modulate the expression of endogenous miRs in cancer.

SOURCE

http://www.nature.com/nmat/journal/v15/n3/abs/nmat4497.html#author-information

 

 

Patch that delivers drug, gene, and light-based therapy to tumor sites shows promising results

In mice, device destroyed colorectal tumors and prevented remission after surgery.

Helen Knight | MIT News Office
July 25, 2016

Approximately one in 20 people will develop colorectal cancer in their lifetime, making it the third-most prevalent form of the disease in the U.S. In Europe, it is the second-most common form of cancer.

The most widely used first line of treatment is surgery, but this can result in incomplete removal of the tumor. Cancer cells can be left behind, potentially leading to recurrence and increased risk of metastasis. Indeed, while many patients remain cancer-free for months or even years after surgery, tumors are known to recur in up to 50 percent of cases.

Conventional therapies used to prevent tumors recurring after surgery do not sufficiently differentiate between healthy and cancerous cells, leading to serious side effects.

In a paper published today in the journal Nature Materials, researchers at MIT describe an adhesive patch that can stick to the tumor site, either before or after surgery, to deliver a triple-combination of drug, gene, and photo (light-based) therapy.

Releasing this triple combination therapy locally, at the tumor site, may increase the efficacy of the treatment, according to Natalie Artzi, a principal research scientist at MIT’s Institute for Medical Engineering and Science (IMES) and an assistant professor of medicine at Brigham and Women’s Hospital, who led the research.

The general approach to cancer treatment today is the use of systemic, or whole-body, therapies such as chemotherapy drugs. But the lack of specificity of anticancer drugs means they produce undesired side effects when systemically administered.

What’s more, only a small portion of the drug reaches the tumor site itself, meaning the primary tumor is not treated as effectively as it should be.

Indeed, recent research in mice has found that only 0.7 percent of nanoparticles administered systemically actually found their way to the target tumor.

“This means that we are treating both the source of the cancer — the tumor — and the metastases resulting from that source, in a suboptimal manner,” Artzi says. “That is what prompted us to think a little bit differently, to look at how we can leverage advancements in materials science, and in particular nanotechnology, to treat the primary tumor in a local and sustained manner.”

The researchers have developed a triple-therapy hydrogel patch, which can be used to treat tumors locally. This is particularly effective as it can treat not only the tumor itself but any cells left at the site after surgery, preventing the cancer from recurring or metastasizing in the future.

Firstly, the patch contains gold nanorods, which heat up when near-infrared radiation is applied to the local area. This is used to thermally ablate, or destroy, the tumor.

These nanorods are also equipped with a chemotherapy drug, which is released when they are heated, to target the tumor and its surrounding cells.

Finally, gold nanospheres that do not heat up in response to the near-infrared radiation are used to deliver RNA, or gene therapy to the site, in order to silence an important oncogene in colorectal cancer. Oncogenes are genes that can cause healthy cells to transform into tumor cells.

The researchers envision that a clinician could remove the tumor, and then apply the patch to the inner surface of the colon, to ensure that no cells that are likely to cause cancer recurrence remain at the site. As the patch degrades, it will gradually release the various therapies.

The patch can also serve as a neoadjuvant, a therapy designed to shrink tumors prior to their resection, Artzi says.

When the researchers tested the treatment in mice, they found that in 40 percent of cases where the patch was not applied after tumor removal, the cancer returned.

But when the patch was applied after surgery, the treatment resulted in complete remission.

Indeed, even when the tumor was not removed, the triple-combination therapy alone was enough to destroy it.

The technology is an extraordinary and unprecedented synergy of three concurrent modalities of treatment, according to Mauro Ferrari, president and CEO of the Houston Methodist Research Institute, who was not involved in the research.

“What is particularly intriguing is that by delivering the treatment locally, multimodal therapy may be better than systemic therapy, at least in certain clinical situations,” Ferrari says.

Unlike existing colorectal cancer surgery, this treatment can also be applied in a minimally invasive manner. In the next phase of their work, the researchers hope to move to experiments in larger models, in order to use colonoscopy equipment not only for cancer diagnosis but also to inject the patch to the site of a tumor, when detected.

“This administration modality would enable, at least in early-stage cancer patients, the avoidance of open field surgery and colon resection,” Artzi says. “Local application of the triple therapy could thus improve patients’ quality of life and therapeutic outcome.”

Artzi is joined on the paper by João Conde, Nuria Oliva, and Yi Zhang, of IMES. Conde is also at Queen Mary University in London.

SOURCE

http://news.mit.edu/2016/patch-delivers-drug-gene-light-based-therapy-tumor-0725

Other related articles published in thie Open Access Online Scientific Journal include the following:

The Development of siRNA-Based Therapies for Cancer

Author: Ziv Raviv, PhD

https://pharmaceuticalintelligence.com/2013/05/09/the-development-of-sirna-based-therapies-for-cancer/

 

Targeted Liposome Based Delivery System to Present HLA Class I Antigens to Tumor Cells: Two papers

Reporter: Stephen J. Williams, Ph.D.

https://pharmaceuticalintelligence.com/2016/07/20/targeted-liposome-based-delivery-system-to-present-hla-class-i-antigens-to-tumor-cells-two-papers/

 

Blast Crisis in Myeloid Leukemia and the Activation of a microRNA-editing Enzyme called ADAR1

Curator: Larry H. Bernstein, MD, FCAP

https://pharmaceuticalintelligence.com/2016/06/10/blast-crisis-in-myeloid-leukemia-and-the-activation-of-a-microrna-editing-enzyme-called-adar1/

 

First challenge to make use of the new NCI Cloud Pilots – Somatic Mutation Challenge – RNA: Best algorithms for detecting all of the abnormal RNA molecules in a cancer cell

Reporter: Aviva Lev-Ari, PhD, RN

https://pharmaceuticalintelligence.com/2016/07/17/first-challenge-to-make-use-of-the-new-nci-cloud-pilots-somatic-mutation-challenge-rna-best-algorithms-for-detecting-all-of-the-abnormal-rna-molecules-in-a-cancer-cell/

 

miRNA Therapeutic Promise

Curator: Larry H. Bernstein, MD, FCAP

https://pharmaceuticalintelligence.com/2016/05/01/mirna-therapeutic-promise/

Read Full Post »

CHI’s  Advancing Bispecifics to the Clinic and Immunotherapy and Bispecifics for Oncology Targets, 3 – 4 November 2016, Lisbon, Portugal

Reporter: Aviva Lev-Ari, PhD, RN

 

Registration

https://chidb.com/reg/pge/reg.asp

 

Eighth Annual

Engineering Bispecifics

New Approaches and Platform Refinements

3 – 4 November 2016   |  Lisbon, Portugal

PEGSummitEurope.com

Register

Register Today for

Savings up to €400!

Engineering for T Cell Engagement at Engineering Bispecifics at PEGS Europe

Cambridge Healthtech’s 8th Annual Engineering Bispecifics conference showcases the latest developments in engineering bispecifics for T-cell engagement with presentations from Rinat Pfizer, Covagen, Immunocore and Glenmark. These focus on selection and validation of target antigens, platform development, tailored architecture for novel modes of action, preclinical and clinical manufacturing, IgG isotype and affinity, efficacy and specificity.

Additional presentations on this topic can be heard at the PEGS Europe tracks on Advancing Bispecifics to the Clinic and Immunotherapy (Chugai, Numab, Pfizer, Xencor and Macrogenics and Roche Innovation Centre. German Cancer Research Center).

Session Preview

High Affinity T Cell Receptor-Based Bi-Functional Biologics for Redirected Tumour Killing

Joseph Dukes, Ph.D., Head, Pre-Clinical Biology, Cell Biology, Immunocore

ImmTACs are bi-specific reagents that target tumours via a soluble monocloncal TCR with exceptionally high sensitivity and specificity and redirect host polyclonal T cells via an anti-CD3 antibody fragment. The selection and validation of appropriate target antigens and the testing of ImmTACs for specificity is critical. Using appropriate tumour and primary human cell lines, the in vitro pre-clinical package can be predictive of in vivo clinical observations.

Pfizer’s T-Cell Engaging Full Length Bispecific Antibody Platform: From Bench to NHS

Javier Chaparro-Riggers, Ph.D., Senior Director, Protein Engineering, Rinat Pfizer, Inc.

The recent clinical success of blinatumomab (antiCD19/CD3) spurred the development of a variety of T cell engaging bispecific antibody architectures. Pfizer developed a T cell engaging antibody platform, which allows the formation of full length human IgG1 and IgG2 antibodies in vitro or in vivo. The effect of IgG isotype and affinities of the T cell- and tumor antigen-targeting arm were explored and optimized.

Engineering of CD3 Bispecific FynomAbs

Julian Bertschinger, Ph.D., VP, Janssen R&D, Managing Director, Covagen

ImmTACs are bi-specific reagents that target tumours via a soluble monocloncal TCR with exceptionally high sensitivity and specificity and redirect host polyclonal T cells via an anti-CD3 antibody fragment. The selection and validation of appropriate target antigens and the testing of ImmTACs for specificity is critical. Using appropriate tumour and primary human cell lines, the in vitro pre-clinical package can be predictive of in vivo clinical observations.

Engineering and Manufacturing of Bispecific Antibodies for T-Cell Redirection

Stanislas Blein, Ph.D., Senior Director, Antibody Engineering, Glenmark Pharmaceuticals

Over the past two decades various functional bispecific antibody formats have been designed with only few molecules reaching clinical trials due to an inherent lack of manufacturability. Herein we describe a versatile bispecific antibody format that fits industrial-scale manufacturing processes and enables the rapid design and making of T-cell redirecting molecules. Engineering, pre-clinical and phase-one manufacturing data will be presented.

Register

[Register Today for Savings up to €400]

Final Agenda  |  Sponsorship & Exhibits |  Present a Poster  |  View Brochure

Short Courses*

DON’T MISS the short course on Engineering of Bispecific Antibodies

To be led by:
Nicolas Fischer, Ph.D., Head, Research, Novimmune SA
Michaela Silacci, Ph.D., Director, Discovery Research, Covagen AG, part of J&J

* separate registration required for short courses

Breakout Discussions

DON’T MISS the in depth Breakout Discussion on Strategy for Engineering and Design of Bispecific TCR-Based Products

To be moderated by:
Julian Bertschinger, Ph.D., VP, Janssen R&D, Managing Director, Covagen

Agenda-At-A-Glance

NEW MODES OF ACTION / T CELL ENGAGEMENT

KEYNOTE PRESENTATION: Protein Engineering for New Modes of Action

Andreas Plückthun, Ph.D., Director and Professor, Biochemistry, University of Zurich

Pfizer’s T Cell Engaging Full Length Bispecific Antibody Platform: From Bench to NHS

Javier Chaparro-Riggers, Ph.D., Senior Director, Protein Engineering, Rinat Pfizer, Inc.

Presentation to be Announced

Sponsored by Schrödinger

ENGINEERING BISPECIFICS FOR T CELL ENGAGEMENT

Engineering of CD3 Bispecific FynomAbs

Julian Bertschinger, Ph.D., Vice President, Janssen R&D, Managing Director, Covagen


High Affinity T Cell Receptor-Based Bifunctional Biologics for Redirected Tumour Killing

Joseph Dukes, Ph.D., Head, Pre-Clinical Biology, Cell Biology, Immunocore

FOCUS ON BISPECIFIC ENGINEERING FOR TARGETING

Epitopes Matter: Strategies to Generate and Analyse Binders to Different Epitopes

Jonas Schaefer, Ph.D., Head, High-Throughput Binder Selection Facility, Biochemistry, University of Zurich

Engineered Fab Domains Promote Efficient Production of Bispecific Antibodies in a Single Cell

Christoph Spiess, Ph.D., Senior Scientist, Antibody Engineering, Genentech, Inc.


A Novel Highly Versatile Multi-Specific Antibody Format

David Urech, Ph.D., CSO and Co-CEO, Numab AG

PLATFORM DEVELOPMENT AND REFINEMENT

CrossMAb Version 2: A Versatile Toolbox for Bispecific Antibody Engineering
Joerg Thomas Regula, Ph.D., Head, Functional Characterisation, Large Molecule Research, Roche Pharmaceutical Research and Early Development

Efficient Generation of Bispecific Mouse Antibodies for Preclinical Investigations

Aran F. Labrijn, Ph.D., Principal Scientist, Antibody Sciences, Genmab BV

Platform Refinements for Bispecifics for Oncology Targets

John de Kruif, Ph.D., CTO, Merus

Sponsored Presentation (Opportunities Available)

NOVEL APPROACHES

Using Alphabodies to Generate Bispecifics with Optimal in vitro and in vivo Characteristics

Yvonne McGrath, Ph.D., CSO, Complix NV

Hapten-Bispecific Antibodies for Drug Discovery and Delivery Applications

Ulrich Brinkmann, Ph.D., Expert Scientist, Roche Innovation Center

Cellular FRET Assay for the Determination of Simultaneous Binding of Bispecific Antibodies

Stefan Seeber, Ph.D., Principal Scientist, Cell Line and Molecule Development, Roche Innovation Center Munich/Large Molecule Research

Novel Strategy for a Bispecific Antibody: Induction of Dual Target Internalisation and Degradation

Ji Min Lee, Ph.D., Principal Scientist, Open Innovation Team, Samsung Bioepis

Engineering and Manufacturing of Bispecific Antibodies for T-Cell Redirection

Stanislas Blein, Ph.D., Senior Director, Antibody Engineering, Glenmark Pharmaceuticals

Event-at-a-Glance

SOURCE

From: Bispecific Antibodies <katiev@pegsummiteurope.com>

Date: Tuesday, July 12, 2016 at 9:22 AM

To: Aviva Lev-Ari <AvivaLev-Ari@alum.berkeley.edu>

Subject: Latest T-Cell Engagement Research from Pfizer, Covagen, Immunocore & Glenmark

Read Full Post »

Older Posts »