Posts Tagged ‘yeast’

Gene Editing with CRISPR gets Crisper

Curators: Larry H. Bernstein, MD, FCAP and Aviva Lev-Ari, PhD, RN



CRISPR Moves from Butchery to Surgery   

More Genomes Are Going Under the CRISPR Knife, So Surgical Standards Are Rising

  • The Dharmacon subsidary of GE Healthcare provides the Edit-R Lentiviral Gene Engineering platform. It is based on the natural S. pyrogenes system, but unlike that system, which uses a single guide RNA (sgRNA), the platform uses two component RNAs, a gene-specific CRISPR RNA (crRNA) and a universal trans-activating crRNA (tracrRNA). Once hybridized to the universal tracrRNA (blue), the crRNA (green) directs the Cas9 nuclease to a specific genomic region to induce a double- strand break.

    Scientists recently convened at the CRISPR Precision Gene Editing Congress, held in Boston, to discuss the new technology. As with any new technique, scientists have discovered that CRISPR comes with its own set of challenges, and the Congress focused its discussion around improving specificity, efficiency, and delivery.

    In the naturally occurring system, CRISPR-Cas9 works like a self-vaccination in the bacterial immune system by targeting and cleaving viral DNA sequences stored from previous encounters with invading phages. The endogenous system uses two RNA elements, CRISPR RNA (crRNA) and trans-activating RNA (tracrRNA), which come together and guide the Cas9 nuclease to the target DNA.

    Early publications that demonstrated CRISPR gene editing in mammalian cells combined the crRNA and tracrRNA sequences to form one long transcript called asingle-guide RNA (sgRNA). However, an alternative approach is being explored by scientists at the Dharmacon subsidiary of GE Healthcare. These scientists have a system that mimics the endogenous system through a synthetic two-component approach thatpreserves individual crRNA and tracrRNA. The tracrRNA is universal to any gene target or species; the crRNA contains the information needed to target the gene of interest.

    Predesigned Guide RNAs

    In contrast to sgRNAs, which are generated through either in vitro transcription of a DNA template or a plasmid-based expression system, synthetic crRNA and tracrRNA eliminate the need for additional cloning and purification steps. The efficacy of guide RNA (gRNA), whether delivered as a sgRNA or individual crRNA and tracrRNA, depends not only on DNA binding, but also on the generation of an indel that will deliver the coup de grâce to gene function.

    “Almost all of the gRNAs were able to create a break in genomic DNA,” said Louise Baskin, senior product manager at Dharmacon. “But there was a very wide range in efficiency and in creating functional protein knock-outs.”

    To remove the guesswork from gRNA design, Dharmacon developed an algorithm to predict gene knockout efficiency using wet-lab data. They also incorporated specificity as a component of their algorithm, using a much more comprehensive alignment tool to predict potential off-target effects caused by mismatches and bulges often missed by other alignment tools. Customers can enter their target gene to access predesigned gRNAs as either two-component RNAs or lentiviral sgRNA vectors for multiple applications.

    “We put time and effort into our algorithm to ensure that our guide RNAs are not only functional but also highly specific,” asserts Baskin. “As a result, customers don’t have to do any design work.”

    Donor DNA Formats
    MilliporeSigma’s CRISPR Epigenetic Activator is based on fusion of a nuclease-deficient Cas9 (dCas9) to the catalytic histone acetyltransferase (HAT) core domain of the human E1A-associated protein p300. This technology allows researchers to target specific DNA regions or gene sequences. Researchers can localize epigenetic changes to their target of interest and see the effects of those changes in gene expression.

    Knockout experiments are a powerful tool for analyzing gene function. However, for researchers who want to introduce DNA into the genome, guide design, donor DNA selection, and Cas9 activity are paramount to successful DNA integration.MilliporeSigma offers two formats for donor DNA: double-stranded DNA (dsDNA) plasmids and single-stranded DNA (ssDNA) oligonucleotides. The most appropriate format depends on cell type and length of the donor DNA. “There are some cell types that have immune responses to dsDNA,” said Gregory Davis, Ph.D., R&D manager, MilliporeSigma.

  • The ssDNA format can save researchers time and money, but it has a limited carrying capacity of approximately 120 base pairs.In addition to selecting an appropriate donor DNA format, controlling where, how, and when the Cas9 enzyme cuts can affect gene-editing efficiency. Scientists are playing tug-of-war, trying to pull cells toward the preferred homology-directed repair (HDR) and away from the less favored nonhomologous end joining (NHEJ) repair mechanism.One method to achieve this modifies the Cas9 enzyme to generate a nickase that cuts only one DNA strand instead of creating a double-strand break. Accordingly, MilliporeSigma has created a Cas9 paired-nickase system that promotes HDR, while also limiting off-target effects and increasing the number of sequences available for site-dependent gene modifications, such as disease-associated single nucleotide polymorphisms (SNPs).“The best thing you can do is to cut as close to the SNP as possible,” advised Dr. Davis. “As you move the double-stranded break away from the site of mutation you get an exponential drop in the frequency of recombination.”


  • Ribonucleo-protein Complexes

    Another strategy to improve gene-editing efficiency, developed by Thermo Fisher, involves combining purified Cas9 protein with gRNA to generate a stable ribonucleoprotein (RNP) complex. In contrast to plasmid- or mRNA-based formats, which require transcription and/or translation, the Cas9 RNP complex cuts DNA immediately after entering the cell. Rapid clearance of the complex from the cell helps to minimize off-target effects, and, unlike a viral vector, the transient complex does not introduce foreign DNA sequences into the genome.

    To deliver their Cas9 RNP complex to cells, Thermo Fisher has developed a lipofectamine transfection reagent called CRISPRMAX. “We went back to the drawing board with our delivery, screened a bunch of components, and got a brand-new, fully  optimized lipid nanoparticle formulation,” explained Jon Chesnut, Ph.D., the company’s senior director of synthetic biology R&D. “The formulation is specifically designed for delivering the RNP to cells more efficiently.”

    Besides the reagent and the formulation, Thermo Fisher has also developed a range of gene-editing tools. For example, it has introduced the Neon® transfection system for delivering DNA, RNA, or protein into cells via electroporation. Dr. Chesnut emphasized the company’s focus on simplifying complex workflows by optimizing protocols and pairing everything with the appropriate up- and downstream reagents.

From Mammalian Cells to Microbes

One of the first sources of CRISPR technology was the Feng Zhang laboratory at the Broad Institute, which counted among its first licensees a company called GenScript. This company offers a gene-editing service called GenCRISPR™ to establish mammalian cell lines with CRISPR-derived gene knockouts.

“There are a lot of challenges with mammalian cells, and each cell line has its own set of issues,” said Laura Geuss, a marketing specialist at GenScript. “We try to offer a variety of packages that can help customers who have difficult-to-work-with cells.” These packages include both viral-based and transient transfection techniques.

However, the most distinctive service offered by GenScript is its microbial genome-editing service for bacteria (Escherichia coli) and yeast (Saccharomyces cerevisiae). The company’s strategy for gene editing in bacteria can enable seamless knockins, knockouts, or gene replacements by combining CRISPR with lambda red recombineering. Traditionally one of the most effective methods for gene editing in microbes, recombineering allows editing without restriction enzymes through in vivo homologous recombination mediated by a phage-based recombination system such as lambda red.

On its own, lambda red technology cannot target multiple genes, but when paired with CRISPR, it allows the editing of multiple genes with greater efficiency than is possible with CRISPR alone, as the lambda red proteins help repair double-strand breaks in E. coli. The ability to knockout different gene combinations makes Genscript’s microbial editing service particularly well suited for the optimization of metabolic pathways.

Pooled and Arrayed Library Strategies

Scientists are using CRISPR technology for applications such as metabolic engineering and drug development. Yet another application area benefitting from CRISPR technology is cancer research. Here, the use of pooled CRISPR libraries is becoming commonplace. Pooled CRISPR libraries can help detect mutations that affect drug resistance, and they can aid in patient stratification and clinical trial design.

Pooled screening uses proliferation or viability as a phenotype to assess how genetic alterations, resulting from the application of a pooled CRISPR library, affect cell growth and death in the presence of a therapeutic compound. The enrichment or depletion of different gRNA populations is quantified using deep sequencing to identify the genomic edits that result in changes to cell viability.

MilliporeSigma provides pooled CRISPR libraries ranging from the whole human genome to smaller custom pools for these gene-function experiments. For pharmaceutical and biotech companies, Horizon Discovery offers a pooled screening service, ResponderSCREEN, which provides a whole-genome pooled screen to identify genes that confer sensitivity or resistance to a compound. This service is comprehensive, taking clients from experimental design all the way through to suggestions for follow-up studies.

Horizon Discovery maintains a Research Biotech business unit that is focused on target discovery and enabling translational medicine in oncology. “Our internal backbone gives us the ability to provide expert advice demonstrated by results,” said Jon Moore, Ph.D., the company’s CSO.

In contrast to a pooled screen, where thousands of gRNA are combined in one tube, an arrayed screen applies one gRNA per well, removing the need for deep sequencing and broadening the options for different endpoint assays. To establish and distribute a whole-genome arrayed lentiviral CRISPR library, MilliporeSigma partnered with the Wellcome Trust Sanger Institute. “This is the first and only arrayed CRISPR library in the world,” declared Shawn Shafer, Ph.D., functional genomics market segment manager, MilliporeSigma. “We were really proud to partner with Sanger on this.”

Pooled and arrayed screens are powerful tools for studying gene function. The appropriate platform for an experiment, however, will be determined by the desired endpoint assay.

Detection and Quantification of Edits


The QX200 Droplet Digital PCR System from Bio-Rad Laboratories can provide researchers with an absolute measure of target DNA molecules for EvaGreen or probe-based digital PCR applications. The system, which can provide rapid, low-cost, ultra-sensitive quantification of both NHEJ- and HDR-editing events, consists of two instruments, the QX200 Droplet Generator and the QX200 Droplet Reader, and their associated consumables.

Finally, one last challenge for CRISPR lies in the detection and quantification of changes made to the genome post-editing. Conventional methods for detecting these alterations include gel methods and next-generation sequencing. While gel methods lack sensitivity and scalability, next-generation sequencing is costly and requires intensive bioinformatics.

To address this gap, Bio-Rad Laboratories developed a set of assay strategies to enable sensitive and precise edit detection with its Droplet Digital PCR (ddPCR) technology. The platform is designed to enable absolute quantification of nucleic acids with high sensitivity, high precision, and short turnaround time through massive droplet partitioning of samples.

Using a validated assay, a typical ddPCR experiment takes about five to six hours to complete. The ddPCR platform enables detection of rare mutations, and publications have reported detection of precise edits at a frequency of <0.05%, and of NHEJ-derived indels at a frequency as low as 0.1%. In addition to quantifying precise edits, indels, and computationally predicted off-target mutations, ddPCR can also be used to characterize the consequences of edits at the RNA level.

According to a recently published Science paper, the laboratory of Charles A. Gersbach, Ph.D., at Duke University used ddPCR in a study of muscle function in a mouse model of Duchenne muscular dystrophy. Specifically, ddPCR was used to assess the efficiency of CRISPR-Cas9 in removing the mutated exon 23 from the dystrophin gene. (Exon 23 deletion by CRISPR-Cas9 resulted in expression of the modified dystrophin gene and significant enhancement of muscle force.)

Quantitative ddPCR showed that exon 23 was deleted in ~2% of all alleles from the whole-muscle lysate. Further ddPCR studies found that 59% of mRNA transcripts reflected the deletion.

“There’s an overarching idea that the genome-editing field is moving extremely quickly, and for good reason,” asserted Jennifer Berman, Ph.D., staff scientist, Bio-Rad Laboratories. “There’s a lot of exciting work to be done, but detection and quantification of edits can be a bottleneck for researchers.”

The gene-editing field is moving quickly, and new innovations are finding their way into the laboratory as researchers lay the foundation for precise, well-controlled gene editing with CRISPR.


Are Current Cancer Drug Discovery Methods Flawed?

GEN May 3, 2016


Researchers utilized a systems biology approach to develop new methods to assess drug sensitivity in cells. [The Institute for Systems Biology]

Understanding how cells respond and proliferate in the presence of anticancer compounds has been the foundation of drug discovery ideology for decades. Now, a new study from scientists at Vanderbilt University casts significant suspicion on the primary method used to test compounds for anticancer activity in cells—instilling doubt on methods employed by the entire scientific enterprise and pharmaceutical industry to discover new cancer drugs.

“More than 90% of candidate cancer drugs fail in late-stage clinical trials, costing hundreds of millions of dollars,” explained co-senior author Vito Quaranta, M.D., director of the Quantitative Systems Biology Center at Vanderbilt. “The flawed in vitro drug discovery metric may not be the only responsible factor, but it may be worth pursuing an estimate of its impact.”

The Vanderbilt investigators have developed what they believe to be a new metric for evaluating a compound’s effect on cell proliferation—called the DIP (drug-induced proliferation) rate—that overcomes the flawed bias in the traditional method.

The findings from this study were published recently in Nature Methods in an article entitled “An Unbiased Metric of Antiproliferative Drug Effect In Vitro.”

For more than three decades, researchers have evaluated the ability of a compound to kill cells by adding the compound in vitro and counting how many cells are alive after 72 hours. Yet, proliferation assays that measure cell number at a single time point don’t take into account the bias introduced by exponential cell proliferation, even in the presence of the drug.

“Cells are not uniform, they all proliferate exponentially, but at different rates,” Dr. Quaranta noted. “At 72 hours, some cells will have doubled three times and others will not have doubled at all.”

Dr. Quaranta added that drugs don’t all behave the same way on every cell line—for example, a drug might have an immediate effect on one cell line and a delayed effect on another.

The research team decided to take a systems biology approach, a mixture of experimentation and mathematical modeling, to demonstrate the time-dependent bias in static proliferation assays and to develop the time-independent DIP rate metric.

“Systems biology is what really makes the difference here,” Dr. Quaranta remarked. “It’s about understanding cells—and life—as dynamic systems.”This new study is of particular importance in light of recent international efforts to generate data sets that include the responses of thousands of cell lines to hundreds of compounds. Using the

  • Cancer Cell Line Encyclopedia (CCLE) and
  • Genomics of Drug Sensitivity in Cancer (GDSC) databases

will allow drug discovery scientists to include drug response data along with genomic and proteomic data that detail each cell line’s molecular makeup.

“The idea is to look for statistical correlations—these particular cell lines with this particular makeup are sensitive to these types of compounds—to use these large databases as discovery tools for new therapeutic targets in cancer,” Dr. Quaranta stated. “If the metric by which you’ve evaluated the drug sensitivity of the cells is wrong, your statistical correlations are basically no good.”

The Vanderbilt team evaluated the responses from four different melanoma cell lines to the drug vemurafenib, currently used to treat melanoma, with the standard metric—used for the CCLE and GDSC databases—and with the DIP rate. In one cell line, they found a glaring disagreement between the two metrics.

“The static metric says that the cell line is very sensitive to vemurafenib. However, our analysis shows this is not the case,” said co-lead study author Leonard Harris, Ph.D., a systems biology postdoctoral fellow at Vanderbilt. “A brief period of drug sensitivity, quickly followed by rebound, fools the static metric, but not the DIP rate.”

Dr. Quaranta added that the findings “suggest we should expect melanoma tumors treated with this drug to come back, and that’s what has happened, puzzling investigators. DIP rate analyses may help solve this conundrum, leading to better treatment strategies.”

The researchers noted that using the DIP rate is possible because of advances in automation, robotics, microscopy, and image processing. Moreover, the DIP rate metric offers another advantage—it can reveal which drugs are truly cytotoxic (cell killing), rather than merely cytostatic (cell growth inhibiting). Although cytostatic drugs may initially have promising therapeutic effects, they may leave tumor cells alive that then have the potential to cause the cancer to recur.

The Vanderbilt team is currently in the process of identifying commercial entities that can further refine the software and make it widely available to the research community to inform drug discovery.


An unbiased metric of antiproliferative drug effect in vitro

Leonard A HarrisPeter L FrickShawn P GarbettKeisha N HardemanB Bishal PaudelCarlos F LopezVito Quaranta & Darren R Tyson
Nature Methods 2 May (2016)

In vitro cell proliferation assays are widely used in pharmacology, molecular biology, and drug discovery. Using theoretical modeling and experimentation, we show that current metrics of antiproliferative small molecule effect suffer from time-dependent bias, leading to inaccurate assessments of parameters such as drug potency and efficacy. We propose the drug-induced proliferation (DIP) rate, the slope of the line on a plot of cell population doublings versus time, as an alternative, time-independent metric.

  1. Zuber, J. et al. Nat. Biotechnol. 29, 7983 (2011).
  2. Berns, K. et al. Nature 428, 431437 (2004).
  3. Bonnans, C., Chou, J. & Werb, Z. Nat. Rev. Mol. Cell Biol. 15, 786801 (2014).
  4. Garnett, M.J. et al. Nature 483, 570575 (2012)


Mapping Traits to Genes with CRISPR

Researchers develop a technique to direct chromosome recombination with CRISPR/Cas9, allowing high-resolution genetic mapping of phenotypic traits in yeast.

By Catherine Offord | May 5, 2016


Researchers used CRISPR/Cas9 to make a targeted double-strand break (DSB) in one arm of a yeast chromosome labeled with a green fluorescent protein (GFP) gene. A within-cell mechanism called homologous repair (HR) mends the broken arm using its homolog, resulting in a recombined region from the site of the break to the chromosome tip. When this cell divides by mitosis, each daughter cell will contain a homozygous section in an outcome known as “loss of heterozygosity” (LOH). One of the daughter cells is detectable because, due to complete loss of the GFP gene, it will no longer be fluorescent.REPRINTED WITH PERMISSION FROM M.J. SADHU ET AL., SCIENCE

When mapping phenotypic traits to specific loci, scientists typically rely on the natural recombination of chromosomes during meiotic cell division in order to infer the positions of responsible genes. But recombination events vary with species and chromosome region, giving researchers little control over which areas of the genome are shuffled. Now, a team at the University of California, Los Angeles (UCLA), has found a way around these problems by using CRISPR/Cas9 to direct targeted recombination events during mitotic cell division in yeast. The team described its technique today (May 5) in Science.

“Current methods rely on events that happen naturally during meiosis,” explained study coauthor Leonid Kruglyak of UCLA. “Whatever rate those events occur at, you’re kind of stuck with. Our idea was that using CRISPR, we can generate those events at will, exactly where we want them, in large numbers, and in a way that’s easy for us to pull out the cells in which they happened.”

Generally, researchers use coinheritance of a trait of interest with specific genetic markers—whose positions are known—to figure out what part of the genome is responsible for a given phenotype. But the procedure often requires impractically large numbers of progeny or generations to observe the few cases in which coinheritance happens to be disrupted informatively. What’s more, the resolution of mapping is limited by the length of the smallest sequence shuffled by recombination—and that sequence could include several genes or gene variants.

“Once you get down to that minimal region, you’re done,” said Kruglyak. “You need to switch to other methods to test every gene and every variant in that region, and that can be anywhere from challenging to impossible.”

But programmable, DNA-cutting champion CRISPR/Cas9 offered an alternative. During mitotic—rather than meiotic—cell division, rare, double-strand breaks in one arm of a chromosome preparing to split are sometimes repaired by a mechanism called homologous recombination. This mechanism uses the other chromosome in the homologous pair to replace the sequence from the break down to the end of the broken arm. Normally, such mitotic recombination happens so rarely as to be impractical for mapping purposes. With CRISPR/Cas9, however, the researchers found that they could direct double-strand breaks to any locus along a chromosome of interest (provided it was heterozygous—to ensure that only one of the chromosomes would be cut), thus controlling the sites of recombination.

Combining this technique with a signal of recombination success, such as a green fluorescent protein (GFP) gene at the tip of one chromosome in the pair, allowed the researchers to pick out cells in which recombination had occurred: if the technique failed, both daughter cells produced by mitotic division would be heterozygous, with one copy of the signal gene each. But if it succeeded, one cell would end up with two copies, and the other cell with none—an outcome called loss of heterozygosity.

“If we get loss of heterozygosity . . . half the cells derived after that loss of heterozygosity event won’t have GFP anymore,” study coauthor Meru Sadhu of UCLA explained. “We search for these cells that don’t have GFP out of the general population of cells.” If these non-fluorescent cells with loss of heterozygosity have the same phenotype as the parent for a trait of interest, then CRISPR/Cas9-targeted recombination missed the responsible gene. If the phenotype is affected, however, then the trait must be linked to a locus in the recombined, now-homozygous region, somewhere between the cut site and the GFP gene.

By systematically making cuts using CRISPR/Cas9 along chromosomes in a hybrid, diploid strain ofSaccharomyces cerevisiae yeast, picking out non-fluorescent cells, and then observing the phenotype, the UCLA team demonstrated that it could rapidly identify the phenotypic contribution of specific gene variants. “We can simply walk along the chromosome and at every [variant] position we can ask, does it matter for the trait we’re studying?” explained Kruglyak.

For example, the team showed that manganese sensitivity—a well-defined phenotypic trait in lab yeast—could be pinpointed using this method to a single nucleotide polymorphism (SNP) in a gene encoding the Pmr1 protein (a manganese transporter).

Jason Moffat, a molecular geneticist at the University of Toronto who was not involved in the work, toldThe Scientist that researchers had “dreamed about” exploiting these sorts of mechanisms for mapping purposes, but without CRISPR, such techniques were previously out of reach. Until now, “it hasn’t been so easy to actually make double-stranded breaks on one copy of a pair of chromosomes, and then follow loss of heterozygosity in mitosis,” he said, adding that he hopes to see the approach translated into human cell lines.

Applying the technique beyond yeast will be important, agreed cell and developmental biologist Ethan Bier of the University of California, San Diego, because chromosomal repair varies among organisms. “In yeast, they absolutely demonstrate the power of [this method],” he said. “We’ll just have to see how the technology develops in other systems that are going to be far less suited to the technology than yeast. . . . I would like to see it implemented in another system to show that they can get the same oomph out of it in, say, mammalian somatic cells.”

Kruglyak told The Scientist that work in higher organisms, though planned, is still in early stages; currently, his team is working to apply the technique to map loci responsible for trait differences between—rather than within—yeast species.

“We have a much poorer understanding of the differences across species,” Sadhu explained. “Except for a few specific examples, we’re pretty much in the dark there.”

M.J. Sadhu, “CRISPR-directed mitotic recombination enables genetic mapping without crosses,” Science, doi:10.1126/science.aaf5124, 2016.


CRISPR-directed mitotic recombination enables genetic mapping without crosses

Meru J Sadhu, Joshua S Bloom, Laura Day, Leonid Kruglyak

Thank you, David, for the kind words and comments. We agree that the most immediate applications of the CRISPR-based recombination mapping will be in unicellular organisms and cell culture. We also think the method holds a lot of promise for research in multicellular organisms, although we did not mean to imply that it “will be an efficient mapping method for all multicellular organisms”. Every organism will have its own set of constraints as well as experimental tools that will be relevant when adapting a new technique. To best help experts working on these organisms, here are our thoughts on your questions.

You asked about mutagenesis during recombination. We Sanger sequenced 72 of our LOH lines at the recombination site and did not observe any mutations, as described in the supplementary materials. We expect the absence of mutagenesis is because we targeted heterozygous sites where the untargeted allele did not have a usable PAM site; thus, following LOH, the targeted site is no longer present and cutting stops. In your experiments you targeted sites that were homozygous; thus, following recombination, the CRISPR target site persisted, and continued cutting ultimately led to repair by NHEJ and mutagenesis.

As to the more general question of the optimal mapping strategies in different organisms, they will depend on the ease of generating and screening for editing events, the cost and logistics of maintaining and typing many lines, and generation time, among other factors. It sounds like in Drosophila today, your related approach of generating markers with CRISPR, and then enriching for natural recombination events that separate them, is preferable. In yeast, we’ve found the opposite to be the case. As you note, even in Drosophila, our approach may be preferable for regions with low or highly non-uniform recombination rates.

Finally, mapping in sterile interspecies hybrids should be straightforward for unicellular hybrids (of which there are many examples) and for cells cultured from hybrid animals or plants. For studies in hybrid multicellular organisms, we agree that driving mitotic recombination in the early embryo may be the most promising approach. Chimeric individuals with mitotic clones will be sufficient for many traits. Depending on the system, it may in fact be possible to generate diploid individuals with uniform LOH genotype, but this is certainly beyond the scope of our paper. The calculation of the number of lines assumes that the mapping is done in a single step; as you note in your earlier comment, mapping sequentially can reduce this number dramatically.

This is a lovely method and should find wide applicability in many settings, especially for microorganisms and cell lines. However, it is not clear that this approach will be, as implied by the discussion, an efficient mapping method for all multicellular organisms. I have performed similar experiments in Drosophila, focused on meiotic recombination, on a much smaller scale, and found that CRISPR-Cas9 can indeed generate targeted recombination at gRNA target sites. In every case I tested, I found that the recombination event was associated with a deletion at the gRNA site, which is probably unimportant for most mapping efforts, but may be a concern in some specific cases, for example for clinical applications. It would be interesting to know how often mutations occurred at the targeted gRNA site in this study.

The wider issue, however, is whether CRISPR-mediated recombination will be more efficient than other methods of mapping. After careful consideration of all the costs and the time involved in each of the steps for Drosophila, we have decided that targeted meiotic recombination using flanking visible markers will be, in most cases, considerably more efficient than CRISPR-mediated recombination. This is mainly due to the large expense of injecting embryos and the extensive effort and time required to screen injected animals for appropriate events. It is both cheaper and faster to generate markers (with CRISPR) and then perform a large meiotic recombination mapping experiment than it would be to generate the lines required for CRISPR-mediated recombination mapping. It is possible to dramatically reduce costs by, for example, mapping sequentially at finer resolution. But this approach would require much more time than marker-assisted mapping. If someone develops a rapid and cheap method of reliably introducing DNA into Drosophila embryos, then this calculus might change.

However, it is possible to imagine situations where CRISPR-mediated mapping would be preferable, even for Drosophila. For example, some genomic regions display extremely low or highly non-uniform recombination rates. It is possible that CRISPR-mediated mapping could provide a reasonable approach to fine mapping genes in these regions.

The authors also propose the exciting possibility that CRISPR-mediated loss of heterozygosity could be used to map traits in sterile species hybrids. It is not entirely obvious to me how this experiment would proceed and I hope the authors can illuminate me. If we imagine driving a recombination event in the early embryo (with maternal Cas9 from one parent and gRNA from a second parent), then at best we would end up with chimeric individuals carrying mitotic clones. I don’t think one could generate diploid animals where all cells carried the same loss of heterozygosity event. Even if we could, this experiment would require construction of a substantial number of stable transgenic lines expressing gRNAs. Mapping an ~20Mbp chromosome arm to ~10kb would require on the order of two-thousand transgenic lines. Not an undertaking to be taken lightly. It is already possible to perform similar tests (hemizygosity tests) using D. melanogaster deficiency lines in crosses with D. simulans, so perhaps CRISPR-mediated LOH could complement these deficiency screens for fine mapping efforts. But, at the moment, it is not clear to me how to do the experiment.


Read Full Post »

Irreconciliable Dissonance in Physical Space and Cellular Metabolic Conception

Irreconciliable Dissonance in Physical Space and Cellular Metabolic Conception

Curator: Larry H. Bernstein, MD, FCAP

Pasteur Effect – Warburg Effect – What its history can teach us today. 

José Eduardo de Salles Roselino

The Warburg effect, in reality the “Pasteur-effect” was the first example of metabolic regulation described. A decrease in the carbon flux originated at the sugar molecule towards the end of the catabolic pathway, with ethanol and carbon dioxide observed when yeast cells were transferred from an anaerobic environmental condition to an aerobic one. In Pasteur´s studies, sugar metabolism was measured mainly by the decrease of sugar concentration in the yeast growth media observed after a measured period of time. The decrease of the sugar concentration in the media occurs at great speed in yeast grown in anaerobiosis (oxygen deficient) and its speed was greatly reduced by the transfer of the yeast culture to an aerobic condition. This finding was very important for the wine industry of France in Pasteur’s time, since most of the undesirable outcomes in the industrial use of yeast were perceived when yeasts cells took a very long time to create, a rather selective anaerobic condition. This selective culture media was characterized by the higher carbon dioxide levels produced by fast growing yeast cells and by a higher alcohol content in the yeast culture media.

However, in biochemical terms, this finding was required to understand Lavoisier’s results indicating that chemical and biological oxidation of sugars produced the same calorimetric (heat generation) results. This observation requires a control mechanism (metabolic regulation) to avoid burning living cells by fast heat released by the sugar biological oxidative processes (metabolism). In addition, Lavoisier´s results were the first indications that both processes happened inside similar thermodynamics limits. In much resumed form, these observations indicate the major reasons that led Warburg to test failure in control mechanisms in cancer cells in comparison with the ones observed in normal cells.

[It might be added that the availability of O2 and CO2 and climatic conditions over 750 million years that included volcanic activity, tectonic movements of the earth crust, and glaciation, and more recently the use of carbon fuels and the extensive deforestation of our land masses have had a large role in determining the biological speciation over time, in sea and on land. O2 is generated by plants utilizing energy from the sun and conversion of CO2. Remove the plants and we tip the balance. A large source of CO2 is from beneath the earth’s surface.]

Biology inside classical thermodynamics places some challenges to scientists. For instance, all classical thermodynamics must be measured in reversible thermodynamic conditions. In an isolated system, increase in P (pressure) leads to increase in V (volume), all this occurring in a condition in which infinitesimal changes in one affects in the same way the other, a continuum response. Not even a quantic amount of energy will stand beyond those parameters.

In a reversible system, a decrease in V, under same condition, will led to an increase in P. In biochemistry, reversible usually indicates a reaction that easily goes either from A to B or B to A. For instance, when it was required to search for an anti-ischemic effect of Chlorpromazine in an extra hepatic obstructed liver, it was necessary to use an adequate system of increased biliary system pressure in a reversible manner to exclude a direct effect of this drug over the biological system pressure inducer (bile secretion) in Braz. J. Med. Biol. Res 1989; 22: 889-893. Frequently, these details are jumped over by those who read biology in ATGC letters.

Very important observations can be made in this regard, when neutral mutations are taken into consideration since, after several mutations (not affecting previous activity and function), a last mutant may provide a new transcript RNA for a protein and elicit a new function. For an example, consider a Prion C from lamb getting similar to bovine Prion C while preserving  its normal role in the lamb when its ability to change Human Prion C is considered (Stanley Prusiner).

This observation is good enough, to confirm one of the most important contributions of Erwin Schrodinger in his What is Life:

“This little book arose from a course of public lectures, delivered by a theoretical physicist to an audience of about four hundred which did not substantially dwindle, though warned at the outset that the subject matter was a difficult one and that the lectures could not be termed popular, even though the physicist’s most dreaded weapon, mathematical deduction, would hardly be utilized. The reason for this was not that the subject was simple enough to be explained without mathematics, but rather that it was much too involved to be fully accessible to mathematics.”

After Hans Krebs, description of the cyclic nature of the citrate metabolism and after its followers described its requirement for aerobic catabolism two major lines of research started the search for the understanding of the mechanism of energy transfer that explains how ADP is converted into ATP. One followed the organic chemistry line of reasoning and therefore, searched for a mechanism that could explain how the breakdown of carbon-carbon link could have its energy transferred to ATP synthesis. One of the major leaders of this research line was Britton Chance. He took into account that relatively earlier in the series of Krebs cycle reactions, two carbon atoms of acetyl were released as carbon dioxide ( In fact, not the real acetyl carbons but those on the opposite side of citrate molecule). In stoichiometric terms, it was not important whether the released carbons were or were not exactly those originated from glucose carbons. His research aimed at to find out an intermediate proteinaceous intermediary that could act as an energy reservoir. The intermediary could store in a phosphorylated amino acid the energy of carbon-carbon bond breakdown. This activated amino acid could transfer its phosphate group to ADP producing ATP. A key intermediate involved in the transfer was identified by Kaplan and Lipmann at John Hopkins as acetyl coenzyme A, for which Fritz Lipmann received a Nobel Prize.

Alternatively, under possible influence of the excellent results of Hodgkin and Huxley a second line of research appears. The work of Hodgkin & Huxley indicated that the storage of electrical potential energy in transmembrane ionic asymmetries and presented the explanation for the change from resting to action potential in excitable cells. This second line of research, under the leadership of Peter Mitchell postulated a mechanism for the transfer of oxide/reductive power of organic molecules oxidation through electron transfer as the key for the energetic transfer mechanism required for ATP synthesis.
This diverted the attention from high energy (~P) phosphate bond to the transfer of electrons. During most of the time the harsh period of the two confronting points of view, Paul Boyer and followers attempted to act as a conciliatory third party, without getting good results, according to personal accounts (in L. A. or Latin America) heard from those few of our scientists who were able to follow the major scientific events held in USA, and who could present to us later. Paul  Boyer could present how the energy was transduced by a molecular machine that changes in conformation in a series of 3 steps while rotating in one direction in order to produce ATP and in opposite direction in order to produce ADP plus Pi from ATP (reversibility).

However, earlier, a victorious Peter Mitchell obtained the result in the conceptual dispute, over the Britton Chance point of view, after he used E. Coli mutants to show H+ gradients in the cell membrane and its use as energy source, for which he received a Nobel Prize. Somehow, this outcome represents such a blow to Chance’s previous work that somehow it seems to have cast a shadow over very important findings obtained during his earlier career that should not be affected by one or another form of energy transfer mechanism.  For instance, Britton Chance got the simple and rapid polarographic assay method of oxidative phosphorylation and the idea of control of energy metabolism that brings us back to Pasteur.

This metabolic alternative result seems to have been neglected in the recent years of obesity epidemics, which led to a search for a single molecular mechanism required for the understanding of the accumulation of chemical (adipose tissue) reserve in our body. It does not mean that here the role of central nervous system is neglected. In short, in respiring mitochondria the rate of electron transport linked to the rate of ATP production is determined primarily by the relative concentrations of ADP, ATP and phosphate in the external media (cytosol) and not by the concentration of respiratory substrate as pyruvate. Therefore, when the yield of ATP is high as it is in aerobiosis and the cellular use of ATP is not changed, the oxidation of pyruvate and therefore of glycolysis is quickly (without change in gene expression), throttled down to the resting state. The dependence of respiratory rate on ADP concentration is also seen in intact cells. A muscle at rest and using no ATP has a very low respiratory rate.   [When skeletal muscle is stressed by high exertion, lactic acid produced is released into the circulation and is metabolized aerobically by the heart at the end of the activity].

This respiratory control of metabolism will lead to preservation of body carbon reserves and in case of high caloric intake in a diet, also shows increase in fat reserves essential for our biological ancestors survival (Today for our obesity epidemics). No matter how important this observation is, it is only one focal point of metabolic control. We cannot reduce the problem of obesity to the existence of metabolic control. There are numerous other factors but on the other hand, we cannot neglect or remove this vital process in order to correct obesity. However, we cannot explain obesity ignoring this metabolic control. This topic is so neglected in modern times that we cannot follow major research lines of the past that were interrupted by the emerging molecular biology techniques and the vain belief that a dogmatic vision of biology could replace all previous knowledge by a new one based upon ATGC readings. For instance, in order to display bad consequences derived from the ignorance of these old scientific facts, we can take into account, for instance, how ion movements across membranes affects membrane protein conformation and therefore contradicts the wrong central dogma of molecular biology. This change in protein conformation (with unchanged amino acid sequence) and/or the lack of change in protein conformation is linked to the factors that affect vital processes as the heart beats. This modern ignorance could also explain some major pitfalls seen in new drugs clinical trials and in a small scale on bad medical practices.

The work of Britton Chance and of Peter Mitchell have deep and sound scientific roots that were made with excellent scientific techniques, supported by excellent scientific reasoning and that were produced in a large series of very important intermediary scientific results. Their sole difference was to aim at very different scientific explanations as their goals (They have different Teleology in their minds made by their previous experiences). When, with the use of mutants obtained in microorganisms P Mitchell´s goal was found to survive and B Chance to succumb to the experimental evidence, all those excellent findings of B Chance and followers were directed to the dustbin of scientific history as an example of lack of scientific consideration.  [On the one hand, the Mitchell model used a unicellular organism; on the other, Chance’s work was with eukaryotic cells, quite relevant to the discussion.]

We can resume the challenge faced by these two great scientists in the following form: The first conceptual unification in bioenergetics, achieved in the 1940s, is inextricably bound up with the name of Fritz Lipmann. Its central feature was the recognition that adenosine triphosphate, ATP, serves as a universal energy  “currency” much as money serves as economic currency. In a nutshell, the purpose of metabolism is to support the synthesis of ATP. In microorganisms, this is perfect! In humans or mammals, or vertebrates, by the same reason that we cannot consider that gene expression is equivalent to protein function (an acceptable error in the case of microorganisms) this oversimplifies the metabolic requirement with a huge error. However, in case our concern is ATP chemistry only, the metabolism produces ATP and the hydrolysis of ATP pays for the performance of almost, all kinds of works. It is possible to presume that to find out how the flow of metabolism (carbon flow) led to ATP production must be considered a major focal point of research of the two contenders. Consequently, what could be a minor fall of one of the contenders, in case we take into account all that was found during their entire life of research, the real failure in B Chance’s final goal was amplified far beyond what may be considered by reason!

Another aspect that must be taken into account: Both contenders have in the scientific past a very sound root. Metabolism may produce two forms of energy currency (I personally don´t like this expression*) and I use it here because it was used by both groups in order to express their findings. Together with simplistic thermodynamics, this expression conveys wrong ideas): The second kind of energy currency is the current of ions passing from one side of a membrane to the other. The P. Mitchell scientific root undoubtedly have the work of Hodgkin & Huxley, Huxley &  Huxley, Huxley & Simmons

*ATP is produced under the guidance of cell needs and not by its yield. When glucose yields only 2 ATPs per molecule it is oxidized at very high speed (anaerobiosis) as is required to match cellular needs. On the other hand, when it may yield (thermodynamic terms) 38 ATP the same molecule is oxidized at low speed. It would be similar to an investor choice its least money yield form for its investment (1940s to 1972) as a solid support. B. Chance had the enzymologists involved in clarifying how ATP could be produced directly from NADH + H+ oxidative reductive metabolic reactions or from the hydrolysis of an enolpyruvate intermediary. Both competitors had their work supported by different but, sound scientific roots and have produced very important scientific results while trying to present their hypothetical point of view.

Before the winning results of P. Mitchell were displayed, one line of defense used by B. Chance followers was to create a conflict between what would be expected by a restrictive role of proteins through its specificity ionic interactions and the general ability of ionic asymmetries that could be associated with mitochondrial ATP production. Chemical catalyzed protein activities do not have perfect specificity but an outstanding degree of selective interaction was presented by the lock and key model of enzyme interaction. A large group of outstanding “mitochondriologists” were able to show ATP synthesis associated with Na+, K+, Ca2+… asymmetries on mitochondrial membranes and any time they did this, P. Mitchell have to display the existence of antiporters that exchange X for hydrogen as the final common source of chemiosmotic energy used by mitochondria for ATP synthesis.

This conceptual battle has generated an enormous knowledge that was laid to rest, somehow discontinued in the form of scientific research, when the final E. Coli mutant studies presented the convincing final evidence in favor of P. Mitchell point of view.

Not surprisingly, a “wise anonymous” later, pointed out: “No matter what you are doing, you will always be better off in case you have a mutant”

(Principles of Medical Genetics T D Gelehrter & F.S. Collins chapter 7, 1990).

However, let’s take the example of a mechanical wristwatch. It clearly indicates when the watch is working in an acceptable way, that its normal functioning condition is not the result of one of its isolated components – or something that can be shown by a reductionist molecular view.  Usually it will be considered that it is working in an acceptable way, in case it is found that its accuracy falls inside a normal functional range, for instance, one or two standard deviations bellow or above the mean value for normal function, what depends upon the rigor wisely adopted. While, only when it has a faulty component (a genetic inborn error) we can indicate a single isolated piece as the cause of its failure (a reductionist molecular view).

We need to teach in medicine, first the major reasons why the watch works fine (not saying it is “automatic”). The functions may cross the reversible to irreversible regulatory limit change, faster than what we can imagine. Latter, when these ideas about normal are held very clear in the mind set of medical doctors (not medical technicians) we may address the inborn errors and what we may have learn from it. A modern medical technician may cause admiration when he uses an “innocent” virus to correct for a faulty gene (a rather impressive technological advance). However, in case the virus, later shows signals that indicate that it was not so innocent, a real medical doctor will be called upon to put things in correct place again.

Among the missing parts of normal evolution in biochemistry a lot about ion fluxes can be found. Even those oscillatory changes in Ca2+ that were shown to affect gene expression (C. De Duve) were laid to rest since, they clearly indicate a source of biological information that despite the fact that it does not change nucleotides order in the DNA, it shows an opposing flux of biological information against the dogma (DNA to RNA to proteins). Another, line has shown a hierarchy, on the use of mitochondrial membrane potential: First the potential is used for Ca2+ uptake and only afterwards, the potential is used for ADP conversion into ATP (A. L. Lehninger). In fact, the real idea of A. L. Lehninger was by far, more complex since according to him, mitochondria works like a buffer for intracellular calcium releasing it to outside in case of a deep decrease in cytosol levels or capturing it from cytosol when facing transient increase in Ca2+ load. As some of Krebs cycle dehydrogenases were activated by Ca2+, this finding was used to propose a new control factor in addition to the one of ADP (B. Chance). All this was discontinued with the wrong use of calculus (today we could indicate bioinformatics in a similar role) in biochemistry that has established less importance to a mitochondrial role after comparative kinetics that today are seen as faulty.

It is important to combat dogmatic reasoning and restore sound scientific foundations in basic medical courses that must urgently reverse the faulty trend that tries to impose a view that goes from the detail towards generalization instead of the correct form that goes from the general finding well understood towards its molecular details. The view that led to curious subjects as bioinformatics in medical courses as training in sequence finding activities can only be explained by its commercial value. The usual form of scientific thinking respects the limits of our ability to grasp new knowledge and relies on reproducibility of scientific results as a form to surpass lack of mathematical equation that defines relationship of variables and the determination of its functional domains. It also uses old scientific roots, as its sound support never replaces existing knowledge by dogmatic and/or wishful thinking. When the sequence of DNA was found as a technical advance to find amino acid sequence in proteins it was just a technical advance. This technical advance by no means could be considered a scientific result presented as an indication that DNA sequences alone have replaced the need to study protein chemistry, its responses to microenvironmental changes in order to understand its multiple conformations, changes in activities and function. As E. Schrodinger correctly describes the chemical structure responsible for the coded form stored of genetic information must have minimal interaction with its microenvironment in order to endure hundreds and hundreds years as seen in Hapsburg’s lips. Only magical reasoning assumes that it is possible to find out in non-reactive chemical structures the properties of the reactive ones.

For instance, knowledge of the reactions of the Krebs cycle clearly indicate a role for solvent that no longer could be considered to be an inert bath for catalytic activity of the enzymes when the transfer of energy include a role for hydrogen transport. The great increase in understanding this change on chemical reaction arrived from conformational energy.

Again, even a rather simplistic view of this atomic property (Conformational energy) is enough to confirm once more, one of the most important contribution of E. Schrodinger in his What is Life:

“This little book arose from a course of public lectures, delivered by a theoretical physicist to an audience of about four hundred which did not substantially dwindle, though warned at the outset that the subject matter was a difficult one and that the lectures could not be termed popular, even though the physicist’s most dreaded weapon, mathematical deduction, would hardly be utilized. The reason for this was not that the subject was simple enough to be explained without mathematics, but rather that it was much too involved to be fully accessible to mathematics.”

In a very simplistic view, while energy manifests itself by the ability to perform work conformational energy as a property derived from our atomic structure can be neutral, positive or negative (no effect, increased or decreased reactivity upon any chemistry reactivity measured as work)


“I mean the fact that we, whose total being is entirely based on a marvelous interplay of this very kind, yet if all possess the power of acquiring considerable knowledge about it. I think it possible that this knowledge may advance to little just a short of a complete understanding -of the first marvel. The second may well be beyond human understanding.”

In fact, scientific knowledge allows us to understand how biological evolution may have occurred or have not occurred and yet does not present a proof about how it would have being occurred. It will be always be an indication of possible against highly unlike and never a scientific proven fact about the real form of its occurrence.

As was the case of B. Chance in its bioenergetics findings, we may get very important findings that indicates wrong directions in the future as was his case, or directed toward our past.

The Skeleton of Physical Time – Quantum Energies in Relative Space of S-labs

By Radoslav S. Bozov  Independent Researcher

WSEAS, Biology and BioSystems of Biomedicine

Space does not equate to distance, displacement of an object by classically defined forces – electromagnetic, gravity or inertia. In perceiving quantum open systems, a quanta, a package of energy, displaces properties of wave interference and statistical outcomes of sums of paths of particles detected by a design of S-labs.

The notion of S-labs, space labs, deals with inherent problems of operational module, R(i+1), where an imagination number ‘struggles’ to work under roots of a negative sign, a reflection of an observable set of sums reaching out of the limits of the human being organ, an eye or other foundational signal processing system.

While heavenly bodies, planets, star systems, and other exotic forms of light reflecting and/or emitting objects, observable via naked eye have been deduced to operate under numerical systems that calculate a periodic displacement of one relative to another, atomic clocks of nanospace open our eyes to ever expanding energy spaces, where matrices of interactive variables point to the problem of infinity of variations in scalar spaces, however, defining properties of minute universes as a mirror image of an astronomical system. The first and furthermost problem is essentially the same as those mathematical methodologies deduced by Isaac Newton and Albert Einstein for processing a surface. I will introduce you to a surface interference method by describing undetermined objective space in terms of determined subjective time.

Therefore, the moment will be an outcome of statistical sums of a numerical system extending from near zero to near one. Three strings hold down a dual system entangled via interference of two waves, where a single wave is a product of three particles (today named accordingly to either weak or strong interactions) momentum.

The above described system emerges from duality into trinity the objective space value of physical realities. The triangle of physical observables – charge, gravity and electromagnetism, is an outcome of interference of particles, strings and waves, where particles are not particles, or are strings strings, or  are waves waves of an infinite character in an open system which we attempt to define to predict outcomes of tomorrow’s parameters, either dependent or independent as well as both subjective to time simulations.

We now know that aging of a biological organism cannot be defined within singularity. Thereafter, clocks are subjective to apparatuses measuring oscillation of defined parameters which enable us to calculate both amplitude and a period, which we know to be dependent on phase transitions.

The problem of phase was solved by the applicability of carbon relative systems. A piece of diamond does not get wet, yet it holds water’s light entangled property. Water is the dark force of light. To formulate such statement, we have been searching truth by examining cooling objects where the Maxwell demon is translated into information, a data complex system.

Modern perspectives in computing quantum based matrices, 0+1 =1 and/or 0+0=1, and/or 1+1 =0, will be reduced by applying a conceptual frame of Aladdin’s flying anti-gravity carpet, unwrapping both past and future by sending a photon to both, placing present always near zero. Thus, each parallel quantum computation of a natural system approaching the limit of a vibration of a string defining 0 does not equal 0, and 1 does not equal 1. In any case, if our method 1+1 = 1, yet, 1 is not 1 at time i+1. This will set the fundamentals of an operational module, called labris operator or in simplicity S-labs. Note, that 1 as a result is an event predictable to future, while interacting parameters of addition 1+1 may be both, 1 as an observable past, and 1 as an imaginary system, or 1+1 displaced interactive parameters of past observable events. This is the foundation of Future Quantum Relative Systems Interference (QRSI), taking analytical technologies of future as a result of data matrices compressing principle relative to carbon as a reference matter rational to water based properties.

Goedel’s concept of loops exist therefore only upon discrete relative space uniting to parallel absolute continuity of time ‘lags’. ( Goedel, Escher and Bach: An Eternal Golden Braid. A Metaphorical Fugue on Minds and Machines in the Spirit of Lewis Carroll. D Hofstadter.  Chapter XX: Strange Loops, Or Tangled Hierarchies. A grand windup of many of the ideas about hierarchical systems and self-reference. It is concerned with the snarls which arise when systems turn back on themselves-for example, science probing science, government investigating governmental wrongdoing, art violating the rules of art, and finally, humans thinking about their own brains and minds. Does Gödel’s Theorem have anything to say about this last “snarl”? Are free will and the sensation of consciousness connected to Gödel’s Theorem? The Chapter ends by tying Gödel, Escher, and Bach together once again.)  The fight struggle in-between time creates dark spaces within which strings manage to obey light properties – entangled bozons of information carrying future outcomes of a systems processing consciousness. Therefore, Albert Einstein was correct in his quantum time realities by rejecting a resolving cube of sugar within a cup of tea (Henri Bergson 19th century philosopher. Bergson’s concept of multiplicity attempts to unify in a consistent way two contradictory features: heterogeneity and continuity. Many philosophers today think that this concept of multiplicity, despite its difficulty, is revolutionary.) However, the unity of time and space could not be achieved by deducing time to charge, gravity and electromagnetic properties of energy and mass.

Charge is further deduced to interference of particles/strings/waves, contrary to the Hawking idea of irreducibility of chemical energy carrying ‘units’, and gravity is accounted for by intrinsic properties of   anti-gravity carbon systems processing light, an electromagnetic force, that I have deduced towards ever expanding discrete energy space-energies rational to compressing mass/time. The role of loops seems to operate to control formalities where boundaries of space fluctuate as a result of what we called above – dark time-spaces.

Indeed, the concept of horizon is a constant due to ever expanding observables. Thus, it fails to acquire a rational approach towards space-time issues.

Richard Feynman has touched on issues of touching of space, sums of paths of particle traveling through time. In a way he has resolved an important paradigm, storing information and possibly studying it by opening a black box. Schroedinger’s cat is alive again, but incapable of climbing a tree when chased by a dog. Every time a cat climbs a garden tree, a fruit falls on hedgehogs carried away parallel to living wormholes whose purpose of generating information lies upon carbon units resolving light.

In order to deal with such a paradigm, we will introduce i+1 under square root in relativity, therefore taking negative one ( -1 = sqrt (i+1), an operational module R dealing with Wheelers foam squeezed by light, releasing water – dark spaces. Thousand words down!

What is a number? Is that a name or some kind of language or both? Is the issue of number theory possibly accountable to the value of the concept of entropic timing? Light penetrating a pyramid holding bean seeds on a piece of paper and a piece of slice of bread, a triple set, where a church mouse has taken a drop of tear, but a blood drop. What an amazing physics! The magic of biology lies above egoism, above pride, and below Saints.

We will set up the twelve parameters seen through 3+1 in classic realities:

–              discrete absolute energies/forces – no contradiction for now between Newtonian and Albert Einstein mechanics

–              mass absolute continuity – conservational law of physics in accordance to weak and strong forces

–              quantum relative spaces – issuing a paradox of Albert Einstein’s space-time resolved by the uncertainty principle

–              parallel continuity of multiple time/universes – resolving uncertainty of united space and energy through evolving statistical concepts of scalar relative space expansion and vector quantum energies by compressing relative continuity of matter in it, ever compressing flat surfaces – finding the inverse link between deterministic mechanics of displacement and imaginary space, where spheres fit within surface of triangles as time unwraps past by pulling strings from future.

To us, common human beings, with an extra curiosity overloaded by real dreams, value happens to play in the intricate foundation of life – the garden of love, its carbon management in mind, collecting pieces of squeezed cooling time.

The infinite interference of each operational module to another composing ever emerging time constrains unified by the Solar system, objective to humanity, perhaps answers that a drop of blood and a drop of tear is united by a droplet of a substance separating negative entropy to time courses of a physical realities as defined by an open algorithm where chasing power subdue to space becomes an issue of time.

Jose Eduardo de Salles Roselino

Some small errors: For intance an increase i P leads to a decrease in V ( not an increase in V)..


Radoslav S. Bozov  Independent Researcher

If we were to use a preventative measures of medical science, instruments of medical science must predict future outcomes based on observable parameters of history….. There are several key issues arising: 1. Despite pinning a difference on genomic scale , say pieces of information, we do not know how to have changed that – that is shift methylome occupying genome surfaces , in a precise manner.. 2. Living systems operational quo DO NOT work as by vector gravity physics of ‘building blocks. That is projecting a delusional concept of a masonry trick, who has not worked by corner stones and ever shifting momenta … Assuming genomic assembling worked, that is dealing with inferences through data mining and annotation, we are not in a position to read future in real time, and we will never be, because of the rtPCR technology self restriction into data -time processing .. We know of existing post translational modalities… 3. We don’t know what we don’t know, and that foundational to future medicine – that is dealing with biological clocks, behavior, and various daily life inputs ranging from radiation to water systems, food quality, drugs…

Read Full Post »