Jul 22

Speeding up single-cell genomics research

0 comments

Scientists at Harvard University have developed a way to vastly accelerate single-cell sequencing, an advance that promises to give a major lift to biomedical genomics research.

The new approach combines microfluidics and novel software to accelerate single-cell ATAC-seq, which identifies parts of the genome that are open and accessible to regulatory proteins. Detailed in a June 24 article in Nature Biotechnology, the innovation was the product of a collaboration between Harvard’s Department of Stem Cell and Regenerative Biology (HSCRB) and researchers at Bio-Rad Laboratories.

 

 

“We’ve scaled it up so we can profile many cells in a human tissue in a straightforward way. The method is simple, so you can run the experiment in a day, start to finish. That will allow researchers to achieve much more in a far shorter time,” said Jason Buenrostro, senior author on the study and an assistant professor in HSCRB.

For many years, scientists have been trying to figure out how genes are turned on and off in different cells at different times in their development, so that cells can carry out specific functions. All life begins, after all, as a single cell, but quickly turns into trillions. The genome in that first cell is copied into all the rest, so how do we end up with so many different cell types?

Single-cell sequencing has transformed this area of research, allowing scientists to study genes in individual cells rather than large pieces of tissue, which contain many different cell types. Studying bigger samples can point researchers in a promising direction but does not give precise insights. Looking at single cells makes it easier to differentiate between cell types and to find individual genes that drive healthy functions or disease processes in the body.

In this study, researchers focused on a type of sequencing called ATAC-seq (Assay for Transposase-Accessible Chromatin using sequencing). Each human cell contains two meters of DNA, tightly packed into the microscopic nucleus. ATAC-seq identifies which parts of the DNA are unwound and accessible to proteins.

“It’s a bit like opening up a multidimensional suitcase to get at the clothes,” said Fabiana Duarte, co-lead author of the study and a postdoctoral fellow in Buenrostro’s lab at Harvard. “If a part of the genome is accessible, an enzyme can make a cut and tag it. Then we find the sequences of all the tagged DNA.”

Genes are controlled by many different proteins. Transcription factors, for example, bind to a piece of DNA and help turn genes on and off. There are countless transcription factors, and each one recognizes and binds to a very specific DNA sequence. Each sequence is called a motif, and because it is so specific it can be flagged in ATAC-seq data.

ATAC-seq was first developed by Buenrostro and colleagues in 2013 and its use is now ubiquitous in genomics — for example in major efforts like the Human Cell Atlas project to understand and map genome function. As a reflection of the impact of this work, MIT’s Technology Review has named Buenrostro as one of its 2019 “Innovators under 35” this week.

Until now, ATAC-seq has been a time-consuming and inefficient process, with only a fraction of cells yielding reliable results. Where previous approaches made it possible to profile 100 cells per reaction, the new method profiles 50,000.

One of the biggest challenges in developing the new method involved isolating the cells to be studied. The new method resolves this problem using microfluidics and droplets.

“We worked with the Bio-Rad team on the method,” Duarte said. “One channel of the device delivers a cell, and another adds a bead — and each bead has a bar-coded tag. Where they meet, there’s oil — so you get droplets. Each droplet has a cell and a bead. You can add a lot of cells into the device to get data on individually labelled cells.”

But the devil is in the details, as any computational biologist will tell you.

“Ideally you’d have one bead in each droplet. But in practice, to make sure that each cell is tagged, droplets can end up having more than one bead,” said Caleb Lareau, co-lead author and a Graduate School of Arts and Sciences graduate student in Buenrostro’s lab. “Our new software identifies those cases and merges them, so you can identify single cells that might at first have looked like many.”

The new approach has paid big dividends. “Because of experimental and computational innovations, we can now get profiles for 95 percent of the cells that are put into the machine. That’s up from around 20 to 30 percent — like night and day,” Lareau said.

“This droplet approach will be available to biomedical researchers in all fields as an off-the-shelf technology, and I’m really proud that we helped it get to this stage,” said Buenrostro, who is also a faculty member of the Harvard Stem Cell Institute. “In this study, we also developed a method within the method: Barcoding added another layer of tagging that allowed us to increase the number of cells we could assay 10-fold or more. It changes the questions you can ask, and how quickly you can find answers. I think we’re going to see research moving much more rapidly now.”

This study received support from the Allen Distinguished Investigator Program through the Paul G. Allen Frontiers Group, the Chan Zuckerberg Initiative, and the National Institutes of Health (NIH F31CA232670).

New Posts
  • An updated analysis from OpenAI shows how dramatically the need for computational resources has increased to reach each new AI breakthrough. In 2018, OpenAI found that the amount of computational power used to train the largest AI models had doubled every 3.4 months since 2012. The San Francisco-based for-profit AI research lab has now added new data to its analysis. This shows how the post-2012 doubling compares to the historic doubling time since the beginning of the field. From 1959 to 2012, the amount of power required doubled every 2 years, following Moore’s Law. This means the doubling time today is more than seven times the previous rate. This dramatic increase in the resources needed underscores just how costly the field’s achievements have become. Keep in mind, the above graph shows a log scale. On a linear scale (below), you can more clearly see how compute usage has increased by 300,000-fold in the last seven years. The chart also notably does not include some of the most recent breakthroughs, including Google’s large-scale language model BERT, OpenAI’s large-scale language model GPT-2,  or DeepMind’s StarCraft II-playing model AlphaStar. In the past year, more and more researchers have sounded the alarm on the exploding costs of deep learning. In June, an analysis from researchers at the University of Massachusetts, Amherst, showed how these increasing computational costs directly translate into carbon emissions. In their paper, they also noted how the trend exacerbates the privatization of AI research because it undermines the ability for academic labs to compete with much more resource-rich private ones. In response to this growing concern, several industry groups have made recommendations. The Allen Institute for Artificial Intelligence, a nonprofit research firm in Seattle, has proposed that researchers always publish the financial and computational costs of training their models along with their performance results, for example. In its own blog, OpenAI suggested policymakers increase funding to academic researchers to bridge the resource gap between academic and industry labs
  • StarckGate is happy to work together with Asimov that will be aiming to radically advance humanity's ability to design living systems. They strive to enable biotechnologies with global benefit by combining synthetic biology and computer science. With their help we will able to grasp the following domains better Synthetic Biology Nature has evolved billions of useful molecular nanotechnology devices in the form of genes, across the tree of life. We catalog, refine, and remix these genetic components to engineer new biological systems. Computational Modeling Biology is complex, and genetic engineering unlocks an unbounded design space. Computational tools are critical to design and model complex biophysical systems and move synthetic biology beyond traditional brute force screening. Cellular Measurement Genome-scale, multi-omics measurement technologies provide deep views into the cell. These techniques permit pathway analysis at the scale of a whole cell, and inspection down at single-nucleotide resolution. Machine Learning We are developing machine learning algorithms that bridge large-scale datasets with mechanistic models of biology. Artificial intelligence can augment human capabilities to design and understand biological complexity.
  • The use of AI (artificial intelligence) in agriculture is not new and has been around for some time with technology spans a wide range of abilities—from that which discriminates between crop seedlings and weeds to greenhouse automation. Indeed, it is easy to think that this is new technology given the way that our culture has distanced so many facets of food production, keeping it far away from urban spaces and our everyday reality. Yet, as our planet reaps the negative repercussions of technological and industrial growth, we must wonder if there are ways that our collective cultures might be able to embrace AI’s use in food production which might include a social response to climate change. Similarly, we might consider if new technology might also be used to educate future generations as to the importance of responsible food production and consumption. While we know that AI can be a force for positive change where, for instance, failures in food growth can be detected and where crops can be analyzed in terms of disease, pests and soil health, we must wonder why food growth has been so divorced from our culture and social reality. In recent years, there has been great pushback within satellite communities and the many creations of villages focussed upon holistic methods of food production. Indeed, RegenVillages is one of many examples where vertical farming, aquaponics, aeroponics and permaculture are part of this community's everyday functioning. Moreover, across the UK are many ecovillages and communities seeking to bring back food production to the core of social life. Lammas is one such ecovillage which I visited seven years ago in Wales which has, as its core concept, the notion of a “collective of eco-smallholdings working together to create and sustain a culture of land-based self-reliance.” And there are thousands of such villagesacross the planet whereby communities are invested in working to reduce their carbon footprint while taking back control of their food production. Even Planet Impact’s reforestation programs are interesting because the links between healthy forests and food production are well known as are the benefits of forest gardening which is widely considered a quite resilient agroecosystem. COO & Founder of Planetimpact.com, Oscar Dalvit, reports that his company’s programs are designed to educate as much as to innovate: “With knowledge, we can fight climate change. Within the for-profit sector, we can win this battle.” Forest gardening is a concept that is not only part of the permaculture practice but is also an ancient tradition still alive and well in places like Kerala, India and Martin Crawford’s forest garden in southwest England where his Agroforestry Research Trust offers courses and serves as a model for such communities across the UK. But how can AI help to make sustainable and local farming practices over and above industrial agriculture? Indeed, one must wonder if it is possible for local communities to take control of their food production. So, how can AI and other new tech interfaces bring together communities and food production methods that might provide a sustainable hybrid model of traditional methods and innovative technology? We know already that the IoT (internet of things) is fast becoming that virtual space where AI is being implemented to include within the latest farming technology. And where businesses invested in robotics are likewise finding that there is no ethical implementation of food technology, we must be mindful of how strategies are implemented which incorporate the best of new tech with the best of old tech. Where AI is helping smaller farms to become more profitable, all sorts of digital interfaces are transmitting knowledge, education and the expansion of local farming methods. This means, for instance, that garden maintenance is continued by others within the community as some members are absent for reasons of vacation or illness. Together with AI, customer experience is as much a business model as it is a local community standard for communication and empowerment. The reality is that industrial farming need not take over local food production and there are myriad ways that communities can directly respond to climate change and the encroachment of big agriculture. The health benefits of local farming practices are already well known as are the many ways that smartphone technology can create high-yield farms within small urban spaces. It is high time that communities reclaim their space within urban centers and that urban dwellers consider their food purchasing and consumption habits while building future sustainability which allows everyone to participate in local food production. As media has recently focussed upon AI and industrial farming, we need to encourage that such technology is used to implement local solutionsthat are far more sustainable and realistic instead of pushing big agriculture.

Proudly created by Starckgate 

© 2020 by Starckgate

  • White Facebook Icon
  • White Twitter Icon
  • White Instagram Icon