Computing Personal Genomics

By Nicole Hemsoth

June 2, 2011

Personal genomics is critical to advancing our ability to treat and preemptively diagnose genetic diseases. However, despite the possibilities of personalizing medicine, it remains tethered, in large part, to the weight of some significant computational-side problems. This includes everything from storage to compute to code, all of which were issues on the table at the National Center for Supercomputing Applications’ (NCSA) Private Sector Program Annual Meeting .

During the event, Dr. Victor Jongeneel, Senior Research Scientist at NCSA and the Institute for Genomic Biology at the University of Illinois detailed some of the bottlenecks and potential solutions that keep expectations for personal genomics grounded.

In the case of personal genomics, the problem is not the scientific understanding of the genome itself, it’s how to reconstruct, compare and make sense of the massive data from sequencers. He claims that the disruptive part of this technology as a whole is rooted in our ability to actually acquire the data. According to Jongeneel, the amount of DNA sequence data generated last year was more than what had been generated over the entire history of sequencing before that.

Personal genomics is anything but a reality right now Jongeneel says. He notes that the range of new services that offer to sequence your genome for a few hundred dollars are far from complete service. These simply take DNA from a saliva kit, probe for a certain number of positions in genomes that are known to be variable and then try to deduce personal characteristics from that information. He claims that this is not personal genomics because in such a case, all you’re examining are known differences between individuals in the population—not your own genome. Besides, to do what is required for a genuine look at one’s personal genomics is far more computationally-intensive and would entail far more than a measly few hundred dollars.

To realize true personal genomics, all differences between individuals need to be analyzed. Jongeneel explained that we are moving toward this more comprehensive genomic sampling via well-funded projects like the 1000 Genomes Initiative, which aims to allow the generation of all necessary data for $1000. He says this soon will be possible but again the computational bottlenecks are the main limitation.

Jongeneel cites three of the main technology vendors that are providing next-generation sequencing and says that while their approaches differ, on average, for a sequenced genome they’re running for 8 days for 200 gigabases worth of information. This translates into well over one terabyte per human genome.

When it’s human genomes sequences are the result of several hundred million (or even a billion) reads—a number that depends on the technology vendor. From there, researchers need to determine where they come from in the genome relative to common reference genomes. This “simple” alignment process whereby the individual genome is compared via alignment with the reference genome is incredibly demanding computationally—as is the next step where one must interpret this alignment to document individual differences and to make sure there is consistency.

Jongeneel says that this alignment step typically takes several days just for the processing of a single sample as it is aligned to the reference genome. To further complicate the process, we all have pieces of DNA that aren’t necessarily found in the DNA of others. While these are small differences he says these can make a very big difference. Analysis of these unique pieces require a complete piecing together of individual reads to allow researchers to see what the larger structure of the genome might look like. And it gets even more demanding.

Rebuilding genomes requires the construction of highly complex graphs, which itself is a strain on computational resources. This is even more demanding when one must disambiguate the graph to make sense of it in terms of an actual genome sequence. After all, there are pieces of sequence rolling off the machines that are on the order of between 75-100 nucleotides long—and you’re trying to reconstitute genomes that are in the millions or billions of nucleotides long. This is the scientific equivalent of fitting a cell-sized piece into a massive tabletop puzzle.

More concretely than the puzzle image, consider this: Jongeneel says that if you wanted to reconstruct an entire genome from this kind of information you’re talking about the construction of a graph would likely have over 3 billon nodes with in excess of 10 billion edges to it. This is, of course, assuming there are no errors in your data which, he apologizes, there probably are. The raw time taken for an algorithm on a medium-sized cluster the assembly properly takes several weeks for each genome.

Jongeneel says that this is the kind of bottleneck that prevents some interesting genomic projects from taking off. For instance, there is currently an effort to sequence the entire range of DNA for several hundred common vertebrates. However, storing that information and spending several weeks for each individual species makes that out of reach—for now, at least. He says that there is hope on the horizon, but it is going to take a rethinking of code and computing.

He says that the problem lies, in large part, in the software itself. His team ran a test on the widely-used genome assembler ABySS, which has broad appeal since it uses MPI and can leverage a much-needed cluster environment. They undertook assembly for a modest-sized genome of a yeast and noted that it was clear, based on wall clock and memory requirements, that this was not a scalable code.

He says this hints at a much deeper problem—many of those developing genomics software aren’t professional developers. Even though they integrate some complex algorithmic ideas, the code they write “isn’t up to the standards of the HPC community.”

He commented on this further, saying that what is needed most is a highly parallel genome assembler. He pointed to some progress in the arena from a group at Iowa State but says that unfortunately, “their software is not in the public domain so it isn’t available, we can’t test it and it’s not in the community.”

A representative from Microsoft in the audience asked Jongeneel about what the solution might be to this problem, inquiring if it was a simple need for more parallel programmers, better tools or languages for developing these, or some other new type of scalable solution. Jongeneel responded that since most of the code being produced is research grade and the technology moves so quickly that it renders “new” code obsolete in very little time. He says that commercial attempts have failed for the same reason—as soon as they’ve produced a viable, scalable solution they’ve been left behind by the swift movement toward new solutions.

Jongeneel said that if you think about personal genomics, if we even wanted to move toward the goal of one million people, we’re going to hit the exabyte range in no time. He feels that in addition these datasets need to be analyzed using workflows with multiple complex steps, thus we require a fundamental rethinking of compute architectures that can enable this kind of research.

That aside, he claims that one side question is what we should do with the massive amount of raw data that is valuable for future research (and sometimes legally sticky to dispose of now anyway). With this raw data in vast volume he says that extraction of ‘relevant’ information is the problem. Jongeneel notes, Data analytics and pattern discovery on large numbers of genomes will be required to produce meaningful results.

View full video from Jongeneel’s talk here.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

TACC Researchers Test AI Traffic Monitoring Tool in Austin

December 13, 2017

Traffic jams and mishaps are often painful and sometimes dangerous facts of life. At this week’s IEEE International Conference on Big Data being held in Boston, researchers from TACC and colleagues will present a new Read more…

AMD Wins Another: Baidu to Deploy EPYC on Single Socket Servers

December 13, 2017

When AMD introduced its EPYC chip line in June, the company said a portion of the line was specifically designed to re-invigorate a single socket segment in what has become an overwhelmingly two-socket landscape in the d Read more…

By John Russell

Microsoft Wants to Speed Quantum Development

December 12, 2017

Quantum computing continues to make headlines in what remains of 2017 as tech giants jockey to establish a pole position in the race toward commercialization of quantum. This week, Microsoft took the next step in advanci Read more…

By Tiffany Trader

HPE Extreme Performance Solutions

Explore the Origins of Space with COSMOS and Memory-Driven Computing

From the formation of black holes to the origins of space, data is the key to unlocking the secrets of the early universe. Read more…

ESnet Now Moving More Than 1 Petabyte/wk

December 12, 2017

Optimizing ESnet (Energy Sciences Network), the world's fastest network for science, is an ongoing process. Recently a two-year collaboration by ESnet users – the Petascale DTN Project – achieved its ambitious goal t Read more…

AMD Wins Another: Baidu to Deploy EPYC on Single Socket Servers

December 13, 2017

When AMD introduced its EPYC chip line in June, the company said a portion of the line was specifically designed to re-invigorate a single socket segment in wha Read more…

By John Russell

Microsoft Wants to Speed Quantum Development

December 12, 2017

Quantum computing continues to make headlines in what remains of 2017 as tech giants jockey to establish a pole position in the race toward commercialization of Read more…

By Tiffany Trader

HPC Iron, Soft, Data, People – It Takes an Ecosystem!

December 11, 2017

Cutting edge advanced computing hardware (aka big iron) does not stand by itself. These computers are the pinnacle of a myriad of technologies that must be care Read more…

By Alex R. Larzelere

IBM Begins Power9 Rollout with Backing from DOE, Google

December 6, 2017

After over a year of buildup, IBM is unveiling its first Power9 system based on the same architecture as the Department of Energy CORAL supercomputers, Summit a Read more…

By Tiffany Trader

Microsoft Spins Cycle Computing into Core Azure Product

December 5, 2017

Last August, cloud giant Microsoft acquired HPC cloud orchestration pioneer Cycle Computing. Since then the focus has been on integrating Cycle’s organization Read more…

By John Russell

GlobalFoundries, Ayar Labs Team Up to Commercialize Optical I/O

December 4, 2017

GlobalFoundries (GF) and Ayar Labs, a startup focused on using light, instead of electricity, to transfer data between chips, today announced they've entered in Read more…

By Tiffany Trader

HPE In-Memory Platform Comes to COSMOS

November 30, 2017

Hewlett Packard Enterprise is on a mission to accelerate space research. In August, it sent the first commercial-off-the-shelf HPC system into space for testing Read more…

By Tiffany Trader

SC17 Cluster Competition: Who Won and Why? Results Analyzed and Over-Analyzed

November 28, 2017

Everyone by now knows that Nanyang Technological University of Singapore (NTU) took home the highest LINPACK Award and the Overall Championship from the recently concluded SC17 Student Cluster Competition. We also already know how the teams did in the Highest LINPACK and Highest HPCG competitions, with Nanyang grabbing bragging rights for both benchmarks. Read more…

By Dan Olds

US Coalesces Plans for First Exascale Supercomputer: Aurora in 2021

September 27, 2017

At the Advanced Scientific Computing Advisory Committee (ASCAC) meeting, in Arlington, Va., yesterday (Sept. 26), it was revealed that the "Aurora" supercompute Read more…

By Tiffany Trader

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

AMD Showcases Growing Portfolio of EPYC and Radeon-based Systems at SC17

November 13, 2017

AMD’s charge back into HPC and the datacenter is on full display at SC17. Having launched the EPYC processor line in June along with its MI25 GPU the focus he Read more…

By John Russell

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Japan Unveils Quantum Neural Network

November 22, 2017

The U.S. and China are leading the race toward productive quantum computing, but it's early enough that ultimate leadership is still something of an open questi Read more…

By Tiffany Trader

GlobalFoundries Puts Wind in AMD’s Sails with 12nm FinFET

September 24, 2017

From its annual tech conference last week (Sept. 20), where GlobalFoundries welcomed more than 600 semiconductor professionals (reaching the Santa Clara venue Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

Leading Solution Providers

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Perspective: What Really Happened at SC17?

November 22, 2017

SC is over. Now comes the myriad of follow-ups. Inboxes are filled with templated emails from vendors and other exhibitors hoping to win a place in the post-SC thinking of booth visitors. Attendees of tutorials, workshops and other technical sessions will be inundated with requests for feedback. Read more…

By Andrew Jones

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

IBM Begins Power9 Rollout with Backing from DOE, Google

December 6, 2017

After over a year of buildup, IBM is unveiling its first Power9 system based on the same architecture as the Department of Energy CORAL supercomputers, Summit a Read more…

By Tiffany Trader

Delays, Smoke, Records & Markets – A Candid Conversation with Cray CEO Peter Ungaro

October 5, 2017

Earlier this month, Tom Tabor, publisher of HPCwire and I had a very personal conversation with Cray CEO Peter Ungaro. Cray has been on something of a Cinderell Read more…

By Tiffany Trader & Tom Tabor

Tensors Come of Age: Why the AI Revolution Will Help HPC

November 13, 2017

Thirty years ago, parallel computing was coming of age. A bitter battle began between stalwart vector computing supporters and advocates of various approaches to parallel computing. IBM skeptic Alan Karp, reacting to announcements of nCUBE’s 1024-microprocessor system and Thinking Machines’ 65,536-element array, made a public $100 wager that no one could get a parallel speedup of over 200 on real HPC workloads. Read more…

By John Gustafson & Lenore Mullin

Flipping the Flops and Reading the Top500 Tea Leaves

November 13, 2017

The 50th edition of the Top500 list, the biannual publication of the world’s fastest supercomputers based on public Linpack benchmarking results, was released Read more…

By Tiffany Trader

Intel Launches Software Tools to Ease FPGA Programming

September 5, 2017

Field Programmable Gate Arrays (FPGAs) have a reputation for being difficult to program, requiring expertise in specialty languages, like Verilog or VHDL. Easin Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Share This