1 / 21

Challenges and Future of Genomic Data

This article discusses the challenges in storing and analyzing genomic data, including the exponential growth in data size, the need for efficient data management, and concerns about information access, data security, and patient privacy. The article also explores potential solutions such as cloud storage, service-oriented architectures, and minimizing data storage through the use of reference genomes.

Download Presentation

Challenges and Future of Genomic Data

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. On the Future of Genomic DataScott D. KahnScience 331, 728 (2011); DOI: 10.1126/science.1197891莊允心(D00921014)、陳亭霓(R01921078)、葉顏偉(R01944043)生醫資訊導論期末報告2013.05.20

  2. Abstract • Challenges in genomics derive from • The informatics needed to store and analyze the raw sequencing data • The need to process terabytes of information has become de rigueur (the regular) • Single week-long sequencing runs today can produce as much data as did entire genome centers a few years ago • The availability of deep (and large) genomic data sets raises concerns • Information access • Data security • Subject/Patient privacy

  3. Study of Genomic • The growth in the size of data • Effectively use • Manage the resulting derived information. • Genomes range from 4000 bases to 670GB • Humans have two copies of their inherited genome of 3.2 Gb each. • New sequencing is expanding at an exponential pace. • Full sequence data has been archived for many thousands of species • More than 3000 humans have been sequenced to some substantial extent and reported in the scientific literature;

  4. First Challenge---Growth • Growth in raw output has outstripped the Moore’s Law advances in information technology and storage capacity, • a standard analysis requires 1 to 2 days on a compute cluster and several weeks on a typical workstation. • next-generation sequencing (NGS) has grown from 10 Mb per day to 40 Gb per day on a single sequencer • 10 to 20 major sequencing labs worldwide that have each deployed more than 10 sequencers • It is driving a discussion about • The value and definition of “raw data” in genomics • The mechanisms for sharing data • The provenance of the tools that effectively define the derived information • The nature of community data repositories in the years ahead (Fig. 1).

  5. Second Challenge---Analyzing • Analyzing all these data effectively • Central to the challenge is the definition of raw data. • Capture image data for each base being sequenced • Must be parsed into a set of intensities as a specific base call and an assigned quality value(the likelihood that the base call is correct)

  6. Second Challenge---Analyzing • The quality values • More storage space than the base. • Greater than 5 TB of information per day • Archiving image data is impractical • Motivated the development of real-time processing of the images • The ability to process the images in near real time has allowed the speed of sequencing to advance

  7. Raw Data • Near real-time analysis • Affords a two-orders of-magnitude reduction in data needing to be stored, archived, and processed further. • Thus, raw data has been redefined to be bases and qualities • Even in “third-generation” sequencing methods • Base quality scores are captured, compressed, and archived will optimize storage and improve analysis

  8. Raw Data • The size of the collective data is becoming a major concern • The current size of the 1000 Genomes Project (www.1000genomes.org) pilot data • A comparative analysis of the full genomes of 629 people, is roughly 7.3 TB of sequence data.

  9. Storage Cloud • Accessing data remotely is limited by network and storage capabilities • The download time for a well-connected site within North America will range between 7 and 20 days. • Having this data reside within a storage cloud does not entirely mitigate the longer term challenges • Data stored within different clouds

  10. Storage Cloud • This fundamental inability to move sequence data in its current form • Considerable changes in format and approach. • Concomitant changes in the definition of data must take place to support movement of the field forward. • Centralization of data on the Cloud is a positive start.

  11. Service-oriented (SOAs) • Service-oriented architectures (SOAs) • encapsulate computation into transportable compute objects that can be run on computers that store targeted data. • SOA compute objects function like applications that are temporarily installed on a remote computer, perform an operation, and then are uninstalled. • This solution poses a challenge • How compute costs are shared • Collaborative compute grids may offer a solution here.

  12. Security and Privacy • Genomic data are inherently identifiable and that additional safeguards are required. • Provide additional regulation so as to discourage inappropriate use of such data • Removing all access to several genomic databases • More regulations

  13. Minimize data storage • Reference genomes • All that needs to be stored in a new analysis are the differences from the reference. • Only base mutations that are distinct from the reference need to be saved; typically • These differences represent just 0.1% of the data. • However, with analysis methods still under active development it may be premature for the transition to referential formats. • Referential formats can also pose problems with capture of data quality throughout a genome.

  14. Data Quality • Knowledge of data quality is most needed when evaluating derived information • To provide a contextual basis for the certainty of the assignment

  15. Metagenomics • Although there is widespread focus on human DNA sequencing • Genomic data can be even more complex • Further problem of the sequencing data • collected at many times, across many tissues, and/or at several collection locations • standards in quality and data vary or evolve • Much sequence data, both affecting humans and not, is not of human • viruses, bacteria, and more

  16. Metagenomic • Metagenomics • Actively engaged in scoping the data • Metadata requirements of the problems are being studied • Standards have not yet been agreed upon.

  17. Standards • Large intersite consortia have begun to develop standard references and protocols • benefit the broader informatics community • several such interdisciplinary workshops and conferences have been organized, and these are having modest success in capturing a shared focus to address the challenges presented.

  18. Electronic Medical Records (EMRs) • Apply sequencing information in order to guide clinical diagnoses and treatment. • Provides timely and efficient access to med records without compromising patient privacy, allow patients to engage in their own health care. • require large cross-functional teams that lack the informatics tools to capture the analysis and diagnostic process (or processes) • Have limited means to build a shared knowledge base.

  19. Future Works • Taking control of the size of data is an ongoing but tractable undertaking. • Use of referential compression will improve data issues

  20. Future Works • More difficult works • Clinical applications raises issues around positive and negative controls and what must be stored as part of the medical record. • The evolution of informatics frameworks • EMRs • SOAs

More Related