This includes raw data and metadata for projects, experiments, and samples submitted by investigators, as well as derived analyses and quality metrics automatically generated from running vetted
secondary analysis pipelines.
Diagram of key components of the open - source data coordination platform, including [1] a data ingestion service, [2] a synchronized data store with multiple cloud replicas [3] a collection of
secondary analysis pipelines for basic data processing and [4] a collection of tertiary portals for analyses, visualizations, and rich forms of data access.
As currently conceived, this data coordination platform will provide four key components: ingestion services for submission of data; synchronized data storage across multiple clouds; standardized
secondary analysis pipelines; and portals for data access, tertiary analysis, and visualization.
Not exact matches
The platform will provide robust, community - vetted
pipelines that run on all newly submitted data and generate
secondary analysis results to be deposited back into the Data Store.
The HCA
Secondary Analysis service will provide
pipeline execution to process raw data using community - vetted algorithms and generate intermediate derived results that will be deposited back into the Data Store.