Technical Architecture » History » Revision 46
Revision 45 (Ward Vandewege, 02/02/2015 08:50 PM) → Revision 46/62 (Ward Vandewege, 02/02/2015 08:51 PM)
h1. Technical Architecture !ArvadosTechnicalDiagramV16_website.png! The technical diagram above represents the basic architecture of Arvados. At the base layer is a "cloud operating system." Arvados does not depend on any particular stack - it runs in the cloud and on bare metal. The platform is currently deployed on Amazon Web Services (AWS) and on Google Compute Platform (GCP), as well as on bare metal. Stack-specific integration - for example for AWS or GCP - is kept to the bare minimum: specifically, the [[Node Manager]] component. h2. Key Components *[[Keep|Content Addressable Distributed File System ("Keep")]]* - Arvados stores files in Keep. Keep is a distributed file system that has been optimized for biomedical data files and write once read many (WORM) scenarios. Keep chunks files into 64MB data blocks and distributes them across physical drives or virtual volumes. Keep writes the data blocks to an underlying file system (e.g. Linux ext4). Keep is also a content addressable store (CAS). When a file is stored, a content address is created for each data block using a cryptographic digest of the contents of block. Then a manifest is created that identifies all of the blocks that make up the file. Each manifest has its own unique content address. This ensures every file can be accurately verified every time it is retrieved from the system. Keep also supports the creation of collections, which include multiple files, as a flexible way to define data sets without re-organizing data on disk. *[[Computation and Pipeline Processing|MapReduce Engine ("Crunch") & Pipeline Management]]* - Crunch manages distributed processing tasks across cores using the MapReduce mode that makes creating algorithms, which use distributed processing to analyze large data sets, much easier. It assigns processing tasks to cores that are physically close to the Keep nodes where data are stored. Crunch is designed to maintain data provenance and pipeline reproducibility. The system supports a flexibly mechanism for defining and invoking pipelines that use common components such as GATK or custom components. It automatically tracks exactly the data inputs and outputs through Keep, the inputs, and the code used for each job through the git repository. *"In-Memory Compact Genome Database 'Lightning'":https://github.com/curoverse/lightning* - Lightning uses a scale-out, open source in-memory database to store genomic data in a compact genome format. VCF files are not suitable for efficient look-ups so we are developing a format to represent variants and other key data for tertiary analysis. Putting this in in a scale-out, in-memory database will make it possible to do very fast queries of these data. This part of the project is in the early development stage. *[[REST API Server]]* - This component provides OAuth2-authenticated REST APIs to Arvados subsystems (metadata database, jobs, etc.) with the notable exception of Keep (which requires direct access to avoid network performance bottlenecks) and VMs and git (which use the SSH protocol and public key authentication). *[[Workbench]]* - Workbench is a set of visual tools for using the underlying Arvados services from a web browser. This is especially helpful for querying and browsing data, visualizing provenance, and monitoring jobs and pipelines. Workbench has a modular architecture designed for seamless integration with other Arvados applications. *[[SDKs|Command Line Interface]]* - The CLI tools provide convenient access to the Arvados API and services in the Arvados platform from the command line. *[[SDKs]]* - Arvados provides native language SDKs for Python, Perl, Ruby, R, and Java to make it easier to work with the REST APIs in common development environments. The SDKs also support the development of clients for Keep, Crunch and Lightning. (Some SDKs have not yet been implemented.) *[[Data Manager]]* - Data Manager helps to orchestrate interactions with data storage. This includes managing rules about permissions, replication, archiving, etc. *[[Node Manager]]* - Node manager manages compute resources in a cloud environment. It starts and stops compute nodes on demand. For a bare metal installation, the number of compute nodes available tends to be static, which means and as such Node Manager is not required. Node Manager currently supports AWS and GCP. *"Documentation":http://doc.arvados.org* - This is the official documentation, which is also included in the Arvados source tree. There is also documentation developer information on the [[Documentation project]] wiki page.