Presentation is loading. Please wait.

Presentation is loading. Please wait.

IU ORE-Chem Update Marlon Pierce, Geoffrey Fox Indiana University.

Similar presentations


Presentation on theme: "IU ORE-Chem Update Marlon Pierce, Geoffrey Fox Indiana University."— Presentation transcript:

1 IU ORE-Chem Update Marlon Pierce, Geoffrey Fox Indiana University

2 IU to lead New US NSF Track 2d $10M Award See http://www.futuregrid.org for more information.http://www.futuregrid.org

3 What We Said We Would Do Apply data-centric workflow technologies (Dryad) – Significant effort Install and run triple store – Done locally. – Need to do this in Azure. Design alternative formats for ORE (JSON, Microformats) – Nothing to report yet Design secure services, compositions, mash-ups – OAuth piece done. – Significant effort on social network interfaces – Nothing to report on ORE-chem enabled services yet Investigate clouds for ORE-Chem – Infrastructure and runtime – Significant effort on virtual data stores, overheads of virtualization.

4 Layer Cake of IU Activities Web 2.0 Research: Security for REST Services Cloud Computing: Infrastructure and Runtimes Infrastructure: Windows HPC Testbeds

5 Hardware Layer

6 Cloud Infrastructure Tempest: HP distributed shared memory cluster with 768 processor cores and 1.5 TB total memory capacity. The cluster includes 13.7 TB of local spinning disk. – Tempest can be dynamically reconfigured to act as either a Windows HPC or Linux cluster. – Smaller versions Madrid and Barcelona Other machines: – The IBM iDataPlex system is an IBM e1350 distributed shared memory cluster with 1024 processor cores and 3 TB total memory capacity. – Cray XT5m distributed shared memory cluster with 672 processor cores and 1.3 TB total memory capacity. – A shared memory system with at least 480 cores and 640 GB of RAM will also be installed at IU as part of the FutureGrid award.

7

8 Data Cloud Infrastructure

9 Triple Store: Intellidimension This has been installed on IU servers. We are ready for data. Efforts to install this on MS Azure were not successful. – Inadequate documentation earlier in the year. – We will revisit this.

10 Open Elastic Block Store Amazon EBS is a way to mount virtual disks in cloud- space. – Empty disk space or archived data stores – ORECHEM enabled data sets, for example. – Clone-able, so keep your own version of community data. We are implementing an open version of this. – Contribute to Nimbus, an open-source EC2 – But independent of Xen, etc. – Would be interesting to do this for Windows Eventual backbone: IU has over a petabyte disk space of lustre file system. – Can be used to load and store VMs. X. Gao won best student poster award at TG09. – Paper accepted to E-Science 2009

11 Block Store Architecture Volume Server Volume Delegate Virtual Machine Manager (Xen Dom 0) VMM Delegate VM instance ( Xen Dom U) VBS Web Service VBS Client VBD iSCSI Create Volume, Export Volume, Create Snapshot,Etc. Import Volume, Attach Device, Detach Device,Etc.

12 Integration with Cloud Computing Systems Volume Server Volume Delegate Xen Dom 0 Xen Delegate Xen Dom U VBS Web Service VBS Client VBD iSCSI Create Volume, Export Volume, Create Snapshot,Etc. Import Volume, Attach Device, Detach Device,Etc. Nimbus Workspace Service VBS_Nimbus Web Service Attach-volume Query for Xen Dom0 Host and DomUId with

13 Programming Clouds

14 Multicore and Cloud Technologies to support Data Intensive applications Using Dryad (Microsoft) and MPI to study structure of Gene Sequences on Tempest Cluster. We are working on PubChem. See http://www. infomall.org/ salsa for lab projects (X. Qiu). http://www. infomall.org/ salsa

15 PubChem dataset consists of binary 166 MACCS keys (fingerprints), which indicate whether a each chemical compound has a special functional molecule or not We have total 26,466,421 chemical compounds. (i.e, the total PubChem dataset has 166 dimensions and 26M records) Randomly selected 50K chemicals to produce 3D GTM map. GTM is an algorithm to find a lower dimension structure from higher dimensional data (3D in this case). http://www.youtube.com/watch?v=nylgjKgnSLg Courtesy of Jong Y. Choi

16 IU’s ORE-CHEM Pipeline Harvest NIH PubChem for 3D Structures Convert PubChem XML to CML Convert CML to Gaussian Input Submit Jobs to TeraGrid with Swarm Convert Gaussian Output to CML Convert CML to RDF->ORE- Chem Insert RDF into RDF Triple Store Conversions are done with Jumbo/CML tools from Peter Murray Rust’s group at Cambridge. Swarm is a Web service capable of managing 10,000’s of jobs on the TeraGrid. We are developing a Dryad version of the pipeline. Goal is to create a public, searchable triple store populated with ORE-CHEM data on drug-like molecules.

17 Iterative MapReduce- Kmeans Clustering and Matrix Multiplication Iterative MapReduce algorithm for Matrix Multiplication Kmeans Clustering implemented as an iterative MapReduce application Overhead of parallel runtimes – Matrix Multiplication Compute intensive application O(n^3) Higher data transfer requirements O(n^2) CGL-MapReduce shows minimal overheads next to MPI Overhead of parallel runtimes – Kmeans Clustering O(n) calculations in each iteration Small data transfer requirements O(1) With large data sets, CGL- MapReduce shows negligible overheads Extremely higher overheads in Hadoop and Dryad Jaliya Ekanayake {jekanaya@cs.indiana.edu}

18 Performance of MPI on virtualized resources – Evaluated using a dedicated private cloud infrastructure – Exactly the same hardware and software configurations in bare-metal and virtual nodes – Applications with different communication: computation ratios – Different virtual machine(VM) allocation strategies {1-VM per node to 8-VMs per node} High Performance Parallel Computing on Cloud Performance of Matrix multiplication under different VM configurations Overhead under different VM configurations for Concurrent Wave Equation Solver O(n^2) communication (n = dimension of a matrix) More susceptible to bandwidth than latency Minimal overheads under virtualized resources O(1) communication (Smaller messages) More susceptible to latency Higher overheads under virtualized resources Jaliya Ekanayake {jekanaya@cs.indiana.edu}

19 Conclusions: Dryad for Scientific Computing Investigated several applications with various computation, communication, and data access requirements All DryadLINQ applications work, and in many cases perform better than Hadoop We can definitely use DryadLINQ (and Hadoop) for scientific analyses We did not implement (find) – Applications that can only be implemented using DryadLINQ but not with typical MapReduce Current release of DryadLINQ has some performance limitations DryadLINQ hides many aspects of parallel computing from user Coding is much simpler in DryadLINQ than Hadoop (provided that the performance issues are fixed) Key issue is support of inhomogeneous data

20 IU’s ORE-CHEM Pipeline Harvest NIH PubChem for 3D Structures Convert PubChem XML to CML Convert CML to Gaussian Input Submit Jobs to TeraGrid with Swarm Convert Gaussian Output to CML Convert CML to RDF->ORE- Chem Insert RDF into RDF Triple Store Conversions are done with Jumbo/CML tools from Peter Murray Rust’s group at Cambridge. Swarm is a Web service capable of managing 10,000’s of jobs on the TeraGrid. We are developing a Dryad version of the pipeline. Goal is to create a public, searchable triple store populated with ORE-CHEM data on drug-like molecules.

21 Architecture of Swarm Service Windows Server Cluster Windows Server Cluster Swarm- Grid Swarm- Dryad Local RDMBS Swarm-Analysis Standard Web Service Interface Large Task Load Optimizer Swarm-Grid Connector Swarm-Dryad Connector Swarm-Hadoop Connector Cloud Comp. Cluster Cloud Comp. Cluster Grid HPC/ Condor Cluster Grid HPC/ Condor Cluster Swarm- Hadoop

22 Swarm-Grid Swarm considers traditional Grid HPC cluster are suitable for the high-throughput jobs. – Parallel jobs (e.g. MPI jobs) – Long running jobs Resource Ranking Manager – Prioritizes the resources with QBETS, INCA Fault Manager – Fatal faults – Recoverable faults Resource Ranking Manager Grid HPC/Condor pool Resource Connector Condor(Grid/Vanilla) with Birdbath Grid HPC Clusters Condor Cluster Standard Web Service Interface Swarm-Grid QBETS Web Service Local RDMBS MyProxy Server MyProxy Server Hosted by TeraGrid Project Hosted by UCSB Request Manager Job Distributor Job Queue Data Model Manager Fault Manager User A’s Job Board

23 Some Details We can submit jobs to 3 different TeraGrid machines – Abe, Mercury, Cobalt (all at NCSA) – IU’s BigRed has some technical problems Can do about 100-200 molecules per day in tests. Approach is fragile because application/system admins have tendency to change things every few months. – Don’t validate Globus invocations of codes

24 Dryad Data Partitioning Two methods: – Manually place the files in every node or – Write a C# code that uses DryadLINQ partitioning operators like Hash Partition or Range Partition A partitioned data set consists of 2 types of files: – A metadata file (.pt as extension) containing metadata that describes the partitions – Set of partition files, one for each data partition. \DryadData\UserName\InputData (file path and name) 4 (number of partitions depending on number of nodes available) 0,2000,NODE01 (Partition files: Partition number, size(in bytes), node name : File path) 1,2000,NODE02,NODE03:FilePath 2,2000,NODE03,NODE04 3,2000,NODE04

25 Programming the Pipeline IQueryable represents query over the data Input data is represented by a PartitionedTable object DryadLINQ programs apply LINQ query operations to PartitionedTable objects. LINQ queries on the PartitionTable object are executed on the Cluster. Jobs will be executed on different nodes and the output would be collected in the outputDirectory. IQueryable filenames = PartitionedTable.Get (filepathuri); IQueryable outputs = filenames.Select(s => XML2CML( execFileName, programSwitches, s.line, outputDirectory));

26 Security in Web 2.0

27 OAuth: REST Security This is actually a Year 2 deliverable but we made progress in Year 1. OAuth is essentially security for REST. – Provide authentication and authorization – Relevant to ORE-CHEM services Use REST URL and HTTP method – Resources are identified by URLs – Access privileges are identified by HTTP methods (GET, POST, PUT, DELETE) Extend OAuth – Add finer-grained authorization information in requests and responses.

28 OAuth Security Status OAuth *Core* Code provides the fundamental piece of OAuth specification 1.0. Includes minimal webapp example – The sample web apps just support shared secret. – We extend to support PKI – Also fixed some bugs in the code. – To support OAuth extensions, more code is needed in OAuth core. For OpenID, we use library OpenID4Java and it seems to offer enough functionalities so far. Tutorial given at TeraGrid09 Slides: http://www.collab- ogce.org/ogce/images/3/39/OAuthOverview-TG09.ppthttp://www.collab- ogce.org/ogce/images/3/39/OAuthOverview-TG09.ppt Code: http://ogce.svn.sourceforge.net/viewvc/ogce/incubator/OGCE- OAuth/

29 Acknowledgments Geoffrey Fox Judy Qiu and SALSA team: data mining – www.infomall.org/salsa www.infomall.org/salsa Jaliya Ekanayake: Dryad and Cloud performance Sangmi Pallickara: Swarm service Xiaoming Gao: Virtual Block Store Zhenhua Guo: OAuth, OpenID, and Social Networks

30 Full Stop

31 Dryad and DryadLINQ Dryad is a high-performance, general-purpose distributed computing engine that simplifies the task of implementing distributed applications on clusters of computers running a windows operating system. DryadLINQ allows us to implement Dryad applications in managed code by using an extended version of the LINQ programming model and API. LINQ was introduced with Microsoft.NET framework version 3.5. DryadLINQ provider translates the application’s LINQ queries into a Dryad job and runs the job as a distributed application on a windows HPC cluster. Dryad and DryadLINQ An Introduction version 1.0. June 30, 2009. by Microsoft Research

32

33 DryadLINQ Programming Guide version 1.0. June 30, 2009. by Microsoft Research

34 Client Workstation: runs DryadLINQ application. DryadLINQ Provider creates a Windows HPC job on the cluster to handle the Dryad processing, receives the results, and returns them to the application. Job Manager: Windows HPC task that manages execution of associated Dryad job on the cluster. Head Node: manages the cluster, hosts the Windows HPC Administration Console and Dryad management service. Vertices: perform the data processing on the compute nodes.

35 java.exe + jumboconverters.jar xml -> cml cml -> gaussian input Local Machine DryadLINQ Provider ( LinqToDryad.dll) Dryad Cluster Distribute gaussian Input files across the cluster and run gaussian.exe on every file at every node in the cluster.

36 Distribute all the initial xml files over the cluster xml to cml conversion Stage1 cml to gaussian conversion Stage2 run gaussian on every file Stage3 Dryad Cluster

37 Drilling Though Data Clouds Bare metal (Computer, network, storage) Bare metal (Computer, network, storage) FutureGrid/VM/Virtual Storage Cloud Technologies (MapReduce, Dryad, Hadoop) Cloud Technologies (MapReduce, Dryad, Hadoop) Classic HPC MPI Classic HPC MPI Applications  Cheminformatics: Mapping PubChem data into low dimensions to aid drug discovery  Biology: Expressed Sequence Tag (EST) sequence assembly (CAP3)  Biology: Pairwise Alu sequence alignment (SW)  Health: Correlating childhood obesity with environmental factors Data mining Algorithm Clustering (Pairwise, Vector) MDS, GTM, PCA, CCA Visualization PlotViz

38 Architecture and Performance of Runtime Environments for Data Intensive Scalable Computing Data/compute intensive applications implemented as MapReduce “filters” Architecture of CGL-MapReduce Measured using 32 Compute nodes each with 8 cores and 16 GB of memory Compute intensive application Embarrassingly parallel operation All runtimes performs equally well Number of Reads processed High Energy Physics Data Analysis CAP3 – Gene Assembly Program Data intensive application MapReduce style parallel operation Both runtimes perform comparably well Jaliya Ekanayake {jekanaya@cs.indiana.edu}

39 Job Execution Time in Swarm-Dryad with Windows HPC 16 nodes

40 Job Execution Time in Swarm-Dryad various number of nodes


Download ppt "IU ORE-Chem Update Marlon Pierce, Geoffrey Fox Indiana University."

Similar presentations


Ads by Google