Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference.

Slides:



Advertisements
Similar presentations
Parallel Programming and Algorithms : A Primer Kishore Kothapalli IIIT-H Workshop on Multi-core Technologies International Institute.
Advertisements

Introductions to Parallel Programming Using OpenMP
Beowulf Supercomputer System Lee, Jung won CS843.
1 DISTRIBUTION STATEMENT XXX– Unclassified, Unlimited Distribution Laser Propagation Modeling Making Large Scale Ultrashort Pulse Laser Simulations Possible.
Ionization of the Hydrogen Molecular Ion by Ultrashort Intense Elliptically Polarized Laser Radiation Ryan DuToit Xiaoxu Guan (Mentor) Klaus Bartschat.
Parallel Computation of the 2D Laminar Axisymmetric Coflow Nonpremixed Flames Qingan Andy Zhang PhD Candidate Department of Mechanical and Industrial Engineering.
Information Technology Center Introduction to High Performance Computing at KFUPM.
AASPI Software Computational Environment Tim Kwiatkowski Welcome Consortium Members November 18, 2008.
Division of Pharmacokinetics and Drug Therapy Department of Pharmaceutical Biosciences Uppsala University Estimating and forecasting in vivo drug disposition.
HIGH PERFORMANCE COMPUTING ENVIRONMENT The High Performance Computing environment consists of high-end systems used for executing complex number crunching.
Reference: Message Passing Fundamentals.
Introduction to Scientific Computing Doug Sondak Boston University Scientific Computing and Visualization.
Hitachi SR8000 Supercomputer LAPPEENRANTA UNIVERSITY OF TECHNOLOGY Department of Information Technology Introduction to Parallel Computing Group.
Sun FIRE Jani Raitavuo Niko Ronkainen. Sun FIRE 15K Most powerful and scalable Up to 106 processors, 576 GB memory and 250 TB online disk storage Fireplane.
Linux clustering Morris Law, IT Coordinator, Science Faculty, Hong Kong Baptist University.
Large-Scale Density Functional Calculations James E. Raynolds, College of Nanoscale Science and Engineering Lenore R. Mullin, College of Computing and.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Bondyakov A.S. Institute of Physics of ANAS, Azerbaijan JINR, Dubna.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
1 Down Place Hammersmith London UK 530 Lytton Ave. Palo Alto CA USA.
Motivation “Every three minutes a woman is diagnosed with Breast cancer” (American Cancer Society, “Detailed Guide: Breast Cancer,” 2006) Explore the use.
Efficient Parallel Implementation of Molecular Dynamics with Embedded Atom Method on Multi-core Platforms Reporter: Jilin Zhang Authors:Changjun Hu, Yali.
1 Programming Multicore Processors Aamir Shafi High Performance Computing Lab
‘Tis not folly to dream: Using Molecular Dynamics to Solve Problems in Chemistry Christopher Adam Hixson and Ralph A. Wheeler Dept. of Chemistry and Biochemistry,
Performance Evaluation of Hybrid MPI/OpenMP Implementation of a Lattice Boltzmann Application on Multicore Systems Department of Computer Science and Engineering,
An approach for solving the Helmholtz Equation on heterogeneous platforms An approach for solving the Helmholtz Equation on heterogeneous platforms G.
Operational computing environment at EARS Jure Jerman Meteorological Office Environmental Agency of Slovenia (EARS)
National Center for Supercomputing Applications University of Illinois at Urbana-Champaign Direct Self-Consistent Field Computations on GPU Clusters Guochun.
© 2008 The MathWorks, Inc. ® ® Parallel Computing with MATLAB ® Silvina Grad-Freilich Manager, Parallel Computing Marketing
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
Computational issues in Carbon nanotube simulation Ashok Srinivasan Department of Computer Science Florida State University.
Sobolev Showcase Computational Mathematics and Imaging Lab.
The WRF Model The Weather Research and Forecasting (WRF) Model is a mesoscale numerical weather prediction system designed for both atmospheric research.
Introduction, background, jargon Jakub Yaghob. Literature T.G.Mattson, B.A.Sanders, B.L.Massingill: Patterns for Parallel Programming, Addison- Wesley,
Rensselaer Why not change the world? Rensselaer Why not change the world? 1.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Porting Irregular Reductions on Heterogeneous CPU-GPU Configurations Xin Huo, Vignesh T. Ravi, Gagan Agrawal Department of Computer Science and Engineering.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
MATRIX MULTIPLY WITH DRYAD B649 Course Project Introduction.
JLab Scientific Computing: Theory HPC & Experimental Physics Thomas Jefferson National Accelerator Facility Newport News, VA Sandy Philpott.
Tests and tools for ENEA GRID Performance test: HPL (High Performance Linpack) Network monitoring A.Funel December 11, 2007.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
CCS Overview Rene Salmon Center for Computational Science.
2009/4/21 Third French-Japanese PAAP Workshop 1 A Volumetric 3-D FFT on Clusters of Multi-Core Processors Daisuke Takahashi University of Tsukuba, Japan.
Computing Resources at Vilnius Gediminas Technical University Dalius Mažeika Parallel Computing Laboratory Vilnius Gediminas Technical University
1 The Performance Analysis of Molecular dynamics RAD GTPase with AMBER application on Cluster computing environtment. The Performance Analysis of Molecular.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
The Distributed Data Interface in GAMESS Brett M. Bode, Michael W. Schmidt, Graham D. Fletcher, and Mark S. Gordon Ames Laboratory-USDOE, Iowa State University.
CS- 492 : Distributed system & Parallel Processing Lecture 7: Sun: 15/5/1435 Foundations of designing parallel algorithms and shared memory models Lecturer/
Efficiency of small size tasks calculation in grid clusters using parallel processing.. Olgerts Belmanis Jānis Kūliņš RTU ETF Riga Technical University.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
Running Mantevo Benchmark on a Bare-metal Server Mohammad H. Mofrad January 28, 2016
A Pattern Language for Parallel Programming Beverly Sanders University of Florida.
Intermediate Parallel Programming and Cluster Computing Workshop Oklahoma University, August 2010 Running, Using, and Maintaining a Cluster From a software.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Scaling up R computation with high performance computing resources.
First INFN International School on Architectures, tools and methodologies for developing efficient large scale scientific computing applications Ce.U.B.
Hybrid Parallel Implementation of The DG Method Advanced Computing Department/ CAAM 03/03/2016 N. Chaabane, B. Riviere, H. Calandra, M. Sekachev, S. Hamlaoui.
HPC need and potential of ANSYS CFD and mechanical products at CERN A. Rakai EN-CV-PJ2 5/4/2016.
HPC Roadshow Overview of HPC systems and software available within the LinkSCEEM project.
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
White Rose Grid Infrastructure Overview
Grid Computing.
WORKFLOW PETRI NETS USED IN MODELING OF PARALLEL ARCHITECTURES
Is System X for Me? Cal Ribbens Computer Science Department
University of Technology
Simulation at NASA for the Space Radiation Effort
Overview of HPC systems and software available within
Hybrid Programming with OpenMP and MPI
Parallel computing in Computational chemistry
Presentation transcript:

Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference in Informatics

PHYSON The culmination of our HPC efforts so far: Joint project between Sofia University and Technical University Assembled by PERSY Ltd Proudly hosted in the department of Theoretical Physics Faculty of Physics, Sofia University “St. Kliment Ohridski” (in Bulgarian only)

Physon – the pink cluster 4 compute nodes: ◦ dual Intel Xeon E5335 (4 2 GHz) ◦ 12 GB (16 node001) ECC DDR RAM ◦ 250 GB SATA2 HDD ◦ 2 x 1 Gbps Fast Ethernet ◦ 1 x 20 Gbps 4x DDR InfiniBand ◦ Scientific Linux bit ◦ Sun N1 Grid Engine executives NFS/NIS server: ◦ Intel Core2 Duo E6600 (2 2,4 GHz) ◦ 2 GB ECC DDR2-667 RAM ◦ 4 x 500 GB SATA2 HDD (total of 1,75 TB in ZFS RAIDZ1 array) ◦ 2 x 1 Gbps Fast Ethernet ◦ Sun Solaris ◦ Sun N1 Grid Engine master

Sun N1 Grid Engine queues Interactive queue ◦ For GUI and interactive programs ◦ Up to 8 cores (1 node) Batch queues ◦ For long running non-interactive tasks ◦ Up to 24 cores (3 nodes) ◦ Interactive node available for serial batch tasks at night Fair share scheduler (to be enabled soon)

Parallel environments ompi – for MPI tasks ◦ OpenMPI – both GNU and Sun Studio compiled ◦ Up to 24 cores per task ◦ Uses SysV IPC on single node and InfiniBand transport between nodes threads – for threaded tasks ◦ OpenMP – Sun Studio only ◦ POSIX Threads ◦ Intel TBB ◦ Up to 8 cores per task (single node only) ◦ Intel Cluster OpenMP considered

Ganglia monitoring system

APPLICATIONS And here they come:

Multicore applications on Physon Hybrid architecture – it is both shared and distributed memory machine Public domain codes: ◦ GROMACS – molecular biology research ◦ Quantum ESPRESSO – electronic structure and nanoscale simulations Custom codes Development tools and libraries: ◦ Sun Studio 12 – C/C++/Fortran compilers and IDE, Sun Performance Library ◦ GNU Compiler Collection – C, C++, FORTRAN77 ◦ OpenMPI

Molecular Dynamics Solves in discrete time steps equations of motion: Force computation is the single time consuming operation – done in parallel using space decomposition techniques CPU 1CPU 2 Exchange area

Molecular Dynamics (2) Requires only local communication between neighboring CPUs for short ranged forces Asynchronous communication leads to good parallel scaling Long ranged forces require global (synchronous) communication – bad scaling 

Density Functional Theory A tool to compute/predict electronic structure and properties of molecules and crystals Minimizes given a set of basis functions, usually planar waves or gaussians Essentially solves a large number of linear equations Requires vast amount of memory

Density Functional Theory (2) Can be combined with Molecular Dynamics – Car-Parrinello MD DFT parallelization: ◦ Use parallel versions of basic linear algebra packages like BLAS and LAPACK ◦ Distribute memory requirements between nodes ◦ Global communication leads to bad parallel scaling

Application: nanotechnology Buckling of carbon nanotube under rotational load – MD simulation using Brenner-Tersoff potential and custom code (BOZA).

BOZA Molecular Dynamics tool for simulations of carbon nanostructures Written in FORTRAN 77 Parallelized using OpenMP Runs on shared memory computers Compiles as pure serial application on uniprocessors or with compilers that do not understand OpenMP

Application: nanotechnology Electron density for cis-[Co(en) 2 Cl 2 ]

Quantum ESPRESSO Swiss army knife for quantum simulations ◦ Electronic structure ◦ Geometry optimization ◦ Car-Parrinello molecular dynamics Written in Fortran 90/95 Parallelized using MPI library calls Runs on both distributed memory (clusters) and shared memory machines

QE: performance # of coresSpeedupParallel efficiency (speedup per core) 21,9899,2% 43,6491,1% 85,5569,3% 125,7748,1% 165,8736,7%

Application: drug design Interferon-gamma alpha-receptor complex studied with MD simulation using GROMACS

GROMACS Molecular Dynamics tool for simulations in molecular biology and biochemistry Written in C Parallelized using MPI library calls Many built-in force fields Runs on both distributed memory (clusters) and shared memory machines

GROMACS: performance # of coresComputation speedComp. speed per core 812,30 GFLOP/s1,54 GFLOP/s 1621,65 GFLOP/s1,35 GFLOP/s 2424,91 GFLOP/s1,04 GFLOP/s

CHALLENGES Propaganda:

Challenges Parallel programming still not popular at Sofia University Most people still using moderate PCs to do research Masters course “Parallel programming and applications”: ◦ Already taught for two years (MPI only) ◦ MPI and OpenMP basics ◦ Performance measurement and optimization ◦ Using batch processing engines and HPC Demands for computational power are growing!

THANK YOU FOR YOUR ATTENTION! And thus it ends.