Presentation is loading. Please wait.

Presentation is loading. Please wait.

Setting up Small Grid Testbed

Similar presentations

Presentation on theme: "Setting up Small Grid Testbed"— Presentation transcript:

1 Setting up Small Grid Testbed
Computational Fluid Dynamics Lab. Setting up Small Grid Testbed & Using Globus, MPICH-G2 Korea Advanced Institute of Science and Technology Div. of Aerospace Engineering Dehee Kim

2 Contents Introduction to GT 2.0 and MPICH-G2 How to install Globus
CFD Lab. Grid Testbed Numerical Test on Testbed About Network Bandwidth Concluding Remarks

3 GT 2.0 Globus Toolkit 2.0 Major improvements over the Globus Toolkit and releases Data Grid Components MDS Components GRAM Components Packaging Technology Security Components Various changes for supporting MPICH-G2

4 MPICH-G2 ● What is MPICH-G2?
grid-enabled implementation of the MPI v1.1 standard converts data in messages sent between machines of different architectures supports multiprotocol communication ● How does MPICH-G2 differ from MPICH-G? Increased bandwidth Reduced latency for intra-machine messaging Increased latency for inter-machine (TCP) messaging

5 Construction of Grid Testbed
Installation Procedure 1. Set up small PC cluster system - rsh, NFS, automount, ntp, … - back end nodes with hard disk 2. Install F77, F90 compiler - Absoft F90, pgf90, etc. - Set environment variables and path

6 Construction of Grid Testbed
If you does not install a Fortran compiler before the installation of GT 2.0, you will see following message ….. Checking for minix/config.h Checking for volatile… yes Running device-specific setup program *#Globus device overrode C compiler setting *#F90 compiler not present; disabling F90 support Disabling long double(not supported by Globus data Conversion library) Checking whether cross-compiling…

7 Construction of Grid Testbed
3. Install Job Queuing System - PBS, CONDOR, LSF, etc. - 2 rpm files(for PBS) - Front end : /usr/spool/pbs/server_priv/nodes - Back end : /usr/spool/pbs/mom_priv/config $clienthost /usr/spool/pbs/default_server 4. Install GT 2.0 - Using simple CA 5. Install MPICH-G2 ./configure –device=globus2 \ -fc=/opt/absoft/bin/f77 \ -f90=/opt/absoft/bin/f90 \ : -flavor=gcc32dbg \ --prefix=/usr/local/mpich g2

8 Construction of Grid Testbed
A trial and error /etc/xinetd.d/globus-gatekeeper Service globus-gatekeeper { socket_type = stream ….. } socket_type=stream O.K. Parsing Error!

9 Construction of Grid Testbed
6. Modify some scripts if necessary - If various environment variables related with jobmanager were not set, set the variables in some files $GLOBUS_LOCATION/libexec/ $GLOBUS_LOCATION/libexec/ ….. GLOBUS_GRAM_JOB_MANAGER_MPIRUN=/usr/local/mpich g2/bin/mpirun GLOBUS_GRAM_JOB_MANAGER_QDEL=/usr/local/bin/qdel GLOBUS_GRAM_JOB_MANAGER_QSTAT=/usr/local/bin/qstat GLOBUS_GRAM_JOB_MANAGER_QSUB=/usr/local/bin/qsub GLOBUS_GRAM_JOB_MANAGER_QSELECT=/usr/local/bin/qselect …… 7. Configure the firewall policy for Globus

10 CFD Lab. Grid Testbed OS-Linux 2.4.x, 2.2.x KAIST CFD Lab.
– 1 Front-end, 4-execution nodes(1.8GHz, 512M RAM) KISTI supercomputing center – 1 Front-end, 4-execution nodes(450MHz, 256M RAM) Globus Toolkit 2.0, MPICH-G2, ABSOFT F90 Job Scheduler – Portable Batch System

11 CFD Lab. Grid Testbed

12 Numerical Test on Testbed
Design Optimization : 2-D design 2-D adjoint sensitivity analysis 2-D airfoil design Design for drag minimization of RAE 2822 airfoil Grid system : 383 x 65 C type Flow conditions : M=0.729, AoA=2.31o, Re = 6.5 x 106 10 Hicks-Henne functions Pressure distribution airfoil before and after design

13 Numerical Test on Testbed
Design Optimization : 3-D design 3-D adjoint sensitivity analysis 3-D wing design Design for drag minimization of ONERA M6 wing Grid system : 193 x 49 x 33 C-O type Flow conditions : M=0.84, AoA=3.06o, Re = 11.7 x 106 50 Hicks-Henne functions ONERA M6 Designed wing

14 Computation Time Design Optimization : Computation Time
Flow analysis around 2-D airfoil and design optimization device resource Flow analysis (ch_p4) Design (ch_p4) Flow analysis (globus2) I 158.0 467.7 158.7 478.9 II 388.8 1166.7 392.9 1170.4 III 410.9 1432.1 I : Pentium GHz CPU, 4 nodes, 512M RAM II : Pentium MHz CPU, 4 nodes, 256M RAM III : I & II Flow analysis around 3-D wing and design optimization(case III) Flow analysis : Design :

15 DFVLR Axial Fan – Dr. J. S. Yoon
3-D Compressible Navier-Stokes Solver k-ω Turbulent Modeling 3 Stage Runge-Kutta Time Marching & Central Scheme 28 Blade(45*19*19) MPICH-G Surface Pressure Contours

16 Computation Time at various Network Bandwidth
Com. Env. Result 1CPU Parallel (2Node) Computation based on Globus(Remote 2 Node, Mbps) 2 5 10 15 25 45 155 Time(Sec) 3773 1998 3367 2488 2235 2147 2078 2047 2038 T/Tp 0.94 1 1.69 1.25 1.12 1.08 1.04 1.02

17 Varying efficiency on time of day Variation of computation time
PC Cluster front ↔ IBM SP2 1:1 CPU, 200 iterations Seriously varying efficiency on time of day Need for proper QoS and CPU Reservation Variation of computation time

18 Concluding Remarks Setup of small testbed Test for design applications
Need for obtaining vast computing resources Implementation to diskless cluster - public IP, private IP(e.g. pacx-mpi)

Download ppt "Setting up Small Grid Testbed"

Similar presentations

Ads by Google