Monte Carlo simulation for radiotherapy in a distributed computing environment S. Chauvie 2,3, S. Guatelli 2, A. Mantero 2, J. Moscicki 1, M.G. Pia 2 CERN.

Slides:



Advertisements
Similar presentations
The GATE-LAB system Sorina Camarasu-Pop, Pierre Gueth, Tristan Glatard, Rafael Silva, David Sarrut VIP Workshop December 2012.
Advertisements

Maria Grazia Pia Simulation in a Distributed Computing Environment Simulation in a Distributed Computing Environment S. Guatelli 1, A. Mantero 1, P. Mendez.
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Dinker Batra CLUSTERING Categories of Clusters. Dinker Batra Introduction A computer cluster is a group of linked computers, working together closely.
Other GEANT4 capabilities Event biasing Parameterisation (fast simulation) Persistency Parallelisation and integration in a distributed computing environment.
A general purpose dosimetric system for brachytherapy
DIANE Project Seminar on Innovative Detectors, Siena Oct 2002 Distributed Computing in Physics Parallel Geant4 Simulation in Medical.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
1 Introduction to Load Balancing: l Definition of Distributed systems. Collection of independent loosely coupled computing resources. l Load Balancing.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Workload Management Massimo Sgaravatto INFN Padova.
The brachytherapy advanced example Susanna Guatelli (CERN/INFN)
DIANE Overview Germán Carrera, Alfredo Solano (CNB/CSIC) EMBRACE COURSE Monday 19th of February to Friday 23th. CNB-CSIC Madrid.
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Grid Data Management A network of computers forming prototype grids currently operate across Britain and the rest of the world, working on the data challenges.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
DIANE Project CHEP 03 DIANE Distributed Analysis Environment for semi- interactive simulation and analysis in Physics Jakub T. Moscicki,
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
Chapter 4 Realtime Widely Distributed Instrumention System.
9 September 2008CIS 340 # 1 Topics reviewTo review the communication needs to support the architectures variety of approachesTo examine the variety of.
Grid Workload Management Massimo Sgaravatto INFN Padova.
Frontiers in Massive Data Analysis Chapter 3.  Difficult to include data from multiple sources  Each organization develops a unique way of representing.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Tools for collaboration How to share your duck tales…
Technological Transfer from HEP to Medical Physics How precise Brachytherapy MonteCarlo simulations can be applied in Clinics Reality Problem: How to achieve.
PARALLEL COMPUTING overview What is Parallel Computing? Traditionally, software has been written for serial computation: To be run on a single computer.
CLUSTER COMPUTING TECHNOLOGY BY-1.SACHIN YADAV 2.MADHAV SHINDE SECTION-3.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
9 Systems Analysis and Design in a Changing World, Fourth Edition.
ATLAS Grid Data Processing: system evolution and scalability D Golubkov, B Kersevan, A Klimentov, A Minaenko, P Nevski, A Vaniachine and R Walker for the.
S. Guatelli, A. Mantero, J. Moscicki, M. G. Pia Geant4 medical simulations in a distributed computing environment 4th Workshop on Geant4 Bio-medical Developments.
1 Cloud Computing Advisor : Cho-Chin Lin Student : Chien-Chen Lai.
IRCC & Mauriziano Hospital & INFN & S Croce e Carle Hospital
7. Grid Computing Systems and Resource Management
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Data Communications and Networks Chapter 9 – Distributed Systems ICT-BVF8.1- Data Communications and Network Trainer: Dr. Abbes Sebihi.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Susanna Guatelli Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Background Computer System Architectures Computer System Software.
Alien and GSI Marian Ivanov. Outlook GSI experience Alien experience Proposals for further improvement.
Next Generation of Apache Hadoop MapReduce Owen
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
CMS Experience with the Common Analysis Framework I. Fisk & M. Girone Experience in CMS with the Common Analysis Framework Ian Fisk & Maria Girone 1.
Fermilab Scientific Computing Division Fermi National Accelerator Laboratory, Batavia, Illinois, USA. Off-the-Shelf Hardware and Software DAQ Performance.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
DIRAC: Workload Management System Garonne Vincent, Tsaregorodtsev Andrei, Centre de Physique des Particules de Marseille Stockes-rees Ian, University of.
Workload Management Workpackage
Clouds , Grids and Clusters
F. Foppiano3, S. Guatelli2, J. Moscicki1, M.G. Pia2 CERN1 INFN Genova2
Design rationale and status of the org.glite.overlay component
Simulation in a Distributed Computing Environment
Grid Computing.
WP1 activity, achievements and plans
F. Foppiano, S. Guatelli, B. Mascialino, M. G. Pia, M. Piergentili
Support for ”interactive batch”
Other GEANT4 capabilities
Simulation in a Distributed Computing Environment
Simulation in a Distributed Computing Environment
Distributed Simulation with Geant4
MapReduce: Simplified Data Processing on Large Clusters
Presentation transcript:

Monte Carlo simulation for radiotherapy in a distributed computing environment S. Chauvie 2,3, S. Guatelli 2, A. Mantero 2, J. Moscicki 1, M.G. Pia 2 CERN 1 INFN 2 S. Croce e Carle Hospital Cuneo 3 Monte Carlo April 2005 Chattanooga, TN, USA

Monte Carlo methods in radiotherapy Monte Carlo methods have been explored for years as a tool for precise dosimetry, in alternative to analytical methods de facto, Monte Carlo simulation is not used in clinical practice (only side studies) speed The major limiting factor is the speed

The reality Treatment planning is performed by means of commercial software The software calculates the dose distribution delivered to the patient Open issues Disadvantages Advantages analytical methods Commercial systems are based on analytical methods Fails in calculate dose in heterogeneities and for small or complex field Quick response specific to one radiotherapic technique Each treatment planning software is specific to one radiotherapic technique

Project Develop a dosimetric system for radiotherapy treatments based on Monte Carlo methodsProject Geant4 as Simulation Toolkit Parallelisation Access to distributed computing resources Calculation precision Quick response

Pilot project: distributed simulation for brachytherapy Explore Geant4-based Monte Carlo simulations in a distributed computing environment –Parallel execution in a local PC farm –Geographically distributed execution on the GRID Pilot project based on an existing simulation for brachytherapy Focus on architectural design –Transparent execution on a single machine, in parallel on a local farm or on the GRID Preliminary evaluation of performance Application to other radiotherapy simulations currently in progress

Brachytherapy Simulation of the energy deposited by a radioactive source in a phantom Requirement from clinical practice: real time response Bebig Isoseed I-125 source source Plan containing the radioactive source Dose Distribution Talk: “A general purpose dosimetric system for brachytherapy”, 20 th April, MC 2005, Room 5

Performance in sequential mode Endocavitary brachytherapy 1M events 61 minutes Interstitial brachytherapy 1M events 67 minutes Superficial brachytherapy 1M events 65 minutes on an “average” PIII machine Monte Carlo simulation is not practically conceivable for clinical application, even if more precise

Speed adequate for clinic use Transparent configuration in sequential or parallel mode Transparent access to the GRID through an intermediate software layer Parallelisation Access to distributed computing resources

Access to distributed computing speed OK but expensive hardware investment + maintenance Hospital LAN SWITCHSWITCH Node01 Node02 Node03 Node04 IMRT Geant4 Simulation and Anaphe analysis on a dedicated Beowulf Cluster S. Chauvie et al., IRCC Torino, Siena 2002

Access to distributed computing Alternative strategy DIANE ParallelisationAccess to the GRID Transparent access to a distributed computing environment

Active Workflow Framework for Parallel Jobs Applications run inside an Active Workflow Framework For applications: –underlying environment is transparent –code changes to use the framework are minimal The Framework provides: –Automatic Communication and Synchronization of tasks –Error recovery –Optimization

DIANE DIstributed ANalysis Environment prototype for an intermediate layer between applications and the GRID Parallel cluster processing – make fine tuning and customisation easy – transparently using GRID technology – application independent Hide complex details of underlying technology Developed by J. Moscicki, CERN

DIANE architecture Master-Worker model Parallel execution of independent tasks Very typical in many scientific applications Usually applied in local clusters R&D in progress for Large Scale Master- Worker Computing

Master - Worker Computing Workers are started up and register to Master Client connects to Master and starts up the job Master controls the execution, dispatches tasks to Workers and combines the result Client receives notifications about the current status of the job and collects the final result

Running in a distributed environment Not affecting the original code of application same code – standalone and distributed case is the same code Good separation of the subsystems – the application does not need to know that it runs in distributed environment – the distributed framework (DIANE) does not need to care about what actions an application performs internally The application developer is shielded from the complexity of underlying technology via DIANE

Distributed environments Different distributed environments: local computing farm GRID

Parallel mode: local cluster / GRID Both applications have the same computing model –a job consists of a number of independent tasks which may be executed in parallel –result of each task is a small data packet (few kilobytes), which is merged as the job runs In a cluster: –computing resources are used for parallel execution –user connects to a possibly remote cluster –input data for the job must be available on the site –typically there is a shared file system and a queuing system –network is fast GRID computing uses resources from multiple computing centres –typically there is no shared file system –(parts of) input data must be replicated in remote sites –network connection is slower than within a cluster

Development costs Strategy to minimise the cost of migrating a Geant4 simulation to a distributed environment for users DIANE Active Workflow framework –provides automatic communication/synchronization mechanisms –application is “glued” to the framework using a small Python module; in most cases no code changes to the original application are required –load balancing and error recovery policies may be plugged in form of simple python functions Transparent adaptation for Clusters/GRIDs, shared/local file systems, shared/private queues Cost in the runtime phase: –near zero (except for loading networking libraries for the first time) Development/modification of application code –original source code unmodified –addition of an interface class which binds together application and M-W framework

Interfacing a Geant4 simulation to DIANE UML Deployment Diagram for Geant4 applications

Practical example: G4 simulation with analysis Each task produces a file with histograms The job result is the sum of histograms produced by tasks Master-worker model – client starts a job – workers perform tasks and produce histograms – master integrates the results Distributed Processing for Geant4 Applications –task = N events –job = M tasks –tasks may be executed in parallel –tasks produce histograms/ntuples –task output is automatically combined (add histograms, append ntuples) Master-Worker Model –Master steers the execution of job, automatically splits the job and merges the results –Worker initializes the Geant4 application and executes macros –Client gets the results

DIANE Prototype and Testing Scalability tests –70 worker nodes –140 milion Geant 4 events

Performance : parallel mode 1M events 4 minutes 34’’ 5M events 4 minutes 36’’ 1M events 4 minutes 25’’ on up to 50 workers, LSF at CERN, PIII machine, MHz Performance adequate for clinical application, but… it is not realistic to expect any hospital to own and maintain a PC farm Endocavitary brachytherapy Interstitial brachytherapy Superficial brachytherapy preliminary: further optimisation in progress

Parallel mode: distributed resources Distributed Geant 4 Simulation: DIANE framework and generic GRID middleware

Grid Wave of interest in grid technology as a basis for “revolution” in e-Science and e-Commerce An infrastructure and standard interfaces capable of providing transparent access to geographically distributed computing power and storage space in a uniform way Ian Foster and Carl Kesselman's book: ”A computational Grid is a hardware and software infrastructure that provides dependable, consistent, pervasive and inexpensive access to high-end computational capabilities”". US projects European projects Many GRID R&D projects, many related to HEP

Large distributed computing resource

Running on the GRID Via DIANE Same application code as running on a sequential machine or on a dedicated cluster –completely transparent to the user A hospital is not required to own and maintain extensive computing resources to exploit the scientific advantages of Monte Carlo simulation for radiotherapy Any hospital even small ones, or in less wealthy countries, that cannot afford expensive commercial software systems – may have access to advanced software technologies and tools for radiotherapy

Traceback from a run on CrossGrid testbed Current #Grid setup (computing elements): 5000 events, 2 workers, 10 tasks (500 events each) - aocegrid.uab.es:2119/jobmanager-pbs-workq - bee001.ific.uv.es:2119/jobmanager-pbs-qgrid - cgnode00.di.uoa.gr:2119/jobmanager-pbs-workq - cms.fuw.edu.pl:2119/jobmanager-pbs-workq - grid01.physics.auth.gr:2119/jobmanager-pbs-workq - xg001.inp.demokritos.gr:2119/jobmanager-pbs-workq - xgrid.icm.edu.pl:2119/jobmanager-pbs-workq - zeus24.cyf-kr.edu.pl:2119/jobmanager-pbs-infinite - zeus24.cyf-kr.edu.pl:2119/jobmanager-pbs-long - zeus24.cyf-kr.edu.pl:2119/jobmanager-pbs-medium - zeus24.cyf-kr.edu.pl:2119/jobmanager-pbs-short - ce01.lip.pt:2119/jobmanager-pbs-qgrid Spain Poland Greece Portugal Resource broker running in Portugal matchmaking CrossGrid computing elements

Study in progress Capability of transparent execution of the radiotherapy simulation on the GRID has been demonstrated Quantitative evaluation of performance speed and stability currently in progress A comprehensive study will be submitted for publication in the coming weeks Optimisation of load balancing, error handling and other issues concerning access to distributed resources currently under study

Application to IMRT simulations Determine the dose distribution in a phantom generated by the head of a linear accelerator Requirement from clinical practice: fast response Without parallelisation:  events  100 CPU days on Pentium IV 3 GHz Talk: “ Geant4 Simulation of an Accelerator Head for Intensity Modulated RadioTherapy ”, 19 th April, MC 2005, Room 6 Lateral profile 6MV, 5x5cm field, 15mm depth

Conclusions Fast performance –parallel processing Access to geographically distributed computing resources –GRID Demonstrated with Geant4 simulation applications + DIANE More information – cern.ch/diane – – – aida.freehep.org