On-Demand Lake Circulation Modeling Chin Wu Nobuaki Kimura Environmental Fluid Mechanics Laboratory University of Wisconsin Paul Hanson Tim Kratz Tim Meinke.

Slides:



Advertisements
Similar presentations
Network II.5 simulator ..
Advertisements

Vega: A Flexible Data Model for Environmental Time Series Data L. A. Winslow, B. J. Benson, K. E. Chiu, P. C. Hanson, T. K. Kratz.
Event detection in ecological sensor networks Owen Langman Center for Limnology University of Wisconsin - Madison GLEON 7 Sept. 29, 2008 Norrtälje, Sweden.
Legacy code support for commercial production Grids G.Terstyanszky, T. Kiss, T. Delaitre, S. Winter School of Informatics, University.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Wofgang Thöne, Institute For Scientific Computing – EGEE-Meeting August 2004 Welcome to the User.
1 Web Servers / Deployment Alastair Dawes Original by Bhupinder Reehal.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Web Servers How do our requests for resources on the Internet get handled? Can they be located anywhere? Global?
GLEON Data Management Luke Winslow PASEO 3/18/09.
Homework 2 In the docs folder of your Berkeley DB, have a careful look at documentation on how to configure BDB in main memory. In the docs folder of your.
1 Software Testing and Quality Assurance Lecture 32 – SWE 205 Course Objective: Basics of Programming Languages & Software Construction Techniques.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
SCADA and Telemetry Presented By:.
A Circulation Model to Investigate the Movement of Wastes from an Open Ocean Aquaculture Site David W. Fredriksson U. S. Naval Academy NOAA Research -
Overview of the ODP Data Provider Sergey Sukhonosov National Oceanographic Data Centre, Russia Expert training on the Ocean Data Portal technology, Buenos.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
ViciDocs for BPO Companies Creating Info repositories from documents.
INTRODUCTION TO WEB DATABASE PROGRAMMING
Apache Airavata GSOC Knowledge and Expertise Computational Resources Scientific Instruments Algorithms and Models Archived Data and Metadata Advanced.
6/1/2001 Supplementing Aleph Reports Using The Crystal Reports Web Component Server Presented by Bob Gerrity Head.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Discussion and conclusion The OGC SOS describes a global standard for storing and recalling sensor data and the associated metadata. The standard covers.
Page 1 CONSULTANCY AND RESEARCH IN AQUACULTURE AND THE AQUATIC ENVIRONMENT A Company in the NIVA-group Methodology for Environmental monitoring of aquaculture.
Introduction to Information Retrieval CS 5604: Information Storage and Retrieval ProjCINETViz by Maksudul Alam, S M Arifuzzaman, and Md Hasanuzzaman Bhuiyan.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
DISTRIBUTED COMPUTING
RDFS Rapid Deployment Forecast System Visit at: Registration required.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Computer and Information Science Ch1.3 Computer Networking Ch1.3 Computer Networking Chapter 1.
1 Overview of the Application Hosting Environment Stefan Zasada University College London.
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
Research and Educational Networking and Cyberinfrastructure Russ Hobby, Internet2 Dan Updegrove, NLR University of Kentucky CI Days 22 February 2010.
Wenjing Wu Computer Center, Institute of High Energy Physics Chinese Academy of Sciences, Beijing BOINC workshop 2013.
OOI Annual Review Year 2 May 16 – 20, 2011 Ocean Observatories Initiative Surface and Subsurface Mooring Telemetry Inductive and acoustic technology and.
CONRAD BLUCHER INSTITUTE ACTIVITIES SUPPORTING TEXAS PORTS AND WATERWAYS OPERATIONS Two Inter-related Services to the Port Community: 1. The Texas Coastal.
Integrated Grid workflow for mesoscale weather modeling and visualization Zhizhin, M., A. Polyakov, D. Medvedev, A. Poyda, S. Berezin Space Research Institute.
PROCESSED RADAR DATA INTEGRATION WITH SOCIAL NETWORKING SITES FOR POLAR EDUCATION Jeffrey A. Wood April 19, 2010 A Thesis submitted to the Graduate Faculty.
Database Design and Management CPTG /23/2015Chapter 12 of 38 Functions of a Database Store data Store data School: student records, class schedules,
Experiences with the Globus Toolkit on AIX and deploying the Large Scale Air Pollution Model as a grid service Ashish Thandavan Advanced Computing and.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
Page 1© Crown copyright 2004 Development of a Ground Based GPS Network for the Near Real Time Measurement of Integrated Water Vapour (IWV) Jonathan Jones.
IODE Ocean Data Portal - ODP  The objective of the IODE Ocean Data Portal (ODP) is to facilitate and promote the exchange and dissemination of marine.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
GEON2 and OpenEarth Framework (OEF) Bradley Wallet School of Geology and Geophysics, University of Oklahoma
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
What is Web Information retrieval from web Search Engine Web Crawler Web crawler policies Conclusion How does a web crawler work Synchronization Algorithms.
Internet Applications (Cont’d) Basic Internet Applications – World Wide Web (WWW) Browser Architecture Static Documents Dynamic Documents Active Documents.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
RTFDDA Engineering System Architecture Input Data Cycle Timing.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
MGRID Architecture Andy Adamson Center for Information Technology Integration University of Michigan, USA.
Distributed Data Servers and Web Interface in the Climate Data Portal Willa H. Zhu Joint Institute for the Study of Ocean and Atmosphere University of.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
Cyberinfrastructure Overview of Demos Townsville, AU 28 – 31 March 2006 CREON/GLEON.
Detecting sensor malfunctions in ecological sensor networks Owen Langman Center for Limnology University of Wisconsin - Madison EIM Conference Sept. 11,
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES Author etc Alarm framework requirements Andrea Sciabà Tony Wildish.
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
SCADA Supervisory Control And Data Acquisition Pantech Solutions Here is the key to learn more.
Sensor Calibration Automation Demonstration Presenters: Barbara Benson and David Balsiger (University of Wisconsin) Collaborators: Laurence Choi, Yu Hen.
MODELING CLIMATE CHANGE EFFECTS ON LAKES USING DISTRIBUTED COMPUTING
Lecture 8 Database Implementation
Module development was supported by NSF DEB and ACI
Lecture 1: Multi-tier Architecture Overview
Laura Bright David Maier Portland State University
Presentation transcript:

On-Demand Lake Circulation Modeling Chin Wu Nobuaki Kimura Environmental Fluid Mechanics Laboratory University of Wisconsin Paul Hanson Tim Kratz Tim Meinke Luke Winslow Center for Limnology, Trout Lake Station University of Wisconsin Kenneth Chiu Yinfei Pan Grid Computing Research Laboratory SUNY-Binghamton

On-Demand Circulation Modeling Hindcast: –Enter a date/time range. –Met data extracted from CFL database. –Model launched. –Results displayed.

Goals Develop a reusable solution for high-quality lake circulation modeling with validation. Provide historical results on demand for analysis. Support data assimilation, coupled models, etc. Support pluggable models. Support out-sourcing most ICT deployment to remote site. –Lake sites do not have the expertise, so they just provide the data.

Data Sources Model inputs –Meterological data –Water temperature etc. Model outputs (validation or assimilation) –Acoustic Doppler Current Profiler (ADCP) Provides 3D flow vectors of a vertical column of water. –One vector for each depth. –Measured waves Subsurface, via high-frequence temperature fluctuations. Surface

Challenges Cyberinfrastructure is a social/institutional problem as much as a technical problem. –E.g., technically, interoperability is a solved problem. Not a human-computer-interaction problem, but a human-institution-computer interaction challenge. (“Institution” in the broad sense.) –Two-way interaction, institutions can themselves be changed. Iterative –Develop rapid prototype, get feedback, repeat –Collaborative –Avoid over-engineering, be pragmatic.

Instrumented buoys and locations ADCP Thermistor chain Dissolve Oxygen sensor Buoy Map of Lake Trout RUSS deep ADCP

FLOW (Data retrieve and conversion) DataflowCAST HIND(HindCasting)NOW NowCasting) (NowCasting)FORE (Forecasting)extension CIRC (Circulation module) ANT (Animation of temp. & Velocity) COMP (Comparison between observation and prediction) Pre-processing Post-processing ADCP (Observed Velocity)

Workflows in WB-CAST Trout Lake Ethernet Radio Serial2Ethernet IP addr. + port # ADCP Unit Data- logger Ethernet Radio Trout Lake Station ADCP computer acqui r Oracle ADCP Binary ADCP Meta Madison CFL CFL SU N ADCP Binary Mod el results Madison EFM EFM Logger Data Logger Data ADCP Binary ADCP Binary Logger Data

Pre- & Post-processing of the circulation model Pre- & Post-processing of the circulation model Matlab-1 Conversion input files Model Compute Velocity W-temp, etc… Output Binary Matlab-2 Creating plot files extract_ adcp ADCP Binary Data base DB-Badger Extract data Logger- net data Linux computer FLOWCIRC ANTCOMP ADCP CFL

The way to drive the 3D circulation model 3DCirculationModel Input data Initial conditions bathymetry, water temprature, velocity surface elevation) Temperature profile Velocity profile Forcing in time serial (MED) wind, wind, Heat flux, …etc. Heat flux, …etc. Characteristics - Non-hydrostatic pressure - Bottom partial cell - Finite volume method OUTPUT

Computation using the Model WI-CAST NOW Time scale (hourly simulation) 9:00 10:00 11:00 Run1 Run2 Run3 Output1Output2Output3 12:00 WI-CAST HIND Time scale (3 hour simulation) 9:00 10:00 11:00 Only one run 12:00 Integration and animation Output1Output2Output3 Integration and animation An output is produced at each hour

Application Server Local Database Server (Mysql) Web Server (tomcat) Request Dir. Web Browser (Firefox, IE, Netscape …) Remote Database Server (Oracle) Internet Firewall Check Requests Find requests (Do Modeling …) Modeling & animation Finished (Update Database) Data acquisition

Client web browser initiates an Modeling & Animation (M&A) Request Server side will generate a request file corresponding to this request (this project using Java Server Pages), putting it into a request directory (Works like a request queue), then update one request record into the local database server. It’s default status is “pending” Client side hold the browser, waiting for M&A finish Server side, a daemon process called “checkrequest” will check if the request directory has requests If has, the request will be processed by the Application Sever (a logical sever here), this include “data acquisition” from a remote database in EFM which stored all the scientific data collected, then running the Modeling process, and finally render he output animation files When M&A finished in Application Server, it will update the request record in Local database server to have a “finish” status The client browser at last will find its request had been fulfilled. And goes into animation shows stage.

Future Work Gather feedback! Package as a toolkit? Service? Opal? Parallelization, increased resolution Job scheduling, Pragma integration? Data assimilation using MPC? Coupled models –Biological, chemical –Fluid-surface interactions Real-time wave reconstruction from captured video

Credits Chin Wu, Nobuaki Kimura (EFM-UWI) –Modeling, output components Paul Hanson, Luke Winslow (CFL-UWI) –Data extraction, processing Tim Kratz, Tim Meinke (TLS-UWI) –Equipment, deployment, sensor network Yinfei Pan, Kenneth Chiu (SUNY-Binghamton) –ADCP acquisition/management, job launching, monitoring, integration, web development

Acknowledgements We’d like to thank the generous support of Moore Foundation, the NSF LTER program, and NSF awards DBI , IIS , CNS , OCI

Application Server Overload Now One day’s prediction needs about one hour for a 2.2GHz Two-CPU machine to do computing … Thus, our further steps will be: –Make the computing available on a whole cluster –Then, make it works as web based computing (deploy web services ) The key problem here –The model computation continuous generate time sequence related meta & data files –It’s hard to make all the computing procedures and functions parallel immediately –Idea (store the modeling output whenever possible) On one cluster, the “checkrequst” daemon works as a job scheduler. On web, employ distributed data hash, also implementing four types of services: job schedule service, data locating service, data acquisition service, and data storage service.

Job Schedule Server Cluster Parallel Analysis Server (with Time Vector Sequence Database) Web Server Requests Central Database (Meta, Data and Animation) Getting the time range available during the time range requested Select a time vector sequence based on the minimum total expected time Form job division and schedule jobs onto computers in cluster Store back each time vector’s delay time corresponding to a time range and update the expected time Getting data from central database Workstation PS: Time Vector here is a set with many (procedure id, expected time) pairs. This procedures are supposed to do at the same time.