Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for.

Slides:



Advertisements
Similar presentations
ESLEA and HEPs Work on UKLight Network. ESLEA Exploitation of Switched Lightpaths in E- sciences Applications Exploitation of Switched Lightpaths in E-
Advertisements

Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
August 10, Circuit TCP (CTCP) Helali Bhuiyan
ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
Restricted Slow-Start for TCP William Allcock 1,2, Sanjay Hegde 3 and Rajkumar Kettimuthu 1,2 1 Argonne National Laboratory 2 The University of Chicago.
Iperf Tutorial Jon Dugan Summer JointTechs 2010, Columbus, OH.
TCP loss sensitivity analysis ADAM KRAJEWSKI, IT-CS-CE.
GridPP meeting Feb 03 R. Hughes-Jones Manchester WP7 Networking Richard Hughes-Jones.
ISCSI Performance in Integrated LAN/SAN Environment Li Yin U.C. Berkeley.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Grid simulation (AliEn) Network data transfer model Eugen Mudnić Technical university Split -FESB.
Grid Computing for High Energy Physics in Japan Hiroyuki Matsunaga International Center for Elementary Particle Physics (ICEPP), The University of Tokyo.
KEK Network Qi Fazhi KEK SW L2/L3 Switch for outside connections Central L2/L3 Switch A Netscreen Firewall Super Sinet Router 10GbE 2 x GbE IDS.
CMS Data Transfer Challenges LHCOPN-LHCONE meeting Michigan, Sept 15/16th, 2014 Azher Mughal Caltech.
1 A Basic R&D for an Analysis Framework Distributed on Wide Area Network Hiroshi Sakamoto International Center for Elementary Particle Physics (ICEPP),
TNC 2007 Bandwidth-on-demand to reach the optimal throughput of media Brecht Vermeulen Stijn Eeckhaut, Stijn De Smet, Bruno Volckaert, Joachim Vermeir,
Site report: Tokyo Tomoaki Nakamura ICEPP, The University of Tokyo 2014/12/10Tomoaki Nakamura1.
IRODS performance test and SRB system at KEK Yoshimi KEK Building data grids with iRODS 27 May 2008.
Large File Transfer on 20,000 km - Between Korea and Switzerland Yusung Kim, Daewon Kim, Joonbok Lee, Kilnam Chon
Profiling Grid Data Transfer Protocols and Servers George Kola, Tevfik Kosar and Miron Livny University of Wisconsin-Madison USA.
Maximizing End-to-End Network Performance Thomas Hacker University of Michigan October 26, 2001.
Grid Applications for High Energy Physics and Interoperability Dominique Boutigny CC-IN2P3 June 24, 2006 Centre de Calcul de l’IN2P3 et du DAPNIA.
High Bandwidth Data Acquisition and Network Streaming in VLBI Jan Wagner, Guifré Molera et al. TKK / Metsähovi Radio Observatory.
Network Tests at CHEP K. Kwon, D. Han, K. Cho, J.S. Suh, D. Son Center for High Energy Physics, KNU, Korea H. Park Supercomputing Center, KISTI, Korea.
Current Testbed : 100 GE 2 sites (NERSC, ANL) with 3 nodes each. Each node with 4 x 10 GE NICs Measure various overheads from protocols and file sizes.
Site Report: Tokyo Tomoaki Nakamura ICEPP, The University of Tokyo 2013/12/13Tomoaki Nakamura ICEPP, UTokyo1.
Wide Area Network Access to CMS Data Using the Lustre Filesystem J. L. Rodriguez †, P. Avery*, T. Brody †, D. Bourilkov *, Y.Fu *, B. Kim *, C. Prescott.
RAC parameter tuning for remote access Carlos Fernando Gamboa, Brookhaven National Lab, US Frederick Luehring, Indiana University, US Distributed Database.
Grid Lab About the need of 3 Tier storage 5/22/121CHEP 2012, The need of 3 Tier storage Dmitri Ozerov Patrick Fuhrmann CHEP 2012, NYC, May 22, 2012 Grid.
High TCP performance over wide area networks Arlington, VA May 8, 2002 Sylvain Ravot CalTech HENP Working Group.
BNL Service Challenge 3 Site Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
High-speed TCP  FAST TCP: motivation, architecture, algorithms, performance (by Cheng Jin, David X. Wei and Steven H. Low)  Modifying TCP's Congestion.
HighSpeed TCP for High Bandwidth-Delay Product Networks Raj Kettimuthu.
Resource Predictors in HEP Applications John Huth, Harvard Sebastian Grinstein, Harvard Peter Hurst, Harvard Jennifer M. Schopf, ANL/NeSC.
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
Analysis in STEP09 at TOKYO Hiroyuki Matsunaga University of Tokyo WLCG STEP'09 Post-Mortem Workshop.
Masaki Hirabaru NICT Koganei 3rd e-VLBI Workshop October 6, 2004 Makuhari, Japan Performance Measurement on Large Bandwidth-Delay Product.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Status SC3 SARA/Nikhef 20 juli Status & results SC3 throughput phase SARA/Nikhef Mark van de Sanden.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
Performance Engineering E2EpiPEs and FastTCP Internet2 member meeting - Indianapolis World Telecom Geneva October 15, 2003
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
Academic and Research Network of Slovenia Static IPv6 in GÉANT and Internet2 LSR a joint effort by ARNES, DANTE, RedIRIS and Juniper TF-NGN meeting Budapest,
Data Transfer Service Challenge Infrastructure Ian Bird GDB 12 th January 2005.
TCP transfers over high latency/bandwidth networks & Grid DT Measurements session PFLDnet February 3- 4, 2003 CERN, Geneva, Switzerland Sylvain Ravot
1.3 ON ENHANCING GridFTP AND GPFS PERFORMANCES A. Cavalli, C. Ciocca, L. dell’Agnello, T. Ferrari, D. Gregori, B. Martelli, A. Prosperini, P. Ricci, E.
Status of Tokyo LCG tier-2 center for atlas / H. Sakamoto / ISGC07 Status of Tokyo LCG Tier 2 Center for ATLAS Hiroshi Sakamoto International Center for.
Final EU Review - 24/03/2004 DataTAG is a project funded by the European Commission under contract IST Richard Hughes-Jones The University of.
Site Throughput Review and Issues Shawn McKee/University of Michigan US ATLAS Tier2/Tier3 Workshop May 27 th, 2008.
B ENCHMARK ON D ELL 2950+MD1000 ATLAS Tier2/Tier3 workshop Wenjing wu AGLT2 / University of Michigan 2008/05/27.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
An Analysis of AIMD Algorithm with Decreasing Increases Yunhong Gu, Xinwei Hong, and Robert L. Grossman National Center for Data Mining.
Hiroyuki Matsunaga (Some materials were provided by Go Iwai) Computing Research Center, KEK Lyon, March
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
Recent experience with PCI-X 2.0 and PCI-E network interfaces and emerging server systems Yang Xia Caltech US LHC Network Working Group October 23, 2006.
G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th The Naples' testbed for the SuperB computing model: first tests G. Russo, D. Del Prete, S.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
The Beijing Tier 2: status and plans
Status report on LHC_2: ATLAS computing
Status Report on LHC_2 : ATLAS computing
TCP loss sensitivity analysis
Service Challenge 3 CERN
The transfer performance of iRODS between CC-IN2P3 and KEK
Project: COMP_01 R&D for ATLAS Grid computing
Wide Area Networking at SLAC, Feb ‘03
FTS Issue in Beijing Erming PEI 2010/06/18.
When to use and when not to use BBR:
Presentation transcript:

Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for Elementary Particle Physics (ICEPP), The University of Tokyo Introduction A WLCG Tier-2 site is running at ICEPP, the University of Tokyo in Japan. This site receives a large amount of data of the ATLAS experiment mainly from the Tier-1 site in Lyon, France. It is important to transfer data in a timely manner to facilitate data analysis at the ICEPP site, but it is not so easy to exploit the available bandwidth for the data transfer due to the large round trip time between the two sites. We performed data transfer tests in various configurations in order to understand the performance of the production system and also possible improvement in future. The tests were carried out by using gridFTP, which is used in the production system for the data transfer over the wide area network (WAN), with test nodes between ICEPP and CERN (near Lyon). Memory-to-memory tests were carried out first by using iperf for varying numbers of streams and TCP window sizes. On sender node at CERN, Linux kernels and were both tried, and was slightly better in performance. Major rate drops (packet losses) were rarely seen. When the aggregated window sizes were large enough, the transfer rates reached the limit of ~1Gbps (~120MB/s). Iperf tests Network Fig._5 GT4.2.1, kernel Fig._3 rate vs. time (typical cases)Fig._2 rate vs. number of streams We have tested data transfer between Europe and Japan via New York. Newer Linux kernel (or TCP implementation) and gridFTP perform better than previous versions. In the nearly optimal configuration, we could use the available bandwidth of ~1Gbps with gridFTP. In the production system, we have observed ~500MB/s data transfer rate for Lyon to ICEPP, while the configuration was not tuned seriously. Based on the test results shown above, we will try to change the system parameters to make better use of the available bandwidth. In future, performance should become better with the new TCP implementation in the recent Linux kernel. Conclusions At CERN and ICEPP, Linux test nodes were setup for the tests. System configuration relevant to the tests is listed below. (sender node at CERN, receiver node at ICEPP) OS: Scientific Linux CERN 4.7 x86_64. kernel: EL.cernsmp (TCP BIC, ICEPP and CERN) (TCP CUBIC, CERN) RAM: 32GB (CERN), 8GB (ICEPP) CPU: Xeon L5420 (CERN), Xeon 5160 (ICEPP) RAID Disk: 3ware (CERN), Infortrend (ICEPP). XFS. >80MB/s for single read/write. NIC: Intel 1Gbps (CERN), Chelsio 10Gbps (ICEPP) Parameters: net.ipv4.tcp_sack = 0 net.ipv4.tcp_dsack = 0 net.ipv4.tcp_timestamps = 0 net.ipv4.tcp_no_metrics_save = 0 net.ipv4.tcp_rmem = net.ipv4.tcp_wmem = net.core.rmem_max = net.core.wmem_wmax = net.core.netdev_max_backlog = txquelelen (of NIC) = Software: iperf gridFTP (Globus Toolkit 3.2.1, 4.2.1) Test setup Tokyo New York SINET3 (10Gbps) GEANT (10Gbps) Fig._1 Network route between Japan and Europe. Figure 1 shows the route for the data transfer. From the ICEPP site, the route to Europe goes through the SINET and GEANT, the Japanese and European academic network, respectively. The bandwidth of the route is 10Gbps, but is shared with other traffic. To the Lyon site, RENATER provides 10Gbps link connected to GEANT, but for this test the bandwidth was limited to 1Gbps at CERN’s HTAR (High Performance Access Route). The round trip time (RTT) is ~290ms for both Lyon-ICEPP and CERN- ICEPP routes, therefore the Bandwidth-Delay Product (BDP) is 1Gbps x 290ms = 36MB, which is needed to fully use 1Gbps bandwidth with a single TCP connection. gridFTP tests Disk-to-disk tests were performed by using gridFTP in Globus Toolkit or (gridFTP server: 3.15 or 1.17, globus-url-copy: 4.14 or 3.6). The time-dependent variation in transfer rates was larger than that of iperf tests due to relatively slow disk IO. Figures 4-7 show throughputs of single file (>1GB) transfers in various configurations. Better performance were seen with more recent versions of the Globus Toolkit and Linux kernel. For the Linux kernel, we have observed more TCP window reductions in (Fig. 8) kernel than in Fig._4 GT4.2.1, kernel Fig._7 GT3.2.1, kernel Fig._6 GT3.2.1, kernel Fig._10 rate vs. concurrent filesFig._9 transfer rate per stream Fig._8 Example of the packet loss during the file transfer Figure 9 shows transfer rates per stream in a file transfer. In this test (GT4.2.1, ), most streams were well balanced. Multiple-file transfers were also tested in a few cases (Figure 10), and the 1Gbps bandwidth was almost saturated. Production system The Disk Pool Manager (DPM) has been deployed as the Storage Element at the ICEPP Tier-2 site. There are now 13 disk servers on which the maximum TCP window size is limited to 2MB in kernel (SLC4 x86_64). For the data transfer from Lyon to ICEPP, numbers of concurrent files and gridFTP streams has been set to 20 and 10 respectively with the File Transfer Service (FTS). The best data transfer rate observed was ~500MB/s for Lyon-ICEPP (Figure 11). At that time (May 2008), we had only 6 disk servers, while >30 disk servers were used in Lyon. Fig._11 best data transfer rate observed for Lyon Tier-1 to ICEPP Tier-2