Trigger Hardware Development Modular Trigger Processing Architecture Matt Stettler, Magnus Hansen CERN Costas Foudas, Greg Iles, John Jones Imperial College.

Slides:



Advertisements
Similar presentations
Digital RF Stabilization System Based on MicroTCA Technology - Libera LLRF Robert Černe May 2010, RT10, Lisboa
Advertisements

ESODAC Study for a new ESO Detector Array Controller.
FPGA-Based System Design: Chapter 3 Copyright  2004 Prentice Hall PTR SRAM-based FPGA n SRAM-based LE –Registers in logic elements –LUT-based logic element.
Phase-0 topological processor Uli Schäfer Johannes Gutenberg-Universität Mainz Uli Schäfer 1.
Level-1 Topology Processor for Phase 0/1 - Hardware Studies and Plans - Uli Schäfer Johannes Gutenberg-Universität Mainz Uli Schäfer 1.
S. Barua – CPSC 440 CHAPTER 8 INTERFACING PROCESSORS AND PERIPHERALS Topics to be covered  How to.
An ATCA and FPGA-Based Data Processing Unit for PANDA Experiment H.XU, Z.-A. LIU,Q.WANG, D.JIN, Inst. High Energy Physics, Beijing, W. Kühn, J. Lang, S.
Interface of DSP to Peripherals of PC Spring 2002 Supervisor: Broodney, Hen | Presenting: Yair Tshop Michael Behar בס " ד.
Uli Schäfer 1 FPGAs for high performance – high density applications Intro Requirements of future trigger systems Features of recent FPGA families 9U *
Using FPGAs with Embedded Processors for Complete Hardware and Software Systems Jonah Weber May 2, 2006.
System Architecture A Reconfigurable and Programmable Gigabit Network Interface Card Jeff Shafer, Hyong-Youb Kim, Paul Willmann, Dr. Scott Rixner Rice.
Module I Overview of Computer Architecture and Organization.
HomeGate Backplane Solutions Brian Von Herzen, Ph.D. Xilinx Consultant June 21, 2004 ISO/IEC JTC1 SC25/WG1 N June 2004 (Von Herzen,
Chassis Architecture Brandon Wagner Office of Information Technology
HyperTransport™ Technology I/O Link Presentation by Mike Jonas.
Lecture 2: Field Programmable Gate Arrays September 13, 2004 ECE 697F Reconfigurable Computing Lecture 2 Field Programmable Gate Arrays.
NETWORK TOPOLOGIES HNC COMPUTING - Network Concepts 1 Network Concepts Topologies.
Prototype of the Global Trigger Processor GlueX Collaboration 22 May 2012 Scott Kaneta Fast Electronics Group.
DLS Digital Controller Tony Dobbing Head of Power Supplies Group.
OPTO Link using Altera Stratix GX transceiver Jerzy Zieliński PERG group Warsaw.
High-Level Interconnect Architectures for FPGAs An investigation into network-based interconnect systems for existing and future FPGA architectures Nick.
SLAC Particle Physics & Astrophysics The Cluster Interconnect Module (CIM) – Networking RCEs RCE Training Workshop Matt Weaver,
Pulsar II Hardware Overview Jamieson Olsen, Fermilab 14 April 2014
U N C L A S S I F I E D FVTX Detector Readout Concept S. Butsyk For LANL P-25 group.
LAN Switching and Wireless – Chapter 1
J. Christiansen, CERN - EP/MIC
FPGA-Based System Design: Chapter 3 Copyright  2004 Prentice Hall PTR FPGA Fabric n Elements of an FPGA fabric –Logic element –Placement –Wiring –I/O.
FPGA-Based System Design: Chapter 3 Copyright  2004 Prentice Hall PTR Topics n FPGA fabric architecture concepts.
Intro to Network Design
Chapter 21 Topologies Chapter 2. 2 Chapter Objectives Explain the different topologies Explain the structure of various topologies Compare different topologies.
CPT Week, April 2001Darin Acosta1 Status of the Next Generation CSC Track-Finder D.Acosta University of Florida.
Frank Lemke DPG Frühjahrstagung 2010 Time synchronization and measurements of a hierarchical DAQ network DPG Conference Bonn 2010 Session: HK 70.3 University.
Lecture 10: Logic Emulation October 8, 2013 ECE 636 Reconfigurable Computing Lecture 13 Logic Emulation.
CMS Global Calorimeter Trigger Hardware Design 21/9/06.
Lecture 13: Logic Emulation October 25, 2004 ECE 697F Reconfigurable Computing Lecture 13 Logic Emulation.
Network Concepts Topologies
NETWORK HARDWARE CABLES NETWORK INTERFACE CARD (NIC)
Field Programmable Port Extender (FPX) 1 Modular Design Techniques for the FPX.
ATLAS Trigger / current L1Calo Uli Schäfer 1 Jet/Energy module calo µ CTP L1.
Development of Programmable Architecture for Base-Band Processing S. Leung, A. Postula, Univ. of Queensland, Australia A. Hemani, Royal Institute of Tech.,
Latest ideas in DAQ development for LHC B. Gorini - CERN 1.
Advanced Computer Networks Lecturer: E EE Eng. Ahmed Hemaid Office: I 114.
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
US Peripheral Crate VMEbus Controller Ben Bylsma EMU – ESR CERN, November 2003.
XLV INTERNATIONAL WINTER MEETING ON NUCLEAR PHYSICS Tiago Pérez II Physikalisches Institut For the PANDA collaboration FPGA Compute node for the PANDA.
SCORES: A Scalable and Parametric Streams-Based Communication Architecture for Modular Reconfigurable Systems Abelardo Jara-Berrocal, Ann Gordon-Ross NSF.
16 February 2011Ian Brawn1 The High Speed Demonstrator and Slice Demonstrator Programme The Proposed High-Speed Demonstrator –Overview –Design Methodology.
Trigger Workshop: Greg Iles Feb Optical Global Trigger Interface Card Dual CMC card with Virtex 5 LX110T 16 bidirectional.
Spring EE 437 Lillevik 437s06-l22 University of Portland School of Engineering Advanced Computer Architecture Lecture 22 Distributed computer Interconnection.
Mohamed ABDELFATTAH Andrew BITAR Vaughn BETZ. 2 Module 1 Module 2 Module 3 Module 4 FPGAs are big! Design big systems High on-chip communication.
Mohamed Younis CMCS 411, Computer Architecture 1 CMCS Computer Architecture Lecture 26 Bus Interconnect May 7,
TBD The Matrix Card Status and Plans Dr. John Jones Princeton University
CMX: Update on status and planning Yuri Ermoline, Wojciech Dan Edmunds, Philippe Laurens, Chip Michigan State University 7-Mar-2012.
E. Hazen1 MicroTCA for HCAL and CMS Review / Status E. Hazen - Boston University for the CMS Collaboration.
E. Hazen -- Upgrade Week1 AMC13 Project Status E. Hazen - Boston University for the CMS Collaboration.
E. Hazen -- xTCA IG1 AMC13 Project Status E. Hazen - Boston University for the CMS Collaboration.
E. Hazen -- Upgrade Week1 HCAL uTCA Readout Plan E. Hazen - Boston University for the CMS HCAL Collaboration.
E. Hazen -- HCAL Upgrade Workshop1 MicroTCA Common Platform For CMS Working Group E. Hazen - Boston University for the CMS Collaboration.
Use of FPGA for dataflow Filippo Costa ALICE O2 CERN
AMC13 Project Status E. Hazen - Boston University
DAQ and TTC Integration For MicroTCA in CMS
AMC13 T1 Rev 2 Preliminary Design Review E. Hazen Boston University
Modeling event building architecture for the triggerless data acquisition system for PANDA experiment at the HESR facility at FAIR/GSI Krzysztof Korcyl.
HyperTransport™ Technology I/O Link
uTCA A Common Hardware Platform for CMS Trigger Upgrades
MicroTCA Common Platform For CMS Working Group
MicroTCA Common Platform For CMS Working Group
Matrix Processor / Backplane Design Details
Multiprocessors and Multi-computers
Presentation transcript:

Trigger Hardware Development Modular Trigger Processing Architecture Matt Stettler, Magnus Hansen CERN Costas Foudas, Greg Iles, John Jones Imperial College 3/2007

Traditional Trigger Processors ● Generally pushed the state of the art  In speed and density  Large hardware designs  This will not change in the future ● Fully custom one-off systems  Complex and difficult to modify ● Difficult to maintain over the life of the experiment ● Designer's knowledge difficult to pass on  Lack clean internal interfaces ● Modular to a degree, but based on ad-hoc bus schemes  Generally don't scale well ● Dataflow defined by hardware architecture ● Inflexible by design 1

Current Systems ● FPGA technology adopted rapidly  Current FPGAs have sufficient performance ● Speed and density  Allow considerable flexibility in logic design ● Very large devices available ● Problem becomes physical concentration of data ● Dataflow is still hardware based  Some designs have used FPGAs as data routers ● Inefficient at this task ● Large number required increases design risk considerably  Many simply define the dataflow in hardware ● Large parallel buses ● Large circuit boards required  Expensive to develop in both time and money 2

Desired Attributes for new Systems ● Modular  Reasonably fine grained ● Smaller circuit boards ● Easier and less expensive to develop  Should scale well ● Well defined internal interfaces  Allows modules to be developed independently ● Perhaps shared across projects  At least electrically compatible ● Flexible in both logic and interconnect  Retain flexibility of FPGAs  Add complementary data routing flexibility ● Modify data flow without altering hardware ● Perhaps give the ability of dynamic reconfiguration 3

Relevant Current Technology ● Large FPGAs with built in SERDES links  Very high density logic  SERDES I/O allows physical concentration of data ● Gbps links available on large devices ● Normal I/O pins support 1+Gbps links ● LVDS/CML crosspoint switches  Up to 144x144 non blocking matrices  Asynchronous, Protocol agnostic ● Multi rate switching supported intrinsically  Currently available at 4+Gbps, 10Gbps announced ● Serial Memory becoming available  FBDIMM modules designed for server market ● Support very large arrays (>100GB) ● Electrically compatible with LVDS/CML 4

Proposal for CMS HCAL Muon and Quiet bit distribution Processing Module 0 Processing Module 1 Processing Module N Fiber I/O Serial Mesh Backplane Based on crosspoint switches 5

Implementation ● Proposed for CMS Global Calorimeter Trigger HCAL muon & quiet bit distribution  This system is ready for layout now ● Micro TCA based, 12 slot chassis  20 high speed links per slot ● Processing module  Based on Xilinx V5LX110T  Mindspeed M x72 crosspoint  MFP Fiber interface (SNAP-12) ● Backplane  Commercial unit possible, but limiting  Custom unit based on Mindspeed M x144 ● Programmable topology ● 500 Gbps maximum bandwidth 6

Micro TCA Crate/Backplane 7

Processing module Block diagram 12 channel MTP receive 12 channel MTP transmit Xilinx V5LX110T (16, 3.2Gbps MGTs) Mindspeed M x72 crosspoint 20 uTCA Gbps Links 12V NXP LPC23xx uController Enet POL supplies 3.3V switcher 1.0V switcher 2.5V linear 8 channel MTP transceiver

Key Features of Processing Module ● Xilinx Virtex 5LX110T FPGA  16 low power MGTs, Superior logic density/speed  Standard I/O 1Gbps capable ● Data I/O direct from fiber  16 channels full 3.2 Gbps  Fiber utilized for inter-crate data links  Same fiber modules used on previous system ● Crosspoint routes to FPGA, backplane, and fiber  1:N data replication supported at wire speed  FPGA output data sent to backplane or fiber  Data sharing between systems intrinsically supported ● Slow control via standard Ethernet  Based on NXP 23xx uController  Standard Ethernet protocols 9

Custom Backplane block diagram uTC A slot 1 uTC A slot 2 uTCA slot 12 uTC A slot 3 Mindspeed M x72 crosspoint NXP LPC23xx uController Ethernet Switch 12 Ethernet from each uTCA slot Uplink Mesh control Mindspeed M x144 crosspoint Control FPGA For Misc Backplane Functions 4444 To Slot 12 To Slot 1 10

Key Features of Custom Backplane ● Active backplane based on crosspoint switches  Mindspeed M21161  Active backplane eliminates congestion ● uTCA Hub slots cannot support full mesh connections ● Programmable topology  Mesh, star, etc based on crosspoint configuration ● 20 links/slot for high speed data communication  M21161 supports 4.1Gbps data rate ● System will be limited to 3.2 Gbps by FPGA/MFP Optics ● 1 link/slot for 100Mb Ethernet control interface  Star topology interface, switch IC on backplane  Backplane configuration via Ethernet also ● Same uController as processing module 11

FBDIMM Architecture (future Interest) 12

Key Features of FBDIMM ● Serial memory is designed for large blocks  Packet based logical model  Optimized for block transfers ● Utilize Advanced Memory Buffers (AMB)  Control local banks of DDR2 devices  Create daisy chain of banks ● Interface to host is via CML links  Currently working at 3.2 Gbps, eventually up to 9.6  Utilize full duplex links ● Up to 8 AMBs can be daisy chained  Achieves 192GB with current technology ● Can be combined with serial mesh  Very large distributed memory arrays possible 13

Serial Link Latency ● Historically the limiting factor  Deserialization has dominated latency  This has improved with new FPGA families ● Current performance (Virtex5)  byte clocks for receive ● This is MHz clocks in our implementation  5-7 byte clocks for transmit ● MHz clocks ● Still significant  Biases FPGA choice to fewer large devices  Number of “hops” needs careful attention ● Crosspoints do not contribute to latency  Nearly wire speed (ps delay) - insignificant  No serialization/deserialization 14