Download presentation
Presentation is loading. Please wait.
Published byElwin Owens Modified over 8 years ago
1
JCOP Workshop September 8th 1999 H.J.Burckhart 1 ATLAS DCS Organization of Detector and Controls Architecture Connection to DAQ Front-end System Practical Points Conclusions
2
JCOP Workshop September 8th 1999 H.J.Burckhart 2 Scope of DCS Detector Sub-detectors Experiment’s infrastructure “External Services” Services of CERN infrastructure LHC accelerator DAQ
3
JCOP Workshop September 8th 1999 H.J.Burckhart 3 ATLAS picture here
4
JCOP Workshop September 8th 1999 H.J.Burckhart 4 Detector Organisation
5
JCOP Workshop September 8th 1999 H.J.Burckhart 5 Detector organisation Hierarchical organisation of quasi independent units (“objects”) Separation for various reasons (organisational, operational, geometrical, etc.) Units have to operate stand-alone and integrated Data flow mainly vertically Common Infrastructure handled like a subdetector
6
JCOP Workshop September 8th 1999 H.J.Burckhart 6
7
7 Hierarchical levels of DCS Hierarchical levels of DCS Supervisor Level operator consoleshift operator, sub-system expert serverdata base, conn. DAQ, External system Subsystem control level Local Ctrl StationGas, HV, endcap -------------------------------------------------------------------------- Device Control Fieldbus nodechamber, sector PLCcooling VME HV Sensors SCADA FE I/O
8
JCOP Workshop September 8th 1999 H.J.Burckhart 8 Consequences of Detector organisation on DCS Subdetector groups choose mapping Unit DCS_level I/F definition SCADA Front-end I/O HW (industry) standards standard SW protocols High numbers of channels (>>10k) multiplexed in F/E Separation Tools Applications work done at different geographical places guaranties homogeneity
9
JCOP Workshop September 8th 1999 H.J.Burckhart 9 External systems
10
JCOP Workshop September 8th 1999 H.J.Burckhart 10 Interaction with External Systems Interaction with External Systems Exchange of (result) data some imported data fully treated in SCADA Possibility to send commands and receive feedback Exchange of (dynamic) Status information ==> Use common mechanism with all External Systems (LDIWG)
11
JCOP Workshop September 8th 1999 H.J.Burckhart 11 Connection DCS - DAQ Connection DCS - DAQ Complete operational independence DCS needed already during production, assembly, installation DCS needed 365/365, 24/24, DAQ during data taking periods Data paths of DAQ and DCS should be separated Independence MUST NOT result in restrictions of functionality Seamless information exchange Sending commands and getting feedback in both directions Common Data Base
12
JCOP Workshop September 8th 1999 H.J.Burckhart 12 Boundary DCS - DAQ Boundary DCS - DAQ DAQ treats all aspects of physics event data (‘event #’): data flow, quality monitoring, storage, etc. DCS treats other data (‘time stamp’) Interaction DAQ LHC via DCS
13
JCOP Workshop September 8th 1999 H.J.Burckhart 13 Interaction with DAQ DAQ control and configuration services organised in “Backend DAQ components” (SCADA has similar components) Information Service Exchange of dynamic information (e.g.status) ==> bi-directional interface with SCADA (filtered) Message Reporting System reports asynchronously events (e.g. error messages, state changes) ==> SCADA has to inject it’s ‘events’ (e.g. actions, state changes)
14
JCOP Workshop September 8th 1999 H.J.Burckhart 14 Interaction with DAQ (cont.) Run Control co-ordinates DAQ sub-systems ==> sending/receiving commands and getting feedback about success/failure Graphical User Interface operator interaction, viewing of DAQ state ==> top level of DCS should be accessible
15
JCOP Workshop September 8th 1999 H.J.Burckhart 15 Interaction with DAQ (cont.) Data Base Configuration (HW, SW, parameters, calibration constants, etc.) command logging (operator, “automatic”, etc.) incident logging (alarms, state changes, etc.) storage of measurement results (currents, temperatures, etc.) ATLAS wide DB for DAQ, DCS and Offline Naming conventions according to HW (PBS,ABS) ==> This DB is the master DB for DCS
16
JCOP Workshop September 8th 1999 H.J.Burckhart 16 Product Breakdown Structure
17
JCOP Workshop September 8th 1999 H.J.Burckhart 17 Examples of operations (‘Use case’) Calibration DCS driven interleaved during data taking DAQ driven Start of run (load, configuration, start) Partitioning Loading parameters
18
JCOP Workshop September 8th 1999 H.J.Burckhart 18 Why separating DAQ and DCS? Time scale DAQ in constant evolution Avoid additional requirements on SCADA platform real time 4 LHC experiments, 4 DAQ systems (?) ATLAS DAQ design already advanced Past experience
19
JCOP Workshop September 8th 1999 H.J.Burckhart 19 Front-end I/O Front-end I/O Fieldbus general purpose node (LMB) dedicated Fieldbus node commercial Fieldbus device PLC SoftPLC VME with RTUnix dedicated processor (e.g. image processing), DSP real time response (e.g. radiation monitor), triggered (e.g. beam dump) high number of channels, high volume I/O
20
JCOP Workshop September 8th 1999 H.J.Burckhart 20 Important SCADA features support of hierarchical objects partitioning (stations loosely coupled) good connection to external DB powerful API driver tool kit capable to support ATLAS naming convention multi-platform
21
JCOP Workshop September 8th 1999 H.J.Burckhart 21 Subdetector Time scale SCT –production 1/01 => –assembly 4/03 => –installation 3/04 => Pixel –mounting 4/01 –assembly 4/03 TileCal –production => 12/01 –assembly + calibration 2002 –installation 2003 TRT –construction =>mid 2002 –assembly + test 2002 –installation 3/04 LAr –EC assembly 1/01 => 10/01 –calibration 12/01 => 6/02 –installation 1/03 => Muon –Decision about SCADA begin 2001
22
JCOP Workshop September 8th 1999 H.J.Burckhart 22 SCADA time scale SCADA is nucleus of DCS (ATLAS, JCOP) subdetector groups want practical experience in their environment start bottom up (Front-end I/O, simple devices, composite systems) –avoid stand-alone solutions –re-use developments
23
JCOP Workshop September 8th 1999 H.J.Burckhart 23 SCADA time scale (cont.) Detector time scale –final elements ready end 2001 –calibration on surface 2002 –installation 2003 ==> Use of SCADA –learning (3 months) –prototype application (6 months) –real application (12 months) ==> some subdetector groups need to start essentially now
24
JCOP Workshop September 8th 1999 H.J.Burckhart 24 Organisation of work SCADA vendor Maintains product (platform, OS, F/E drivers) CERN Controls Group connection to CERN Infrastructure and LHC, drivers, generic applications (e.g. gas, HT) LHC Detector Controls Groups integration of subdetectors (rules), general purpose I/O, operations, subdetector “experiment infrastructure”
25
JCOP Workshop September 8th 1999 H.J.Burckhart 25 Organisation of work (cont.) Sub-detector Controls expert(s) integration of units (states, info exchange), control procedures sub-system expert control algorithms, HW functions
26
JCOP Workshop September 8th 1999 H.J.Burckhart 26 To do now... Start field work with subdetectors make organizational plan (course, license, etc.) make prioritized list of common developments and applications start engineering final system
27
JCOP Workshop September 8th 1999 H.J.Burckhart 27 Conclusions ATLAS wants to continue the common approach SCADA system looks promising (many thanks to IT/CO for evaluation) implement now real subdetector application(s) with the most promising (and affordable) product
Similar presentations
© 2024 SlidePlayer.com Inc.
All rights reserved.