An industrial SCADA system will be used for the development of the controls of the four LHC experiments. This paper describes the SCADA systems in terms of their architecture, their interface to the process hardware, the functionality and the application development facilities they provide. Some attention is also paid to the industrial standards to which they abide, their planned evolution as well as the potential benefits of their use.
On 20 Sept. 2000, the Finance Committee approved the proposal to negotiate a contract with ETM A.G. (Eisenstadt, Austria) for the supply of PVSS - ETM’s SCADA - for developing the control systems of ALICE, ATLAS, CMS and LHCb. In addition the SCADA Working Group, that was set up by the CERN Controls Board, recommends PVSS as one of the SCADA products for the development of future control systems at CERN.
These decisions are the accomplishment of around thirteen person-years (FTE) of effort - spanning over more than three years - to identify and evaluate a proper industrial control system that copes with the extreme requirements of high energy particle physics experiments such as those of the LHC.
Widely used in industry for Supervisory Control and Data Acquisition of industrial processes, SCADA systems are now also penetrating the experimental physics laboratories for the controls of ancillary systems such as cooling, ventilation, power distribution, etc. More recently they were also applied for the controls of smaller size particle detectors such as the L3 muon detector and the NA48 experiment, to name just two examples at CERN.
SCADA systems have made substantial progress over the recent years in terms of functionality, scalability, performance and openness such that they are an alternative to in house development even for very demanding and complex control systems as those of physics experiments.
SCADA stands for Supervisory Control And Data Acquisition. As the name indicates, it is not a full control system, but rather focuses on the supervisory level. As such, it is a purely software package that is positioned on top of hardware to which it is interfaced, in general via Programmable Logic Controllers (PLCs), or other commercial hardware modules.
SCADA systems are used not only in industrial processes: e.g. steel making, power generation (conventional and nuclear) and distribution, chemistry, but also in some experimental facilities such as nuclear fusion. The size of such plants range from a few 1000 to several 10 thousands input/output (I/O) channels. However, SCADA systems evolve rapidly and are now penetrating the market of plants with a number of I/O channels of several 100 K: we know of two cases of near to 1 M I/O channels currently under development.
SCADA systems used to run on DOS, VMS and UNIX; in recent years all SCADA vendors have moved to NT and some also to Linux.
This section describes the common features of the SCADA products that have been evaluated at CERN in view of their possible application to the control systems of the LHC detectors , .
3.1 Hardware Architecture
One distinguishes two basic layers in a SCADA system: the “client layer” which caters for the man machine interaction and the “data server layer” which handles most of the process data control activities. The data servers communicate with devices in the field through process controllers. Process controllers, e.g. PLCs, are connected to the data servers either directly or via networks or fieldbuses that are proprietary (e.g. Siemens H1), or non-proprietary (e.g. Profibus). Data servers are connected to each other and to client stations via an Ethernet LAN. The data servers and client stations are NT platforms but for many products the client stations may also be W95 machines. Fig.1. shows typical hardware architecture.
Figure 1: Typical Hardware Architecture
3.2 Software Architecture
The products are multi-tasking and are based upon a real-time database (RTDB) located in one or more servers. Servers are responsible for data acquisition and handling (e.g. polling controllers, alarm checking, calculations, logging and archiving) on a set of parameters, typically those they are connected to.
Figure 2: Generic Software Architecture
However, it is possible to have dedicated servers for particular tasks, e.g. historian, datalogger, alarm handler. Fig. 2 shows a SCADA architecture that is generic for the products that were evaluated.
Server-client and server-server communication is in general on a publish-subscribe and event-driven basis and uses a TCP/IP protocol, i.e., a client application subscribes to a parameter which is owned by a particular server application and only changes to that parameter are then communicated to the client application.
The data servers poll the controllers at a user defined polling rate. The polling rate may be different for different parameters. The controllers pass the requested parameters to the data servers. Time stamping of the process parameters is typically performed in the controllers and this time-stamp is taken over by the data server. If the controller and communication protocol used support unsolicited data transfer then the products will support this too.
The products provide communication drivers for most of the common PLCs and widely used field-buses, e.g., Modbus. Of the three fieldbuses that are recommended at CERN, both Profibus and Worldfip are supported but CANbus often not . Some of the drivers are based on third party products (e.g., Applicom cards) and therefore have additional cost associated with them. VME on the other hand is generally not supported.
A single data server can support multiple communications protocols: it can generally support as many such protocols as it has slots for interface cards.
The effort required to develop new drivers is typically in the range of 2-6 weeks depending on the complexity and similarity with existing drivers, and a driver development toolkit is provided for this.
Application Interfaces / Openness
The provision of OPC client functionality for SCADA to access devices in an open and standard manner is developing. There still seems to be a lack of devices/controllers, which provide OPC server software, but this improves rapidly as most of the producers of controllers are actively involved in the development of this standard. OPC has been evaluated by the CERN-IT-CO group .
The products also provide
an Open Data Base Connectivity (ODBC) interface to the data in the archive/logs, but not to the configuration database,
an ASCII import/export facility for configuration data,
a library of APIs supporting C, C++, and Visual Basic (VB) to access data in the RTDB, logs and archive. The API often does not provide access to the product’s internal features such as alarm handling, reporting, trending, etc.
The PC products provide support for the Microsoft standards such as Dynamic Data Exchange (DDE) which allows e.g. to visualise data dynamically in an EXCEL spreadsheet, Dynamic Link Library (DLL) and Object Linking and Embedding (OLE).
The configuration data are stored in a database that is logically centralised but physically distributed and that is generally of a proprietary format.
For performance reasons, the RTDB resides in the memory of the servers and is also of proprietary format.
The archive and logging format is usually also proprietary for performance reasons, but some products do support logging to a Relational Data Base Management System (RDBMS) at a slower rate either directly or via an ODBC interface.
Scalability is understood as the possibility to extend the SCADA based control system by adding more process variables, more specialised servers (e.g. for alarm handling) or more clients. The products achieve scalability by having multiple data servers connected to multiple controllers. Each data server has its own configuration database and RTDB and is responsible for the handling of a sub-set of the process variables (acquisition, alarm handling, archiving).
The products often have built in software redundancy at a server level, which is normally transparent to the user. Many of the products also provide more complete redundancy solutions if required.
Users are allocated to groups, which have defined read/write access privileges to the process parameters in the system and often also to specific product functionality.
The products support multiple screens, which can contain combinations of synoptic diagrams and text.
They also support the concept of a “generic” graphical object with links to process variables. These objects can be “dragged and dropped” from a library and included into a synoptic diagram.
Most of the SCADA products that were evaluated decompose the process in “atomic” parameters (e.g. a power supply current, its maximum value, its on/off status, etc.) to which a Tag-name is associated. The Tag-names used to link graphical objects to devices can be edited as required. The products include a library of standard graphical symbols, many of which would however not be applicable to the type of applications encountered in the experimental physics community.
Standard windows editing facilities are provided: zooming, re-sizing, scrolling… On-line configuration and customisation of the MMI is possible for users with the appropriate privileges. Links can be created between display pages to navigate from one view to another.
The products all provide trending facilities and one can summarise the common capabilities as follows:
the parameters to be trended in a specific chart can be predefined or defined on-line
a chart may contain more than 8 trended parameters or pens and an unlimited number of charts can be displayed (restricted only by the readability)
real-time and historical trending are possible, although generally not in the same chart
historical trending is possible for any archived parameter
zooming and scrolling functions are provided
parameter values at the cursor position can be displayed
The trending feature is either provided as a separate module or as a graphical object (ActiveX), which can then be embedded into a synoptic display. XY and other statistical analysis plots are generally not provided.
Alarm handling is based on limit and status checking and performed in the data servers. More complicated expressions (using arithmetic or logical expressions) can be developed by creating derived parameters on which status or limit checking is then performed. The alarms are logically handled centrally, i.e., the information only exists in one place and all users see the same status (e.g., the acknowledgement), and multiple alarm priority levels (in general many more than 3 such levels) are supported.
It is generally possible to group alarms and to handle these as an entity (typically filtering on group or acknowledgement of all alarms in a group). Furthermore, it is possible to suppress alarms either individually or as a complete group. The filtering of alarms seen on the alarm page or when viewing the alarm log is also possible at least on priority, time and group. However, relationships between alarms cannot generally be defined in a straightforward manner. E-mails can be generated or predefined actions automatically executed in response to alarm conditions.
The terms logging and archiving are often used to describe the same facility. However, logging can be thought of as medium-term storage of data on disk, whereas archiving is long-term storage of data either on disk or on another permanent storage medium. Logging is typically performed on a cyclic basis, i.e., once a certain file size, time period or number of points is reached the data is overwritten. Logging of data can be performed at a set frequency, or only initiated if the value changes or when a specific predefined event occurs. Logged data can be transferred to an archive once the log is full. The logged data is time-stamped and can be filtered when viewed by a user. The logging of user actions is in general performed together with either a user ID or station ID. There is often also a VCR facility to play back archived data.
One can produce reports using SQL type queries to the archive, RTDB or logs. Although it is sometimes possible to embed EXCEL charts in the report, a “cut and paste” capability is in general not provided. Facilities exist to be able to automatically generate, print and archive reports.
The majority of the products allow actions to be automatically triggered by events. A scripting language provided by the SCADA products allows these actions to be defined. In general, one can load a particular display, send an Email, run a user defined application or script and write to the RTDB.
The concept of recipes is supported, whereby a particular system configuration can be saved to a file and then re-loaded at a later date.
Sequencing is also supported whereby, as the name indicates, it is possible to execute a more complex sequence of actions on one or more devices. Sequences may also react to external events.
Some of the products do support an expert system but none has the concept of a Finite State Machine (FSM).
5. Application Development
The development of the applications is typically done in two stages. First the process parameters and associated information (e.g. relating to alarm conditions) are defined through some sort of parameter definition template and then the graphics, including trending and alarm displays are developed, and linked where appropriate to the process parameters. The products also provide an ASCII Export/Import facility for the configuration data (parameter definitions), which enables large numbers of parameters to be configured in a more efficient manner using an external editor such as Excel and then importing the data into the configuration database.
However, many of the PC tools now have a Windows Explorer type development studio. The developer then works with a number of folders, which each contains a different aspect of the configuration, including the graphics.
The facilities provided by the products for configuring very large numbers of parameters are not very strong. However, this has not really been an issue so far for most of the products to-date, as large applications are typically about 50K I/O points and database population from within an ASCII editor such as Excel is still a workable option.
On-line modifications to the configuration database and the graphics is generally possible with the appropriate level of privileges.
The following development tools are provided as standard:
a graphics editor, with standard drawing facilities including freehand, lines, squares circles, etc. It is possible to import pictures in many formats as well as using predefined symbols including e.g. trending charts, etc. A library of generic symbols is provided that can be linked dynamically to variables and animated as they change. It is also possible to create links between views so as to ease navigation at run-time.
a data base configuration tool (usually through parameter templates). It is in general possible to export data in ASCII files so as to be edited through an ASCII editor or Excel.
a scripting language
an Application Program Interface (API) supporting C, C++, VB
a Driver Development Toolkit to develop drivers for hardware that is not supported by the SCADA product.
The products in general have the concept of graphical object classes, which support inheritance. In addition, some of the products have the concept of an object within the configuration database. In general the products do not handle objects, but rather handle individual parameters, e.g., alarms are defined for parameters, logging is performed on parameters, and control actions are performed on parameters. The support of objects is therefore fairly superficial.
SCADA vendors release one major version and one to two additional minor versions once per year. These products evolve thus very rapidly so as to take advantage of new market opportunities, to meet new requirements of their customers and to take advantage of new technologies.
As was already mentioned, most of the SCADA products that were evaluated decompose the process in “atomic” parameters to which a Tag-name is associated. This is impractical in the case of very large processes when very large sets of Tags need to be configured. As the industrial applications are increasing in size, new SCADA versions are now being designed to handle devices and even entire systems as full entities (classes) that encapsulate all their specific attributes and functionality. In addition, they will also support multi-team development.
As far as new technologies are concerned, the SCADA products are now adopting:
Web technology, ActiveX, Java, etc.
OPC as a means for communicating internally between the client and server modules. It should thus be possible to connect OPC compliant third party modules to that SCADA product.
Whilst one should rightly anticipate significant development and maintenance savings by adopting a SCADA product for the implementation of a control system, it does not mean a “no effort” operation. The need for proper engineering can not be sufficiently emphasised to reduce development effort and to reach a system that complies with the requirements, that is economical in development and maintenance and that is reliable and robust. Examples of engineering activities specific to the use of a SCADA system are the definition of:
a library of objects (PLC, device, subsystem) complete with standard object behaviour (script, sequences, …), graphical interface and associated scripts for animation,
templates for different types of “panels”, e.g. alarms,
instructions on how to control e.g. a device …,
a mechanism to prevent conflicting controls (if not provided with the SCADA),
alarm levels, behaviour to be adopted in case of specific alarms, …
8. Potential benefits of SCADA
The benefits one can expect from adopting a SCADA system for the control of experimental physics facilities can be summarised as follows:
a rich functionality and extensive development facilities. The amount of effort invested in SCADA product amounts to 50 to 100 p-years!
the amount of specific development that needs to be performed by the end-user is limited, especially with suitable engineering.
reliability and robustness. These systems are used for mission critical industrial processes where reliability and performance are paramount. In addition, specific development is performed within a well-established framework that enhances reliability and robustness.
technical support and maintenance by the vendor.
For large collaborations, as for the CERN LHC experiments, using a SCADA system for their controls ensures a common framework not only for the development of the specific applications but also for operating the detectors. Operators experience the same “look and feel” whatever part of the experiment they control. However, this aspect also depends to a significant extent on proper engineering