MMA Project Book, Chapter 8, Section 2
Mick Brooks
Brian Glendenning
Larry D’Addario
Last revised 1999 -April-8
Revision History:
1998-4-8: Add in new detail on M&C data sizes and rates; minor changes to discussion points
1998-11-16: Combine previous contributions – favor distributed computing model
Summary
This section describes hardware and communications considerations for the Monitor and Control (M&C) system, and some consideration of task structure. Other software considerations are described in Chapter 12.
Data rates for M&C system are modest in most situations, ~2500 Bytes/second/antenna. Sampling of total power data, video data from an optical telescope, and FPGA downloads are some of the possible situations where the data load is significantly larger. An architecture in which as-dumb-as-possible devices communicate with M&C computers distributed throughout the array is described, making use of separate, dedicated communications media for the higher throughput subsystems.
Table 8.2.1 Principal M&C milestones, D&D phase
M&C draft interface specifications |
1999/06/01 |
Preliminary software design review |
1999/06/30 |
Standard bus interface circuit prototype |
1999/09/01 |
Critical design review (M&C) |
2000/03/31 |
Deliver single dish antenna test system |
2001/03/01 |
8.2.1 Design Considerations
This is a list of high-level issues that affect the choice of communication hardware and protocols; the type and distribution of computers; and the design of real-time control software. Each is followed here by some discussion.
Comment: This is a crucial decision that must be made early, since it sets important aspects of the design of many devices, not just of the monitor/control system. It determines, for example, whether a given piece of hardware requires an embedded computer or instead can rely on a higher level device (or the master computer) to handle any complex control. If a local computer is required, the overall policy determines whether it can be a simple microprocessor running one fixed task, or a "real computer" that requires an operating system.
Comment: Note that it is not sufficient to have enough capacity to handle the long-term-average aggregate data rate. Each control signal must be delivered on time, within a maximum latency and with minimal jitter. If the requirements are tight, then one way to meet them is to have a hierarchy of control computers, where local ones have little to do but can meet close timing tolerances and higher level ones operate leisurely on big buffers. However, such a system is complex. At the other extreme, a single master computer might be able to meet all requirements. The decision is a trade-off between simplicity and flexibility.
Comment: If the system design assumes a sufficiently high minimum level of intelligence in all devices, then it relieves all computers (except those embedded in devices) of the burden of dealing with low-level, device-specific functions. This makes the design and maintenance of the software in those computers much easier. However, it imposes a heavy burden on the device designer to build the embedded processor and its software, no matter how simple the device might be. Since devices will be designed by various engineers at different laboratories, a wide variety of implementations may result, and all must be maintained. More intelligence is required to close control loops at high rates; slow control loops may be closed over the communications link by a master computer.
Comment: This is a practical problem which has plagued several of our large telescope projects. There are several possible approaches. One is to provide duplicates of the M/C computer system and its software (all levels) at each development laboratory, and to provide at least one duplicate for each level at the site along with switching that allows the duplicate to be used for parts of the system and the main for other parts. Another approach is to support testing via separate computers (e.g., laptops) that need not have the same architecture, operating system, nor code as the main M/C system, but that support the same physical interface(s) to devices, subsystems, and antennas. Such computers can then be substituted for the M/C system whenever it is convenient to do so, and their software can be tailored to the testing requirements rather than to operational requirements.
However, there are new questions that will arise in this process. A device or logical function can be classified according to whether it is associated with one antenna, associated with a subset of antennas, or common to all antennas. Some things (like the correlator outputs) are organized by interferometer baseline; others may be associated with a subarray of antennas whose membership is time variable.
Comment: The proper handling of subarrays has been one of the most difficult issues in the design of the VLA and VLBA control systems. For the MMA, even greater flexibility is required. It is therefore important that this be taken into account early in the design.
Comment: It's assumed that the master computer will be in a control building of some sort. Should there be a separate path from there to each antenna (star configuration), or a single party line for all antennas (linear or ring configuration), or something in between? Should every station have its own connection, whether occupied by an antenna or not? Even if there is a separate physical path to each antenna, it is still possible to have a single logical network, such that all elements receive all messages but they respond only to those appropriately addressed. This approach requires much more communication bandwidth and is in this sense wasteful, but it may result in substantial simplification of software and some hardware. Another assumption is that devices on one antenna do not need to communicate with devices at another antenna.
Comment: Distributed control systems are common in the industrial process control and factory automation industries. In the past, NRAO has developed communications protocols and interfaces itself. Commercial products offer some advantages in terms of cost, development time and fault tolerance.
8.2.2 Data Rates
The rate at which devices will have to be monitored or controlled is only roughly known at this time. Current best guess estimates follow. The first table summarizes the average and peak data rates at each antenna, according to information at the time of writing. The remaining tables breakdown data rates by subsystem at each antenna, and at the central building.
Table 8.2.2.1 Average and Peak Data Rates at Each Antenna
Mode |
Average Data Rate (B/s) |
Peak Data Rate (B/s) |
Normal Observing |
2500 |
10000 |
Total Power OTF |
3600 |
10000 |
Frequency Switched |
848 |
10000 |
Video Data |
250000 |
250000 |
Holography |
8000 |
8000 |
Note that the net data rates are low (excluding science data of course) if we exclude the possibility of video data. For example, the aggregate throughput to the antenna is less than 3000 Bytes/s. Sporadic data such as FPGA downloads may be quite large (20 Mbytes) but require only soft delivery deadlines; these account for the peak rates.
Table 8.2.2.2 Devices at each antenna
Item |
Control |
Monitor |
||
Size (B) |
Time (s) |
Size (B) |
Time (s) |
|
ACU |
10 |
0.1 |
10 |
0.01 |
Subreflector focus adj. |
4 |
120 |
2 |
600 |
Subreflector nutation control |
4 |
60 |
2 |
60 |
Cryogenics |
2 |
Rare |
100 |
600 |
HFET Receivers (3) |
10 |
Rare |
180 |
60 |
SIS Receivers (7) |
10 |
Rare |
180 |
60 |
Optical Telescope |
8 |
Rare |
250 000 |
0.5 |
Total Power |
0 |
8 |
0.002 |
|
1st LO switching |
4 |
60 |
4 |
60 |
1st LO tuning (if conventional) |
20 |
60 |
30 |
300 |
Fringe rotation (if conventional) |
10 |
1 |
10 |
60 |
IF switching |
4 |
60 |
4 |
60 |
2nd, 3rd LO |
8 |
600 |
||
IF level attenuators, detectors |
16 |
60 |
16 |
10 |
Optical transmitters |
8 |
Rare |
12 |
Rare |
Other (environmental, safety, etc). |
32 |
600 |
Table 8.2.2.3 Devices for each antenna, at central building
Item |
Control |
Monitor |
||
Size (B) |
Time (s) |
Size (B) |
Time (s) |
|
1st LO tuning, status |
20 |
60 |
30 |
300 |
Fringe rotation (if photonic) |
10 |
1 |
10 |
60 |
Cable length monitor |
2 |
Rare |
4 |
10 |
Last LO tuning (BBC) |
32 |
60 |
32 |
60 |
Bandwidth selection (BBC) |
16 |
60 |
16 |
60 |
Digitizer mode |
16 |
60 |
16 |
60 |
Table 8.2.2.4 Common devices, at central building
Item |
Control |
Monitor |
||
Size (B) |
Time (s) |
Size (B) |
Time (s) |
|
Timing standard |
20 |
600 |
||
LO System |
6 |
Rare |
||
Optical Transmitters |
8 |
Rare |
18 |
Rare |
Reference signal generation |
2 |
Rare |
10 |
600 |
Weather instruments |
2 |
Rare |
80 |
60 |
Table 8.2.2.5 Correlator related
Item |
Control |
Monitor |
||
Size (B) |
Time (s) |
Size (B) |
Time (s) |
|
Input configuration |
64 |
60 |
64 |
600 |
Delay tracking |
32 |
1 |
||
Output configuration |
4096 |
60 |
4096 |
600 |
8.2.3 Conceptual Design
The design presented here should not be considered final. Detailed timing or other requirements could change it.
This design essentially uses a computer to couple a local intra-antenna bus to a wider MMA network. The aggregate computing power at each antenna exceeds that which is required for its coordination role. On the other hand this design allows for much flexibility in handling antennas with special instrumentation, implementing high-speed sampling for debugging devices at the antenna remotely ("virtual oscilloscope"), and other requirements which are unknown now but will inevitably become important later.
Testing outside of the M&C system can be accomplished in two ways. First, the entire antenna may be unplugged from the M&C system and into another computer (e.g., a technician’s laptop or a computer on the antenna transporter). Similarly, a particular device may be plugged into a local bus that is attached to some other computer. This allows for testing during development when a full M&C software system is not available. For example, a PC with a commercial field-bus interface card could be used to act as a bus master in the lab. It is anticipated that LabView will be the defacto standard for test software in this phase.
Tasks are organized into three levels: One Master Task, which is concerned with issues common to all elements of the array; a Subarray Task for each active subarray, which includes a group of antennas currently participating in the same observation; and an Antenna Task for each antenna. Some of the jobs performed by each type of task are listed in the figure.
The controlled hardware is divided into three classes: antenna-related, baseline-related, and common. "Antenna-related" means that there is exactly one copy for each antenna; some of these devices are located at the central building and some are located at the antenna. The correlator is treated specially; it contains the only baseline-related hardware, but it also contains antenna-related hardware (like delay lines). "Common" devices include such things as weather monitoring, central building environmental control, and the LO reference system.
It is presently undecided whether or not total power data will be transmitted by the monitor and control system.
Figure 8.2.3.1 – Monitor and control tasks and communication paths