In the automotive domain, the following use cases are defined:
- Touch and Go Assistant
With the arrival of an airplane at the airport, lots of manual work is performed, e.g. attaching and running ground units like power, fuel, water, cooling or heating. In frame of this project the assistant will be applied to a specific ground unit that supplies pre-conditioned air to the airplane. Operating these devices at high efficiency currently requires two operators present the whole time – one in the airplane and one on ground. The operators communicate the necessary adjustment of the air flow via walky-talky. This “human control loop” is too cumbersome for some airlines, hence the devices run in an un-supervised and much less efficient mode. The Touch and Go Assistant will simplify this work: An initial association of the airplane and its ground unit – in line with current procedures – is sufficient. After that the machines will communicate and work independently at high efficiency. No need for further interaction between ground and airplane personal. The result of the Touch and Go Assistant use case will be a detailed description of the applied mechanisms and tools to verify the secure behavior and automated security checking of aerospace systems.
- Reconfigurable Video Processor for Space
In this use case the system under test is an industrial system that deals with these two drawbacks by means of including multicore architectures capable of in-flight reconfiguration in present actual payload data processing equipment, for video processing, and navigation sensors based on camera systems. These video processors are in charge of conditioning, processing, compressing and ciphering the images acquired by an Earth Observation Satellite before their transmission to ground or by any kind of spacecraft for navigation purposes. The systems under test will be multicore architectures presently available in the market capable of withstanding the space environment and that can follow the stringent design rules specified for Space equipment. The result of the Reconfigurable Video Processor for Space use case will be a detailed description of the condition and requirements for a system used for navigation or observation under environmental, reliability and dependability space conditions.
The final objective of intermediate demonstrator is to show different capabilities of the FIE and by doing so, provoke various effects on the SRAM-based FPGA similar to those caused by radiation in space environment. Several types of faults are injected using the engine in order to demonstrate the behavior of the implemented design. This demonstrator consequently proves the fact that it is possible to reproduce several radiation induced faults without the need for an expensive beam testing of the component.
This will help the designers to get a better understanding of how their hardware will fail in space in order to achieve a higher reliability. Apart from that, FIE features the so called “smart” injection allowing the designer to alter different parts of the FPGA in a controlled fashion. Consequently, the designer shall be able to affect different parts of the design which should significantly reduce the debugging process and help in the understanding the radiation effects on the SRAM-based FPGAs.
Finally, it is also shown how the cyclic redundancy checker can heal some of the faults that the radiation can produce on a device.
The demo consists of three main elements:
- ZCU102 evaluation board: For this demonstrator a Multi-Processor System on Chip (MPSoC) is going to be used. The target MPSoC will be the Zynq UltraScale+ from Xilinx.
- OV7670 Camera: The OV7670 is a CMOS image sensor that provides the full functionality of a single-chip VGA camera and image processor in a small footprint package.
- Computer: Finally, a computer is needed in order to communicate with the FIE and to receive the captured video signal.