Spie2006 Paperpdf

1
Avionics sensor simulation and prototype design workstation using COTS reconfigurable computing technology James Falasco, GE Fanuc Embedded Systems 1240 Campbell Road Richardson, Texas 75081 Abstract This paper reviews hardware and software solutions that allow for rapid prototyping of new or modified embedded avionics sensor designs, mission payloads and functional sub assemblies. We define reconfigurable computing in the context of being able to place various PMC modules depending upon mission scenarios onto a base SBC (Single Board Computer). This SBC could be either a distributed or shared memory archi- tecture concept and have either two or four PPC7447 A/7448 processor clusters. In certain scenarios, various combinations of boards could be combined in order to provide a heterogeneous computing environment. Keywords: Sensor simulation, reconfigurable computing, video compres- sion, video, reflective memory, payload integration, sensor fusion, high speed data acquisition, video surveillance 1. Introduction The design of avionics sensors and mission payloads can be a complex and time-consuming process. However, the design cycle can be significantly shortened if the system designer has access to a flexible, reconfigurable development environment that closely mimics the capabilities and tech- nologies of the deployed system. By integrating various PMC modules with a scalable, reconfigurable multi- processor architecture, it is possible to create a development tool that will allow the system designer to rather quickly and accurately simulate and test with real data sets the various sensor designs and mission compo- nents to be fielded. Specifically, we present a rapid prototyping and rapid evaluation system that will simplify the establishment of performance requirements, and allow the quick evaluation of hardware and software components being considered for inclusion in a new sensor system design or a legacy platform upgrade. 2. An Open Systems, COTS Platform The system hardware and software used to evaluate sensor designs and mission payload components and algorithms should be open and recon- figurable to allow for the mixing and matching of various vendor offerings. Provision for hardware independence is critical, since the hardware is very likely to rapidly evolve at the pace of new computer technology. The software infrastructure should be scalable and flexible allowing the algo- rithm developers the ability to spend their time and budget addressing the important functionality and usability aspects of the systems design. The system proposed here, a test and evaluation workstation built around reconfigurable hardware and a component-based software toolset, provides the necessary tools to ensure the success and cost effectiveness of initial sensor design and payload development. At the center of any scalable prototyping system is a reconfigurable multi- processing CPU engine with associated memory. The system depicted in Fig. 1 provides developers a scalable environment for application design and test. A COTS single board computer (SBC) tightly integrated to a PMC FPGA card for design experimentation forms the core system. Other PMC modules can then be selected depending on the type of sensor input that needs to be processed. One of the main advantages of this approach is the ability to rapidly proto- type mockups for test and evaluation without a concern for the limitations of embedded development at this early stage. The COTS SBC shown has the ability to host two PMC modules per card. The Video card can efficiently manage real sensor input coming in from an E/O sensor/camera, and also display that data in a fused fashion. This approach allows the system builder to work in the lab and the field using the same hardware/software environment. This system configuration also provides the developer with a test bed to define/design new hardware requirements and processing streams. 3. Prototyping System Structure Modern embedded avionics sensor design is primarily driven by the goal of providing a net-centric flow of data from platform to platform. The achieve- ment of this vision depends on transferring and processing vast amounts of data from multiple sources. Let’s examine how one could use this approach to control and manage various sensors in a rapid prototyping environment. As depicted in Figure 1, the host server for the embedded avionics sensor design workstation is a COTS Dual 7447A PowerPC VMEbus SBC. The board’s architecture provides a distributed processing environment that allows scaling of multiple processing nodes to achieve high performance for the most demanding signal and imaging applications associated with various sensor and mission payload design requirements. Fig. 1. FPGA vision and graphics platform The COTS SBC architecture combined with the FPGA/Video cards allows for seamless mapping of imaging applications oriented toward change detec- tion and sensor fusion which will allow the systems designer to view multiple data streams in a simulations real time display environment. The SBC implements processing nodes using the latest Motorola 7447A/7448 PowerPC® processors running at up to a 1.4 GHz clock rate. Using the distributed processing architecture of the SBC, one bridge chip per node allows the Front Side Bus (FSB) of each PowerPC to run in MPX mode up to its maximum rate of 133 MHz. The high performance data transfer mechanism facilitated by the built-in 64-bit full duplex crossbar in the Discovery II bridge permits concurrent data transfers between different interfaces as well as transaction pipelining for same source and destination transactions. The prototyping architecture outlined here is based on a combination of tightly integrated reconfigurable computing, video and graphics. It will place the embedded avionics sensor design community in position to utilize the proposed system architecture in development of the actual controller for payload packages as well as the sensor itself and in its actual deploy- ment integration. Most importantly it will allow for cost-effectively maintaining and extending the system when new technologies are avail- able in the future. 3.1 An Example: Infrared (IR) scene projection Increased sensor performance and bandwidth has begun to overwhelm existing backend processing capability. Current test and evaluation methods are not adequate for fully assessing the operational performance of imaging infrared sensors while they are installed on the weapon system platform. However, the use of infrared (IR) scene projection in test and evaluation can augment and redefine test methodologies currently being used to test and evaluate forward looking infrared (FLIR) and imaging IR sensors. One could project accurate, dynamic and realistic IR imagery into the entrance aperture of the sensor, such that the sensor would perceive and respond to the imagery as it would to the real-world scenario. This approach includes development, analysis, integration, exploitation, training, and test and evaluation of ground and aviation based imaging IR sensors/subsystems/systems. This applies to FLIR systems, imaging IR missile seekers/guidance sections, as well as non-imaging thermal sensors. The systems approach proposed in this paper has the scalability to accom- plish this type of reconfigurable sensor mix and match. Algorithms such as the one depicted in Figure 2 could be transformed into a Mathlab® Simulink® Blockset for easy execution in a reconfigurable system utilizing FPGA- and PPC-based computing elements. Embedded avionics HWIL Simulation demands include low latency, high data rates and interfacing. It is essential to have a capable platform for handling and processing of the data streams. Tools must also complement this so that a systems designer is able to construct the final system leveraging design tools, such as Mathlab and Simulink, with a reconfigurable computing platform. Fig. 2. Example: sensor images This approach allows one to demonstrate how algorithms can be imple- mented and simulated in a familiar rapid application development environment before they are automatically transposed for downloading directly to the distributed multiprocessing computing platform. This complements the established control tools, which usually handle the configuration and control of the processing systems leading to a tool suite for system development and implementation. 3.2 The advantages of FPGA Computing As the development tools have evolved, the core-processing platform has also been enhanced. These improved platforms are based on dynamically reconfigurable computing, utilizing FPGA technologies and parallel processing methods that more than double the performance and data bandwidth capabilities. This offers support for processing of images in Infrared Scene Projectors with 1024 X 1024 resolutions at 400 Hz frame rates. Simulink Blocksets could add the programming involved to organize algorithms that could then be partitioned to operate on a multiprocessor based FPGA/PPC based configuration. Another key component of reconfigurable scalable embedded avionics sensor design and prototyping capability is access to FPGA based processing. FPGA (Field Programmable Gate Array) is defined as an array of logic blocks that can be ‘glued’ together (or configured) to produce higher level functions in hardware. Based on SRAM technology, i.e., configurations are defined on power up and when power is removed the configuration is lost – until it is ‘reconfigured’ again. Since an FPGA is a hardware device, it is faster than software. The FPGA can best be described as a parallel device that makes it faster than software. FPGAs as programmable “ASICs” can be configured for high performance processing, excelling at continuous, high bandwidth applica- tions. FPGAs can provide inputs from digital and analog sensors —LVDS, Camerink, RS170 — with which the designer can interactively apply filters, do processing ,compression, image reconstruction and encryption time of applications. Examples of the flexibility of this approach using COTS PMC Modules hosted by a COTS multiprocessing base platform are shown in Figures 3 & 4. Fig. 3. Example: FPGA application. PMC FPGA processor for capture and compression. PMC module for graphics and display. Fig. 4. RS-170/MPEG-4 video PMC. Integrated bundle of PMC, mezzanine and MPEG-4. 3.3 The ‘Mix and Match’ Platform Advantage Today’s soldier, who is the ultimate customer of the embedded systems designer, is faced with a continuously fluid chain of world events. These changing events are closely mapped into the deployment of various air, sea and land platforms which contain the reconfigurable embedded systems architecture under discussion in this paper. For example, based on changing mission requirements, any of the following three different areas might be needed by the soldier: sonar processing; SIGINT; or video surveillance. The system designer addressing these defined application areas could place various PMC modules together in the prototyping system with relative ease. Specifically, these applications might collectively require the following list of PMC modules: PMC #1—Graphics; PMC#2—Video Compression; PMC#3—Reflective Memory; PMC#4—1553; PMC#5—High Speed Data Acqui- sition; PMC#6—Race++. A key cornerstone of the ‘mix and match’ strategy is that while the PMC module may change from application to application, the core SBC ‘multiprocessor’ engine and its associated software tool chain remains constant. The same can be said for the VME enclosure that contains the processing cards. Let’s now examine the application areas and understand how we could place various PMC combinations together to address the processing requirements of each application. 3.4 Sonar Processing: PMC #1—Graphics, PMC #5 —High Speed Data Acquisition In this example the PMC Graphics card would be utilized to display sonar waterfall display data perhaps with an overlay of tactical positions. The High Speed Data Acquisition PMC could be used to facilitate the information coming in from a high speed sensor interface. The combination of the two modules hosted by the multiprocessor-enabled SBC could move from this application area to another type of sonar processing, triggered by software or the base modules could be switched to those shown in the SIGINT example below. 3.5 SIGINT: PMC #6—Race ++, PMC #3—Reflective memory In a SIGINT application scenario a RACE ++ PMC module is tied to other Race++ peripherals allowing for interfacing while the reflective memory module stores real time acquired data that could be used for post processing data analysis. 3.6 Video Surveillance: PMC #2—Video Compression, PMC #4—1553 In the video surveillance application area, the video compression PMC module would pre-process and reduce the incoming data stream and pass it to the multiprocessor base system for potential change detection analysis. Then using the PMC 1553 module, one could communicate to an external avionics platform to perhaps control or guide ordnance being placed upon the target under surveillance. In each of the three examples above, the modules are interchangeable depending upon the specific mission. This approach would allow the soldier to move module packages between platforms achieving different mission scenarios through software loading while maintaining the core base multi- processing and packaged environment. 4. Conclusions The goal of integrating a COTS system such as the one depicted here is to allow designers to use the same environment in the lab that they could then take to the field for live data collection activity. This is of particular value to a system engineer who could use such an environment to perform new development activities. Because of these efficiencies, the outlined prototyping system would pay off in accelerated product develop- ment time. Systems designers are traditionally faced with the challenge that today’s sensors are generating data at a rate far faster than the backend end systems are configured to process. Combine this fact with the reality that a sensor fusion “paradigm” is now mandatory for designers wishing to turn concepts into reality rapidly. A key component to a flexible hardware system, of course, is a software structure that enables designers to go from their ideas and algorithmic concepts to code or HDL. Packaging of the system is largely dependent on the user’s requirements for flexibility. Should the user desire a system that can be scaled up by adding additional cards, then a larger slot chassis could be configured to allow for the addition of other cards. The key point is that the core hardware outlined not only has the potential for scalability by adding addi- tional modules to the base processing units, but the entire system has scalability as well. PPC to PCI-X Bridge (64360) PPC to PCI-X Bridge (64360) 1 GHz 7448 or 7447A PowerPC 32 MB User Flash 32 MB User Flash VME Bus P1 Up to 320 MB/s Peak Up to 1064 MB/s Peak PMC 1 PMC 2 PCI-X Bridge PCI-X Bridge NVRAM NVRAM 256 MB DDR 256 MB DDR 64-bit/133 MHz 64-bit/133 MHz Gigabit Ethernet Multifunction Serial Gigabit Ethernet Multifunction Serial Up to 1064 MB/s Peak 64-bit/ 133 MHz 64-bit/ 133 MHz 2eSST VME Bridge 1 GHz 7448 or 7447A PowerPC OPTICAL FILTER FILTER WHEEL INTENSIFIER FOCAL BACKPLANE INTENSIFIER HIGH VOLTAGE POWER SUPPLY MICROCONTROLLER ELECTRONICS CCD IMAGING CHIP AND VIDEO ELECTRONICS Video Rate Camera • 30 frames per second • Average image size (after registration): 642 x 343 pixels (8 bit) Spinning Filter Wheel • Creates 6-band multispectral image Representative Sensor Graphics PMC Module High Speed Data Acquisition PMC Module Multiprocessor BSP Graphics SW Layer (OpenGL) Adding Applications Layer Mix & Match Platform Advantage Example 1 Sonar Processing SIGINT Reconfigurable System Environment RF Receiver Real-Time Capabilities Real-time Signal Detection Signal Tracking Operator Event Marking Signal Classification Fusion of Data Incoming Information Pre-Processed- Processed & Post Processed and stored for future analysis Digital Capture of Signal Data Near-Real-Time Capabilities Automatic Detection Signal Localization Display and Analyze Results Operator in the Loop Data Tapes Processing & analysis may be performed on the ground at near-real-time rates by transfer of information from platform to ground station. Signal Analysis Function (SAF) Mix & Match Platform Advantage Example 3 Video Surveillance 1553 PMC Video Compression PMC Multiprocessor BSP Adding Applications Layer Battle Damage Assessment Electronic Surveillance Cross-cueing Sensors FOPEN Radar Minefield Detection LADAR Data Link Compression All Weather Target Acquisition Sensor Fusion Mission Planning UAV Application Examples Existing Platforms & New Designs Will Continue Flying For Decades Incremental Mission Payload Responsibilities Will Be Added Requirements For Interoperability Continue To Increase Developers Will Need To Shrink Payload Processing System Conclusions Overcoming Data Compression Bottlenecks Pentium enabled SBC’s VMIC Product Line PPC enabled SBC’s single,dual & quad configurations Richardson Product Line Display Capability CDI Product Line Storage Capability Camarillo Product Line 1553 Data link Step 3 The process of the first two steps allows the system developer to combine real world data with simulated scenarios and compare and contrast in order to hone overall system design efficiency Step 1 Data acquired in real time Step 2 Data is passed to PMC module to be processed & on Nexus Single Board Computer for backend processing & matching for decisions (mission driven scenarios) PMC Module Collect & Process Data from suspension subassembly Data Acquisition Sensor Solving The Bandwidth Problem PPC to PCI-X Bridge (64360) PPC to PCI-X Bridge (64360) 1 GHz 7448 or 7447A PowerPC 32 MB User Flash 32 MB User Flash VME Bus P1 Up to 320 MB/s Peak Up to 1064 MB/s Peak PMC 1 PMC 2 PCI-X Bridge PCI-X Bridge NVRAM NVRAM 256 MB DDR 256 MB DDR 64-bit/133 MHz 64-bit/133 MHz Gigabit Ethernet Multifunction Serial Gigabit Ethernet Multifunction Serial Up to 1064 MB/s Peak 64-bit/ 133 MHz 64-bit/ 133 MHz 2eSST VME Bridge 1 GHz 7448 or 7447A PowerPC RS170 PMC Graphics PMC FPGA PMC 1553 Technology Chassis Technology Display Technology Video Compression PMC Input Format Matrix Processor Demean Window Averager Target Window Average Background/ Guard Window Averager Cross Product Generator Band Interleaved Image Input RX Image Output Video Input Processing Video Server Application WAVE Software Drivers (PCI Bus) MPEG-4 Bit-stream Input 1 MPEG-4 Compression Input 2 MPEG-4 Compression TS-PMC with Stratix FPGA RTP Encapsulated MPEG-4 Bit-stream Server Configuration via TCP/IP Video Server CPU Dual RS-170 Mezzanine FPGA PMC Video Compression PMC Integrated Bundle Video Compression Encoder with Host Decode Multiprocessor SBC + = + Video Camera Audio Input • Capture • Filter, Format • Encrypt • Compress GPU PMC Overlay and Display TS-PMC Capture and Compress Video Camera Audio Input TS-PMC Capture and Compress TS-PMC Capture and Compress TS-PMC Capture and Compress • Capture • Filter, Format • Encrypt • Compress • Graphic Generation • Overlay • Display • Graphic Generation • Overlay • Display Thin Pipe Network Encrypter Wireless Network Multiprocessor SBC Race ++ PMC Module Reflective Memory PMC Module Multiprocessor BSP Adding Applications Layer Mix & Match Platform Advantage Example 2 SIGINT

description

This is a paper I presented at the SPIE Conference in 2006. The SPIE Defense Conference has become THE show for the sensor world.

Transcript of Spie2006 Paperpdf

Page 1: Spie2006 Paperpdf

Avionics sensor simulation and prototype design workstationusing COTS reconfigurable computing technology

James Falasco, GE Fanuc Embedded Systems1240 Campbell Road

Richardson, Texas 75081

Abstract

This paper reviews hardware and software solutions that allow for rapid

prototyping of new or modified embedded avionics sensor designs, mission

payloads and functional sub assemblies. We define reconfigurable

computing in the context of being able to place various PMC modules

depending upon mission scenarios onto a base SBC (Single Board

Computer). This SBC could be either a distributed or shared memory archi-

tecture concept and have either two or four PPC7447 A/7448 processor

clusters. In certain scenarios, various combinations of boards could be

combined in order to provide a heterogeneous computing environment.

Keywords: Sensor simulation, reconfigurable computing, video compres-

sion, video, reflective memory, payload integration, sensor fusion, high

speed data acquisition, video surveillance

1. Introduction

The design of avionics sensors and mission payloads can be a complex and

time-consuming process. However, the design cycle can be significantly

shortened if the system designer has access to a flexible, reconfigurable

development environment that closely mimics the capabilities and tech-

nologies of the deployed system.

By integrating various PMC modules with a scalable, reconfigurable multi-

processor architecture, it is possible to create a development tool that will

allow the system designer to rather quickly and accurately simulate and

test with real data sets the various sensor designs and mission compo-

nents to be fielded.

Specifically, we present a rapid prototyping and rapid evaluation system

that will simplify the establishment of performance requirements, and allow

the quick evaluation of hardware and software components being

considered for inclusion in a new sensor system design or a legacy

platform upgrade.

2. An Open Systems, COTS Platform

The system hardware and software used to evaluate sensor designs and

mission payload components and algorithms should be open and recon-

figurable to allow for the mixing and matching of various vendor offerings.

Provision for hardware independence is critical, since the hardware is very

likely to rapidly evolve at the pace of new computer technology. The

software infrastructure should be scalable and flexible allowing the algo-

rithm developers the ability to spend their time and budget addressing the

important functionality and usability aspects of the systems design. The

system proposed here, a test and evaluation workstation built around

reconfigurable hardware and a component-based software toolset,

provides the necessary tools to ensure the success and cost effectiveness

of initial sensor design and payload development.

At the center of any scalable prototyping system is a reconfigurable multi-

processing CPU engine with associated memory. The system depicted in

Fig. 1 provides developers a scalable environment for application design

and test. A COTS single board computer (SBC) tightly integrated to a PMC

FPGA card for design experimentation forms the core system. Other PMC

modules can then be selected depending on the type of sensor input that

needs to be processed.

One of the main advantages of this approach is the ability to rapidly proto-

type mockups for test and evaluation without a concern for the limitations

of embedded development at this early stage. The COTS SBC shown has the

ability to host two PMC modules per card. The Video card can efficiently

manage real sensor input coming in from an E/O sensor/camera, and also

display that data in a fused fashion. This approach allows the system

builder to work in the lab and the field using the same hardware/software

environment. This system configuration also provides the developer with a

test bed to define/design new hardware requirements and processing

streams.

3. Prototyping System Structure

Modern embedded avionics sensor design is primarily driven by the goal of

providing a net-centric flow of data from platform to platform. The achieve-

ment of this vision depends on transferring and processing vast amounts of

data from multiple sources. Let’s examine how one could use this approach

to control and manage various sensors in a rapid prototyping environment.

As depicted in Figure 1, the host server for the embedded avionics sensor

design workstation is a COTS Dual 7447A PowerPC VMEbus SBC. The board’s

architecture provides a distributed processing environment that allows

scaling of multiple processing nodes to achieve high performance for the

most demanding signal and imaging applications associated with various

sensor and mission payload design requirements.

Fig. 1. FPGA vision and graphics platform

The COTS SBC architecture combined with the FPGA/Video cards allows for

seamless mapping of imaging applications oriented toward change detec-

tion and sensor fusion which will allow the systems designer to view

multiple data streams in a simulations real time display environment.

The SBC implements processing nodes using the latest Motorola

7447A/7448 PowerPC® processors running at up to a 1.4 GHz clock rate.

Using the distributed processing architecture of the SBC, one bridge chip

per node allows the Front Side Bus (FSB) of each PowerPC to run in MPX

mode up to its maximum rate of 133 MHz. The high performance data

transfer mechanism facilitated by the built-in 64-bit full duplex crossbar in

the Discovery II bridge permits concurrent data transfers between different

interfaces as well as transaction pipelining for same source and

destination transactions.

The prototyping architecture outlined here is based on a combination of

tightly integrated reconfigurable computing, video and graphics. It will

place the embedded avionics sensor design community in position to utilize

the proposed system architecture in development of the actual controller

for payload packages as well as the sensor itself and in its actual deploy-

ment integration. Most importantly it will allow for cost-effectively

maintaining and extending the system when new technologies are avail-

able in the future.

3.1 An Example: Infrared (IR) scene projection

Increased sensor performance and bandwidth has begun to overwhelm

existing backend processing capability. Current test and evaluation

methods are not adequate for fully assessing the operational performance

of imaging infrared sensors while they are installed on the weapon system

platform. However, the use of infrared (IR) scene projection in test and

evaluation can augment and redefine test methodologies currently being

used to test and evaluate forward looking infrared (FLIR) and imaging

IR sensors.

One could project accurate, dynamic and realistic IR imagery into the

entrance aperture of the sensor, such that the sensor would perceive and

respond to the imagery as it would to the real-world scenario. This

approach includes development, analysis, integration, exploitation, training,

and test and evaluation of ground and aviation based imaging IR

sensors/subsystems/systems. This applies to FLIR systems, imaging IR

missile seekers/guidance sections, as well as non-imaging thermal sensors.

The systems approach proposed in this paper has the scalability to accom-

plish this type of reconfigurable sensor mix and match. Algorithms such as

the one depicted in Figure 2 could be transformed into a Mathlab®

Simulink® Blockset for easy execution in a reconfigurable system utilizing

FPGA- and PPC-based computing elements. Embedded avionics

HWIL Simulation demands include low latency, high data rates and

interfacing. It is essential to have a capable platform for handling and

processing of the data streams. Tools must also complement this so that a

systems designer is able to construct the final system leveraging design

tools, such as Mathlab and Simulink, with a reconfigurable

computing platform.

Fig. 2. Example: sensor images

This approach allows one to demonstrate how algorithms can be imple-

mented and simulated in a familiar rapid application development

environment before they are automatically transposed for downloading

directly to the distributed multiprocessing computing platform. This

complements the established control tools, which usually handle the

configuration and control of the processing systems leading to a tool suite

for system development and implementation.

3.2 The advantages of FPGA Computing

As the development tools have evolved, the core-processing platform has

also been enhanced. These improved platforms are based on dynamically

reconfigurable computing, utilizing FPGA technologies and parallel

processing methods that more than double the performance and data

bandwidth capabilities. This offers support for processing of images in

Infrared Scene Projectors with 1024 X 1024 resolutions at 400 Hz frame

rates. Simulink Blocksets could add the programming involved to organize

algorithms that could then be partitioned to operate on a multiprocessor

based FPGA/PPC based configuration.

Another key component of reconfigurable scalable embedded avionics

sensor design and prototyping capability is access to FPGA based

processing. FPGA (Field Programmable Gate Array) is defined as an array of

logic blocks that can be ‘glued’ together (or configured) to produce higher

level functions in hardware. Based on SRAM technology, i.e., configurations

are defined on power up and when power is removed the configuration is

lost – until it is ‘reconfigured’ again. Since an FPGA is a hardware device, it is

faster than software.

The FPGA can best be described as a parallel device that makes it faster

than software. FPGAs as programmable “ASICs” can be configured for high

performance processing, excelling at continuous, high bandwidth applica-

tions. FPGAs can provide inputs from digital and analog sensors —LVDS,

Camerink, RS170 — with which the designer can interactively apply filters,

do processing ,compression, image reconstruction and encryption time of

applications. Examples of the flexibility of this approach using COTS PMC

Modules hosted by a COTS multiprocessing base platform are shown in

Figures 3 & 4.

Fig. 3. Example: FPGA application. PMC FPGA processor for capture and

compression. PMC module for graphics and display.

Fig. 4. RS-170/MPEG-4 video PMC. Integrated bundle of PMC, mezzanine and

MPEG-4.

3.3 The ‘Mix and Match’ Platform Advantage

Today’s soldier, who is the ultimate customer of the embedded systems

designer, is faced with a continuously fluid chain of world events. These

changing events are closely mapped into the deployment of various air, sea

and land platforms which contain the reconfigurable embedded systems

architecture under discussion in this paper. For example, based on

changing mission requirements, any of the following three different areas

might be needed by the soldier: sonar processing; SIGINT; or

video surveillance.

The system designer addressing these defined application areas could

place various PMC modules together in the prototyping system with relative

ease. Specifically, these applications might collectively require the following

list of PMC modules: PMC #1—Graphics; PMC#2—Video Compression;

PMC#3—Reflective Memory; PMC#4—1553; PMC#5—High Speed Data Acqui-

sition; PMC#6—Race++. A key cornerstone of the ‘mix and match’ strategy is

that while the PMC module may change from application to application, the

core SBC ‘multiprocessor’ engine and its associated software tool chain

remains constant. The same can be said for the VME enclosure that

contains the processing cards.

Let’s now examine the application areas and understand how we could

place various PMC combinations together to address the processing

requirements of each application.

3.4 Sonar Processing: PMC #1—Graphics, PMC #5

—High Speed Data Acquisition

In this example the PMC Graphics card would be utilized to display sonar

waterfall display data perhaps with an overlay of tactical positions. The

High Speed Data Acquisition PMC could be used to facilitate the information

coming in from a high speed sensor interface. The combination of the two

modules hosted by the multiprocessor-enabled SBC could move

from this application area to another type of sonar processing, triggered by

software or the base modules could be switched to those shown in the

SIGINT example below.

3.5 SIGINT: PMC #6—Race ++, PMC #3—Reflective memory

In a SIGINT application scenario a RACE ++ PMC module is tied to other

Race++ peripherals allowing for interfacing while the reflective memory

module stores real time acquired data that could be used for post

processing data analysis.

3.6 Video Surveillance: PMC #2—Video Compression, PMC #4—1553

In the video surveillance application area, the video compression PMC

module would pre-process and reduce the incoming data stream and pass

it to the multiprocessor base system for potential change detection

analysis. Then using the PMC 1553 module, one could communicate to an

external avionics platform to perhaps control or guide ordnance

being placed upon the target under surveillance.

In each of the three examples above, the modules are interchangeable

depending upon the specific mission. This approach would allow the soldier

to move module packages between platforms achieving different mission

scenarios through software loading while maintaining the core base multi-

processing and packaged environment.

4. Conclusions

The goal of integrating a COTS system such as the one depicted here is to

allow designers to use the same environment in the lab that they could

then take to the field for live data collection activity. This is of particular

value to a system engineer who could use such an environment to perform

new development activities. Because of these efficiencies, the

outlined prototyping system would pay off in accelerated product develop-

ment time.

Systems designers are traditionally faced with the challenge that today’s

sensors are generating data at a rate far faster than the backend end

systems are configured to process. Combine this fact with the reality that a

sensor fusion “paradigm” is now mandatory for designers wishing to turn

concepts into reality rapidly. A key component to a flexible hardware

system, of course, is a software structure that enables designers to go from

their ideas and algorithmic concepts to code or HDL.

Packaging of the system is largely dependent on the user’s requirements

for flexibility. Should the user desire a system that can be scaled up by

adding additional cards, then a larger slot chassis could be configured to

allow for the addition of other cards. The key point is that the core

hardware outlined not only has the potential for scalability by adding addi-

tional modules to the base processing units, but the entire system has

scalability as well.

PPC toPCI-X Bridge

(64360)

PPC toPCI-X Bridge

(64360)

1 GHz7448 or7447A

PowerPC

32 MBUser Flash

32 MBUser Flash

VME BusP1

Up to 320 MB/s

Peak

Up to 1064 MB/s

Peak

PMC 1

PMC 2

PCI-XBridge

PCI-XBridge

NVRAM NVRAM

256 MB DDR 256 MB DDR

64-bit/133 MHz

64-bit/133 MHz

GigabitEthernet

MultifunctionSerial

GigabitEthernet

MultifunctionSerial

Up to 1064 MB/s

Peak

64-bit/133 MHz

64-bit/133 MHz

2eSSTVME Bridge

1 GHz7448 or7447A

PowerPC

OPTICAL FILTER

FILTER WHEEL

INTENSIFIER

FOCAL BACKPLANE

INTENSIFIERHIGH VOLTAGE

POWER SUPPLY

MICROCONTROLLERELECTRONICS

CCD IMAGING CHIPAND VIDEO ELECTRONICS

Video Rate Camera• 30 frames per second

• Average image size (after registration):

642 x 343 pixels (8 bit)

Spinning Filter Wheel• Creates 6-band multispectral image

Representative Sensor

Graphics PMC ModuleHigh Speed Data Acquisition PMC ModuleMultiprocessor BSPGraphics SW Layer (OpenGL)Adding Applications Layer

Mix & Match Platform Advantage Example 1Sonar Processing

SIGINT Reconfigurable System Environment

RF Receiver

Real-Time Capabilities

Real-time Signal DetectionSignal TrackingOperator Event MarkingSignal ClassificationFusion of Data

Incoming Information Pre-Processed-Processed & Post Processed and stored for future analysis

Digital Capture ofSignal Data

Near-Real-Time CapabilitiesAutomatic DetectionSignal Localization

Display and Analyze ResultsOperator in the Loop

Data Tapes

Processing & analysis may be performed on the ground at near-real-time rates by transfer of information from platform to ground station.

Signal Analysis Function (SAF)

Mix & Match Platform Advantage Example 3Video Surveillance

1553 PMC Video Compression PMCMultiprocessor BSPAdding Applications Layer

Battle Damage Assessment

Electronic Surveillance

Cross-cueing Sensors

FOPEN Radar

Minefield Detection

LADAR

Data Link Compression

All Weather Target Acquisition

Sensor Fusion

Mission Planning

UAV Application Examples

Existing Platforms & New Designs Will Continue Flying For Decades

Incremental Mission Payload Responsibilities Will Be Added

Requirements For Interoperability Continue To Increase

Developers Will Need To Shrink Payload Processing System

Conclusions

Overcoming Data Compression Bottlenecks

Pentium enabled SBC’sVMIC Product Line

PPC enabled SBC’s single,dual & quad configurationsRichardson Product Line

Display Capability CDI Product Line

Storage CapabilityCamarillo Product Line

1553 Data link

Step 3

The process of the first two steps allows the system developer to combine real world data with simulated scenarios and compare and contrast in order to hone overall system design efficiency

Step 1

Data acquired in real time

Step 2

Data is passed to PMC module to be

processed & on Nexus Single Board

Computer for backend processing &

matching for decisions (mission driven

scenarios)

PMC Module

Collect & Process Data from suspension subassembly

Data Acquisition Sensor

Solving The Bandwidth Problem

PPC toPCI-X Bridge

(64360)

PPC toPCI-X Bridge

(64360)

1 GHz7448 or7447A

PowerPC

32 MBUser Flash

32 MBUser Flash

VME BusP1

Up to 320 MB/s

Peak

Up to 1064 MB/s

Peak

PMC 1

PMC 2

PCI-XBridge

PCI-XBridge

NVRAM NVRAM

256 MB DDR 256 MB DDR

64-bit/133 MHz

64-bit/133 MHz

GigabitEthernet

MultifunctionSerial

GigabitEthernet

MultifunctionSerial

Up to 1064 MB/s

Peak

64-bit/133 MHz

64-bit/133 MHz

2eSSTVME Bridge

1 GHz7448 or7447A

PowerPC

RS170 PMC

Graphics PMC

FPGA PMC

1553 Technology

Chassis Technology

Display Technology

VideoCompression

PMC

InputFormat

MatrixProcessor

DemeanWindowAverager

TargetWindowAverage

Background/Guard

WindowAverager

CrossProduct

Generator

BandInterleaved

ImageInput

RX ImageOutput

Video InputProcessing

Video Server Application

WAVESoftwareDrivers

(PCI Bus)

MPEG-4Bit-stream

Input 1MPEG-4 Compression

Input 2MPEG-4 Compression

TS-PMC with Stratix FPGA

RTP Encapsulated MPEG-4 Bit-stream

Server Configuration via TCP/IP

Video Server CPU

Dual RS-170MezzanineFPGA PMC

Video Compression PMCIntegrated Bundle

Video Compression Encoderwith Host Decode

Multiprocessor SBC

+ =+

Video CameraAudio Input

• Capture• Filter, Format• Encrypt• Compress

GPU PMCOverlay

andDisplay

TS-PMCCapture

andCompress

Video CameraAudio Input

TS-PMCCapture

andCompress

TS-PMCCapture

andCompress

TS-PMCCapture

andCompress

• Capture• Filter, Format• Encrypt• Compress

• Graphic Generation• Overlay• Display

• Graphic Generation• Overlay• Display

Thin Pipe Network

Encrypter WirelessNetwork

Multiprocessor SBC

Race ++ PMC ModuleReflective Memory PMC ModuleMultiprocessor BSPAdding Applications Layer

Mix & Match Platform Advantage Example 2SIGINT