• Keine Ergebnisse gefunden

Fusion of IR/CCD video streams and digital terrain models for multi target tracking

N/A
N/A
Protected

Academic year: 2022

Aktie "Fusion of IR/CCD video streams and digital terrain models for multi target tracking"

Copied!
9
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Fusion of IR/CCD video streams and digital terrain models for multi target tracking

Kaeye Dästner , Bastian Köhler, Felix Opitz Defence & Communications Systems

Integrated Systems / Air Dominance & Sensor Data Fusion Engineering (ISEG42) Wörthstraße 85

89077 Ulm

{kaeye.daestner, bastian.koehler, felix.opitz}@eads.com

Abstract: Video Streams of optical camera system are analysed by a video processing software which detects moving objects in the video. They are presented in image pixel coordinates. With the knowledge of the direction of the camera and the field of view this information can be georeferenced using maps of Digital Terrain Elevation Data (DTED). The output are plots in polar coordinates which are processed with a sensor tracker using IMM techniques.

1 Introduction

In military and civil surveillance systems IR and CCD sensors have become more and more important. Such systems allow the operator to observe the area of interest visually at day and at night and allows him to steer the camera to a region of interest manually.

So optical sensors can be used to verify tracking results of other sensor systems like radars and give the operator the ability to classify the object on his own to evaluate the current situation.

But for an efficient usage of these surveillance sources an automation is required which helps the operator to manage the sensors and give him more accurate information. Since image information from a streamed video is just presented as the camera direction only, several different objects in the image can have very different distances from the camera position. With the knowledge of the direction of the camera and the field of view this information can be georeferenced using maps of Digital Terrain Elevation Data (DTED).

The output are plots in polar coordinates which are processed with a sensor tracker using IMM techniques. Finally, synergies are established between video and DTED information which helps the video processing to increase its performance due to the availability of range information.

Working with an intelligent sensor management radars and optical sensor systems can complement one another and therefore receive best results of the surveillance ground, naval and air picture and to generate a joint operational picture [OFD07], [OKD07], [Op08], [ODK09].

(2)

This paper presents the fusion of video stream information with digital terrain data and is structured as follows: section 2 describes the software architecture of the system and section 3 presents the first test results.

2 Architecture

Four main software components are used to create a georeferenced view of the camera image. Fig. 1 gives an overview of the architecture and the used components:

1. video processing, detection and tracking 2. georeferencing

3. sensor tracking

4. visualisation with a display area

Fig. 1: Software architecture of the georeferenced tracking system.

2.1 Video Streaming

There are two ways to consume a video signal before it is prepared for further video processing and analysis:

• Digital: the video stream is MPEG encoded and provided by a server via Real- time Transport Protocol (RTP)

• Analogue: the video stream is provided as analogue signal

If the stream is consumed as MPEG from a RTP server the stream has to be decoded again after reception. The decoding extracts single frames from the stream which are analysed afterwards.

But to ensure real time tracking the second solution was preferred. MPEG encoding and decoding can cost up to several seconds and also a loss of quality has to be accepted then.

(3)

Before the video processing is able to work with the analogue signal it has to be digitalised first. Therefore a FireWire interface was used to handle the conversion to a digital signal and Video4Linux to extract PAL frames from the video stream [Wa04], [WOZ01].

2.2 Video based Detection and Tracking

A video based detection and tracking software is used to identify moving objects in a scene, e.g. a landscape with moving vehicles and pedestrians, that is observed by an IR sensor. The software detects these objects automatically and delivers their pixel coordinates in the video picture.

The software operates with detections of changes between at least two image frames, which are substracted from each other pixel wise. The result is an image again with the difference as content. Also denoising techniques e.g. Random Markov Fields are applied to improve the original and meanwhile the difference image [Br01], [LMR98].

First the software has to identify the changes in a frame. Optimization can be done by integrating several frames to learn the background, which can then be substracted from an image instead. The Bi-level thresholding method improves the results where a low threshold neglects a detected change and high threshold amplifies it. So a better contrast is received and edges become more clear. [AR05]

Second the software has to cluster pixel level changes to form the objects of interest. An iterative process with minimization technique segmentises these cluster to segments.

Parameter as minimum estimated object size in pixel is essential for tuning purposes.

[AR05], [WOZ01].

Fig. 2: Principle of video based detection & tracking. From left to right: raw image, difference image, segmentation image, track image.

(4)

2.3 Image Georeferencing

Fig. 3: Image with vehicle and its pixel coordinates. The software returns the information of the (x1, y1) and (x2, y2) coordinates.

If an object is detected within one video frame it is described in image coordinates of the centre of this object in pixel. With respect to the current platform parameter, given by the azimuth φ, the elevation ε and the current field of view (FOV) α, the real target direction can be calculated with:





 −

+

= 2

' 1

xmax

x αϕ

φ φ

where x is the horizontal image position of the image with the width xmax. The same equation can be used for the elevation vector with the values ε, y and ymax respectively.

Fig. 4: Mathematical schema for the determination of the azimuth.

(5)

2.4 Image Georeferencing with DTED

Digital Terrain Elevation Data allows the range estimation of an object that is moving on the ground. One has to know the sensor position in WGS84 and the target direction. By calculating the intersection between the direction and a DTED layer the range can be found [Ba09]. Fig. 5 shows this principle.

Fig. 5: Principle of Georeferencing a direction vector using DTED.

The elevation accuracy has a great influence on the range estimation. The following estimation gives an impression of the idea to obtain range information by calculating the cross section between the direction vector and a planar ground. The distance d can be calculated using 

 

=  d tan h

ε where h is the camera height and ε the elevation. In particular since the elevation is very low (d >> h) small elevation differences can lead to great range differences. Fig. 6 describes this dependency between elevation and range.

Therefore measurements in the near field are more accurate and the camera system has to be aligned very exact to achieve similar results for the wide field.

0 0,5 1 1,5 2

0 2000 4000 6000 8000 10000 12000 14000 16000 18000 Distance [m]

Elevation [ °]

Fig. 6: Elevation vs. Distance calculated with a camera position height 35 m.

(6)

2.5 Sensor Tracking

Once the extracted image is georeferenced the object can be tracked. The tracker input is given as a three dimensional plot in polar coordinates. The update rate is coupled with the frame rate which is up to 25 frames per second for the PAL format.

Fig. 7: Screenshot of the Fusion++ Tracker Configuration

We used a sensor tracking configuration of our fusion system Fusion++ [DKO07a], [DKO07b], [DKO07c], with a nearest neighbour algorithm for association [BP99]. To support the association algorithm and prevent a false plot to track correlation each track was locked at least for 100 ms between two possible updates. After 2s a not updated track was deleted.

Ground surveillance has to deal with pedestrians, animals and vehicles. Their dynamics are rather difficult to describe because they can make a complete turn or jump to the sides instantaneously. This is a behaviour that is not covered by the standard extrapolation models using constant accelerations that are used in normal filter techniques.

To handle such motion models we developed an Interactive Multiple Model (IMM) with two filter [BP99]. One with a normal Kalman Filter (KF) that is responsible for vehicle movement and a Low Velocity Filter (LVF) for random movements of pedestrians.

Fig. 8: Screenshot of the Fusion++ IMM configuration with a Kalman Filter and Low Velocity Filter.

(7)

3 Evaluation

Fig. 9: Left: Image of the test scenario with two detected objects. Right: Tracking results in relative x-y coordinates.

We tested the software under real environmental conditions and with real sensor systems. The tracking itself works good if the target is moving radial but has great fluctuations while moving tangential to the camera direction (Fig. 9). The reason for this behaviour is obvious: the detected objects of the video processing fluctuate in the pixel coordinates, but the influence of the y-coordinate is major because it is used to calculate the elevation. As discussed in 2.4 small changes in the elevation lead to great changes in the range estimation.

Fig. 10: Left: IMM Model Probability: Model 0 is the LVF, Model 1 the KF. Right: Speed of one tracked vehicle. In the beginning of the scenario the vehicle slowed down.

Near the sharp bend the vehicles slowed down and then speed up again, so the velocity got in the range of the LVF. The probabilities of the two models of the IMM demonstrate this very well (Fig. 10). During the complete scenario the Kalman Filter (Model 1) is the dominating model but in the curve the probability of the LVF model

(8)

increases rapidly. This corresponds with the recorded speed in Fig. 10 and is a confirmation of our expectations.

4 Conclusion

The tests of the software were very successful and proved that it is possible to obtain 3D information from an image in general as long as the camera parameter are known and sufficient DTED material is available. But track fluctuations of tangential moving targets have to be reduced in future researches with respect to jumps of the y-coordinates in the video processor software.

Also the usage of an IMM with the Low Velocity Filter turned out as an effective approach which needs more fine tuning investigations in the future. But in first runs it transpired that a simple filter like the LVF is the better solution to track random moving targets than only one Kalman Filter.

These first tests were made at a near distance of 3 km. The next step is to test this system for a wide field of e.g. 10 km in the future. But this is more difficult for two reasons: first one has to find an area where the camera can observe this distance and second the camera has to be aligned very exactly to achieve similar good results.

References

[AR05] Tinku Acharya and Ajoy K. Ray, Image Processing Principles and Applications, Wiley, 2005

[Ba09] Norbert Bartelme, Geoinformatik: Modelle, Strukturen, Funktionen, Springer, Berlin, 2009

[Br01] Pierre Bremaud, Markov Chains - Gibbs Fields, Monte Carlo Simulation, and Queues, Texts in Applied Mathematics , Vol. 31, Springer, 2001.

[BP99] Samuel Blackman and Robert Popoli, Design and Analysis of Modern Tracking Systems, Artech, Norwood, MA, 1999

[DKO07a] Kaeye Dästner, Thomas Kausch, Felix Opitz, An Object Oriented Development Suite for Data Fusion: Design, Generation, Simulation and Testing, Fusion 2007, Québec, Canada, 9-12 July 2007

[DKO07b] Kaeye Dästner, Thomas Kausch, Felix Opitz, A Development Suite for Complex Data Fusion Systems Sensor Suites and Data Fusion in Anti Asymmetric Warfare, International Radar Syposium IRS 2007, Köln

[DKO07c] Kaeye Dästner, Thomas Kausch, Felix Opitz: "An Object Oriented Approach for Data Fusion". GI Jahrestagung (2) 2007, Bremen

(9)

[LMR98] Alfred K. Louis, Peter Maaß, Andreas Rieder, Wavelets. Theorie und Anwendungen, Teubner Verlag,1998

[OFD07] Felix Opitz, Josef Filusch, Kaeye Dästner, Thomas Kausch, Information Fusion in Anti Asymmetric Warfare and Low Intensity Conflicts, Fusion 2007, Québec, Canada, 9-12 July 2007

[OKD07] Felix Opitz, Guy Kouemou, Kaeye Dästner, Thomas Kausch, Sensor Suites and Data Fusion in Anti Asymmetric Warfare, International Radar Syposium IRS 2007, Köln [Op08] Opitz, F. Sensor Suite and Data Fusion in Security Applications. SET-125 Symposium

on “Sensors and Technology for Defence Against Terrorism (DAT)” Bundesakademie für Wehrverwaltung und Wehrtechnik, Mannheim, Germany, 22-25 April 2008

[ODK09] Felix Opitz, Kaeye Dästner, Bastian Köhler, Guy Kouemou. Sensor Integration in the Security domain, GfI 2009

[Wa04] John Watkinson, The MPEG Handbook. MPEG-1, MPEG-2, MPEG-4 (MPEG-4 Part 10/H.264/AVC included), Butterworth Heinemann, 2004

[WOZ01]Yao Wang, Jörn Ostermann, YA-Qin Zhang, Video Processing and Communications, Prentice-Hall Signal Processing Series, Prentice Hall, 2001

Referenzen

ÄHNLICHE DOKUMENTE

In  the  sec tions  that  fol low,  we  be gin  by  outlining  the  con cep tual 

Polygons resulting from the intersection of terrain, soils, and inland water bodies were deleted when the corresponding area was less than 625 km 2 , i.e., less than 1cm 2 on the

The implications of the study are ethical, since they display participants' expectations regarding informed consent, and how they continuously achieve it in their interactions as

En las dinámicas socioculturales de la Región de Antofagasta, se materializan conjuntos de signos y símbolos que adoptan representaciones interesantes sobre su connotación

The data points which scatter within our experimental resolution are in excellent agreement with the above results obtained from calculated particle images and show the same

Nonaka and Takeuchi (1995) have developed the knowledge spiral model in 1995 to show the interaction between tacit and explicit knowledge in an organization for socialization,

All these systems should not be considered alone, but in their role in the agricultural production process on the field and in the digital agricultural ecosystem: what

The alternative commercial offer preferred by the industry, namely the purchase of music titles (especially downloads) in legal online music stores such as iTunes, was