• Keine Ergebnisse gefunden

1.4 Overview

2.1.1 Detectable Coverage

In this section, the fused coverage exemplarily resembles the combined field of view of several cameras.

The fused coverage is defined in Definition 1.2.1 as the union or intersection of the camera coverages.

The coverage of a single camera is a preimage of the sensor mapσ : E×P×A → S. First, each of the following sets is specified: the set of the environmentsE, the set of sensor network parametersPthat are utilized as variables in an optimization, the set of surveillance partsAwhich contains the points that need to be monitored, and the set of sensor labelsS. In the end, we show how to assign a sensor label to each point in the surveillance area.

Environment and Parameter Space of a Single Camera

An environment E ∈ EwithE ⊂ P(R3) is the space where the camera will be placed, it contains the parts that need to be monitored and models visible restrictions like walls and furniture. In this thesis, the environment resembles theclosed empty space of a room, which is not occupied by furniture or walls but only by transparent, non-solid matter. The model may seem contra-intuitive, however, there are two reasons for it: Firstly, a camera can just be placed in an empty space of a room. Secondly, the boundary between empty space and non-empty space of a room, i.e. informally the surfaces of furniture and walls, defines whether a point in the environment is visible from camera point of view.

At first, the empty space is modeled as a polyhedral area. Polyhedral Terrain Models (PTM) are often used to represent a terrain or topographic surface in a visibility analysis, [31, 101]. In [44] an informal definition of a polyhedron is given: Aconvex polyhedronis an intersection of finitely many closed half spaces. Apolyhedronis the union of finitely many convex polyhedra. In this thesis, such a polyhedron is considered in the context of vertices, edges, and faces:

Definition 2.1.1

Letn,m∈Nbe integers withn≥m≥1. LetH(q,p) ⊂Rn,q=1, . . . ,Q1,p=1, . . . ,Q2be half-spaces.

1. A finite concatenationP:= S

p=1,...,Q1

T

q=1,...,Q2

H(q,p)is called

(a) mD-polyhedral areainnD if anm-dimensional affine subspaceA⊂Rnexists withP ⊂A.

(b) mD-polyhedroninnD ifPis anmD-polyhedral area and anm-manifold with boundary P.

2. LetFiP,i∈ {1, ...,I},I ∈Nbe









points in casem=1

(m−1)D-polyhedra in casem>1 If P=S

i∈{1,...,I}Fiand if for all the pairs (Fi,Fj),i,j=1, . . . ,I the intersectionFi∩ Fj ⊂ Ais maximally a (m−2)D polyhedron then theFiare calledfacesof themD-polyhedral area.

3. A 2D-polyhedron is calledpolygon, its faces are callededges. The face of an edge is calledvertex.

4. The set of all 3D-polyhedral areas in 3D is defined asE.

A polyhedron is am-manifold, which means each point has a neighborhood that is homeomorphic to the Euclidean space of dimensionmor a half-space of dimensionm, again this means that it is nowhere flat or pinched [44] and neither is one of its faces. In contrast to the polyhedron, anmD-polyhedral area is not necessarily a manifold. A polyhedral area is not necessarily convex, it can have holes of dimension mand it can be distributed to several regions. The “mD” stands for its dimension in the spaceRn, which will only be needed for polyhedra with a lower dimensionm ≤ n, e.g., a polygon in a 3D space. The environment in which the cameras of the network are placed is 3D-polyhedral:

Notation 2.1.2

In this thesis, theenvironment Ewhere the camera network is placed constitutes a 3D-polyhedral area:

E∈E.

In other words, an environment is a not necessarily convex polyhedron potentially containing three-dimensional holes, in fact not necessarily being connected, at all. Be aware that the boundary of the environment does not contain two-dimensional holes, it is a two-dimensional manifold without boundary.

The cameras are placed in such an environment.

Definition 2.1.3

LetE∈Ebe an environment and letp∈E. Leto,u∈ B31(0) be orthonormal unit vectors. Thecamera parameters in the environmentE are defined as a 5-tupel (p,o,u, θu, θo×u) ∈ (E × B31(0)× B31(0)× [0, π]×[0, π]) consisting of the following:

• The point pis calledcamera position, sometimes alsoviewpoint.

• The vectorois calledcamera orientation.

• The vectoruis orthonormal too. Ifuis a linear combination ofoandez = (0 0 1)T it is called view-upvector.

2.1. COVERAGE OF MULTIPLE CAMERAS 21

• The scalarsθu, θo×u∈[0, π] are called theopening anglesin directionowithopening vector uand the vector orthogonal tooanduwhich is (o×u).

• The tuple (λo, λu, λo×u) withλo, λu, λo×u ∈ Ris called camera coordinatesof the pointx = p+ λoo+λuu+λo×u(o×u).

Figure 2.2 (left) illustrates the camera coordinate system with the parameter vectors p,u ando of the camera. In this thesis, the position p ∈ Eand orientationo,u ∈ B31(0) of each camera are considered as variables to the objective function in an optimization. The opening angles θo×u ∈ [0, π] and θu ∈ [0, π] are fixed. In case of u being the view-up vector, u points to the ceiling and can be derived by u= o1+1o2

o1o3 o2o3 (o1+o2)2

with (o1+o2),0 from the orientationo=(o1,o2,o3)T. Otherwise, the rotation between the view-up vector andudefines the rotation of the image plane of the camera about the orientation and can be determined by a single angle. The rotation angle of the image plane is called roll. In our investigations, however, we assume the roll to be fixed at 0 in favor of a clearer notation.

p

o

θu

θo×u

u o×u

d θ

v ρ

Td(θ,v)

Figure 2.2: Left: Illustration of parameters of a camera in 3D: Position p, orientation o, and opening anglesθuo×ufor the opening vectoruando×u; The camera frustum is an intersection ofθ-spaces. Right: Td(θ,v) is theθ-space with angleθ < π, directiond, and openingv.

Thus, the investigated space of parameters of a single camera that serve as variables for an objective function is

P⊂ E× B31(0). (2.1)

The camera parameters are now used to define the visibility as already promised in the last paragraph.

Thereby, we use the term θ-space as a halfspace with an angle smaller than πillustrated in Figure 2.2 (right), to define the camera frustum in 3D.

Definition 2.1.4

LetE∈Ebe an environment. Leta=(p,o)∈Pbe camera parameters as in Inclusion (2.1).

1. A pointy∈R3is calledvisible in Efrom pointp∈Eif theline of sight[p,y] holds [p,y]⊂E.

2. V(p,E) :={y∈E |yvisible inEfromp} ⊂Eis called theset of all visible pointsfromp∈E.

3. Letθ ∈ [0, π] be an angle. Furthermore, letd,v ∈ B1(0) be orthogonal to each other. And let ρ(d,v)be the projection onto thev-d-hyperplane. Then aθ-spaceis defined as

Td(θ,v) :={x∈R3| |](ρ(d,v)(x),d)| ≤ θ 2} dis called thedirectionandvtheopeningof theθ-space.

4. Letθu, θo×u∈[0, π] opening angles of the camera. Thefrustumof the camera with parametersais the set

Fa :={y∈R3|(y−p)∈(Tou,u)∩Too×u,(o×u))} 5. Va(E) :=V(p,E)∩Fais called thefield of viewof the camera with parametersa.

In literature, the setVa(E) has several names, e.g., it is called “field of view” (FoV) by [2] or “viewshed”

by [101]. Also further names like “camera beam” are thinkable. Nevertheless, it is not to be mistaken for the “depth of field”, giving the distance between the furthest and nearest objects that appear in acceptable sharp focus, [48]. For the setFathe name “camera cone” [70] exists next to the name “viewing frustum”

[48]. The opening angles of a camera frustum are not limited to values smaller than π. However, if opening angles larger thanπare used thenFais not an intersection but a union ofθ-spaces.

Set of Sensor Labels and Surveillance Parts

Usually, not all the points in an environment need to be monitored by the sensor network. Additionally, the environment is sometimes rasterized into small cubes, calledvoxels. By using voxels the set of points that need to be covered can be discretized, e.g. the set encompasses the center or corners of each voxel.

LetA⊂Edenote the surveillance area of the environment, i.e. the points that need to be covered.

With a given position and orientation of the camera (p,o)∈Pwe are able to partition the environment as well as the surveillance area into detectable and undetectable regions. This is done by assigning a label ofSto each point inEby the mappingσ. The following definition shows which labels inSneed to be used.

Definition 2.1.5

Let E ∈ E be an environment and A ⊂ E be the surveillance area. Let a = (p,o) ∈ P be camera parameters as in Inclusion (2.1).

1. A pointy∈Ais calledout of rangeof a camera with parametersaify<Fa.

2. A pointy ∈ Ais calleddetectableby a camera with parameters aify ∈ Va(E). Otherwise it is undetectableby the camera with parameter vectora.

3. A pointy∈Ais calledoccludedify<V(p,E) andy∈Fa.

Let the set of sensor labels beS:={detectable, occluded, out o f range}. With the given environmentE and given camera parametersa∈Pincluding the position of the camerap, we can classify whether a part y∈Ais detectable, occluded, and out of range by the mapσ:E×P×A→S. This map is illustrated in Figure 2.3.

2.1. COVERAGE OF MULTIPLE CAMERAS 23

Inside frustum?

In front of all faces of the environment?

detectable yes

occluded no

yes

out of range no

| {z }

undetectable

Figure 2.3: Tree of surveillance area division: The gray nodes categorize a point in the surveillance area according to the camera position an orientation. The colored nodes correspond to the labels inS ⊂S. Blue corresponds to a label which a point outside of the viewing frustumFais marked with. A red labeled point is inFabut not in the set of visible points V(p,E). A point labeled in white is in the field of view of the camera.

Let N ∈ Nbe the number of cameras in the network and let x = (a1, . . . ,aN) ∈ P1 ×. . .×PN be the variable vector of the camera network. In an optimization, the detectable regions of the surveillance area can be maximized in order to increase the efficiency of a camera network. One example of sensor network optimization which can be derived from Equation (1.1) and Definition 1.2.1 can be denoted by

x∈Pmax1×...×PN

λ(Cx(S)) withS ={detectable}.

where the fused coverageCis defined by the preimageσ−1(S) of the sensor map as in Definition 1.2.1 andλ(·) denotes the volume of a set. The setCx({detectable}) is called thedetectable coverage, further on. In addition to the detectable coverage, in the next section, the notion of the fused coverage is used to express the silhouette of a human.

2.1.2 3D Background-Subtraction Method

In this section, the fused coverage exemplarily resembles a human silhouette. For constructing a sil-houette of a human by one camera, theBackground Subtraction Method(BG subtraction) can be used.

This is a method to detect changes in an environment. The method distinguishes between regions of the environment which have changed since the moment a reference image has been taken and regions that remained unchanged. Additionally, regions are considered which are out of range or occluded behind objects, like walls, doors, tables, and racks. Therefore, let the set of sensor labels of Definition 1.2.1 become

S:={identical,changed,occluded,out o f range}.

In order to explain the method further, it is necessary to distinguish the objects in an environment, [64]:

Definition 2.1.6

LetE∈Ebe an environment.

• A set of faces ofEis calledobjectif it is the boundary of a polyhedral area.

• Faces ofEthat do not change in time are calledstatic, all other faces are calleddynamic. The set of static faces will be denoted byEsEand the set of dynamic faces byEdE.

In the following investigation, assume that Es and Ed are static and dynamic objects. With this defi-nition, the changes of an environment are directly depending on the dynamic objects. The method of background subtraction detects these dynamic objects in the following way: The camera takes an image of the environment in absence of any dynamic objects in the environment, this image is calledreference image at timet0 ∈ R, say. Such an image is illustrated in Figure 2.4 (left). The image is rasterized into small squares calledpixel. The boundary of the environment at this time is just consisting of static objects Es. The pixel value of an image that is taken at any moment of timet> t0 aftert0is subtracted from the value of the pixel in the reference image at the same spot. The result of each pixel can be saved in an additional image, calledsilhouette imagein [107]. In Figure 2.4, an image showing the additional dynamic objects (middle) and a silhouette image (right) are illustrated.

If a pixel value of the silhouette image is zero, the part of environment boundary which is mapped onto the pixel has not changed, otherwise a dynamic object has appeared. The pixel is called background pixelif it has the value zero, andforegroundif it has not. The constant assumption is made that dynamic objects in the second picture differ from the static objects in the first picture (e.g., in color).

Figure 2.4: Illustration of the reference image (left) showing the static objects of an environment at timet0and an image showing the additional dynamic objects (middle) aftert0∈Rat timet>t0. The pixel value of the image at timet>t0is subtracted from the value of the pixel in the reference image at the same spot. This results in the silhouette image (right) with pixel values larger (red) and equal zero (white).

− =

In the context of BG subtraction, each point of the surveillance area can be marked as either one of the labelsidentical, changed, occluded, andout of range. Thereby, the camera sensor in the original sense becomes a sensor for classifying identical, changed, occluded and out of range regions of the surveillance area. Given the cameras position, a pointy ∈ Acan be classified as one of the labelsoccluded orout of range just as in Section 2.1.1 with one exception: The second choice, whether or not the point is occluded, only regards the static objects and not the dynamic objects.

Points that are inside the viewing frustum of the camera and in front of all static objects are detectable points. A detectable point is mapped onto a specific pixel by projecting it onto the image plane and assigning a specific pixel, e.g., by rounding. It is labeled the following way:

Definition 2.1.7

LetI ≡ I(p,o) ⊂ Ebe theimage planeof the camera andρ :I → N2denote the rasterization of the

2.1. COVERAGE OF MULTIPLE CAMERAS 25

image plane into pixels. Let the functionπ:Va(E)→ Idenote the projection of a detectable point onto the image plane.

• A pointy∈Va(E) is calledidenticalifρ(π(y)) is a background pixel.

• Otherwise it is calledchanged.

The proposedthree-dimensional background-subtractionmethod (3DBGS) uses the image space to de-termine regions with the labelschangedandidenticalin the surveillance area. The procedure is illustrated in the Figures 2.5 (left).

Inside frustum?

In front of all static objects?

Behind background pixels?

identical yes

changed no

yes

occluded no

yes

out of range no

E A

Figure 2.5: By the 3DBGS, the surveillance area is subdivided intoidentical (yellow), changed (orange),out of range(blue), andoccluded(red) regions. This is illustrated by a tree (left) and a 2D environment including a pinhole camera model (right) with corresponding colors.

Left: The colors correspond to Figure 2.3. An orange marked point may have changed during surveillance at timet>t0and yellow labeled points definitely have not changed.

Right: Illustration of an environmentEwith static (gray) and dynamic (red) objects with a pinhole camera model (green). The surveillance area (contoured, dotted) is part of the environmentA⊂ E.

Assuming the camera is a pinhole camera, the 3DBGS labels the detectable points of the surveillance area as illustrated in Figure 2.5 (right). The identical parts (yellow) definitely do not include any dynamic objects: Let p∈E be the position of the camera’s pinhole. Letx⊂ Ibe a point on a background pixel in the image planeI. The ray [p,x) will not intersect any objects, dynamic objects in particular. Thus, the region projected by a background pixel will always beidenticalto the region at the momentt0. The remaining points are marked aschanged.

The 3DBGS has been discussed for a single camera. It will be generalized for other types of sensors and more than one sensor in the next section.