lecture1.pdf
-
Upload
vnod-rathode-b -
Category
Documents
-
view
6 -
download
1
description
Transcript of lecture1.pdf
![Page 1: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/1.jpg)
Sensor Fusion
Sensor FusionFredrik Gustafsson
Lecture Content Chapters
1 Course overview. Estimation theory for linear models. 1–2
2 Estimation theory for nonlinear models 3
3 Sensor networks and detection theory 4–5
4 Nonlinear filter theory. The Kalman filter. Filter banks. 6–7, 10
5 Kalman filter approximation for nonlinear models (EKF, UKF). 8
6 The point-mass filter and the particle filter. 9
7 The particle filter theory. The marginalized particle filter. 9
8 Simultaneous localization and mapping (SLAM). 11
9 Modeling and motion models. 12–13
10 Sensors and filter validation. 14–15
Literature: Statistical Sensor Fusion. Studentlitteratur, 2010 or 2012 (updatedversion).
Lecture 1, 2012 1
![Page 2: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/2.jpg)
Sensor Fusion
Exercises: compendium sold by Bokadademin. Software: Signals and Systems Labfor Matlab.
Lecture 1, 2012 2
![Page 3: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/3.jpg)
Sensor Fusion
Lecture 1: Estimation theory in linear models
Whiteboard:
• The weighted least squares (WLS) method.
• The maximum likelihood (ML) method.
• The Cramer-Rao lower bound (CRLB).
• Fusion algorithms.
Slides:
• Examples
• Code examples
• Algorithms
Lecture 1, 2012 3
![Page 4: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/4.jpg)
Sensor Fusion
Example 1: sensor network
[m]
[m]
−150 −100 −50 0 50 100 150 200 250
−50
0
50
100
150
200
250
12 sensor nodes, each one with microphone, geophone and magnetometer.One moving target.Detect, localize and track/predict the target.
Lecture 1, 2012 4
![Page 5: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/5.jpg)
Sensor Fusion
Example 2: fusion of GPS and IMU
GPS gives good position. IMU gives accurate accelerations.Combine these to get even better position, velocity and acceleration.
Lecture 1, 2012 5
![Page 6: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/6.jpg)
Sensor Fusion
Example 3: fusion of camera and radar images
Radar gives range and range rate with good horizontal angle resolution, but novertical resolution.Camera gives very good angular resolution, and color, but no range.Combined, they have a great potential for situation awareness.
Lecture 1, 2012 6
![Page 7: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/7.jpg)
Sensor Fusion
Chapter 2: estimation for linear models
• Least squares and likelihood methods.
• Sensor network example.
• Fusion and safe fusion in distributed algorithms
Lecture 1, 2012 7
![Page 8: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/8.jpg)
Sensor Fusion
Code for Signals and Systems Lab:
p1=[0;0];p2=[2;0];x=[1;1];X1=ndist(x,0.1*[1 -0.8;-0.8 1]);X2=ndist(x,0.1*[1 0.8;0.8 1]);plot2(X1,X2)
−0.5 0 0.5 1 1.5 2 2.5−0.5
0
0.5
1
1.5
x1
x2
S1 S2
Lecture 1, 2012 8
![Page 9: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/9.jpg)
Sensor Fusion
Sensor network example, cont’d
X3=fusion(X1,X2); % WLSX4=0.5*X1+0.5*X2; % LSplot2(X3,X4)
−0.5 0 0.5 1 1.5 2 2.5−0.5
0
0.5
1
1.5
x1
x2
S1 S2
Lecture 1, 2012 9
![Page 10: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/10.jpg)
Sensor Fusion
Information loops in sensor networks
• Information and sufficient statistics should be communicated in sensor networks.
• In sensor networks with untagged observations, our own observations may beincluded in the information we receive.
• Information loops (updating with the same sensor reading several times) giverise to optimistic covariances.
• Safe fusion algorithms (or covariance intersection techniques) give conservativecovariances, using a worst case way of reasoning.
Lecture 1, 2012 10
![Page 11: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/11.jpg)
Sensor Fusion
Safe fusionGiven two unbiased estimates x1, x2 with information I1 = P−1
1 and I2 = P−12
(pseudo-inverses if singular covariances), respectively. Compute the following:
1. SVD: I1 = U1D1UT1 .
2. SVD:D−1/21 UT
1 I2U1D−1/21 = U2D2U
T2 .
3. Transformation matrix: T = UT2 D
1/21 U1.
4. State transformation: ˆx1 = T x1 and ˆx2 = T x2. The covariances of these areCOV(ˆx1) = I and COV(ˆx2) = D−1
2 , respectively.
5. For each component i = 1, 2, . . . , nx, let
ˆxi = ˆxi1, Dii = 1 if Dii
2 < 1,
ˆxi = ˆxi2, Dii = Dii
2 if Dii2 > 1.
6. Inverse state transformation:
x = T−1 ˆx, P = T−1D−1T−T
Lecture 1, 2012 11
![Page 12: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/12.jpg)
Sensor Fusion
Transformation steps
x1 x2 x1
x2
x1
x2
Lecture 1, 2012 12
![Page 13: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/13.jpg)
Sensor Fusion
Sensor network example, cont’d
X3=fusion(X1,X2); % WLSX4=fusion(X1,X3); % X1 used twiceX5=safefusion(X1,X3);plot2(X3,X4,X5)
Lecture 1, 2012 13
![Page 14: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/14.jpg)
Sensor Fusion
Sequential WLS
The WLS estimate can be computed recursively in the space/time sequence yk .Suppose the estimate xk−1 with covariance Pk based on observations y1:k−1. Anew observation is fused using
xk = xk−1 + Pk−1HTk
(HkPk−1H
Tk +Rk
)−1(yk −Hkxk−1),
Pk = Pk−1 − Pk−1HTk
(HkPk−1H
Tk +Rk
)−1HkPk−1.
Note that the fusion formula can be used alternatively. In fact, the derivation isbased on the information fusion formula applying the matrix inversion lemma.
Lecture 1, 2012 14
![Page 15: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/15.jpg)
Sensor Fusion
Batch vs sequential evaluation of loss function
The minimizing loss function can be computed in two ways using batch andsequential computations, respectively,
V WLS(xN ) =N∑
k=1
(yk −HkxN )TR−1k (yk −HkxN )
=N∑
k=1
(yk −Hkxk−1)T (HkPk−1H
Tk +Rk)
−1(yk −Hkxk−1)
− (x0 − xN )TP−10 (x0 − xN )
The second expression should be used in decentralized sensor networkimplementations and on-line algorithms.
The last correction term to de-fuse the influence of the initial values is needed onlywhen this initialization is used.
Lecture 1, 2012 15
![Page 16: lecture1.pdf](https://reader031.fdocuments.in/reader031/viewer/2022020207/552a036a4a795931778b4671/html5/thumbnails/16.jpg)
Sensor Fusion
Batch vs sequential evaluation of likelihood
The Gaussian likelihood of data is important in model validation, change detectionand diagnosis. Generally, Bayes formula gives
p(y1:N ) = p(y1)
N∏
k=2
p(yk|y1:k−1).
For Gaussian noise and using the sequential algorithm, this is
p(y1:N ) =
N∏
k=1
γ(yk;Hkxk−1, HkPk−1HTk +Rk)
The off-line form is:
p(y1:N ) = γ(xN ;x0, P0)√det(PN )
N∏
k=1
γ(yk;HkxN , Rk).
Prefered for de-centralized computation in sensor networks or on-line algorithms.
Lecture 1, 2012 16