ppt

Distributed Optimization in Sensor
Networks
Mike Rabbat & Rob Nowak
Monday, April 26, 2004
IPSN’04, Berkeley, CA
A Motivating Example
“What is µ(x)?”
(E.g.,
“It’s
)
x7
x4
x1
x…
x5
x2
x6
x3
!”
xn
Two Extreme Approaches
n sensors distributed uniformly
over a square region
1) Transmit Data
2) Transmit A Result
x7
x4
x1
x6
x3
x7
x4
x…
x5
x2
compute
x1
xn
x…
x5
x2
x6
x3
xn
Energy-Accuracy Tradeoffs
• Multi-hop communication
b(n) = total number of bits
h(n) = avg number of hops per bit
e(n) = avg energy per hop
• Total Energy Consumption
• Consider two situations
1. Sensors transmit data, result computed at destination
2. Sensors process in-network, result transmitted to destination
Distributed Iterative Optimization
n sensors
Minimize w.r.t. µ
x1
xn
x…
x5
x7
x4
x6
E.g., fi(µ; xi) = (µ – xi)2
x2
x3
Strategy:
• Cycle over each sensor
• Update using previous
value and local data
Distributed Iterative Optimization
n sensors
Minimize w.r.t. µ
#1(1)
x1
xn
x…
x5
x7
x4
x6
Energy used:
E.g., fi(µ; xi) = (µ – xi)2
x2
x3
Strategy:
• Cycle over each sensor
• Update using previous
value and local data
Distributed Iterative Optimization
n sensors
Minimize w.r.t. µ
x1
xn
x…
x4
x6
Energy used:
E.g., fi(µ; xi) = (µ – xi)2
x2
x5
x7
#1(1)
#2(1)
x3
Strategy:
• Cycle over each sensor
• Update using previous
value and local data
Distributed Iterative Optimization
n sensors
Minimize w.r.t. µ
#n(1)
x1
xn
x…
x4
x6
Energy used:
E.g., fi(µ; xi) = (µ – xi)2
x2
x5
x7
#1(1)
#2(1)
x3
Strategy:
• Cycle over each sensor
• Update using previous
value and local data
Distributed Iterative Optimization
n sensors
Minimize w.r.t. µ
#n(K)
x1
xn
x…
x4
x6
Energy used:
Compared with:
E.g., fi(µ; xi) = (µ – xi)2
x2
x5
x7
#1(K)
#2(K)
x3
Strategy:
• Cycle over each sensor
• Update using previous
value and local data
• Repeat K times
Incremental Subgradient Methods
• Have data xi at sensor i, for i=1,2,…,n
• Find µ that minimizes
• Distributed iterative procedure
small positive step size
Incremental Subgradient Methods
• Have data xi at sensor i, for i=1,2,…,n
• Find µ that minimizes
• Distributed iterative procedure
use subgradient of fi is non-differentiable
Convergence
Theorem (Nedić & Bertsekas, ’01):
Assume the fi are convex on a convex set ,
µ*2 , and krfi(µ)k<C. Define D = diam().
Then after K cycles, with
we are guaranteed that
Comparing Resource Usage
1) Transmit Data
x7
x4
x1
2) Transmit A Result
x2
x6
x3
x4
x…
x5
x7
x1
xn
x…
x5
x2
x6
x3
xn
Energy Savings
• When does distributed processing use less energy?
Energy
vs.
n
Robust Estimation
• Estimate the mean (squared
error loss)
4
3.5
3
2.5
(z)
• Robust estimate
(robust loss function)
2
1.5
1
0.5
0
-2
-1
0
z
1
2
Robust Estimates and “Bad Sensors”
• E.g., Monitoring ozone levels, µ = average level
• Normal sensor measures µ ± 1
200 sensors
10 measurements each
10% “bad”
• Damaged sensor measures µ ± 10
10
• Energy used
(K = 25 iterations)
5
Residual Error
compared to
robust loss
squared error loss
0
-5
-10
0
50
100
Iteration (k)
150
200
Source Localization
• Isotropic energy source located at µ
• Sensor i at location ri,
measures received signal strength
• Local cost functions
Source Localization
100 sensors
50 £ 50 square
10 measurements
Avg SNR = 3dB
Converged in 45 cycles
Compare with
In Conclusion
• When is in-network processing more energy efficient?
• Incremental subgradient optimization
– Simple to implement
– Applicable to a general class of problems
– Analyzable rate of convergence
• Distributed in-network processing uses less energy:
vs.
Ongoing & Future Work
x2
x1
xn
x3
x6
x…
x5
x7
[email protected]
www.cae.wisc.edu/~rabbat
x4