Distributed Optimization in Sensor Networks Mike Rabbat & Rob Nowak Monday, April 26, 2004 IPSN’04, Berkeley, CA A Motivating Example “What is µ(x)?” (E.g., “It’s ) x7 x4 x1 x… x5 x2 x6 x3 !” xn Two Extreme Approaches n sensors distributed uniformly over a square region 1) Transmit Data 2) Transmit A Result x7 x4 x1 x6 x3 x7 x4 x… x5 x2 compute x1 xn x… x5 x2 x6 x3 xn Energy-Accuracy Tradeoffs • Multi-hop communication b(n) = total number of bits h(n) = avg number of hops per bit e(n) = avg energy per hop • Total Energy Consumption • Consider two situations 1. Sensors transmit data, result computed at destination 2. Sensors process in-network, result transmitted to destination Distributed Iterative Optimization n sensors Minimize w.r.t. µ x1 xn x… x5 x7 x4 x6 E.g., fi(µ; xi) = (µ – xi)2 x2 x3 Strategy: • Cycle over each sensor • Update using previous value and local data Distributed Iterative Optimization n sensors Minimize w.r.t. µ #1(1) x1 xn x… x5 x7 x4 x6 Energy used: E.g., fi(µ; xi) = (µ – xi)2 x2 x3 Strategy: • Cycle over each sensor • Update using previous value and local data Distributed Iterative Optimization n sensors Minimize w.r.t. µ x1 xn x… x4 x6 Energy used: E.g., fi(µ; xi) = (µ – xi)2 x2 x5 x7 #1(1) #2(1) x3 Strategy: • Cycle over each sensor • Update using previous value and local data Distributed Iterative Optimization n sensors Minimize w.r.t. µ #n(1) x1 xn x… x4 x6 Energy used: E.g., fi(µ; xi) = (µ – xi)2 x2 x5 x7 #1(1) #2(1) x3 Strategy: • Cycle over each sensor • Update using previous value and local data Distributed Iterative Optimization n sensors Minimize w.r.t. µ #n(K) x1 xn x… x4 x6 Energy used: Compared with: E.g., fi(µ; xi) = (µ – xi)2 x2 x5 x7 #1(K) #2(K) x3 Strategy: • Cycle over each sensor • Update using previous value and local data • Repeat K times Incremental Subgradient Methods • Have data xi at sensor i, for i=1,2,…,n • Find µ that minimizes • Distributed iterative procedure small positive step size Incremental Subgradient Methods • Have data xi at sensor i, for i=1,2,…,n • Find µ that minimizes • Distributed iterative procedure use subgradient of fi is non-differentiable Convergence Theorem (Nedić & Bertsekas, ’01): Assume the fi are convex on a convex set , µ*2 , and krfi(µ)k<C. Define D = diam(). Then after K cycles, with we are guaranteed that Comparing Resource Usage 1) Transmit Data x7 x4 x1 2) Transmit A Result x2 x6 x3 x4 x… x5 x7 x1 xn x… x5 x2 x6 x3 xn Energy Savings • When does distributed processing use less energy? Energy vs. n Robust Estimation • Estimate the mean (squared error loss) 4 3.5 3 2.5 (z) • Robust estimate (robust loss function) 2 1.5 1 0.5 0 -2 -1 0 z 1 2 Robust Estimates and “Bad Sensors” • E.g., Monitoring ozone levels, µ = average level • Normal sensor measures µ ± 1 200 sensors 10 measurements each 10% “bad” • Damaged sensor measures µ ± 10 10 • Energy used (K = 25 iterations) 5 Residual Error compared to robust loss squared error loss 0 -5 -10 0 50 100 Iteration (k) 150 200 Source Localization • Isotropic energy source located at µ • Sensor i at location ri, measures received signal strength • Local cost functions Source Localization 100 sensors 50 £ 50 square 10 measurements Avg SNR = 3dB Converged in 45 cycles Compare with In Conclusion • When is in-network processing more energy efficient? • Incremental subgradient optimization – Simple to implement – Applicable to a general class of problems – Analyzable rate of convergence • Distributed in-network processing uses less energy: vs. Ongoing & Future Work x2 x1 xn x3 x6 x… x5 x7 [email protected] www.cae.wisc.edu/~rabbat x4
© Copyright 2025 Paperzz