Distributed Optimization in Sensor Networks

Report 3 Downloads 212 Views
Distributed Optimization in Sensor Networks Michael Rabbat, Robert Nowak. Information Processing in Sensor Networks, 2004

October 14, 2011

Meng Guo

Automatic Control Lab, KTH

Introduction

Introduction

Wireless sensor Network (WSN) collects enormous amount of data over space and time. Objective: to derive an optimal estimate of certain parameters from these “raw” data. How? To minimize the energy consumption. Simple example: n sensors uniformly distributed over 1 m2 . Each collects m measurements. Task: what is the average value?

Meng Guo

Automatic Control Lab, KTH

Introduction

Possible Approaches

Possible Approaches

1

Sensors transmit all data to a central processor. O(mn) bits over O(1) meter.

2

Sensors first compute a local average and then transmit to the central unit.O(n) bits over O(1) meter.

3

Construct a path passing through all sensors once. Each node adds its local average to the total along the way. An accumulation process from start node to finish node. O(n) bits over O(n−1/2 ) meter. Far less energy and communication needed for large (m, n).

Meng Guo

Automatic Control Lab, KTH

General Setup

General Setup

So-called Incremental Subgradient Optimization. An estimate of the parameter is passed from node to node. Each node adjusts the estimate based on local data, then passes the update to the next node. Several “cycles” through the network may be needed.

It is distributed, “in-network” optimization!

Meng Guo

Automatic Control Lab, KTH

Formulation

Formulation A network of n sensors in which each sensor collects m measurements. Denote xi,j the jth measurement at the ith sensor. Goal: 1 θˆ = arg min θ∈Θ n θ∈

Rd

n X

fi ({xi,j }m j=1 , θ)

i=1

is the set of parameters to be estimated.

Cost function fi : Rd → R, are convex and Θ is a nonempty, closed, convex subset of Rd . Denote fi (θ) = fi ({xi,j }m j=1 , θ), which depends on sensor i’s measurements.

Meng Guo

Automatic Control Lab, KTH

Formulation

Centralized Method

Centralized Version Centralized subgradient descent approach: θˆ(k+1) = θˆ(k) − α

n X

gi,k ,

i=1

where gi,k ∈ ∂fi (θˆ(k) ), α is a positive step size, and k is the iteration number. Note! each update step in this approach uses data from all of the sensors. ∂f (θ0 ) is the subdifferential of f at θ0 , the set of all subgradients of f at θ0 . The subgradient of f at θ0 is any direction g satisfying f (θ) ≥ f (θ0 ) + (θ − θ0 )T g ,

Meng Guo

∀θ ∈ Θ

Automatic Control Lab, KTH

Formulation

Decenralized Incremental Method

Decenralized Incremental Method

Each update: a cycle of n subiterations (k)

θi is the estimate of θ at sensor i during the kth cycle. At each subiteration sensor i minimize the cost function fi (θ), based on {xi,j }m j=1 : (k)

θi (k)

(k)

where gi,k ∈ ∂fi (θi−1 ) and θ0 Assumptions: 1 2

(k)

= θi−1 − αgi,k (k−1)

= θn

optimal solution θ∗ exists. kgi,k k ≤ ζ for all fi (θ) and θ ∈ Θ

Convergence guaranteed.

Meng Guo

Automatic Control Lab, KTH

Formulation

Convergence Speed

Convergence Speed After K cycles, K =b

kθˆ(0) − θ∗ k , α2 ζ 2

it is guaranteed that min f (θˆ(0) ) ≤ f (θ∗ ) + αζ 2

0≤k≤K

Set  = αζ 2 . Convergence to -ball of the optimal value after at most K≤

c0 ζ 2 = O(−2 ) 2

Time varying or fixed step size α.

Meng Guo

Automatic Control Lab, KTH

Formulation

Energy Features

Energy Features

Energy for communication >> local computation. How the amount of communication required scales with the size (m, n)of the network? Total energy used for communication grows linearly with the size of the network: Eincr (n) ≤ c2 n −2 e(n) Energy saving ratio (w.r.t centralized data fusion) R=

Eincr (n) = c3 m n1/d ε2 Ecent (n)

Meng Guo

Automatic Control Lab, KTH

Applications

Example 1

Example 1 Problem: Identify and discard “bad” measurements from the estimation process. Solution: Incremental algorithm with modified local cost function (instead of kx − θk2 ). Less weight to points that deviate greatly from the parameter θ: ( kx − θk2 /2, for kx − θk ≤ r ρh (x, θ) = 2 γkx − θk − γ /2, for kx − θk > r Distributed cost function: m

fi (θ) =

1 X ρ(xi, j , θ) m j=1

Meng Guo

Automatic Control Lab, KTH

Applications

Meng Guo

Example 1

Automatic Control Lab, KTH

Applications

Example 2

Example 2

Problem: Estimate the location of an acoustic source using the signal strength received at each sensor. Based on the isotropic energy propagation model, the local cost function is m 1 X A fi (θ) = (xi,j − ) m kθ − ri kβ j=1

Fits into the general incremental subgradient framework.

Meng Guo

Automatic Control Lab, KTH

Applications

Meng Guo

Example 2

Automatic Control Lab, KTH

Applications

Example 3

Example 3 Problem: Cluster measurements from a mixture of Gaussian distributions with unknown means and covariances (different weights). Distributed Expectation-Maximization (DEM) Algorithms fi (θ) =

n X m X

J X log( αi,k N (yi,j |µk , Σk )))

i=1 j=1

k=1

Generally hard: 1

2

Run DEM to estimate the mean and covariance of each distribution (how many). Determine which class it belongs to.

More assumptions needed to implement the Decenralized Incremental Method

Meng Guo

Automatic Control Lab, KTH

Applications

Meng Guo

Example 3

Automatic Control Lab, KTH

Conclusion

Summary

Incremental subgradient optimization 1 2 3 4

Determine the cost function Circulate a parameter estimate through the network. Small adjustments at each node based on local data. Terminate after enough cycles

Enormous energy saving with large systems. Constraints? 1 2 3

Path routing, centralized planning? Optimal step size α. Robustness to communication failures?

Meng Guo

Automatic Control Lab, KTH