Compressive sensing

(Oszczędne próbkowanie)

Emmanuel Candès, 2004 ($500k) 

  • by coincidence (Shepp–Logan phantom)
  • l1 minimalization
  • “It was as if you gave me the first three digits of a 10-digit bank account number — and then I was able to guess the next seven,”

So what and why?

Compressive sensing is a mathematical tool that creates hi-res data sets from lo-res samples. 

 

It can be used to:
  - resurrect old musical recordings, 
  - find enemy radio signals, 
  - and generate MRIs much more quickly.

Here’s how it would work with a photograph.

Undersample

Fill in the dots

(l1)

Add shapes

(sparsity)

Add smaller shapes

Achieve clarity

What is the key?

Compressive sensing

  • magic of notion called sparsity
  • pic made of solid blocks of color or wiggly lines is sparse, 
  • a screenful of random and chaotic dots is not
  • out of all the bazillion possible reconstructions, the simplest, or sparsest, image is almost always the right one or very close to it
  • how to find the sparsest image quickly?

Candès and Tao have shown mathematically that

the l1 minimization is all we need

R example

library(R1magic)
N <- 100 
# Sparse components 
K <- 4 
#  Up to Measurements  > K LOG (N/K)
M <- 40
# Measurement Matrix (Random Sampling Sampling)
phi <- GaussianMatrix(N,M)
# R1magic generate random signal
xorg <- sparseSignal(N, K, nlev=1e-3)
y <- phi %*% xorg ;# generate measurement
T <- diag(N) ;# Do identity transform
p <- matrix(0, N, 1) ;# initial guess

# R1magic Convex Minimization
ll <- solveL1(phi, y, T, p)
x1 <- ll$estimate

plot( 1:100, seq(0.011,1.1,0.011), type = "n",xlab="",ylab="")
title(main="Random Sparse Signal Recovery",
      xlab="Signal Component",ylab="Spike Value")
lines(1:100, xorg , col = "red")
lines(1:100, x1, col = "blue", cex = 1.5) 

Bibliography

  • http://www.wired.com/2010/02/ff_algorithm/all/1
  • http://www.londonr.org/msuzen.pdf
  • http://people.csail.mit.edu/indyk/princeton.pdf
Made with Slides.com