Initializing Kernel Adaptive Filters via Probabilistic Inference

Iván Castro, Cristóbal Silva and Felipe Tobar

IEEE-DSP XII August 2017

Overview

  • Non linear time series prediction through Kernel Adaptive Filters (KAFs)
  • Probabilistic framework for KAF parameters:
    • Determine initial conditions
    • Construct a fully adaptive parameter-wise KAF
  • Prior distributions over parameters:
    • Dictionary elements
    • Filter weights
  • Sparsity focus in parameter estimation
  • Non linearity of parameter search is overcome through MCMC optimization
  • Validation over synthetic and real world data outperforms standard KAF

Why is non-linear modeling useful?

Linearly distributed data

  • Well studied problems
  • It exists a tremendous variety of tools for different learning topics

Non linearly distributed data

  • The problem arises: should we develop new tools, or tinker the existing ones?
  • For practical purposes, is easier to adjust the data to the existing tools
  • Therefore, feature or space transformations are the answer

Linear and non-linear estimation

KAFs and PI in a nutshell

Probabilistic Model for KAFs

Probabilistic Model

$$p(Y) = \prod_{i=d}^{N} \frac{1}{2\pi\sigma_{\epsilon}^2} \exp \left( \frac{\left(y - {\alpha}^{\texttt{T}} K_{\sigma_k}\left(x_i, \mathcal{D}\right)\right)^2}{2\sigma_{\epsilon}^2} \right)$$

Model Likelihood

Weights prior

$$p(\alpha) = \frac{1}{\sqrt{2\pi l_{\alpha}^2}} \exp \left( - \frac{\left\| \alpha \right\|^2}{2 l_{\alpha}^2} \right) $$

Sparsity Inducing Prior

$$ p(\mathcal{D}) = \frac{1}{\sqrt{2\pi l_{\mathcal{D}}^2}} \exp \left( - \frac{\left\| K_{\sigma}(\mathcal{D}, \mathcal{D}) \right\|^2}{2 l_{\mathcal{D}}^2} \right)$$

$$ y_i = \sum_{j=1}^{N_i} \alpha_{j} K_{\sigma_k}(i, j) + \epsilon_i $$

Simulations: Lorentz time series offline

\( x[i + 1] = x[i] + dt(\alpha(y[i] - x[i])) \)

\( y[i + 1] = y[i] + dt(x[i](\rho - z[i]) - y[i]) \)

\( z[i + 1] = z[i] + dt(x[i]y[i] - \beta z[i]) \)

Results: anemometer wind series offline

Fully Adaptive Kernel Filtering Online

Analysis of main findings

  • Pre-training

 

  • Sparsity inducing prior contribution

 

  • Fully Adaptive KAF

Concluding Remarks

  • Probabilistic model improved KAF performance in offline and online applications.
  • Further development of online application looks promising.
  • MCMC related faster approaches could be explored.

References

[1] W. Liu, P. Pokharel, and J. Principe, “The kernel least-mean-square algorithm,” IEEE Trans. on Signal Process., vol. 56, no. 2, pp. 543–554, 2008.

 

[2] C. Richard, J. Bermudez, and P. Honeine, “Online prediction of time series data with kernels,” IEEE Trans. on Signal Process., vol. 57, no. 3, pp. 1058 –1067, 2009.

Initializing Kernel Adaptive Filters via Probabilistic Inference

By crsilva

Initializing Kernel Adaptive Filters via Probabilistic Inference

  • 372