Optimally Compressed Nonparametric Online Learning: Tradeoffs between memory and consistency

You are here

Top Reasons to Join SPS Today!

1. IEEE Signal Processing Magazine
2. Signal Processing Digital Library*
3. Inside Signal Processing Newsletter
4. SPS Resource Center
5. Career advancement & recognition
6. Discounts on conferences and publications
7. Professional networking
8. Communities for students, young professionals, and women
9. Volunteer opportunities
10. Coming soon! PDH/CEU credits
Click here to learn more.

Optimally Compressed Nonparametric Online Learning: Tradeoffs between memory and consistency

By: 
Alec Koppel; Amrit Singh Bedi; Ketan Rajawat; Brian M. Sadler;

Batch training of machine learning models based on neural networks is well established, whereas, to date, streaming methods are largely based on linear models. To go beyond linear in the online setting, nonparametric methods are of interest due to their universality and ability to stably incorporate new information via convexity or Bayes's rule. Unfortunately, when applied online, nonparametric methods suffer a "curse of dimensionality," which precludes their use: their complexity scales at least with the time index. We survey online compression tools that bring their memory under control and attain approximate convergence. The asymptotic bias depends on a compression parameter that trades off memory and accuracy. Applications to robotics, communications, economics, and power are discussed as well as extensions to multiagent systems.

SPS on Twitter

SPS Videos


Signal Processing in Home Assistants

 


Multimedia Forensics


Careers in Signal Processing             

 


Under the Radar