Learning, Duality, and Algorithms
Victor Sanches Portella
Advisor: Marcel K. de Carli Silva
IME - USP
May, 2019
At each round
Player chooses a point
Enemy chooses a function
Player suffers a loss
SIMULTANEOUSLY
Player
Enemy
!
!
CONVEX
Player and Enemy see
An Online Convex Optimization Problem
convex set
set of convex functions
Player
Enemy
Player
Enemy
Experts
0.5
0.1
0.3
0.1
1
0
-1
1
Probabilities
Costs
Online Linear Regression
Player
Enemy
Regression Function
Query & Answer
Loss
Cost of always choosing
Player's Loss
Sublinear regret under mild conditions
Focus of this talk: algorithms for the Player
Hupefully efficiently implementable
Enemy
Player
Enemy
Player
Fixed Regularizer
At round use regularizer
Regularizer Increment
Convex Function
Not clear in general
projection
Representation of derivative
What is
?
direction
Online Gradient Descent Update
point
functional
(Riesz Repr. Theorem)
functional
functional
Directional derivative of at
strictly convex and differentiable on
For every
there is
such that
Bregman Projections onto attained by
Bregman Projector
Bregman
Projection
First round
Round
for
Mirror Map Increments
Bregman
Projection
First round
First round
For
For
inside
outside
inside
outside
inside
outside
Limited Feedback: Bandit, two-point Bandit feedback
Special Cases: Combinatorial, other specific settings
Player
Drop or Add Hypotheses: Convexity, adversarial enemies,
Hypercube
L2-Ball
Change Metric: Policy Regret, Raw Loss
side information
Quantum Computing
Approximately Maximum Flow
Robust Optimization
Competitive Analysis
Spectral Sparsification
SDP Solver
Oracle Boosting
Ideas
New Setting
Variational Perspective
Learning, Duality, and Algorithms
Victor Sanches Portella
Advisor: Marcel K. de Carli Silva
IME - USP
May, 2019