Machine Learning Center Seminar | Vidya Muthukumar – Data-adaptivity and model selection in online decision-making

*********************************
There is now a CONTENT FREEZE for Mercury while we switch to a new platform. It began on Friday, March 10 at 6pm and will end on Wednesday, March 15 at noon. No new content can be created during this time, but all material in the system as of the beginning of the freeze will be migrated to the new platform, including users and groups. Functionally the new site is identical to the old one. webteam@gatech.edu
*********************************

Event Details
  • Date/Time:
    • Wednesday September 14, 2022
      12:15 pm - 1:00 pm
  • Location: Marcus Nano building, Rooms 1116-1118
  • Phone:
  • URL:
  • Email:
  • Fee(s):
    N/A
  • Extras:
Contact

Lia Namirr
Machine Learning Center at Georgia Tech

Summaries

Summary Sentence: This talk will describe two research vignettes in full-information and bandit learning.

Full Summary: This talk will describe two research vignettes in full-information and bandit learning.

Media
  • Vidya Muthukumar Vidya Muthukumar
    (image/jpeg)

Abstract: Classical online learning algorithms make a static assumption on the nature of the data generating process (either stochastic or adversarial) and the nature of the offline benchmark to measure performance. Neither of these assumptions are well-justified in practice. While assuming a probability model on the data could lead to suboptimal performance in practice, worst-case adversarially robust algorithms may be unnecessarily pessimistic. More subtly, since our objective in online learning is to maximize reward rather than regret, the choice of offline benchmark (i.e. model) matters as much as the choice of online algorithm.  

 This talk will describe two research vignettes in full-information and bandit learning. The first research vignette will motivate and describe the design of online learning algorithms in the “tree-expert” setting that are computationally efficient and adapt both to underlying stochasticity and the minimal model complexity. These algorithms achieve reward that is almost as good as an “oracle” algorithm that has access to all of this information beforehand. The second research vignette will describe approaches to data-adaptive model selection in the more challenging limited-information feedback paradigm, which includes contextual bandits and reinforcement learning. I will conclude with a discussion of open problems involving computational efficiency and non-adversarial strategic behavior, and, time permitting, mention some initial work in these directions.  

  

Bio: Vidya Muthukumar is an Assistant Professor in the Schools of Electrical and Computer Engineering and Industrial and Systems Engineering at Georgia Institute of Technology. Her broad interests are in game theory, online and statistical learning. She is particularly interested in designing learning algorithms that provably adapt in strategic environments, fundamental properties of overparameterized models, and algorithmic foundations of multi-agent reinforcement learning.  

Vidya received the PhD degree in Electrical Engineering and Computer Sciences from the University of California, Berkeley. She is the recipient of the Adobe Data Science Research Award, a Simons-Berkeley-Google Research Fellowship (for the Fall 2020 program on “Theory of Reinforcement Learning”), IBM Science for Social Good Fellowship and a Georgia Tech Class of 1969 Teaching Fellowship for the academic year 2021-2022.  

Additional Information

In Campus Calendar
Yes
Groups

ML@GT

Invited Audience
Faculty/Staff, Postdoc, Public, Graduate students, Undergraduate students
Categories
Seminar/Lecture/Colloquium
Keywords
No keywords were submitted.
Status
  • Created By: Joshua Preston
  • Workflow Status: Published
  • Created On: Sep 12, 2022 - 9:32am
  • Last Updated: Sep 12, 2022 - 9:40am