16699

Online Learning via Offline Greedy Algorithms: Applications in Market Design and Optimization

APA

(2020). Online Learning via Offline Greedy Algorithms: Applications in Market Design and Optimization. The Simons Institute for the Theory of Computing. https://simons.berkeley.edu/talks/online-learning-offline-greedy-algorithms-applications-market-design-and-optimization

MLA

Online Learning via Offline Greedy Algorithms: Applications in Market Design and Optimization. The Simons Institute for the Theory of Computing, Oct. 29, 2020, https://simons.berkeley.edu/talks/online-learning-offline-greedy-algorithms-applications-market-design-and-optimization

BibTex

          @misc{ scivideos_16699,
            doi = {},
            url = {https://simons.berkeley.edu/talks/online-learning-offline-greedy-algorithms-applications-market-design-and-optimization},
            author = {},
            keywords = {},
            language = {en},
            title = {Online Learning via Offline Greedy Algorithms: Applications in Market Design and Optimization},
            publisher = {The Simons Institute for the Theory of Computing},
            year = {2020},
            month = {oct},
            note = {16699 see, \url{https://scivideos.org/Simons-Institute/16699}}
          }
          
Negin Golrezaei (MIT)
Talk number16699
Source RepositorySimons Institute

Abstract

Motivated by online decision-making in time-varying combinatorial environments, we study the problem of transforming offline algorithms to their online counterparts. We focus on offline combinatorial problems that are amenable to a constant factor approximation using a greedy algorithm that is robust to local errors. For such problems, we provide a general framework that efficiently transforms offline robust greedy algorithms to online ones using Blackwell approachability. We show that the resulting online algorithms have O(T^{1/2}) (approximate) regret under the full information setting. We further introduce a bandit extension of Blackwell approachability that we call Bandit Blackwell approachability. We leverage this notion to transform greedy robust offline algorithms into an O(T^{2/3})(approximate) regret in the bandit setting. Demonstrating the flexibility of our framework, we apply our offline-to-online transformation to several problems at the intersection of revenue management, market design, and online optimization, including product ranking optimization in online platforms, reserve price optimization in auctions, and submodular maximization. We show that our transformation, when applied to these applications, leads to new regret bounds or improves the current known bounds.