Bandit

Results: 280



#Item
1Online Learning for Personalized Room-Level Thermal Control: A Multi-Armed Bandit Framework Parisa Mansourifard Farrokh Jazizadeh

Online Learning for Personalized Room-Level Thermal Control: A Multi-Armed Bandit Framework Parisa Mansourifard Farrokh Jazizadeh

Add to Reading List

Source URL: anrg.usc.edu

Language: English - Date: 2013-12-06 19:17:18
    2The N-Tuple Bandit Evolutionary Algorithm for Automatic Game Improvement Kamolwan Kunanusont, Raluca D. Gaina, Jialin Liu, Diego Perez-Liebana and Simon M. Lucas University of Essex, Colchester, UK Email: {kkunan, rdgain

    The N-Tuple Bandit Evolutionary Algorithm for Automatic Game Improvement Kamolwan Kunanusont, Raluca D. Gaina, Jialin Liu, Diego Perez-Liebana and Simon M. Lucas University of Essex, Colchester, UK Email: {kkunan, rdgain

    Add to Reading List

    Source URL: www.diego-perez.net

    Language: English - Date: 2017-03-23 19:41:34
      3Supplementary Material for ”Combinatorial multi-armed bandit: general framework, results and applications”, by Wei Chen, Yajun Wang, and Yang Yuan. A. Full proof of Theorem 1 We use the following two well known bound

      Supplementary Material for ”Combinatorial multi-armed bandit: general framework, results and applications”, by Wei Chen, Yajun Wang, and Yang Yuan. A. Full proof of Theorem 1 We use the following two well known bound

      Add to Reading List

      Source URL: proceedings.mlr.press

      Language: English - Date: 2018-07-16 03:38:06
        4Journal of Machine Learning Research1755  Submitted 6/15; Published 9/15 Batch Learning from Logged Bandit Feedback through Counterfactual Risk Minimization

        Journal of Machine Learning Research1755 Submitted 6/15; Published 9/15 Batch Learning from Logged Bandit Feedback through Counterfactual Risk Minimization

        Add to Reading List

        Source URL: jmlr.csail.mit.edu

        Language: English - Date: 2017-07-22 15:41:54
          5Using Bandit Algorithms on Changing Reward Rates Introduction One of the problems we have at System1 is updating our estimate of a feature’s performance over time. Even if our initial estimate is correct, the performan

          Using Bandit Algorithms on Changing Reward Rates Introduction One of the problems we have at System1 is updating our estimate of a feature’s performance over time. Even if our initial estimate is correct, the performan

          Add to Reading List

          Source URL: www.system1.com

          Language: English - Date: 2018-07-13 19:30:08
            6Combinatorial Multi-Armed Bandit: General Framework, Results and Applications Wei Chen Microsoft Research Asia, Beijing, China

            Combinatorial Multi-Armed Bandit: General Framework, Results and Applications Wei Chen Microsoft Research Asia, Beijing, China

            Add to Reading List

            Source URL: proceedings.mlr.press

            Language: English - Date: 2018-07-16 03:38:06
              7THE NON-BAYESIAN RESTLESS MULTI-ARMED BANDIT: A CASE OF NEAR-LOGARITHMIC REGRET Wenhan Dai†∗ , Yi Gai‡ , Bhaskar Krishnamachari‡ , Qing Zhao§ †  School of Information Science and Technology, Tsinghua Universit

              THE NON-BAYESIAN RESTLESS MULTI-ARMED BANDIT: A CASE OF NEAR-LOGARITHMIC REGRET Wenhan Dai†∗ , Yi Gai‡ , Bhaskar Krishnamachari‡ , Qing Zhao§ † School of Information Science and Technology, Tsinghua Universit

              Add to Reading List

              Source URL: ceng.usc.edu

              Language: English - Date: 2011-10-16 14:12:22
                8Multi-Bandit Best Arm Identification Victor Gabillon Mohammad Ghavamzadeh Alessandro Lazaric INRIA Lille - Nord Europe, Team SequeL

                Multi-Bandit Best Arm Identification Victor Gabillon Mohammad Ghavamzadeh Alessandro Lazaric INRIA Lille - Nord Europe, Team SequeL

                Add to Reading List

                Source URL: papers.nips.cc

                Language: English - Date: 2014-02-24 03:34:19
                  9Online Algorithms for the Multi-Armed Bandit Problem with Markovian Rewards arXiv:1007.2238v2 [math.OC] 26 JulCem Tekin, Mingyan Liu

                  Online Algorithms for the Multi-Armed Bandit Problem with Markovian Rewards arXiv:1007.2238v2 [math.OC] 26 JulCem Tekin, Mingyan Liu

                  Add to Reading List

                  Source URL: arxiv.org

                  Language: English - Date: 2010-07-26 20:13:34
                    10Dex-Net 1.0: A Cloud-Based Network of 3D Objects for Robust Grasp Planning Using a Multi-Armed Bandit Model with Correlated Rewards Jeffrey Mahler1 , Florian T. Pokorny1 , Brian Hou1 , Melrose Roderick1 , Michael Laskey1

                    Dex-Net 1.0: A Cloud-Based Network of 3D Objects for Robust Grasp Planning Using a Multi-Armed Bandit Model with Correlated Rewards Jeffrey Mahler1 , Florian T. Pokorny1 , Brian Hou1 , Melrose Roderick1 , Michael Laskey1

                    Add to Reading List

                    Source URL: goldberg.berkeley.edu

                    Language: English - Date: 2016-02-17 17:39:47