ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2108.13703
  4. Cited By
Evaluating the Robustness of Off-Policy Evaluation

Evaluating the Robustness of Off-Policy Evaluation

31 August 2021
Yuta Saito
Takuma Udagawa
Haruka Kiyohara
Kazuki Mogi
Yusuke Narita
Kei Tateno
    ELMOffRL
ArXiv (abs)PDFHTML

Papers citing "Evaluating the Robustness of Off-Policy Evaluation"

21 / 21 papers shown
Title
Clustering Context in Off-Policy Evaluation
Clustering Context in Off-Policy Evaluation
Daniel Guzman-Olivares
Philipp Schmidt
Jacek Golebiowski
Artur Bekasov
CMLOffRL
76
0
0
28 Feb 2025
Effective Off-Policy Evaluation and Learning in Contextual Combinatorial
  Bandits
Effective Off-Policy Evaluation and Learning in Contextual Combinatorial Bandits
Tatsuhiro Shimizu
Koichi Tanaka
Ren Kishimoto
Haruka Kiyohara
Masahiro Nomura
Yuta Saito
CMLOffRL
88
1
0
20 Aug 2024
AutoOPE: Automated Off-Policy Estimator Selection
AutoOPE: Automated Off-Policy Estimator Selection
Nicolò Felicioni
Michael Benigni
Maurizio Ferrari Dacrema
OffRL
51
1
0
26 Jun 2024
Cross-Validated Off-Policy Evaluation
Cross-Validated Off-Policy Evaluation
Matej Cief
Branislav Kveton
Michal Kompan
OffRL
58
1
0
24 May 2024
Optimal Baseline Corrections for Off-Policy Contextual Bandits
Optimal Baseline Corrections for Off-Policy Contextual Bandits
Shashank Gupta
Olivier Jeunen
Harrie Oosterhuis
Maarten de Rijke
85
7
0
09 May 2024
Hyperparameter Optimization Can Even be Harmful in Off-Policy Learning
  and How to Deal with It
Hyperparameter Optimization Can Even be Harmful in Off-Policy Learning and How to Deal with It
Yuta Saito
Masahiro Nomura
OffRL
92
3
0
23 Apr 2024
POTEC: Off-Policy Learning for Large Action Spaces via Two-Stage Policy
  Decomposition
POTEC: Off-Policy Learning for Large Action Spaces via Two-Stage Policy Decomposition
Yuta Saito
Jihan Yao
Thorsten Joachims
OffRL
102
9
0
09 Feb 2024
Off-Policy Evaluation of Slate Bandit Policies via Optimizing
  Abstraction
Off-Policy Evaluation of Slate Bandit Policies via Optimizing Abstraction
Haruka Kiyohara
Masahiro Nomura
Yuta Saito
89
8
0
03 Feb 2024
Towards Assessing and Benchmarking Risk-Return Tradeoff of Off-Policy
  Evaluation
Towards Assessing and Benchmarking Risk-Return Tradeoff of Off-Policy Evaluation
Haruka Kiyohara
Ren Kishimoto
K. Kawakami
Ken Kobayashi
Kazuhide Nakata
Yuta Saito
OffRL
72
9
0
30 Nov 2023
SCOPE-RL: A Python Library for Offline Reinforcement Learning and
  Off-Policy Evaluation
SCOPE-RL: A Python Library for Offline Reinforcement Learning and Off-Policy Evaluation
Haruka Kiyohara
Ren Kishimoto
K. Kawakami
Ken Kobayashi
Kazuhide Nakata
Yuta Saito
OffRLELM
91
4
0
30 Nov 2023
On (Normalised) Discounted Cumulative Gain as an Off-Policy Evaluation
  Metric for Top-$n$ Recommendation
On (Normalised) Discounted Cumulative Gain as an Off-Policy Evaluation Metric for Top-nnn Recommendation
Olivier Jeunen
Ivan Potapov
Aleksei Ustimenko
ELMOffRL
111
12
0
27 Jul 2023
Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Off-Policy Evaluation of Ranking Policies under Diverse User Behavior
Haruka Kiyohara
Masatoshi Uehara
Yusuke Narita
N. Shimizu
Yasuo Yamamoto
Yuta Saito
OffRLCML
83
9
0
26 Jun 2023
Off-Policy Evaluation for Large Action Spaces via Conjunct Effect
  Modeling
Off-Policy Evaluation for Large Action Spaces via Conjunct Effect Modeling
Yuta Saito
Qingyang Ren
Thorsten Joachims
CMLOffRL
67
26
0
14 May 2023
Variational Latent Branching Model for Off-Policy Evaluation
Variational Latent Branching Model for Off-Policy Evaluation
Qitong Gao
Ge Gao
Min Chi
Miroslav Pajic
OffRL
84
6
0
28 Jan 2023
Local Policy Improvement for Recommender Systems
Local Policy Improvement for Recommender Systems
Dawen Liang
N. Vlassis
OffRL
35
5
0
22 Dec 2022
Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Policy-Adaptive Estimator Selection for Off-Policy Evaluation
Takuma Udagawa
Haruka Kiyohara
Yusuke Narita
Yuta Saito
Keisuke Tateno
OffRL
53
25
0
25 Nov 2022
Off-Policy Evaluation for Large Action Spaces via Embeddings
Off-Policy Evaluation for Large Action Spaces via Embeddings
Yuta Saito
Thorsten Joachims
OffRL
82
46
0
13 Feb 2022
Doubly Robust Off-Policy Evaluation for Ranking Policies under the
  Cascade Behavior Model
Doubly Robust Off-Policy Evaluation for Ranking Policies under the Cascade Behavior Model
Haruka Kiyohara
Yuta Saito
Tatsuya Matsuhiro
Yusuke Narita
N. Shimizu
Yasuo Yamamoto
OffRL
81
43
0
03 Feb 2022
Accelerating Offline Reinforcement Learning Application in Real-Time
  Bidding and Recommendation: Potential Use of Simulation
Accelerating Offline Reinforcement Learning Application in Real-Time Bidding and Recommendation: Potential Use of Simulation
Haruka Kiyohara
K. Kawakami
Yuta Saito
OffRL
71
13
0
17 Sep 2021
Open Bandit Dataset and Pipeline: Towards Realistic and Reproducible
  Off-Policy Evaluation
Open Bandit Dataset and Pipeline: Towards Realistic and Reproducible Off-Policy Evaluation
Yuta Saito
Shunsuke Aihara
Megumi Matsutani
Yusuke Narita
OffRL
201
75
0
17 Aug 2020
Towards Resolving Propensity Contradiction in Offline Recommender
  Learning
Towards Resolving Propensity Contradiction in Offline Recommender Learning
Yuta Saito
Masahiro Nomura
OffRL
81
13
0
16 Oct 2019
1