ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.11642
  4. Cited By
Off-Policy Evaluation and Learning for External Validity under a
  Covariate Shift
v1v2v3 (latest)

Off-Policy Evaluation and Learning for External Validity under a Covariate Shift

Neural Information Processing Systems (NeurIPS), 2020
26 February 2020
Masahiro Kato
Masatoshi Uehara
Shota Yasui
    OffRL
ArXiv (abs)PDFHTML

Papers citing "Off-Policy Evaluation and Learning for External Validity under a Covariate Shift"

28 / 28 papers shown
Semi-Supervised Treatment Effect Estimation with Unlabeled Covariates via Generalized Riesz Regression
Semi-Supervised Treatment Effect Estimation with Unlabeled Covariates via Generalized Riesz Regression
Masahiro Kato
CML
288
1
0
11 Nov 2025
Nearest Neighbor Matching as Least Squares Density Ratio Estimation and Riesz Regression
Nearest Neighbor Matching as Least Squares Density Ratio Estimation and Riesz Regression
Masahiro Kato
125
7
0
28 Oct 2025
Direct Debiased Machine Learning via Bregman Divergence Minimization
Direct Debiased Machine Learning via Bregman Divergence Minimization
Masahiro Kato
CML
419
6
0
27 Oct 2025
Two-sample comparison through additive tree models for density ratios
Two-sample comparison through additive tree models for density ratios
Naoki Awaya
Yuliang Xu
Li Ma
239
1
0
05 Aug 2025
Doubly Robust Alignment for Large Language Models
Doubly Robust Alignment for Large Language Models
Erhan Xu
Kai Ye
Hongyi Zhou
Luhan Zhu
Francesco Quinzan
Chengchun Shi
347
7
0
01 Jun 2025
Active Adaptive Experimental Design for Treatment Effect Estimation with
  Covariate Choices
Active Adaptive Experimental Design for Treatment Effect Estimation with Covariate Choices
Masahiro Kato
Akihiro Oga
Wataru Komatsubara
Ryo Inokuchi
375
11
0
06 Mar 2024
Distributionally Robust Policy Evaluation under General Covariate Shift
  in Contextual Bandits
Distributionally Robust Policy Evaluation under General Covariate Shift in Contextual Bandits
Yi Guo
Hao Liu
Yisong Yue
Anqi Liu
OffRL
296
3
0
21 Jan 2024
Sample Complexity of Preference-Based Nonparametric Off-Policy
  Evaluation with Deep Networks
Sample Complexity of Preference-Based Nonparametric Off-Policy Evaluation with Deep Networks
Zihao Li
Xiang Ji
Minshuo Chen
Mengdi Wang
OffRL
292
0
0
16 Oct 2023
Federated Learning under Covariate Shifts with Generalization Guarantees
Federated Learning under Covariate Shifts with Generalization Guarantees
Ali Ramezani-Kebrya
Fanghui Liu
Thomas Pethick
Grigorios G. Chrysos
Volkan Cevher
FedMLOOD
367
12
0
08 Jun 2023
MAGDiff: Covariate Data Set Shift Detection via Activation Graphs of
  Deep Neural Networks
MAGDiff: Covariate Data Set Shift Detection via Activation Graphs of Deep Neural Networks
Charles Arnal
Felix Hensel
Mathieu Carrière
Théo Lacombe
Hiroaki Kurihara
Yuichi Ike
Frédéric Chazal
172
2
0
22 May 2023
Learning Complementary Policies for Human-AI Teams
Learning Complementary Policies for Human-AI Teams
Ruijiang Gao
M. Saar-Tsechansky
Maria De-Arteaga
349
11
0
06 Feb 2023
Offline Minimax Soft-Q-learning Under Realizability and Partial Coverage
Offline Minimax Soft-Q-learning Under Realizability and Partial CoverageNeural Information Processing Systems (NeurIPS), 2023
Masatoshi Uehara
Nathan Kallus
Jason D. Lee
Wen Sun
OffRL
372
8
0
05 Feb 2023
Counterfactual Learning with Multioutput Deep Kernels
Counterfactual Learning with Multioutput Deep Kernels
A. Caron
G. Baio
I. Manolopoulou
BDLCMLOffRL
226
2
0
20 Nov 2022
Bayesian Counterfactual Mean Embeddings and Off-Policy Evaluation
Bayesian Counterfactual Mean Embeddings and Off-Policy Evaluation
Diego Martinez-Taboada
Dino Sejdinovic
CMLOffRL
166
0
0
02 Nov 2022
Unified Perspective on Probability Divergence via Maximum Likelihood
  Density Ratio Estimation: Bridging KL-Divergence and Integral Probability
  Metrics
Unified Perspective on Probability Divergence via Maximum Likelihood Density Ratio Estimation: Bridging KL-Divergence and Integral Probability Metrics
Masahiro Kato
Masaaki Imaizumi
Kentaro Minami
222
0
0
31 Jan 2022
Generalizing Off-Policy Learning under Sample Selection Bias
Generalizing Off-Policy Learning under Sample Selection Bias
Tobias Hatt
D. Tschernutter
Stefan Feuerriegel
OffRL
198
21
0
02 Dec 2021
Rejoinder: Learning Optimal Distributionally Robust Individualized
  Treatment Rules
Rejoinder: Learning Optimal Distributionally Robust Individualized Treatment Rules
Weibin Mo
Zhengling Qi
Yufeng Liu
125
0
0
17 Oct 2021
Evaluating the Robustness of Off-Policy Evaluation
Evaluating the Robustness of Off-Policy EvaluationACM Conference on Recommender Systems (RecSys), 2021
Yuta Saito
Takuma Udagawa
Haruka Kiyohara
Kazuki Mogi
Yusuke Narita
Kei Tateno
ELMOffRL
278
47
0
31 Aug 2021
Combining Online Learning and Offline Learning for Contextual Bandits
  with Deficient Support
Combining Online Learning and Offline Learning for Contextual Bandits with Deficient Support
Hung The Tran
Sunil R. Gupta
Thanh Nguyen-Tang
Santu Rana
Svetha Venkatesh
OffRL
174
6
0
24 Jul 2021
Optimal Uniform OPE and Model-based Offline Reinforcement Learning in
  Time-Homogeneous, Reward-Free and Task-Agnostic Settings
Optimal Uniform OPE and Model-based Offline Reinforcement Learning in Time-Homogeneous, Reward-Free and Task-Agnostic SettingsNeural Information Processing Systems (NeurIPS), 2021
Ming Yin
Yu Wang
OffRL
313
19
0
13 May 2021
Learning Under Adversarial and Interventional Shifts
Learning Under Adversarial and Interventional Shifts
Harvineet Singh
Shalmali Joshi
Finale Doshi-Velez
Himabindu Lakkaraju
OOD
193
4
0
29 Mar 2021
Reliable Off-policy Evaluation for Reinforcement Learning
Reliable Off-policy Evaluation for Reinforcement Learning
Jie Wang
Rui Gao
H. Zha
OffRL
445
13
0
08 Nov 2020
Off-Policy Evaluation of Bandit Algorithm from Dependent Samples under
  Batch Update Policy
Off-Policy Evaluation of Bandit Algorithm from Dependent Samples under Batch Update Policy
Masahiro Kato
Yusuke Kaneko
OffRL
232
4
0
23 Oct 2020
A Practical Guide of Off-Policy Evaluation for Bandit Problems
A Practical Guide of Off-Policy Evaluation for Bandit Problems
Masahiro Kato
Kenshi Abe
Kaito Ariu
Shota Yasui
OffRL
212
3
0
23 Oct 2020
Open Bandit Dataset and Pipeline: Towards Realistic and Reproducible
  Off-Policy Evaluation
Open Bandit Dataset and Pipeline: Towards Realistic and Reproducible Off-Policy Evaluation
Yuta Saito
Shunsuke Aihara
Megumi Matsutani
Yusuke Narita
OffRL
711
91
0
17 Aug 2020
Confidence Interval for Off-Policy Evaluation from Dependent Samples via
  Bandit Algorithm: Approach from Standardized Martingales
Confidence Interval for Off-Policy Evaluation from Dependent Samples via Bandit Algorithm: Approach from Standardized Martingales
Masahiro Kato
OffRL
195
3
0
12 Jun 2020
Non-Negative Bregman Divergence Minimization for Deep Direct Density
  Ratio Estimation
Non-Negative Bregman Divergence Minimization for Deep Direct Density Ratio EstimationInternational Conference on Machine Learning (ICML), 2020
Masahiro Kato
Takeshi Teshima
432
51
0
12 Jun 2020
Counterfactual Mean Embeddings
Counterfactual Mean Embeddings
Krikamol Muandet
Motonobu Kanagawa
Sorawit Saengkyongam
S. Marukatat
CMLOffRL
346
47
0
22 May 2018
1
Page 1 of 1