Nathan Kallus

Orcid: 0000-0003-1672-0507

According to our database1, Nathan Kallus authored at least 106 papers between 2014 and 2024.

Collaborative distances:
  • Dijkstra number2 of four.
  • Erdős number3 of four.

Timeline

Legend:

Book 
In proceedings 
Article 
PhD thesis 
Dataset
Other 

Links

On csauthors.net:

Bibliography

2024
Hessian-Free Laplace in Bayesian Deep Learning.
CoRR, 2024

Risk-Sensitive RL with Optimized Certainty Equivalents via Reduction to Standard RL.
CoRR, 2024

Is Cosine-Similarity of Embeddings Really About Similarity?
CoRR, 2024

Switching the Loss Reduces the Cost in Batch Reinforcement Learning.
CoRR, 2024

Applied Causal Inference Powered by ML and AI.
CoRR, 2024

More Benefits of Being Distributional: Second-Order Bounds for Reinforcement Learning.
CoRR, 2024

Peeking with PEAK: Sequential, Nonparametric Composite Hypothesis Tests for Means of Multiple Data Streams.
CoRR, 2024

Multi-Armed Bandits with Interference.
CoRR, 2024

2023
Treatment Effect Risk: Bounds and Inference.
Manag. Sci., August, 2023

Stochastic Optimization Forests.
Manag. Sci., April, 2023

The Power and Limits of Predictive Approaches to Observational Data-Driven Optimization: The Case of Pricing.
INFORMS J. Optim., January, 2023

Faster Rates for Switchback Experiments.
CoRR, 2023

Low-Rank MDPs with Continuous Action Spaces.
CoRR, 2023

Off-Policy Evaluation for Large Action Spaces via Policy Convolution.
CoRR, 2023

Source Condition Double Robust Inference on Functionals of Inverse Problems.
CoRR, 2023

JoinGym: An Efficient Query Optimization Environment for Reinforcement Learning.
CoRR, 2023

Provable Offline Reinforcement Learning with Human Feedback.
CoRR, 2023

Minimax Instrumental Variable Regression and $L_2$ Convergence Guarantees without Identification or Closedness.
CoRR, 2023

Refined Value-Based Offline RL under Realizability and Partial Coverage.
CoRR, 2023

The Benefits of Being Distributional: Small-Loss Bounds for Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Offline Minimax Soft-Q-learning Under Realizability and Partial Coverage.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Future-Dependent Value-Based Off-Policy Evaluation in POMDPs.
Proceedings of the Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, 2023

Near-Minimax-Optimal Risk-Sensitive Reinforcement Learning with CVaR.
Proceedings of the International Conference on Machine Learning, 2023

Computationally Efficient PAC RL in POMDPs with Latent Determinism and Conditional Embeddings.
Proceedings of the International Conference on Machine Learning, 2023

B-Learner: Quasi-Oracle Bounds on Heterogeneous Causal Effects Under Hidden Confounding.
Proceedings of the International Conference on Machine Learning, 2023

Smooth Non-stationary Bandits.
Proceedings of the International Conference on Machine Learning, 2023

Minimax Instrumental Variable Regression and L<sub>2</sub> Convergence Guarantees without Identification or Closedness.
Proceedings of the Thirty Sixth Annual Conference on Learning Theory, 2023

Inference on Strongly Identified Functionals of Weakly Identified Functions.
Proceedings of the Thirty Sixth Annual Conference on Learning Theory, 2023

Large Language Models as Zero-Shot Conversational Recommenders.
Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, 2023

Robust and Agnostic Learning of Conditional Distributional Treatment Effects.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2023

Provable Safe Reinforcement Learning with Binary Feedback.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2023

2022
Efficiently Breaking the Curse of Horizon in Off-Policy Evaluation with Double Reinforcement Learning.
Oper. Res., November, 2022

Smooth Contextual Bandits: Bridging the Parametric and Nondifferentiable Regret Regimes.
Oper. Res., November, 2022

Assessing Algorithmic Fairness with Unobserved Protected Class Using Data Combination.
Manag. Sci., 2022

Fast Rates for Contextual Linear Optimization.
Manag. Sci., 2022

Data Pooling in Stochastic Optimization.
Manag. Sci., 2022

Generalization Bounds and Representation Learning for Estimation of Potential Outcomes and Causal Effects.
J. Mach. Learn. Res., 2022

A Review of Off-Policy Evaluation in Reinforcement Learning.
CoRR, 2022

Provably Efficient Reinforcement Learning in Partially Observable Dynamical Systems.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

The Implicit Delta Method.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

What's the Harm? Sharp Bounds on the Fraction Negatively Affected by Treatment.
Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, 2022

Doubly Robust Distributionally Robust Off-Policy Evaluation and Learning.
Proceedings of the International Conference on Machine Learning, 2022

Learning Bellman Complete Representations for Offline Policy Evaluation.
Proceedings of the International Conference on Machine Learning, 2022

Estimating Structural Disparities for Face Models.
Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022

Stateful Offline Contextual Policy Evaluation and Learning.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2022

2021
Minimax-Optimal Policy Learning Under Unobserved Confounding.
Manag. Sci., 2021

Doubly-Valid/Doubly-Sharp Sensitivity Analysis for Causal Inference with Unmeasured Confounding.
CoRR, 2021

An Empirical Evaluation of the Impact of New York's Bail Reform on Crime Using Synthetic Controls.
CoRR, 2021

Proximal Reinforcement Learning: Efficient Off-Policy Evaluation in Partially Observed Markov Decision Processes.
CoRR, 2021

Residual Overfit Method of Exploration.
CoRR, 2021

Causal Inference Under Unmeasured Confounding With Negative Controls: A Minimax Learning Approach.
CoRR, 2021

Finite Sample Analysis of Minimax Offline Reinforcement Learning: Completeness, Fast Rates and First-Order Efficiency.
CoRR, 2021

Control Variates for Slate Off-Policy Evaluation.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Risk Minimization from Adaptively Collected Data: Guarantees for Supervised and Policy Learning.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Post-Contextual-Bandit Inference.
Proceedings of the Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, 2021

Optimal Off-Policy Evaluation from Multiple Logging Policies.
Proceedings of the 38th International Conference on Machine Learning, 2021

Fairness, Welfare, and Equity in Personalized Pricing.
Proceedings of the FAccT '21: 2021 ACM Conference on Fairness, 2021

Fast Rates for the Regret of Offline Reinforcement Learning.
Proceedings of the Conference on Learning Theory, 2021

Off-policy Evaluation in Infinite-Horizon Reinforcement Learning with Latent Confounders.
Proceedings of the 24th International Conference on Artificial Intelligence and Statistics, 2021

2020
From Predictive to Prescriptive Analytics.
Manag. Sci., 2020

Double Reinforcement Learning for Efficient Off-Policy Evaluation in Markov Decision Processes.
J. Mach. Learn. Res., 2020

Generalized Optimal Matching Methods for Causal Inference.
J. Mach. Learn. Res., 2020

Dynamic Assortment Personalization in High Dimensions.
Oper. Res., 2020

The Variational Method of Moments.
CoRR, 2020

Rejoinder: New Objectives for Policy Learning.
CoRR, 2020

Efficient Evaluation of Natural Stochastic Policies in Offline Reinforcement Learning.
CoRR, 2020

DTR Bandit: Learning to Make Response-Adaptive Decisions With Low Regret.
CoRR, 2020

Comment: Entropy Learning for Dynamic Treatment Regimes.
CoRR, 2020

On the role of surrogates in the efficient estimation of treatment effects with limited outcome data.
CoRR, 2020

Confounding-Robust Policy Evaluation in Infinite-Horizon Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Doubly Robust Off-Policy Value and Gradient Estimation for Deterministic Policies.
Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, 2020

Statistically Efficient Off-Policy Policy Gradients.
Proceedings of the 37th International Conference on Machine Learning, 2020

Double Reinforcement Learning for Efficient and Robust Off-Policy Evaluation.
Proceedings of the 37th International Conference on Machine Learning, 2020

DeepMatch: Balancing Deep Covariate Representations for Causal Inference Using Adversarial Training.
Proceedings of the 37th International Conference on Machine Learning, 2020

Efficient Policy Learning from Surrogate-Loss Classification Reductions.
Proceedings of the 37th International Conference on Machine Learning, 2020

Smooth Contextual Bandits: Bridging the Parametric and Non-differentiable Regret Regimes.
Proceedings of the Conference on Learning Theory, 2020

2019
Localized Debiased Machine Learning: Efficient Estimation of Quantile Treatment Effects, Conditional Value at Risk, and Beyond.
CoRR, 2019

Efficiently Breaking the Curse of Horizon: Double Reinforcement Learning in Infinite-Horizon Processes.
CoRR, 2019

More Efficient Policy Learning via Optimal Retargeting.
CoRR, 2019

Assessing Disparate Impacts of Personalized Interventions: Identifiability and Bounds.
CoRR, 2019

The Fairness of Risk Scores Beyond Classification: Bipartite Ranking and the XAUC Metric.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Assessing Disparate Impact of Personalized Interventions: Identifiability and Bounds.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Intrinsically Efficient, Stable, and Bounded Off-Policy Evaluation for Reinforcement Learning.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Deep Generalized Method of Moments for Instrumental Variable Analysis.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Policy Evaluation with Latent Confounders via Optimal Balance.
Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, 2019

Classifying Treatment Responders Under Causal Effect Monotonicity.
Proceedings of the 36th International Conference on Machine Learning, 2019

Fairness Under Unawareness: Assessing Disparity When Protected Class Is Unobserved.
Proceedings of the Conference on Fairness, Accountability, and Transparency, 2019

Interval Estimation of Individual-Level Causal Effects Under Unobserved Confounding.
Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics, 2019

2018
Robust sample average approximation.
Math. Program., 2018

Data-driven robust optimization.
Math. Program., 2018

Confounding-Robust Policy Improvement.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Removing Hidden Confounding by Experimental Grounding.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Causal Inference with Noisy and Missing Covariates via Matrix Factorization.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Balanced Policy Evaluation and Learning.
Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, 2018

Residual Unfairness in Fair Machine Learning from Prejudiced Data.
Proceedings of the 35th International Conference on Machine Learning, 2018

Instrument-Armed Bandits.
Proceedings of the Algorithmic Learning Theory, 2018

Policy Evaluation and Optimization with Continuous Treatments.
Proceedings of the International Conference on Artificial Intelligence and Statistics, 2018

2017
Recursive Partitioning for Personalization using Observational Data.
Proceedings of the 34th International Conference on Machine Learning, 2017

A Framework for Optimal Matching for Causal Inference.
Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, 2017

2016
Causal Inference by Minimizing the Dual Norm of Bias: Kernel Matching & Weighting Estimators for Causal Effects.
Proceedings of the UAI 2016 Workshop on Causation: Foundation to Application co-located with the 32nd Conference on Uncertainty in Artificial Intelligence (UAI 2016), 2016

Revealed Preference at Scale: Learning Personalized Preferences from Assortment Choices.
Proceedings of the 2016 ACM Conference on Economics and Computation, 2016

2015
The Power of Optimization Over Randomization in Designing Experiments Involving Small Samples.
Oper. Res., 2015

Learning Preferences from Assortment Choices in a Heterogeneous Population.
CoRR, 2015

2014
From Predictions to Data-Driven Decisions Using Machine Learning.
CoRR, 2014

Predicting crowd behavior with big public data.
Proceedings of the 23rd International World Wide Web Conference, 2014

On the Predictive Power of Web Intelligence and Social Media - The Best Way to Predict the Future Is to tweet It.
Proceedings of the Big Data Analytics in the Social and Ubiquitous Context, 2014


  Loading...