Alert button
Picture for Marek Petrik

Marek Petrik

Alert button

Percentile Criterion Optimization in Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Apr 07, 2024
Elita A. Lobo, Cyrus Cousins, Yair Zick, Marek Petrik

Viaarxiv icon

Data Poisoning Attacks on Off-Policy Policy Evaluation Methods

Add code
Bookmark button
Alert button
Apr 06, 2024
Elita Lobo, Harvineet Singh, Marek Petrik, Cynthia Rudin, Himabindu Lakkaraju

Viaarxiv icon

A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits

Add code
Bookmark button
Alert button
Jun 02, 2023
Mohammad Ghavamzadeh, Marek Petrik, Guy Tennenholtz

Figure 1 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Figure 2 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Figure 3 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Figure 4 for A Convex Relaxation Approach to Bayesian Regret Minimization in Offline Bandits
Viaarxiv icon

On Dynamic Program Decompositions of Static Risk Measures

Add code
Bookmark button
Alert button
Apr 24, 2023
Jia Lin Hau, Erick Delage, Mohammad Ghavamzadeh, Marek Petrik

Figure 1 for On Dynamic Program Decompositions of Static Risk Measures
Figure 2 for On Dynamic Program Decompositions of Static Risk Measures
Viaarxiv icon

Reducing Blackwell and Average Optimality to Discounted MDPs via the Blackwell Discount Factor

Add code
Bookmark button
Alert button
Jan 31, 2023
Julien Grand-Clément, Marek Petrik

Figure 1 for Reducing Blackwell and Average Optimality to Discounted MDPs via the Blackwell Discount Factor
Figure 2 for Reducing Blackwell and Average Optimality to Discounted MDPs via the Blackwell Discount Factor
Viaarxiv icon

On the Convergence of Policy Gradient in Robust MDPs

Add code
Bookmark button
Alert button
Dec 20, 2022
Qiuhao Wang, Chin Pang Ho, Marek Petrik

Figure 1 for On the Convergence of Policy Gradient in Robust MDPs
Figure 2 for On the Convergence of Policy Gradient in Robust MDPs
Figure 3 for On the Convergence of Policy Gradient in Robust MDPs
Viaarxiv icon

On the convex formulations of robust Markov decision processes

Add code
Bookmark button
Alert button
Sep 21, 2022
Julien Grand-Clément, Marek Petrik

Figure 1 for On the convex formulations of robust Markov decision processes
Figure 2 for On the convex formulations of robust Markov decision processes
Figure 3 for On the convex formulations of robust Markov decision processes
Figure 4 for On the convex formulations of robust Markov decision processes
Viaarxiv icon

RASR: Risk-Averse Soft-Robust MDPs with EVaR and Entropic Risk

Add code
Bookmark button
Alert button
Sep 14, 2022
Jia Lin Hau, Marek Petrik, Mohammad Ghavamzadeh, Reazul Russel

Figure 1 for RASR: Risk-Averse Soft-Robust MDPs with EVaR and Entropic Risk
Figure 2 for RASR: Risk-Averse Soft-Robust MDPs with EVaR and Entropic Risk
Figure 3 for RASR: Risk-Averse Soft-Robust MDPs with EVaR and Entropic Risk
Figure 4 for RASR: Risk-Averse Soft-Robust MDPs with EVaR and Entropic Risk
Viaarxiv icon

Robust Phi-Divergence MDPs

Add code
Bookmark button
Alert button
May 27, 2022
Chin Pang Ho, Marek Petrik, Wolfram Wiesemann

Figure 1 for Robust Phi-Divergence MDPs
Figure 2 for Robust Phi-Divergence MDPs
Figure 3 for Robust Phi-Divergence MDPs
Figure 4 for Robust Phi-Divergence MDPs
Viaarxiv icon