From 6c77c2e36c49233994707409809993e99affa05f Mon Sep 17 00:00:00 2001 From: Joshua Spear <52503016+joshuaspear@users.noreply.github.com> Date: Mon, 22 Jul 2024 17:06:51 +0100 Subject: [PATCH] Update README.md --- README.md | 6 +----- 1 file changed, 1 insertion(+), 5 deletions(-) diff --git a/README.md b/README.md index ec852cc..fd1cab7 100644 --- a/README.md +++ b/README.md @@ -65,18 +65,14 @@ Basic unit testing has been implemented for all the core functionality of the pa * PyPi: https://pypi.org/project/offline-rl-ope/ * To install from sourse using pip, clone this repository and run ```pip install .``` in the top level of the repo. -### Limitations -* When using DM methods with the d3rlpy training API, the FQE model is required to be refitted for every epoch. This could be improved by trainin the FQE model once, before training of the offline RL model commences - ### Future work * Async/multithread support * Additional estimators: * DualDICE * MAGIC - * Extended DR estimator as per equation 12 in https://arxiv.org/pdf/1511.03722.pdf + * State importance sampling * APIs * Add additional apis e.g. for stable baselines - * Address limitations with existing DM integration with d3rlpy ### Tips