Skip to content

Commit

Permalink
Add a link to Pgx - a collection of 20+ JAX environments.
Browse files Browse the repository at this point in the history
Fixes #65.

PiperOrigin-RevId: 569229226
  • Loading branch information
fidlej authored and MctxDev committed Sep 28, 2023
1 parent 2ff5048 commit 021f72a
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 4 deletions.
9 changes: 6 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ pip install mctx
or you can install the latest development version from GitHub:

```sh
pip install git+https://github.com/deepmind/mctx.git
pip install git+https://github.com/google-deepmind/mctx.git
```

## Motivation
Expand Down Expand Up @@ -93,7 +93,7 @@ new_embedding)` with a `RecurrentFnOutput` and the embedding of the next state.
The `RecurrentFnOutput` contains the `reward` and `discount` for the transition,
and `prior_logits` and `value` for the new state.

In [`examples/visualization_demo.py`](https://github.com/deepmind/mctx/blob/main/examples/visualization_demo.py), you can
In [`examples/visualization_demo.py`](https://github.com/google-deepmind/mctx/blob/main/examples/visualization_demo.py), you can
see calls to a policy:

```python
Expand All @@ -110,11 +110,14 @@ We recommend to use the `gumbel_muzero_policy`.
[Gumbel MuZero](https://openreview.net/forum?id=bERaNdoegnO) guarantees a policy
improvement if the action values are correctly evaluated. The policy improvement
is demonstrated in
[`examples/policy_improvement_demo.py`](https://github.com/deepmind/mctx/blob/main/examples/policy_improvement_demo.py).
[`examples/policy_improvement_demo.py`](https://github.com/google-deepmind/mctx/blob/main/examples/policy_improvement_demo.py).

### Example projects
The following projects demonstrate the Mctx usage:

- [Pgx](https://github.com/sotetsuk/pgx) — A collection of 20+ vectorized
JAX environments, including backgammon, chess, shogi, Go, and an AlphaZero
example.
- [Basic Learning Demo with Mctx](https://github.com/kenjyoung/mctx_learning_demo)
AlphaZero on random mazes.
- [a0-jax](https://github.com/NTT123/a0-jax) — AlphaZero on Connect Four,
Expand Down
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -44,7 +44,7 @@ def _parse_requirements(path):
setup(
name='mctx',
version=_get_version(),
url='https://github.com/deepmind/mctx',
url='https://github.com/google-deepmind/mctx',
license='Apache 2.0',
author='DeepMind',
description=('Monte Carlo tree search in JAX.'),
Expand Down

0 comments on commit 021f72a

Please sign in to comment.