Skip to content

Commit

Permalink
update examples
Browse files Browse the repository at this point in the history
  • Loading branch information
jakegrigsby committed Nov 3, 2023
1 parent b355caa commit 08b27a9
Show file tree
Hide file tree
Showing 5 changed files with 28 additions and 14 deletions.
2 changes: 1 addition & 1 deletion _site/feed.xml
Original file line number Diff line number Diff line change
@@ -1 +1 @@
<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="4.2.2">Jekyll</generator><link href="http://localhost:4000/feed.xml" rel="self" type="application/atom+xml" /><link href="http://localhost:4000/" rel="alternate" type="text/html" /><updated>2023-11-02T23:11:02-05:00</updated><id>http://localhost:4000/feed.xml</id><title type="html">AMAGO</title><subtitle>A simple and scalable agent for sequence-based RL</subtitle></feed>
<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="4.2.2">Jekyll</generator><link href="http://localhost:4000/feed.xml" rel="self" type="application/atom+xml" /><link href="http://localhost:4000/" rel="alternate" type="text/html" /><updated>2023-11-02T23:51:57-05:00</updated><id>http://localhost:4000/feed.xml</id><title type="html">AMAGO</title><subtitle>A simple and scalable agent for sequence-based RL</subtitle></feed>
21 changes: 14 additions & 7 deletions _site/index.html
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@


<!-- <meta property="og:image" content="src/figure/approach.png"> -->
<meta property="og:title" content="TRILL" />
<meta property="og:title" content="AMAGO" />

<script src="./src/popup.js" type="text/javascript"></script>

Expand Down Expand Up @@ -157,7 +157,6 @@ <h2>
<a href="https://www.utexas.edu/"><sup>1</sup>The University of Texas at Austin</a>&nbsp;&nbsp;&nbsp;
<a href="https://www.nvidia.com/en-us/research/"><sup>2</sup>NVIDIA Research</a>&nbsp;&nbsp;&nbsp;
</h2>
<img src="./src/logos/amago_logo_2.png" alt="amagologo" width="250" align="center" />
<h2><a href="https://arxiv.org/abs/2310.09971v1">Paper</a>&nbsp; | &nbsp;<a href="https://github.com/UT-Austin-RPL/amago">Code</a></h2>
</center>

Expand All @@ -169,10 +168,13 @@ <h2><a href="https://arxiv.org/abs/2310.09971v1">Paper</a>&nbsp; | &nbsp;<a href
<table align="center" width="800px">
<tr>
<td>
<p align="center" width="20%">
<p align="justify" width="20%">
<img src="./src/logos/amago_logo_2.png" alt="amagologo" width="25%" align="right" />
<div style="display:inline">
<h3>
"In-context" RL trains memory-equipped agents to adapt to new environments from test-time experience and unifies meta-RL, zero-shot generalization, and long-term memory into a single problem. While this technique was one of the first approaches to deep meta-RL <a href="https://arxiv.org/abs/1611.02779">[1]</a>, it is often outperformed by more complicated methods. Fortunately, the right off-policy implementation details and tuning can make in-context RL stable and competitive <a href="https://arxiv.org/abs/2110.05038">[2]</a>. Off-policy in-context RL creates a tradeoff because it is conceptually simple but hard to use, and agents are limited by their model size, memory length, and planning horizon. <b>AMAGO</b> redesigns off-policy sequence-based RL to break these bottlenecks and stably train long-context Transformers with end-to-end RL. AMAGO is open-source and designed to require minimal tuning with the goal of making in-context RL an easy-to-use default in new research on adaptive agents. <br /><br />
</h3>
</div>
</p></td></tr></table>
</p>
</div>
Expand Down Expand Up @@ -315,10 +317,15 @@ <h4>
<h1 align="center">Using AMAGO</h1>

<h4>
In-context RL is applicable to any memory, generalization, or meta-learning problem, and we have designed AMAGO to be flexible enough to support all of these cases. Our code is fully open-source and includes examples of how to apply AMAGO to new domains. We hope our agent can serve as a strong baseline in the development of new benchmarks that require long-term memory and adaptation. <a href="https://github.com/UT-Austin-RPL/amago">Check it out on GitHub here</a>.
</h4>
In-context RL is applicable to any memory, generalization, or meta-learning problem, and we have designed AMAGO to be flexible enough to support all of these cases. Our code is fully open-source <a href="https://github.com/UT-Austin-RPL/amago">and available on GitHub</a>. We hope our agent can serve as a strong baseline in the development of new benchmarks that require long-term memory and adaptation, and include many <a href="https://github.com/UT-Austin-RPL/amago/tree/main/examples">examples</a> of how to apply AMAGO to:
<ul>
<li> Standard (Memory-Free) MDPs/gym Environments </li>
<li> POMDPs and Long-Term Memory Tasks </li>
<li> K-Shot Meta-RL </li>
<li> Goal-Conditioned Environment Adaptation </li>
<li> Multi-task Learning from Pixels </li>
</ul>

<a href="https://github.com/UT-Austin-RPL/amago"><img src="./src/logos/rpl_logo.png" style="width:30%;" /> </a>



Expand All @@ -342,4 +349,4 @@ <h4>
</tr>
</table>

</div></body>
</h4></div></body>
Binary file added _site/src/logos/amago_logo_white_bg.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
19 changes: 13 additions & 6 deletions index.markdown
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ usemathjax: true


<!-- <meta property="og:image" content="src/figure/approach.png"> -->
<meta property="og:title" content="TRILL">
<meta property="og:title" content="AMAGO">

<script src="./src/popup.js" type="text/javascript"></script>

Expand Down Expand Up @@ -168,7 +168,6 @@ highlight {
<a href="https://www.utexas.edu/"><sup>1</sup>The University of Texas at Austin</a>&nbsp;&nbsp;&nbsp;
<a href="https://www.nvidia.com/en-us/research/"><sup>2</sup>NVIDIA Research</a>&nbsp;&nbsp;&nbsp;
</h2>
<img src="./src/logos/amago_logo_2.png" alt="amagologo" width="250" align="center"/>
<h2><a href="https://arxiv.org/abs/2310.09971v1">Paper</a>&nbsp; | &nbsp;<a href="https://github.com/UT-Austin-RPL/amago">Code</a></h2>
</center>

Expand All @@ -180,10 +179,13 @@ highlight {
<table align=center width=800px>
<tr>
<td>
<p align="center" width="20%">
<p align="justify" width="20%">
<img src="./src/logos/amago_logo_2.png" alt="amagologo" width="25%" align="right"/>
<div style="display:inline">
<h3>
"In-context" RL trains memory-equipped agents to adapt to new environments from test-time experience and unifies meta-RL, zero-shot generalization, and long-term memory into a single problem. While this technique was one of the first approaches to deep meta-RL <a href="https://arxiv.org/abs/1611.02779">[1]</a>, it is often outperformed by more complicated methods. Fortunately, the right off-policy implementation details and tuning can make in-context RL stable and competitive <a href="https://arxiv.org/abs/2110.05038">[2]</a>. Off-policy in-context RL creates a tradeoff because it is conceptually simple but hard to use, and agents are limited by their model size, memory length, and planning horizon. <b>AMAGO</b> redesigns off-policy sequence-based RL to break these bottlenecks and stably train long-context Transformers with end-to-end RL. AMAGO is open-source and designed to require minimal tuning with the goal of making in-context RL an easy-to-use default in new research on adaptive agents. <br><br>
</h3>
</div>
</p></td></tr></table>
</p>
</div>
Expand Down Expand Up @@ -326,10 +328,15 @@ Above, we use several single-task instructions to evaluate the exploration capab
<h1 align="center">Using AMAGO</h1>

<h4>
In-context RL is applicable to any memory, generalization, or meta-learning problem, and we have designed AMAGO to be flexible enough to support all of these cases. Our code is fully open-source and includes examples of how to apply AMAGO to new domains. We hope our agent can serve as a strong baseline in the development of new benchmarks that require long-term memory and adaptation. <a href="https://github.com/UT-Austin-RPL/amago">Check it out on GitHub here</a>.
</h4>
In-context RL is applicable to any memory, generalization, or meta-learning problem, and we have designed AMAGO to be flexible enough to support all of these cases. Our code is fully open-source <a href="https://github.com/UT-Austin-RPL/amago">and available on GitHub</a>. We hope our agent can serve as a strong baseline in the development of new benchmarks that require long-term memory and adaptation, and include many <a href="https://github.com/UT-Austin-RPL/amago/tree/main/examples">examples</a> of how to apply AMAGO to:
<ul>
<li> Standard (Memory-Free) MDPs/gym Environments </li>
<li> POMDPs and Long-Term Memory Tasks </li>
<li> K-Shot Meta-RL </li>
<li> Goal-Conditioned Environment Adaptation </li>
<li> Multi-task Learning from Pixels </li>
</ul>

<a href="https://github.com/UT-Austin-RPL/amago"><img src="./src/logos/rpl_logo.png" style="width:30%;"> </a>



Expand Down
Binary file added src/logos/amago_logo_white_bg.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.

0 comments on commit 08b27a9

Please sign in to comment.