Skip to content

Model or optimizer wrapping? #3

Answered by andreped
andreped asked this question in Q&A
Discussion options

You must be logged in to vote

In the latest release v0.3.0 now we support both approaches:
https://github.com/andreped/GradientAccumulator/releases/tag/v0.3.0

The main reason why optimizer wrapping is a better solution with the current state of TF2, is that multi-GPU distribute strategy is incompatible with our train_step approach. However, it should work with the optimizer wrapper approach. To be added in the future.

Replies: 1 comment

Comment options

andreped
Jan 29, 2023
Maintainer Author

You must be logged in to vote
0 replies
Answer selected by andreped
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
1 participant