This repository has been archived by the owner on Jun 2, 2023. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 14
unifying multi-task implementation #106
Comments
I'm good with the RGCN approach given the exploration you've already done. |
When adding in the state updating functionality, I thought the RGCN code was easier to follow. If the gradient correction doesn't improve performance, then I'd vote for the RGCN approach |
jsadler2
added a commit
that referenced
this issue
Jun 4, 2021
(just add losses together for multitask)
jsadler2
added a commit
that referenced
this issue
Jun 4, 2021
jsadler2
added a commit
that referenced
this issue
Jun 4, 2021
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Currently, the LSTM and GRU implementation of multitask learning is different than the RGCN implementation.
train_step
functionI think it makes sense to either choose one implementation or the other so as to simplify.
I lean toward the RGCN approach and just add the losses together. It is simpler and I think it will be sufficient for our needs. One drawback is that it makes it so you can't do the fancy gradient correction approach I was using before. But that didn't actually improve performance, and if we want to we can always go back and look at the code and resurrect it.
The text was updated successfully, but these errors were encountered: