You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello author, I have a question regarding your code. Why isn't the hidden state of the GRU detached from the computational graph? This could lead to exploding/vanishing gradients. I've seen other code using RNNs that seems to detach the hidden state from the previous step. It appears that only PyMARL and its various improved extensions don't do this. I checked https://github.com/oxwhirl/pymarl and it seems PyMARL is also written this way, and all the various improved repositories based on PyMARL handle it similarly. I hope to get an answer. Thank you.