You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jan 22, 2022. It is now read-only.
Hi! I'm a student learning CS285 online. Thank you for your great and generous work!
When I'm doing homework1 and running the same code in two different machines, one Linux and one Windows, I got two different actor result (but the expert results are the same).
Aftrer looking into details, because of random seed, the data batches used to update parameters in every training iteration are exactly the same between two machines. Differences start to show up after running gradient descent even the first time.
So my question is just if the differences comes from different machine situations or there are some other reasons? What do you guys think?
The text was updated successfully, but these errors were encountered:
Hi! I'm a student learning CS285 online. Thank you for your great and generous work!
When I'm doing homework1 and running the same code in two different machines, one Linux and one Windows, I got two different actor result (but the expert results are the same).
Aftrer looking into details, because of random seed, the data batches used to update parameters in every training iteration are exactly the same between two machines. Differences start to show up after running gradient descent even the first time.
So my question is just if the differences comes from different machine situations or there are some other reasons? What do you guys think?
The text was updated successfully, but these errors were encountered: