Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Default hyper Parameters #52

Open
JonasGrutter opened this issue Apr 15, 2024 · 3 comments
Open

Default hyper Parameters #52

JonasGrutter opened this issue Apr 15, 2024 · 3 comments

Comments

@JonasGrutter
Copy link

Hey, I found out that the default shape of the latent space shape for DMC Proprio is 1024 which is way bigger the the shape of the observations. Can you explain me why ?

@NM512
Copy link
Owner

NM512 commented Apr 16, 2024

Thank you for bringing this up. The network size for each task is detailed in the paper, particularly in Appendices A and B.

Originally, I intended for the hidden unit size for the size S, including the encoder and decoder, to be 512. However, upon reviewing the original implementation, it appears that a configuration of 1024 x 5 layers was used for the encoder and decoder(not latent space), which I followed from this specific commit.

@JonasGrutter
Copy link
Author

Thank you for your answer! I am still wondering why (not on your side but in the paper), when we don't train on images, we are encoding an observation that is low dimensional into a higher dimensional space. Do you think there would be any disadvantage in dropping the auto-encoder in that case ? I am doing some experiments right now and will come back to you when I ll have results

@NM512
Copy link
Owner

NM512 commented Apr 16, 2024

It seems like if we're solely focusing on state-based model-free RL, a large encoder might not be necessary. I guess in the model-based RL the dynamic function being just a few layers of NN would indeed require a larger size MLP encoder for disentangled features, especially for complex nonlinear dynamics models(depending on the task).
When I tested with DMC tasks, there was no big difference between 512x2 vs 1024x5 settings.
I look forward the results!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants