Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to prevent caching? #1263

Open
keithachorn-intel opened this issue Jun 24, 2024 · 2 comments
Open

How to prevent caching? #1263

keithachorn-intel opened this issue Jun 24, 2024 · 2 comments

Comments

@keithachorn-intel
Copy link

I am using cm to download the MLPerf DLRM model (~100G) using cm. However, I want to specify the final location of this dataset. By default, it resides in a 'cache' directory with a pseudo-random key in the filepath, so I cannot predict the final location beforehand. Ideally, I want to simply specify the output directory or prevent caching so that it will land in the local dir.

However, despite searching for a way to do this with the documentation in this repo (and trying '--no-cache') the model continues to be cached. Any guidance here?

@arjunsuresh
Copy link
Contributor

Hi @keithachorn-intel we'll add the --no-cache option soon. But you can use --to=<download path> option to change the location of the model download. Please let us know if this works for you.

https://github.com/GATEOverflow/cm4mlops/blob/mlperf-inference/script/get-ml-model-dlrm-terabyte/_cm.json#L21

@anandhu-eng we can follow up our discussion for --no-cache

@anandhu-eng
Copy link
Contributor

Sure @arjunsuresh 🤝

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants