Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Explicitly state Databricks runtime version #247

Draft
wants to merge 2 commits into
base: main
Choose a base branch
from

Conversation

jacobtomlinson
Copy link
Member

Closes #246

The 12.2 LTS runtime doesn't seem to work with the databricksruntime/gpu-conda:cuda11 base image. Updating the instructions to explicitly recommend users choose the 10.4 LTS runtime.

@jacobtomlinson
Copy link
Member Author

I tried creating a notebook using the cluster and am now seeing another error.

java.lang.Exception: Cannot run program "virtualenv" (in directory "/local_disk0/pythonVirtualEnvDirs"): error=2, No such file or directory
	at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
	at org.apache.spark.util.Utils$.executeCommand(Utils.scala:1507)
	at org.apache.spark.util.Utils$.executeAndGetOutput(Utils.scala:1533)
	at org.apache.spark.util.DatabricksUtils$.createPythonVirtualEnv(DatabricksUtils.scala:602)
	at org.apache.spark.util.DatabricksUtils$.$anonfun$getOrCreatePythonVirtualEnvDir$1(DatabricksUtils.scala:722)
	at org.apache.spark.util.DatabricksUtils$.withStripedLock(DatabricksUtils.scala:411)
	at org.apache.spark.util.DatabricksUtils$.getOrCreatePythonVirtualEnvDir(DatabricksUtils.scala:708)
	at com.databricks.backend.daemon.driver.DBUtilsPythonEnvManager.getPythonExecutable(DBUtilsPythonEnvManager.scala:27)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.launchPython(PythonDriverLocal.scala:315)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.startPython0(PythonDriverLocal.scala:956)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.startPython(PythonDriverLocal.scala:1009)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.<init>(PythonDriverLocal.scala:1032)
	at com.databricks.backend.daemon.driver.PythonDriverWrapper.instantiateDriver(DriverWrapper.scala:697)
	at com.databricks.backend.daemon.driver.DriverWrapper.setupRepl(DriverWrapper.scala:335)
	at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:224)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.Throwable: error=2, No such file or directory
	at java.lang.UNIXProcess.forkAndExec(Native Method)
	at java.lang.UNIXProcess.<init>(UNIXProcess.java:247)
	at java.lang.ProcessImpl.start(ProcessImpl.java:134)
	at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
	... 15 more

@jacobtomlinson
Copy link
Member Author

I'm seeing that same error with the databricksruntime/gpu-conda:cuda11 base image to perhaps it just isn't compatible with Databricks anymore?

@jacobtomlinson
Copy link
Member Author

Things still aren't working with the changes, it looks like Databricks is looking for virtualenv but isn't finding it in the place it is expecting. Converting this to a draft while we explore how to resolve.

@jacobtomlinson jacobtomlinson marked this pull request as draft June 14, 2023 20:01
@skirui-source
Copy link
Contributor

@jacobtomlinson I was able to reproduce the databricks instructions on dt08 (not sure why workstation was taking so long).
But In testing Rapids notebook, I am also seeing the same error as yours:

Failure starting repl. Try detaching and re-attaching the notebook.

java.lang.Exception: Cannot run program "virtualenv" (in directory "/local_disk0/pythonVirtualEnvDirs"): error=2, No such file or directory
	at java.lang.ProcessBuilder.start(ProcessBuilder.java:1048)
	at org.apache.spark.util.Utils$.executeCommand(Utils.scala:1507)
	at org.apache.spark.util.Utils$.executeAndGetOutput(Utils.scala:1533)
	at org.apache.spark.util.DatabricksUtils$.createPythonVirtualEnv(DatabricksUtils.scala:602)
	at org.apache.spark.util.DatabricksUtils$.$anonfun$getOrCreatePythonVirtualEnvDir$1(DatabricksUtils.scala:722)
	at org.apache.spark.util.DatabricksUtils$.withStripedLock(DatabricksUtils.scala:411)
	at org.apache.spark.util.DatabricksUtils$.getOrCreatePythonVirtualEnvDir(DatabricksUtils.scala:708)
	at com.databricks.backend.daemon.driver.DBUtilsPythonEnvManager.getPythonExecutable(DBUtilsPythonEnvManager.scala:27)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.launchPython(PythonDriverLocal.scala:315)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.startPython0(PythonDriverLocal.scala:956)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.startPython(PythonDriverLocal.scala:1009)
	at com.databricks.backend.daemon.driver.PythonDriverLocal.<init>(PythonDriverLocal.scala:1032)
	at com.databricks.backend.daemon.driver.PythonDriverWrapper.instantiateDriver(DriverWrapper.scala:697)
	at com.databricks.backend.daemon.driver.DriverWrapper.setupRepl(DriverWrapper.scala:335)
	at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:224)
	at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.Throwable: error=2, No such file or directory
	at java.lang.UNIXProcess.forkAndExec(Native Method)
	at java.lang.UNIXProcess.<init>(UNIXProcess.java:247)
	at java.lang.ProcessImpl.start(ProcessImpl.java:134)
	at java.lang.ProcessBuilder.start(ProcessBuilder.java:1029)
	... 15 more

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working platform/databricks
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Databricks instructions not working
2 participants