Basic tour of Cloudera Data Science Workbench.
There are 4 scripts provided which walk through the interactive capabilities of Cloudera Data Science Workbench.
- Basic Python visualizations (Python 2). Demonstrates:
- Markdown via comments
- Jupyter-compatible visualizations
- Simple console sharing
- PySpark (Python 2). Demonstrates:
- Easy connectivity to (kerberized) Spark in YARN client mode.
- Access to Hadoop HDFS CLI (e.g.
hdfs dfs -ls /
).
- Tensorflow (Python 2). Demonstrates:
- Ability to install and use custom packages (e.g.
pip search tensorflow
)
- R on Spark via Sparklyr (R). Demonstrates:
- Use familiar dplyr with Spark using Sparklyr
- Advanced visualization with Shiny (R) Demonstrates:
- Use of 'shiny' to provide interactive graphics inside CDSW
We recommend setting up a "Nightly Analysis" job to illustrate how data scientists can easily automate their projects.
Note: You only need to do this once.
- In a Python 2 Session:
! pip2 install --upgrade dask keras matplotlib pandas_highcharts protobuf tensorflow seaborn
Note, you must then stop the session and start a new Python session in order for all the packages to be seen.
- In an R Session:
install.packages('sparklyr')
install.packages('plotly')
install.packages("nycflights13")
install.packages("Lahman")
install.packages("mgcv")
install.packages('shiny')
- Stop all sessions, then proceed.
‹