Skip to content

Commit

Permalink
Fix some link error in databricks doc (NVIDIA#44)
Browse files Browse the repository at this point in the history
Remove the optional step in databricks doc
Remove $ in databricks doc
Correct the kernel from "PySpark" to "Spark" in AWS Sagemaker
  • Loading branch information
GaryShen2008 authored and tgravescs committed Sep 26, 2019
1 parent 57d5199 commit 12b6610
Show file tree
Hide file tree
Showing 3 changed files with 9 additions and 9 deletions.
2 changes: 1 addition & 1 deletion getting-started-guides/csp/aws/sagemaker.md
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ Your output should look like the following:

If you get an error, it likely means that your ports have not been opened in the security group, so I would recommend going back and checking those settings!

Let's close the terminal. Type exit and then close the browser tab with the terminal. Open the tab with Jupyter, and choose New and then Sparkmagic (PySpark) to open a PySpark notebook. Just to be sure, let's re-start the kernel by choosing Kernel and then Restart.
Let's close the terminal. Type exit and then close the browser tab with the terminal. Open the tab with Jupyter, and choose New and then Sparkmagic (Spark) to open a Spark notebook. Just to be sure, let's re-start the kernel by choosing Kernel and then Restart.

![Kernel Restart](pics/sagemaker-kernel-restart.png)

Expand Down
10 changes: 5 additions & 5 deletions getting-started-guides/csp/azure/azure-databricks.md
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,6 @@ $ find . -type f -print|sort

* Go to "databricks" menu in top left bar
* Go to "Import and Explore Data"
* Create New Table - default location is /FileStore/tables
* Select mortgage_eval_merged.csv and mortgage_train_merged.csv to upload.

* Upload the XGBoost-4j Spark jars
Expand All @@ -66,7 +65,7 @@ In a new shell, create a directory and an init script that has commands to copy

For DataBricks Runtime 5.3 and 5.4:
```
$ dbutils.fs.mkdirs("dbfs:/databricks/init_scripts/")
dbutils.fs.mkdirs("dbfs:/databricks/init_scripts/")
dbutils.fs.put("/databricks/init_scripts/init.sh","""
#!/bin/bash
Expand All @@ -78,7 +77,7 @@ sudo cp /dbfs/FileStore/jars/[dbfs uploaded xgboost4j_spark_2.11 1.0.0_Beta jar]

For DataBricks Runtime 5.5:
```
$ dbutils.fs.mkdirs("dbfs:/databricks/init_scripts/")
dbutils.fs.mkdirs("dbfs:/databricks/init_scripts/")
dbutils.fs.put("/databricks/init_scripts/init.sh","""
#!/bin/bash
Expand All @@ -102,10 +101,11 @@ Start A Databricks Cluster
Import the GPU Mortgage Example Notebook
---------------------------
1. See [Managing Notebooks](https://docs.databricks.com/user-guide/notebooks/notebook-manage.html) on how to import a notebook.
2. Import the notebook: [XGBoost4j-Spark mortgage notebook](../notebook/databricks/mortgage-gpu.scala)
2. Import the notebook: [XGBoost4j-Spark mortgage notebook](/examples/notebooks/scala/mortgage-gpu.ipynb)
3. Change the dataset path like "/data/mortgage/csv/train/" to "dbfs:/FileStore/tables/mortgage_train_merged.csv"

The example notebook comes with the following configuration, you can adjust this according to your setup.
See supported configuration options here: [xgboost parameters](supported_xgboost_parameters.md)
See supported configuration options here: [xgboost parameters](/examples/app-parameters/supported_xgboost_parameters_scala.md)
```
val commParamMap = Map(
"eta" -> 0.1,
Expand Down
6 changes: 3 additions & 3 deletions getting-started-guides/csp/databricks/databricks.md
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,6 @@ $ find . -type f -print|sort

* Go to "databricks" menu in top left bar
* Go to "Import and Explore Data"
* Create New Table - default location is /FileStore/tables
* Select mortgage_eval_merged.csv and mortgage_train_merged.csv to upload.

* Upload the XGBoost-4j Spark jars
Expand Down Expand Up @@ -93,10 +92,11 @@ Start A Databricks Cluster
Import the GPU Mortgage Example Notebook
---------------------------
1. See [Managing Notebooks](https://docs.databricks.com/user-guide/notebooks/notebook-manage.html) on how to import a notebook.
2. Import the notebook: [XGBoost4j-Spark mortgage notebook](../notebook/databricks/mortgage-gpu.scala)
2. Import the notebook: [XGBoost4j-Spark mortgage notebook](/examples/notebooks/scala/mortgage-gpu.ipynb)
3. Change the dataset path like "/data/mortgage/csv/train/" to "dbfs:/FileStore/tables/mortgage_train_merged.csv"

The example notebook comes with the following configuration, you can adjust this according to your setup.
See supported configuration options here: [xgboost parameters](supported_xgboost_parameters.md)
See supported configuration options here: [xgboost parameters](/examples/app-parameters/supported_xgboost_parameters_scala.md)
```
val commParamMap = Map(
"eta" -> 0.1,
Expand Down

0 comments on commit 12b6610

Please sign in to comment.