Skip to content

Commit a510e4d

Browse files
authored
update doc for rapids v2202 release (#106)
* update databricks init 7.3&9.1 scripts Signed-off-by: liyuan <[email protected]> * update release version from 21.12 to 22.02 Signed-off-by: liyuan <[email protected]> * revert databricks init scripts changes, dont need to remove the xgboost jars in script, we already overwrite them Signed-off-by: liyuan <[email protected]>
1 parent 195c92a commit a510e4d

File tree

14 files changed

+42
-42
lines changed

14 files changed

+42
-42
lines changed

docs/get-started/xgboost-examples/csp/databricks/databricks.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ CUDA 11.0 toolkit on the cluster. This can be done with the [generate-init-scri
4848
Spark plugin and the CUDA 11 toolkit.
4949
- [Databricks 9.1 LTS
5050
ML](https://docs.databricks.com/release-notes/runtime/9.1ml.html#system-environment) has CUDA 11
51-
installed. Users will need to use 21.12.0 or later on Databricks 9.1 LTS ML. In this case use
51+
installed. Users will need to use 22.02.0 or later on Databricks 9.1 LTS ML. In this case use
5252
[generate-init-script.ipynb](generate-init-script.ipynb) which will install
5353
the RAPIDS Spark plugin.
5454
2. Once you are in the notebook, click the “Run All” button.

docs/get-started/xgboost-examples/csp/databricks/generate-init-script-7.3.ipynb

+5-5
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,8 @@
2424
"source": [
2525
"%sh\n",
2626
"cd ../../dbfs/FileStore/jars/\n",
27-
"sudo wget -O cudf-21.12.0-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.0/cudf-21.12.0-cuda11.jar\n",
28-
"sudo wget -O rapids-4-spark_2.12-21.12.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar\n",
27+
"sudo wget -O cudf-22.02.0-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/cudf-22.02.0-cuda11.jar\n",
28+
"sudo wget -O rapids-4-spark_2.12-22.02.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar\n",
2929
"sudo wget -O xgboost4j_3.0-1.4.2-0.2.0.jar https://repo1.maven.org/maven2/com/nvidia/xgboost4j_3.0/1.4.2-0.2.0/xgboost4j_3.0-1.4.2-0.2.0.jar\n",
3030
"sudo wget -O xgboost4j-spark_3.0-1.4.2-0.2.0.jar https://repo1.maven.org/maven2/com/nvidia/xgboost4j-spark_3.0/1.4.2-0.2.0/xgboost4j-spark_3.0-1.4.2-0.2.0.jar\n",
3131
"ls -ltr\n",
@@ -58,8 +58,8 @@
5858
"dbutils.fs.put(\"/databricks/init_scripts/init.sh\",\"\"\"\n",
5959
"#!/bin/bash\n",
6060
"sudo cp /dbfs/FileStore/jars/xgboost4j_3.0-1.4.2-0.2.0.jar /databricks/jars/spark--maven-trees--ml--7.x--xgboost--ml.dmlc--xgboost4j_2.12--ml.dmlc__xgboost4j_2.12__1.0.0.jar\n",
61-
"sudo cp /dbfs/FileStore/jars/cudf-21.12.0-cuda11.jar /databricks/jars/\n",
62-
"sudo cp /dbfs/FileStore/jars/rapids-4-spark_2.12-21.12.0.jar /databricks/jars/\n",
61+
"sudo cp /dbfs/FileStore/jars/cudf-22.02.0-cuda11.jar /databricks/jars/\n",
62+
"sudo cp /dbfs/FileStore/jars/rapids-4-spark_2.12-22.02.0.jar /databricks/jars/\n",
6363
"sudo cp /dbfs/FileStore/jars/xgboost4j-spark_3.0-1.4.2-0.2.0.jar /databricks/jars/spark--maven-trees--ml--7.x--xgboost--ml.dmlc--xgboost4j-spark_2.12--ml.dmlc__xgboost4j-spark_2.12__1.0.0.jar\n",
6464
"sudo wget -O /etc/apt/preferences.d/cuda-repository-pin-600 https://developer.download.nvidia.com/compute/cuda/repos/ubuntu1804/x86_64/cuda-ubuntu1804.pin\n",
6565
"sudo wget -O ~/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb https://developer.download.nvidia.com/compute/cuda/11.0.3/local_installers/cuda-repo-ubuntu1804-11-0-local_11.0.3-450.51.06-1_amd64.deb\n",
@@ -139,7 +139,7 @@
139139
"1. Edit your cluster, adding an initialization script from `dbfs:/databricks/init_scripts/init.sh` in the \"Advanced Options\" under \"Init Scripts\" tab\n",
140140
"2. Reboot the cluster\n",
141141
"3. Go to \"Libraries\" tab under your cluster and install `dbfs:/FileStore/jars/xgboost4j-spark_3.0-1.4.2-0.2.0.jar` in your cluster by selecting the \"DBFS\" option for installing jars\n",
142-
"4. Import the mortgage example notebook from `https://github.com/NVIDIA/spark-rapids-examples/blob/branch-21.12/examples/Spark-ETL+XGBoost/mortgage/notebooks/python/mortgage-gpu.ipynb`\n",
142+
"4. Import the mortgage example notebook from `https://github.com/NVIDIA/spark-rapids-examples/blob/branch-22.02/examples/Spark-ETL+XGBoost/mortgage/notebooks/python/mortgage-gpu.ipynb`\n",
143143
"5. Inside the mortgage example notebook, update the data paths\n",
144144
" `train_data = reader.schema(schema).option('header', True).csv('/data/mortgage/csv/small-train.csv')`\n",
145145
" `trans_data = reader.schema(schema).option('header', True).csv('/data/mortgage/csv/small-trans.csv')`"

docs/get-started/xgboost-examples/csp/databricks/generate-init-script.ipynb

+5-5
Original file line numberDiff line numberDiff line change
@@ -24,8 +24,8 @@
2424
"source": [
2525
"%sh\n",
2626
"cd ../../dbfs/FileStore/jars/\n",
27-
"sudo wget -O cudf-21.12.0-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.0/cudf-21.12.0-cuda11.jar\n",
28-
"sudo wget -O rapids-4-spark_2.12-21.12.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar\n",
27+
"sudo wget -O cudf-22.02.0-cuda11.jar https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/cudf-22.02.0-cuda11.jar\n",
28+
"sudo wget -O rapids-4-spark_2.12-22.02.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar\n",
2929
"sudo wget -O xgboost4j_3.0-1.4.2-0.2.0.jar https://repo1.maven.org/maven2/com/nvidia/xgboost4j_3.0/1.4.2-0.2.0/xgboost4j_3.0-1.4.2-0.2.0.jar\n",
3030
"sudo wget -O xgboost4j-spark_3.0-1.4.2-0.2.0.jar https://repo1.maven.org/maven2/com/nvidia/xgboost4j-spark_3.0/1.4.2-0.2.0/xgboost4j-spark_3.0-1.4.2-0.2.0.jar\n",
3131
"ls -ltr\n",
@@ -58,8 +58,8 @@
5858
"dbutils.fs.put(\"/databricks/init_scripts/init.sh\",\"\"\"\n",
5959
"#!/bin/bash\n",
6060
"sudo cp /dbfs/FileStore/jars/xgboost4j_3.0-1.4.2-0.2.0.jar /databricks/jars/spark--maven-trees--ml--9.x--xgboost-gpu--ml.dmlc--xgboost4j-gpu_2.12--ml.dmlc__xgboost4j-gpu_2.12__1.4.1.jar\n",
61-
"sudo cp /dbfs/FileStore/jars/cudf-21.12.0-cuda11.jar /databricks/jars/\n",
62-
"sudo cp /dbfs/FileStore/jars/rapids-4-spark_2.12-21.12.0.jar /databricks/jars/\n",
61+
"sudo cp /dbfs/FileStore/jars/cudf-22.02.0-cuda11.jar /databricks/jars/\n",
62+
"sudo cp /dbfs/FileStore/jars/rapids-4-spark_2.12-22.02.0.jar /databricks/jars/\n",
6363
"sudo cp /dbfs/FileStore/jars/xgboost4j-spark_3.0-1.4.2-0.2.0.jar /databricks/jars/spark--maven-trees--ml--9.x--xgboost-gpu--ml.dmlc--xgboost4j-spark-gpu_2.12--ml.dmlc__xgboost4j-spark-gpu_2.12__1.4.1.jar\"\"\", True)"
6464
]
6565
},
@@ -132,7 +132,7 @@
132132
"1. Edit your cluster, adding an initialization script from `dbfs:/databricks/init_scripts/init.sh` in the \"Advanced Options\" under \"Init Scripts\" tab\n",
133133
"2. Reboot the cluster\n",
134134
"3. Go to \"Libraries\" tab under your cluster and install `dbfs:/FileStore/jars/xgboost4j-spark_3.0-1.4.2-0.2.0.jar` in your cluster by selecting the \"DBFS\" option for installing jars\n",
135-
"4. Import the mortgage example notebook from `https://github.com/NVIDIA/spark-rapids-examples/blob/branch-21.12/examples/Spark-ETL+XGBoost/mortgage/notebooks/python/mortgage-gpu.ipynb`\n",
135+
"4. Import the mortgage example notebook from `https://github.com/NVIDIA/spark-rapids-examples/blob/branch-22.02/examples/Spark-ETL+XGBoost/mortgage/notebooks/python/mortgage-gpu.ipynb`\n",
136136
"5. Inside the mortgage example notebook, update the data paths\n",
137137
" `train_data = reader.schema(schema).option('header', True).csv('/data/mortgage/csv/small-train.csv')`\n",
138138
" `trans_data = reader.schema(schema).option('header', True).csv('/data/mortgage/csv/small-trans.csv')`"

docs/get-started/xgboost-examples/on-prem-cluster/kubernetes-scala.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ export SPARK_DOCKER_IMAGE=<gpu spark docker image repo and name>
4040
export SPARK_DOCKER_TAG=<spark docker image tag>
4141

4242
pushd ${SPARK_HOME}
43-
wget https://github.com/NVIDIA/spark-rapids-examples/raw/branch-21.12/dockerfile/Dockerfile
43+
wget https://github.com/NVIDIA/spark-rapids-examples/raw/branch-22.02/dockerfile/Dockerfile
4444

4545
# Optionally install additional jars into ${SPARK_HOME}/jars/
4646

docs/get-started/xgboost-examples/prepare-package-data/preparation-python.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -9,11 +9,11 @@ For simplicity export the location to these jars. All examples assume the packag
99
* [XGBoost4j-Spark Package](https://repo1.maven.org/maven2/com/nvidia/xgboost4j-spark_3.0/1.4.2-0.2.0/)
1010

1111
2. Download the RAPIDS Accelerator for Apache Spark plugin jar
12-
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar)
12+
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar)
1313

1414
Then download the version of the cudf jar that your version of the accelerator depends on.
1515

16-
* [cuDF Package](https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.2/cudf-21.12.2-cuda11.jar)
16+
* [cuDF Package](https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/cudf-22.02.0-cuda11.jar)
1717

1818
### Build XGBoost Python Examples
1919

@@ -29,8 +29,8 @@ You need to download Mortgage dataset to `/opt/xgboost` from this [site](https:/
2929

3030
``` bash
3131
export SPARK_XGBOOST_DIR=/opt/xgboost
32-
export CUDF_JAR=${SPARK_XGBOOST_DIR}/cudf-21.12.2-cuda11.jar
33-
export RAPIDS_JAR=${SPARK_XGBOOST_DIR}/rapids-4-spark_2.12-21.12.0.jar
32+
export CUDF_JAR=${SPARK_XGBOOST_DIR}/cudf-22.02.0-cuda11.jar
33+
export RAPIDS_JAR=${SPARK_XGBOOST_DIR}/rapids-4-spark_2.12-22.02.0.jar
3434
export XGBOOST4J_JAR=${SPARK_XGBOOST_DIR}/xgboost4j_3.0-1.4.2-0.2.0.jar
3535
export XGBOOST4J_SPARK_JAR=${SPARK_XGBOOST_DIR}/xgboost4j-spark_3.0-1.4.2-0.2.0.jar
3636
export SAMPLE_ZIP=${SPARK_XGBOOST_DIR}/samples.zip

docs/get-started/xgboost-examples/prepare-package-data/preparation-scala.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -5,11 +5,11 @@ For simplicity export the location to these jars. All examples assume the packag
55
### Download the jars
66

77
1. Download the RAPIDS Accelerator for Apache Spark plugin jar
8-
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar)
8+
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar)
99

1010
Then download the version of the cudf jar that your version of the accelerator depends on.
1111

12-
* [cuDF Package](https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.2/cudf-21.12.2-cuda11.jar)
12+
* [cuDF Package](https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/cudf-22.02.0-cuda11.jar)
1313

1414
### Build XGBoost Scala Examples
1515

@@ -25,7 +25,7 @@ You need to download mortgage dataset to `/opt/xgboost` from this [site](https:/
2525

2626
``` bash
2727
export SPARK_XGBOOST_DIR=/opt/xgboost
28-
export CUDF_JAR=${SPARK_XGBOOST_DIR}/cudf-21.12.2-cuda11.jar
29-
export RAPIDS_JAR=${SPARK_XGBOOST_DIR}/rapids-4-spark_2.12-21.12.0.jar
28+
export CUDF_JAR=${SPARK_XGBOOST_DIR}/cudf-22.02.0-cuda11.jar
29+
export RAPIDS_JAR=${SPARK_XGBOOST_DIR}/rapids-4-spark_2.12-22.02.0.jar
3030
export SAMPLE_JAR=${SPARK_XGBOOST_DIR}/sample_xgboost_apps-0.2.2-jar-with-dependencies.jar
3131
```

examples/Spark-ETL+XGBoost/mortgage/notebooks/python/MortgageETL.ipynb

+3-3
Original file line numberDiff line numberDiff line change
@@ -9,16 +9,16 @@
99
"All data could be found at https://docs.rapids.ai/datasets/mortgage-data\n",
1010
"\n",
1111
"### 2. Download needed jars\n",
12-
"* [cudf-21.12.2-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.0/)\n",
13-
"* [rapids-4-spark_2.12-21.12.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar)\n",
12+
"* [cudf-22.02.0-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/)\n",
13+
"* [rapids-4-spark_2.12-22.02.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar)\n",
1414
"\n",
1515
"\n",
1616
"### 3. Start Spark Standalone\n",
1717
"Before running the script, please setup Spark standalone mode\n",
1818
"\n",
1919
"### 4. Add ENV\n",
2020
"```\n",
21-
"$ export SPARK_JARS=cudf-21.12.2-cuda11.jar,rapids-4-spark_2.12-21.12.0.jar\n",
21+
"$ export SPARK_JARS=cudf-22.02.0-cuda11.jar,rapids-4-spark_2.12-22.02.0.jar\n",
2222
"$ export PYSPARK_DRIVER_PYTHON=jupyter \n",
2323
"$ export PYSPARK_DRIVER_PYTHON_OPTS=notebook\n",
2424
"```\n",

examples/Spark-ETL+XGBoost/mortgage/notebooks/scala/mortgage-ETL.ipynb

+3-3
Original file line numberDiff line numberDiff line change
@@ -19,15 +19,15 @@
1919
"All data could be found at https://docs.rapids.ai/datasets/mortgage-data\n",
2020
"\n",
2121
"### 2. Download needed jars\n",
22-
"* [cudf-21.12.2-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.2/)\n",
23-
"* [rapids-4-spark_2.12-21.12.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar)\n",
22+
"* [cudf-22.02.0-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/)\n",
23+
"* [rapids-4-spark_2.12-22.02.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar)\n",
2424
"\n",
2525
"### 3. Start Spark Standalone\n",
2626
"Before Running the script, please setup Spark standalone mode\n",
2727
"\n",
2828
"### 4. Add ENV\n",
2929
"```\n",
30-
"$ export SPARK_JARS=cudf-21.12.2-cuda11.jar,rapids-4-spark_2.12-21.12.0.jar\n",
30+
"$ export SPARK_JARS=cudf-22.02.0-cuda11.jar,rapids-4-spark_2.12-22.02.0.jar\n",
3131
"\n",
3232
"```\n",
3333
"\n",

examples/Spark-ETL+XGBoost/mortgage/notebooks/scala/mortgage_gpu_crossvalidation.ipynb

+2-2
Original file line numberDiff line numberDiff line change
@@ -160,10 +160,10 @@
160160
"```scala\n",
161161
"import org.apache.spark.sql.SparkSession\n",
162162
"val spark = SparkSession.builder().appName(\"Taxi-GPU\").getOrCreate\n",
163-
"%AddJar file:/data/libs/cudf-21.12.2-cuda11.jar\n",
163+
"%AddJar file:/data/libs/cudf-22.02.0-cuda11.jar\n",
164164
"%AddJar file:/data/libs/xgboost4j_3.0-1.4.2-0.2.0.jar\n",
165165
"%AddJar file:/data/libs/xgboost4j-spark_3.0-1.4.2-0.2.0.jar\n",
166-
"%AddJar file:/data/libs/rapids-4-spark_2.12-21.12.0.jar\n",
166+
"%AddJar file:/data/libs/rapids-4-spark_2.12-22.02.0.jar\n",
167167
"// ...\n",
168168
"```"
169169
]

examples/Spark-ETL+XGBoost/taxi/notebooks/python/taxi-ETL.ipynb

+3-3
Original file line numberDiff line numberDiff line change
@@ -19,15 +19,15 @@
1919
"All data could be found at https://www1.nyc.gov/site/tlc/about/tlc-trip-record-data.page\n",
2020
"\n",
2121
"### 2. Download needed jars\n",
22-
"* [cudf-21.12.2-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.2/)\n",
23-
"* [rapids-4-spark_2.12-21.12.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar)\n",
22+
"* [cudf-22.02.0-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/)\n",
23+
"* [rapids-4-spark_2.12-22.02.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar)\n",
2424
"\n",
2525
"### 3. Start Spark Standalone\n",
2626
"Before running the script, please setup Spark standalone mode\n",
2727
"\n",
2828
"### 4. Add ENV\n",
2929
"```\n",
30-
"$ export SPARK_JARS=cudf-21.12.2-cuda11.jar,rapids-4-spark_2.12-21.12.0.jar\n",
30+
"$ export SPARK_JARS=cudf-22.02.0-cuda11.jar,rapids-4-spark_2.12-22.02.0.jar\n",
3131
"$ export PYSPARK_DRIVER_PYTHON=jupyter \n",
3232
"$ export PYSPARK_DRIVER_PYTHON_OPTS=notebook\n",
3333
"```\n",

examples/Spark-ETL+XGBoost/taxi/notebooks/scala/taxi-ETL.ipynb

+3-3
Original file line numberDiff line numberDiff line change
@@ -19,15 +19,15 @@
1919
"All data could be found at https://www1.nyc.gov/site/tlc/about/tlc-trip-record-data.page\n",
2020
"\n",
2121
"### 2. Download needed jars\n",
22-
"* [cudf-21.12.2-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/21.12.2/)\n",
23-
"* [rapids-4-spark_2.12-21.12.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/21.12.0/rapids-4-spark_2.12-21.12.0.jar)\n",
22+
"* [cudf-22.02.0-cuda11.jar](https://repo1.maven.org/maven2/ai/rapids/cudf/22.02.0/)\n",
23+
"* [rapids-4-spark_2.12-22.02.0.jar](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/22.02.0/rapids-4-spark_2.12-22.02.0.jar)\n",
2424
"\n",
2525
"### 3. Start Spark Standalone\n",
2626
"Before running the script, please setup Spark standalone mode\n",
2727
"\n",
2828
"### 4. Add ENV\n",
2929
"```\n",
30-
"$ export SPARK_JARS=cudf-21.12.2-cuda11.jar,rapids-4-spark_2.12-21.12.0.jar\n",
30+
"$ export SPARK_JARS=cudf-22.02.0-cuda11.jar,rapids-4-spark_2.12-22.02.0.jar\n",
3131
"\n",
3232
"```\n",
3333
"\n",

examples/Spark-ETL+XGBoost/taxi/notebooks/scala/taxi_gpu_crossvalidation.ipynb

+2-2
Original file line numberDiff line numberDiff line change
@@ -150,10 +150,10 @@
150150
"```scala\n",
151151
"import org.apache.spark.sql.SparkSession\n",
152152
"val spark = SparkSession.builder().appName(\"Taxi-GPU\").getOrCreate\n",
153-
"%AddJar file:/data/libs/cudf-21.12.2-cuda11.jar\n",
153+
"%AddJar file:/data/libs/cudf-22.02.0-cuda11.jar\n",
154154
"%AddJar file:/data/libs/xgboost4j_3.0-1.4.2-0.2.0.jar\n",
155155
"%AddJar file:/data/libs/xgboost4j-spark_3.0-1.4.2-0.2.0.jar\n",
156-
"%AddJar file:/data/libs/rapids-4-spark-21.12.0.jar\n",
156+
"%AddJar file:/data/libs/rapids-4-spark-22.02.0.jar\n",
157157
"// ...\n",
158158
"```"
159159
]

examples/Spark-cuML/pca/spark-submit.sh

+4-4
Original file line numberDiff line numberDiff line change
@@ -15,9 +15,9 @@
1515
# limitations under the License.
1616
#
1717

18-
ML_JAR=/root/.m2/repository/com/nvidia/rapids-4-spark-ml_2.12/21.12.0-SNAPSHOT/rapids-4-spark-ml_2.12-21.12.0-SNAPSHOT.jar
19-
CUDF_JAR=/root/.m2/repository/ai/rapids/cudf/21.12.0-SNAPSHOT/cudf-21.12.0-SNAPSHOT-cuda11.jar
20-
PLUGIN_JAR=/root/.m2/repository/com/nvidia/rapids-4-spark_2.12/21.12.0-SNAPSHOT/rapids-4-spark_2.12-21.12.0-SNAPSHOT.jar
18+
ML_JAR=/root/.m2/repository/com/nvidia/rapids-4-spark-ml_2.12/22.02.0-SNAPSHOT/rapids-4-spark-ml_2.12-22.02.0-SNAPSHOT.jar
19+
CUDF_JAR=/root/.m2/repository/ai/rapids/cudf/22.02.0-SNAPSHOT/cudf-22.02.0-SNAPSHOT-cuda11.jar
20+
PLUGIN_JAR=/root/.m2/repository/com/nvidia/rapids-4-spark_2.12/22.02.0-SNAPSHOT/rapids-4-spark_2.12-22.02.0-SNAPSHOT.jar
2121

2222
$SPARK_HOME/bin/spark-submit \
2323
--master spark://127.0.0.1:7077 \
@@ -39,4 +39,4 @@ $SPARK_HOME/bin/spark-submit \
3939
--conf spark.network.timeout=1000s \
4040
--jars $ML_JAR,$CUDF_JAR,$PLUGIN_JAR \
4141
--class com.nvidia.spark.examples.pca.Main \
42-
/workspace/target/PCAExample-21.12.0-SNAPSHOT.jar
42+
/workspace/target/PCAExample-22.02.0-SNAPSHOT.jar

examples/micro-benchmarks/notebooks/micro-benchmarks-gpu.ipynb

+2-2
Original file line numberDiff line numberDiff line change
@@ -22,8 +22,8 @@
2222
"import os\n",
2323
"# Change to your cluster ip:port and directories\n",
2424
"SPARK_MASTER_URL = os.getenv(\"SPARK_MASTER_URL\", \"spark:your-ip:port\")\n",
25-
"CUDF_JAR = os.getenv(\"CUDF_JAR\", \"/your-path/cudf-21.12.2-cuda11.jar\")\n",
26-
"RAPIDS_JAR = os.getenv(\"RAPIDS_JAR\", \"/your-path/rapids-4-spark_2.12-21.12.0.jar\")\n"
25+
"CUDF_JAR = os.getenv(\"CUDF_JAR\", \"/your-path/cudf-22.02.0-cuda11.jar\")\n",
26+
"RAPIDS_JAR = os.getenv(\"RAPIDS_JAR\", \"/your-path/rapids-4-spark_2.12-22.02.0.jar\")\n"
2727
]
2828
},
2929
{

0 commit comments

Comments
 (0)