Skip to content

Commit 735cd73

Browse files
committed
Merge branch 'branch-25.02' into main-2502-release
2 parents 8a23765 + efd0ed6 commit 735cd73

File tree

113 files changed

+14003
-11390
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

113 files changed

+14003
-11390
lines changed

.github/workflows/add-to-project.yml

+6-8
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# Copyright (c) 2024, NVIDIA CORPORATION.
1+
# Copyright (c) 2024-2025, NVIDIA CORPORATION.
22
#
33
# Licensed under the Apache License, Version 2.0 (the "License");
44
# you may not use this file except in compliance with the License.
@@ -23,13 +23,11 @@ on:
2323
- opened
2424

2525
jobs:
26-
add-to-project:
27-
if: github.repository == 'NVIDIA/spark-rapids-examples'
28-
name: Add new issues and pull requests to project
26+
Add-to-project:
27+
if: github.repository_owner == 'NVIDIA' # avoid adding issues from forks
2928
runs-on: ubuntu-latest
3029
steps:
31-
- uses: actions/[email protected]
30+
- name: add-to-project
31+
uses: NVIDIA/spark-rapids-common/add-to-project@main
3232
with:
33-
project-url: https://github.com/orgs/NVIDIA/projects/4
34-
github-token: ${{ secrets.PROJECT_TOKEN }}
35-
33+
token: ${{ secrets.PROJECT_TOKEN }}
+54
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,54 @@
1+
# Copyright (c) 2024, NVIDIA CORPORATION.
2+
#
3+
# Licensed under the Apache License, Version 2.0 (the "License");
4+
# you may not use this file except in compliance with the License.
5+
# You may obtain a copy of the License at
6+
#
7+
# http://www.apache.org/licenses/LICENSE-2.0
8+
#
9+
# Unless required by applicable law or agreed to in writing, software
10+
# distributed under the License is distributed on an "AS IS" BASIS,
11+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12+
# See the License for the specific language governing permissions and
13+
# limitations under the License.
14+
15+
# A workflow to check copyright/license header
16+
name: license header check
17+
18+
on:
19+
pull_request:
20+
types: [opened, synchronize, reopened]
21+
22+
jobs:
23+
license-header-check:
24+
runs-on: ubuntu-latest
25+
if: "!contains(github.event.pull_request.title, '[bot]')"
26+
steps:
27+
- name: Get checkout depth
28+
run: |
29+
echo "PR_FETCH_DEPTH=$(( ${{ github.event.pull_request.commits }} + 10 ))" >> $GITHUB_ENV
30+
31+
- name: Checkout code
32+
uses: actions/checkout@v4
33+
with:
34+
fetch-depth: ${{ env.PR_FETCH_DEPTH }}
35+
36+
- name: license-header-check
37+
uses: NVIDIA/spark-rapids-common/license-header-check@main
38+
with:
39+
included_file_patterns: |
40+
*.sh,
41+
*.java,
42+
*.py,
43+
*.pbtxt,
44+
*Dockerfile*,
45+
*Jenkinsfile*,
46+
*.yml,
47+
*.yaml,
48+
*.cpp,
49+
*.hpp,
50+
*.txt,
51+
*.cu,
52+
*.scala,
53+
*.ini,
54+
*.xml

dockerfile/gpu_executor_template.yaml

+14
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,17 @@
1+
# Copyright (c) 2024, NVIDIA CORPORATION.
2+
#
3+
# Licensed under the Apache License, Version 2.0 (the "License");
4+
# you may not use this file except in compliance with the License.
5+
# You may obtain a copy of the License at
6+
#
7+
# http://www.apache.org/licenses/LICENSE-2.0
8+
#
9+
# Unless required by applicable law or agreed to in writing, software
10+
# distributed under the License is distributed on an "AS IS" BASIS,
11+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12+
# See the License for the specific language governing permissions and
13+
# limitations under the License.
14+
115
apiVersion: v1
216
kind: Pod
317
spec:

docs/get-started/xgboost-examples/csp/databricks/databricks.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ Navigate to your home directory in the UI and select **Create** > **File** from
2121
create an `init.sh` scripts with contents:
2222
```bash
2323
#!/bin/bash
24-
sudo wget -O /databricks/jars/rapids-4-spark_2.12-24.12.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/24.12.0/rapids-4-spark_2.12-24.12.0.jar
24+
sudo wget -O /databricks/jars/rapids-4-spark_2.12-25.02.1.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/25.02.1/rapids-4-spark_2.12-25.02.1.jar
2525
```
2626
1. Select the Databricks Runtime Version from one of the supported runtimes specified in the
2727
Prerequisites section.
@@ -68,7 +68,7 @@ create an `init.sh` scripts with contents:
6868
```bash
6969
spark.rapids.sql.python.gpu.enabled true
7070
spark.python.daemon.module rapids.daemon_databricks
71-
spark.executorEnv.PYTHONPATH /databricks/jars/rapids-4-spark_2.12-24.12.0.jar:/databricks/spark/python
71+
spark.executorEnv.PYTHONPATH /databricks/jars/rapids-4-spark_2.12-25.02.1.jar:/databricks/spark/python
7272
```
7373
Note that since python memory pool require installing the cudf library, so you need to install cudf library in
7474
each worker nodes `pip install cudf-cu11 --extra-index-url=https://pypi.nvidia.com` or disable python memory pool

docs/get-started/xgboost-examples/csp/databricks/init.sh

+15-1
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,21 @@
1+
# Copyright (c) 2024, NVIDIA CORPORATION.
2+
#
3+
# Licensed under the Apache License, Version 2.0 (the "License");
4+
# you may not use this file except in compliance with the License.
5+
# You may obtain a copy of the License at
6+
#
7+
# http://www.apache.org/licenses/LICENSE-2.0
8+
#
9+
# Unless required by applicable law or agreed to in writing, software
10+
# distributed under the License is distributed on an "AS IS" BASIS,
11+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12+
# See the License for the specific language governing permissions and
13+
# limitations under the License.
14+
115
sudo rm -f /databricks/jars/spark--maven-trees--ml--10.x--xgboost-gpu--ml.dmlc--xgboost4j-gpu_2.12--ml.dmlc__xgboost4j-gpu_2.12__1.5.2.jar
216
sudo rm -f /databricks/jars/spark--maven-trees--ml--10.x--xgboost-gpu--ml.dmlc--xgboost4j-spark-gpu_2.12--ml.dmlc__xgboost4j-spark-gpu_2.12__1.5.2.jar
317

4-
sudo wget -O /databricks/jars/rapids-4-spark_2.12-24.12.0.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/24.12.0/rapids-4-spark_2.12-24.12.0.jar
18+
sudo wget -O /databricks/jars/rapids-4-spark_2.12-25.02.1.jar https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/25.02.1/rapids-4-spark_2.12-25.02.1.jar
519
sudo wget -O /databricks/jars/xgboost4j-gpu_2.12-1.7.1.jar https://repo1.maven.org/maven2/ml/dmlc/xgboost4j-gpu_2.12/1.7.1/xgboost4j-gpu_2.12-1.7.1.jar
620
sudo wget -O /databricks/jars/xgboost4j-spark-gpu_2.12-1.7.1.jar https://repo1.maven.org/maven2/ml/dmlc/xgboost4j-spark-gpu_2.12/1.7.1/xgboost4j-spark-gpu_2.12-1.7.1.jar
721
ls -ltr

docs/get-started/xgboost-examples/on-prem-cluster/kubernetes-scala.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@ export SPARK_DOCKER_IMAGE=<gpu spark docker image repo and name>
4040
export SPARK_DOCKER_TAG=<spark docker image tag>
4141

4242
pushd ${SPARK_HOME}
43-
wget https://github.com/NVIDIA/spark-rapids-examples/raw/branch-24.12/dockerfile/Dockerfile
43+
wget https://github.com/NVIDIA/spark-rapids-examples/raw/branch-25.02/dockerfile/Dockerfile
4444

4545
# Optionally install additional jars into ${SPARK_HOME}/jars/
4646

docs/get-started/xgboost-examples/prepare-package-data/preparation-python.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ For simplicity export the location to these jars. All examples assume the packag
55
### Download the jars
66

77
Download the RAPIDS Accelerator for Apache Spark plugin jar
8-
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/24.12.0/rapids-4-spark_2.12-24.12.0.jar)
8+
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/25.02.1/rapids-4-spark_2.12-25.02.1.jar)
99

1010
### Build XGBoost Python Examples
1111

docs/get-started/xgboost-examples/prepare-package-data/preparation-scala.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ For simplicity export the location to these jars. All examples assume the packag
55
### Download the jars
66

77
1. Download the RAPIDS Accelerator for Apache Spark plugin jar
8-
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/24.12.0/rapids-4-spark_2.12-24.12.0.jar)
8+
* [RAPIDS Spark Package](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/25.02.1/rapids-4-spark_2.12-25.02.1.jar)
99

1010
### Build XGBoost Scala Examples
1111

examples/ML+DL-Examples/Optuna-Spark/README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -147,8 +147,8 @@ We use [RAPIDS](https://docs.rapids.ai/install/#get-rapids) for GPU-accelerated
147147
``` shell
148148
sudo apt install libmysqlclient-dev
149149

150-
conda create -n rapids-24.12 -c rapidsai -c conda-forge -c nvidia \
151-
cudf=24.12 cuml=24.12 python=3.10 'cuda-version>=12.0,<=12.5'
150+
conda create -n rapids-25.02 -c rapidsai -c conda-forge -c nvidia \
151+
cudf=25.02 cuml=25.02 python=3.10 'cuda-version>=12.0,<=12.5'
152152
conda activate optuna-spark
153153
pip install mysqlclient
154154
pip install optuna joblib joblibspark ipywidgets

examples/ML+DL-Examples/Optuna-Spark/optuna-examples/databricks/init_optuna.sh

+2-2
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ fi
4141

4242

4343
# rapids import
44-
SPARK_RAPIDS_VERSION=24.12.0
44+
SPARK_RAPIDS_VERSION=25.02.1
4545
curl -L https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/${SPARK_RAPIDS_VERSION}/rapids-4-spark_2.12-${SPARK_RAPIDS_VERSION}.jar -o \
4646
/databricks/jars/rapids-4-spark_2.12-${SPARK_RAPIDS_VERSION}.jar
4747

@@ -54,7 +54,7 @@ ln -s /usr/local/cuda-11.8 /usr/local/cuda
5454

5555
sudo /databricks/python3/bin/pip3 install \
5656
--extra-index-url=https://pypi.nvidia.com \
57-
"cudf-cu11==24.12.*" "cuml-cu11==24.12.*"
57+
"cudf-cu11==25.02.*" "cuml-cu11==25.02.*"
5858

5959
# setup python environment
6060
sudo apt clean && sudo apt update --fix-missing -y

examples/ML+DL-Examples/Optuna-Spark/optuna-examples/databricks/start_cluster.sh

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ json_config=$(cat <<EOF
1212
"spark_version": "13.3.x-gpu-ml-scala2.12",
1313
"spark_conf": {
1414
"spark.task.resource.gpu.amount": "1",
15-
"spark.executorEnv.PYTHONPATH": "/databricks/jars/rapids-4-spark_2.12-24.12.0.jar:/databricks/spark/python:/databricks/python3",
15+
"spark.executorEnv.PYTHONPATH": "/databricks/jars/rapids-4-spark_2.12-25.02.1.jar:/databricks/spark/python:/databricks/python3",
1616
"spark.executor.cores": "8",
1717
"spark.rapids.memory.gpu.minAllocFraction": "0.0001",
1818
"spark.plugins": "com.nvidia.spark.SQLPlugin",

examples/ML+DL-Examples/Optuna-Spark/optuna-examples/optuna-dataframe.ipynb

+2-2
Original file line numberDiff line numberDiff line change
@@ -444,14 +444,14 @@
444444
"24/12/11 23:47:52 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable\n",
445445
"Setting default log level to \"WARN\".\n",
446446
"To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).\n",
447-
"24/12/11 23:47:52 WARN RapidsPluginUtils: RAPIDS Accelerator 24.12.0 using cudf 24.12.0, private revision bd4e99e18e20234ee0c54f95f4b0bfce18a6255e\n",
447+
"24/12/11 23:47:52 WARN RapidsPluginUtils: RAPIDS Accelerator 25.02.1 using cudf 25.02.1, private revision bd4e99e18e20234ee0c54f95f4b0bfce18a6255e\n",
448448
"24/12/11 23:47:52 WARN RapidsPluginUtils: RAPIDS Accelerator is enabled, to disable GPU support set `spark.rapids.sql.enabled` to false.\n"
449449
]
450450
}
451451
],
452452
"source": [
453453
"def get_rapids_jar():\n",
454-
" SPARK_RAPIDS_VERSION = \"24.12.0\"\n",
454+
" SPARK_RAPIDS_VERSION = \"25.02.1\"\n",
455455
" rapids_jar = f\"rapids-4-spark_2.12-{SPARK_RAPIDS_VERSION}.jar\"\n",
456456
" if not os.path.exists(rapids_jar):\n",
457457
" print(\"Downloading Spark Rapids jar\")\n",

0 commit comments

Comments
 (0)