From 972837567ea0dadbd593247aec2b5cef5a0b1dc2 Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 15:25:31 +0800 Subject: [PATCH 1/8] fix deadlink in tool notebook Signed-off-by: liyuan --- ...tor for Apache Spark] Profiling Tool Notebook Template.ipynb | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tools/databricks/[RAPIDS Accelerator for Apache Spark] Profiling Tool Notebook Template.ipynb b/tools/databricks/[RAPIDS Accelerator for Apache Spark] Profiling Tool Notebook Template.ipynb index 2824bd1ec..3cb822824 100644 --- a/tools/databricks/[RAPIDS Accelerator for Apache Spark] Profiling Tool Notebook Template.ipynb +++ b/tools/databricks/[RAPIDS Accelerator for Apache Spark] Profiling Tool Notebook Template.ipynb @@ -1 +1 @@ -{"cells":[{"cell_type":"markdown","metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"5156a76c-7af7-465d-aff4-41a2e54e3595","showTitle":false,"title":""}},"source":["# Welcome to the Profiling Tool for the RAPIDS Accelerator for Apache Spark\n","To run the tool, you need to enter a log path that represents the DBFS location for your Spark GPU event logs. Then you can select \"Run all\" to execute the notebook. After the notebook completes, you will see various output tables show up below. More options for running the profiling tool can be found here: https://nvidia.github.io/spark-rapids/docs/spark-profiling-tool.html#profiling-tool-options.\n","\n","## GPU Job Tuning Recommendations\n","This has general suggestions for tuning your applications to run optimally on GPUs.\n","\n","## Per-Job Profile\n","The profiler output includes information about the application, data sources, executors, SQL stages, Spark properties, and key application metrics at the job and stage levels."]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"53b4d770-9db6-4bd7-9b93-d036d375eac5","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
Out[59]: 2011685
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
Out[59]: 2011685
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["import json\n","import requests\n","import base64\n","import shlex\n","import subprocess\n","import pandas as pd\n","\n","TOOL_JAR_URL = 'https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark-tools_2.12/23.12.2/rapids-4-spark-tools_2.12-23.12.2.jar'\n","TOOL_JAR_LOCAL_PATH = '/tmp/rapids-4-spark-tools.jar'\n","\n","# Profiling tool output directory.\n","OUTPUT_DIR = '/tmp' \n","\n","response = requests.get(TOOL_JAR_URL)\n","open(TOOL_JAR_LOCAL_PATH, \"wb\").write(response.content)"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"f0e4371a-d2d9-4449-81ed-8f6c61ae8f80","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["dbutils.widgets.text(\"log_path\", \"\")\n","eventlog_string=dbutils.widgets.get(\"log_path\") \n","\n","dbutils.widgets.text(\"output_path\", \"\")\n","outputpath_string=dbutils.widgets.get(\"output_path\")"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"6c35e478-abe6-49b7-97f9-a8aba71f11d3","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["worker_info_path = \"/tmp/worker_info.yaml\"\n","\n","worker_info = \"\"\"\n"," system:\n"," numCores: 32\n"," memory: 212992MiB\n"," numWorkers: 5\n"," gpu:\n"," memory: 15109MiB\n"," count: 4\n"," name: T4\n"," softwareProperties:\n"," spark.driver.maxResultSize: 7680m\n"," spark.driver.memory: 15360m\n"," spark.executor.cores: '8'\n"," spark.executor.instances: '2'\n"," spark.executor.memory: 47222m\n"," spark.executorEnv.OPENBLAS_NUM_THREADS: '1'\n"," spark.scheduler.mode: FAIR\n"," spark.sql.cbo.enabled: 'true'\n"," spark.ui.port: '0'\n"," spark.yarn.am.memory: 640m\n","\"\"\"\n","\n","with open(worker_info_path, 'w') as f:\n"," f.write(worker_info)"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"e9e7cecf-c2dc-4a0f-aea1-61a323e4ccc4","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["!java -Xmx10g -cp /tmp/rapids-4-spark-tools.jar:/databricks/jars/* com.nvidia.spark.rapids.tool.profiling.ProfileMain --csv --worker-info $worker_info_path --auto-tuner -o $outputpath_string $eventlog_string &> /tmp/prof_debug.log"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"be0a2da7-1ee3-475e-96f9-303779edfd85","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["import os\n","\n","app_df = pd.DataFrame(columns = ['appId', 'appName'])\n","\n","for x in os.scandir(outputpath_string + \"/rapids_4_spark_profile/\"):\n"," tmp_df = pd.read_csv(x.path + \"/application_information.csv\")\n"," app_df = app_df.append(tmp_df[['appId', 'appName']])"]},{"cell_type":"markdown","metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"a1e326ec-5701-4b08-ae0f-7df0c8440038","showTitle":false,"title":""}},"source":["## GPU Job Tuning Recommendations"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"4979f78c-44a0-4e54-b803-e5e194b71104","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["
apprecommendations
app-20220210005817-0212\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1197m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004538-0189\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000414-0117\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2353m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005713-0210\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000744-0123\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002521-0154\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004801-0193\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3158m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002620-0156\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001501-0135\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1365m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001417-0134\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1365m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001930-0143\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005502-0206\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002316-0150\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004324-0185\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005039-0198\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004834-0194\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2099m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004011-0180\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004656-0191\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001324-0133\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2225m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000856-0125\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000241-0114\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002105-0146\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000312-0115\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003325-0169\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002654-0157\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005425-0205\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1509m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000700-0122\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001109-0129\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002138-0147\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001717-0139\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000018-0110\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002725-0158\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000933-0126\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004617-0190\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005846-0213\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1197m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002757-0159\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001038-0128\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000628-0121\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3404m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001959-0144\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000556-0120\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004727-0192\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3354m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005222-0201\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000825-0124\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005536-0207\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001139-0130\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000343-0116\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003705-0176\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2337m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003359-0170\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005611-0208\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2974m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000119-0112\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3061m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005322-0203\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003900-0179\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004355-0186\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001648-0138\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002938-0162\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005354-0204\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001821-0141\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2530m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210005252-0202\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002452-0153\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001617-0137\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003254-0168\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004507-0188\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001751-0140\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005642-0209\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004905-0195\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003751-0177\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002030-0145\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001854-0142\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002549-0155\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001239-0132\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2225m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004428-0187\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3796m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002240-0149\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005147-0200\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2974m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220523230623-0000\n","Spark Properties:\n","--conf spark.executor.instances=20\n","--conf spark.executor.memory=16384m\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=6\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' should be increased since spilling occurred.\n","
app-20220210000207-0113\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220209235945-0109\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002349-0151\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002421-0152\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000448-0118\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001005-0127\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003149-0166\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003635-0175\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003223-0167\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004041-0181\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003532-0173\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003607-0174\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003048-0164\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004145-0183\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003825-0178\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003118-0165\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000049-0111\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000519-0119\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3061m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003015-0163\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005744-0211\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210005111-0199\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004217-0184\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002904-0161\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3796m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001546-0136\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003503-0172\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001209-0131\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004114-0182\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002208-0148\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002833-0160\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003431-0171\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"aggData":[],"aggError":"","aggOverflow":false,"aggSchema":[],"aggSeriesLimitReached":false,"aggType":"","arguments":{},"columnCustomDisplayInfos":{},"data":[["app-20220210005817-0212","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1197m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004538-0189","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000414-0117","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2353m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005713-0210","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000744-0123","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002521-0154","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004801-0193","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3158m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002620-0156","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001501-0135","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1365m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001417-0134","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1365m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001930-0143","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005502-0206","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002316-0150","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004324-0185","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005039-0198","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004834-0194","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2099m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004011-0180","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004656-0191","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001324-0133","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2225m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000856-0125","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000241-0114","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002105-0146","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000312-0115","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003325-0169","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002654-0157","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005425-0205","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1509m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000700-0122","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001109-0129","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002138-0147","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001717-0139","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000018-0110","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002725-0158","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000933-0126","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004617-0190","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005846-0213","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1197m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002757-0159","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001038-0128","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000628-0121","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3404m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001959-0144","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000556-0120","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004727-0192","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3354m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005222-0201","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000825-0124","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005536-0207","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001139-0130","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000343-0116","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003705-0176","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2337m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003359-0170","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005611-0208","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2974m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000119-0112","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3061m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005322-0203","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003900-0179","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004355-0186","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001648-0138","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002938-0162","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005354-0204","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001821-0141","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2530m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210005252-0202","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002452-0153","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001617-0137","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003254-0168","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004507-0188","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001751-0140","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005642-0209","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004905-0195","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003751-0177","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002030-0145","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001854-0142","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002549-0155","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001239-0132","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2225m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004428-0187","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3796m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002240-0149","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005147-0200","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2974m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220523230623-0000","\nSpark Properties:\n--conf spark.executor.instances=20\n--conf spark.executor.memory=16384m\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=6\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' should be increased since spilling occurred.\n"],["app-20220210000207-0113","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220209235945-0109","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002349-0151","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002421-0152","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000448-0118","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001005-0127","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003149-0166","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003635-0175","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003223-0167","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004041-0181","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003532-0173","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003607-0174","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003048-0164","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004145-0183","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003825-0178","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003118-0165","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000049-0111","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000519-0119","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3061m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003015-0163","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005744-0211","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210005111-0199","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004217-0184","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002904-0161","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3796m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001546-0136","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003503-0172","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001209-0131","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004114-0182","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002208-0148","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002833-0160","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003431-0171","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"]],"datasetInfos":[],"dbfsResultPath":null,"isJsonSchema":true,"metadata":{},"overflow":false,"plotOptions":{"customPlotOptions":{},"displayType":"table","pivotAggregation":null,"pivotColumns":null,"xColumns":null,"yColumns":null},"removedWidgets":[],"schema":[{"metadata":"{}","name":"app","type":"\"string\""},{"metadata":"{}","name":"recommendations","type":"\"string\""}],"type":"table"}},"output_type":"display_data"}],"source":["app_list = app_df[\"appId\"].tolist()\n","app_recommendations = pd.DataFrame(columns=['app', 'recommendations'])\n","\n","for app in app_list:\n"," app_file = open(outputpath_string + \"/rapids_4_spark_profile/\" + app + \"/profile.log\")\n"," recommendations_start = 0\n"," recommendations_str = \"\"\n"," for line in app_file:\n"," if recommendations_start == 1:\n"," recommendations_str = recommendations_str + line\n"," if \"### D. Recommended Configuration ###\" in line:\n"," recommendations_start = 1\n"," app_recommendations = app_recommendations.append({'app': app, 'recommendations': recommendations_str}, ignore_index=True)\n"," \n","display(app_recommendations)"]}],"metadata":{"application/vnd.databricks.v1+notebook":{"dashboards":[{"elements":[{"dashboardResultIndex":0,"elementNUID":"be0a2da7-1ee3-475e-96f9-303779edfd85","elementType":"command","guid":"05eef9d3-7c55-4e26-8d1f-fa80338359e6","options":null,"position":{"height":6,"width":24,"x":0,"y":0,"z":null},"resultIndex":null}],"globalVars":{},"guid":"a9ea7799-040a-484e-a59d-c3cdf5072953","layoutOption":{"grid":true,"stack":true},"nuid":"91c1bfb2-695a-4e5c-8a25-848a433108dc","origId":2690941040041430,"title":"Executive View","version":"DashboardViewV1","width":1600},{"elements":[],"globalVars":{},"guid":"0896a45f-af1b-4849-b6c2-2b6abcb8b97b","layoutOption":{"grid":true,"stack":true},"nuid":"62243296-4562-4f06-90ac-d7a609f19c16","origId":2690941040041431,"title":"App View","version":"DashboardViewV1","width":1920}],"language":"python","notebookMetadata":{"pythonIndentUnit":2,"widgetLayout":[{"breakBefore":false,"name":"log_path","width":562},{"breakBefore":false,"name":"output_path","width":511}]},"notebookName":"[RAPIDS Accelerator for Apache Spark] Profiling Tool Notebook Template","notebookOrigID":2690941040041407,"widgets":{"log_path":{"currentValue":"/dbfs/user1/profiler_logs","nuid":"c7ce3870-db19-4813-b1cb-cead3f4c36f1","widgetInfo":{"defaultValue":"","label":null,"name":"log_path","options":{"validationRegex":null,"widgetType":"text"},"widgetType":"text"}},"output_path":{"currentValue":"/tmp","nuid":"a7d1d293-d8c3-452b-9ffb-786ea7a28843","widgetInfo":{"defaultValue":"","label":null,"name":"output_path","options":{"validationRegex":null,"widgetType":"text"},"widgetType":"text"}}}},"language_info":{"name":"python"}},"nbformat":4,"nbformat_minor":0} +{"cells":[{"cell_type":"markdown","metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"5156a76c-7af7-465d-aff4-41a2e54e3595","showTitle":false,"title":""}},"source":["# Welcome to the Profiling Tool for the RAPIDS Accelerator for Apache Spark\n","To run the tool, you need to enter a log path that represents the DBFS location for your Spark GPU event logs. Then you can select \"Run all\" to execute the notebook. After the notebook completes, you will see various output tables show up below. More options for running the profiling tool can be found here: https://docs.nvidia.com/spark-rapids/user-guide/latest/spark-profiling-tool.html#profiling-tool-options.\n","\n","## GPU Job Tuning Recommendations\n","This has general suggestions for tuning your applications to run optimally on GPUs.\n","\n","## Per-Job Profile\n","The profiler output includes information about the application, data sources, executors, SQL stages, Spark properties, and key application metrics at the job and stage levels."]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"53b4d770-9db6-4bd7-9b93-d036d375eac5","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
Out[59]: 2011685
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
Out[59]: 2011685
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["import json\n","import requests\n","import base64\n","import shlex\n","import subprocess\n","import pandas as pd\n","\n","TOOL_JAR_URL = 'https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark-tools_2.12/23.12.2/rapids-4-spark-tools_2.12-23.12.2.jar'\n","TOOL_JAR_LOCAL_PATH = '/tmp/rapids-4-spark-tools.jar'\n","\n","# Profiling tool output directory.\n","OUTPUT_DIR = '/tmp' \n","\n","response = requests.get(TOOL_JAR_URL)\n","open(TOOL_JAR_LOCAL_PATH, \"wb\").write(response.content)"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"f0e4371a-d2d9-4449-81ed-8f6c61ae8f80","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["dbutils.widgets.text(\"log_path\", \"\")\n","eventlog_string=dbutils.widgets.get(\"log_path\") \n","\n","dbutils.widgets.text(\"output_path\", \"\")\n","outputpath_string=dbutils.widgets.get(\"output_path\")"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"6c35e478-abe6-49b7-97f9-a8aba71f11d3","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["worker_info_path = \"/tmp/worker_info.yaml\"\n","\n","worker_info = \"\"\"\n"," system:\n"," numCores: 32\n"," memory: 212992MiB\n"," numWorkers: 5\n"," gpu:\n"," memory: 15109MiB\n"," count: 4\n"," name: T4\n"," softwareProperties:\n"," spark.driver.maxResultSize: 7680m\n"," spark.driver.memory: 15360m\n"," spark.executor.cores: '8'\n"," spark.executor.instances: '2'\n"," spark.executor.memory: 47222m\n"," spark.executorEnv.OPENBLAS_NUM_THREADS: '1'\n"," spark.scheduler.mode: FAIR\n"," spark.sql.cbo.enabled: 'true'\n"," spark.ui.port: '0'\n"," spark.yarn.am.memory: 640m\n","\"\"\"\n","\n","with open(worker_info_path, 'w') as f:\n"," f.write(worker_info)"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"e9e7cecf-c2dc-4a0f-aea1-61a323e4ccc4","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["!java -Xmx10g -cp /tmp/rapids-4-spark-tools.jar:/databricks/jars/* com.nvidia.spark.rapids.tool.profiling.ProfileMain --csv --worker-info $worker_info_path --auto-tuner -o $outputpath_string $eventlog_string &> /tmp/prof_debug.log"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"be0a2da7-1ee3-475e-96f9-303779edfd85","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"arguments":{},"data":"
","datasetInfos":[],"metadata":{},"removedWidgets":[],"type":"html"}},"output_type":"display_data"}],"source":["import os\n","\n","app_df = pd.DataFrame(columns = ['appId', 'appName'])\n","\n","for x in os.scandir(outputpath_string + \"/rapids_4_spark_profile/\"):\n"," tmp_df = pd.read_csv(x.path + \"/application_information.csv\")\n"," app_df = app_df.append(tmp_df[['appId', 'appName']])"]},{"cell_type":"markdown","metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"a1e326ec-5701-4b08-ae0f-7df0c8440038","showTitle":false,"title":""}},"source":["## GPU Job Tuning Recommendations"]},{"cell_type":"code","execution_count":null,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{},"inputWidgets":{},"nuid":"4979f78c-44a0-4e54-b803-e5e194b71104","showTitle":false,"title":""}},"outputs":[{"data":{"text/html":["
apprecommendations
app-20220210005817-0212\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1197m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004538-0189\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000414-0117\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2353m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005713-0210\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000744-0123\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002521-0154\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004801-0193\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3158m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002620-0156\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001501-0135\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1365m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001417-0134\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1365m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001930-0143\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005502-0206\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002316-0150\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004324-0185\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005039-0198\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004834-0194\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2099m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004011-0180\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004656-0191\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001324-0133\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2225m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000856-0125\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000241-0114\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002105-0146\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000312-0115\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003325-0169\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002654-0157\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005425-0205\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1509m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000700-0122\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001109-0129\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002138-0147\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001717-0139\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000018-0110\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002725-0158\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000933-0126\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004617-0190\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005846-0213\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=1197m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002757-0159\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001038-0128\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000628-0121\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3404m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001959-0144\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000556-0120\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004727-0192\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3354m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005222-0201\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210000825-0124\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005536-0207\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001139-0130\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000343-0116\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003705-0176\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2337m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003359-0170\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005611-0208\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2974m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000119-0112\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3061m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005322-0203\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003900-0179\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004355-0186\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001648-0138\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002938-0162\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005354-0204\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001821-0141\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2530m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210005252-0202\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002452-0153\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001617-0137\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003254-0168\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004507-0188\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001751-0140\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005642-0209\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004905-0195\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003751-0177\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002030-0145\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001854-0142\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002549-0155\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210001239-0132\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2225m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004428-0187\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3796m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002240-0149\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005147-0200\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=2974m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220523230623-0000\n","Spark Properties:\n","--conf spark.executor.instances=20\n","--conf spark.executor.memory=16384m\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.shuffle.partitions=6\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' should be increased since spilling occurred.\n","
app-20220210000207-0113\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220209235945-0109\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002349-0151\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002421-0152\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000448-0118\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001005-0127\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003149-0166\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003635-0175\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003223-0167\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004041-0181\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003532-0173\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003607-0174\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003048-0164\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210004145-0183\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003825-0178\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003118-0165\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000049-0111\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210000519-0119\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3061m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210003015-0163\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210005744-0211\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210005111-0199\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004217-0184\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210002904-0161\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=3796m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001546-0136\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003503-0172\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210001209-0131\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","
app-20220210004114-0182\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002208-0148\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210002833-0160\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
app-20220210003431-0171\n","Spark Properties:\n","--conf spark.executor.cores=8\n","--conf spark.executor.instances=20\n","--conf spark.executor.memoryOverhead=5734m\n","--conf spark.rapids.memory.pinnedPool.size=4096m\n","--conf spark.rapids.sql.concurrentGpuTasks=2\n","--conf spark.sql.files.maxPartitionBytes=4096m\n","--conf spark.sql.shuffle.partitions=200\n","--conf spark.task.resource.gpu.amount=0.125\n","\n","Comments:\n","- 'spark.executor.memoryOverhead' was not set.\n","- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n","- 'spark.sql.shuffle.partitions' was not set.\n","- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n","
"]},"metadata":{"application/vnd.databricks.v1+output":{"addedWidgets":{},"aggData":[],"aggError":"","aggOverflow":false,"aggSchema":[],"aggSeriesLimitReached":false,"aggType":"","arguments":{},"columnCustomDisplayInfos":{},"data":[["app-20220210005817-0212","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1197m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004538-0189","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000414-0117","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2353m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005713-0210","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000744-0123","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002521-0154","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004801-0193","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3158m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002620-0156","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001501-0135","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1365m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001417-0134","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1365m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001930-0143","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005502-0206","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002316-0150","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004324-0185","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005039-0198","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004834-0194","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2099m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004011-0180","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004656-0191","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001324-0133","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2225m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000856-0125","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000241-0114","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002105-0146","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000312-0115","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003325-0169","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002654-0157","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005425-0205","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1509m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000700-0122","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001109-0129","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002138-0147","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001717-0139","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000018-0110","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002725-0158","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000933-0126","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004617-0190","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005846-0213","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=1197m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002757-0159","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001038-0128","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000628-0121","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3404m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001959-0144","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000556-0120","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004727-0192","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3354m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005222-0201","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210000825-0124","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005536-0207","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001139-0130","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000343-0116","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003705-0176","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2337m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003359-0170","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005611-0208","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2974m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000119-0112","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3061m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005322-0203","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003900-0179","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004355-0186","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001648-0138","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002938-0162","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005354-0204","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001821-0141","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2530m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210005252-0202","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002452-0153","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001617-0137","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003254-0168","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004507-0188","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001751-0140","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005642-0209","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004905-0195","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003751-0177","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002030-0145","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001854-0142","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002549-0155","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210001239-0132","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2225m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004428-0187","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3796m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002240-0149","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005147-0200","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=2974m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220523230623-0000","\nSpark Properties:\n--conf spark.executor.instances=20\n--conf spark.executor.memory=16384m\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.shuffle.partitions=6\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' should be increased since spilling occurred.\n"],["app-20220210000207-0113","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220209235945-0109","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002349-0151","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002421-0152","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000448-0118","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001005-0127","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003149-0166","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003635-0175","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003223-0167","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004041-0181","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003532-0173","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003607-0174","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003048-0164","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210004145-0183","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003825-0178","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003118-0165","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000049-0111","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210000519-0119","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3061m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210003015-0163","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210005744-0211","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210005111-0199","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004217-0184","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210002904-0161","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=3796m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001546-0136","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003503-0172","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210001209-0131","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n"],["app-20220210004114-0182","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002208-0148","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210002833-0160","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"],["app-20220210003431-0171","\nSpark Properties:\n--conf spark.executor.cores=8\n--conf spark.executor.instances=20\n--conf spark.executor.memoryOverhead=5734m\n--conf spark.rapids.memory.pinnedPool.size=4096m\n--conf spark.rapids.sql.concurrentGpuTasks=2\n--conf spark.sql.files.maxPartitionBytes=4096m\n--conf spark.sql.shuffle.partitions=200\n--conf spark.task.resource.gpu.amount=0.125\n\nComments:\n- 'spark.executor.memoryOverhead' was not set.\n- 'spark.executor.memoryOverhead' must be set if using 'spark.rapids.memory.pinnedPool.size\n- 'spark.sql.shuffle.partitions' was not set.\n- Average JVM GC time is very high. Other Garbage Collectors can be used for better performance.\n"]],"datasetInfos":[],"dbfsResultPath":null,"isJsonSchema":true,"metadata":{},"overflow":false,"plotOptions":{"customPlotOptions":{},"displayType":"table","pivotAggregation":null,"pivotColumns":null,"xColumns":null,"yColumns":null},"removedWidgets":[],"schema":[{"metadata":"{}","name":"app","type":"\"string\""},{"metadata":"{}","name":"recommendations","type":"\"string\""}],"type":"table"}},"output_type":"display_data"}],"source":["app_list = app_df[\"appId\"].tolist()\n","app_recommendations = pd.DataFrame(columns=['app', 'recommendations'])\n","\n","for app in app_list:\n"," app_file = open(outputpath_string + \"/rapids_4_spark_profile/\" + app + \"/profile.log\")\n"," recommendations_start = 0\n"," recommendations_str = \"\"\n"," for line in app_file:\n"," if recommendations_start == 1:\n"," recommendations_str = recommendations_str + line\n"," if \"### D. Recommended Configuration ###\" in line:\n"," recommendations_start = 1\n"," app_recommendations = app_recommendations.append({'app': app, 'recommendations': recommendations_str}, ignore_index=True)\n"," \n","display(app_recommendations)"]}],"metadata":{"application/vnd.databricks.v1+notebook":{"dashboards":[{"elements":[{"dashboardResultIndex":0,"elementNUID":"be0a2da7-1ee3-475e-96f9-303779edfd85","elementType":"command","guid":"05eef9d3-7c55-4e26-8d1f-fa80338359e6","options":null,"position":{"height":6,"width":24,"x":0,"y":0,"z":null},"resultIndex":null}],"globalVars":{},"guid":"a9ea7799-040a-484e-a59d-c3cdf5072953","layoutOption":{"grid":true,"stack":true},"nuid":"91c1bfb2-695a-4e5c-8a25-848a433108dc","origId":2690941040041430,"title":"Executive View","version":"DashboardViewV1","width":1600},{"elements":[],"globalVars":{},"guid":"0896a45f-af1b-4849-b6c2-2b6abcb8b97b","layoutOption":{"grid":true,"stack":true},"nuid":"62243296-4562-4f06-90ac-d7a609f19c16","origId":2690941040041431,"title":"App View","version":"DashboardViewV1","width":1920}],"language":"python","notebookMetadata":{"pythonIndentUnit":2,"widgetLayout":[{"breakBefore":false,"name":"log_path","width":562},{"breakBefore":false,"name":"output_path","width":511}]},"notebookName":"[RAPIDS Accelerator for Apache Spark] Profiling Tool Notebook Template","notebookOrigID":2690941040041407,"widgets":{"log_path":{"currentValue":"/dbfs/user1/profiler_logs","nuid":"c7ce3870-db19-4813-b1cb-cead3f4c36f1","widgetInfo":{"defaultValue":"","label":null,"name":"log_path","options":{"validationRegex":null,"widgetType":"text"},"widgetType":"text"}},"output_path":{"currentValue":"/tmp","nuid":"a7d1d293-d8c3-452b-9ffb-786ea7a28843","widgetInfo":{"defaultValue":"","label":null,"name":"output_path","options":{"validationRegex":null,"widgetType":"text"},"widgetType":"text"}}}},"language_info":{"name":"python"}},"nbformat":4,"nbformat_minor":0} From a7b6d920d7e5e7f1f93b336a222f3b8398d948d3 Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 15:31:57 +0800 Subject: [PATCH 2/8] check all links and update the stable to legacy page Signed-off-by: liyuan --- .github/workflows/markdown-links-check.yml | 1 - README.md | 4 ++-- .../UDF-Examples/RAPIDS-accelerated-UDFs/README.md | 14 +++++++------- 3 files changed, 9 insertions(+), 10 deletions(-) diff --git a/.github/workflows/markdown-links-check.yml b/.github/workflows/markdown-links-check.yml index 8bba68ac4..8cad61d76 100644 --- a/.github/workflows/markdown-links-check.yml +++ b/.github/workflows/markdown-links-check.yml @@ -30,6 +30,5 @@ jobs: with: max-depth: -1 use-verbose-mode: 'yes' - check-modified-files-only: 'yes' config-file: '.github/workflows/markdown-links-check/markdown-links-check-config.json' base-branch: 'main' \ No newline at end of file diff --git a/README.md b/README.md index 6c4df4ca5..a75e6dacc 100644 --- a/README.md +++ b/README.md @@ -37,7 +37,7 @@ can be built for running on GPU with RAPIDS Accelerator in this repo: | 3 | XGBoost | Taxi (Scala) | End-to-end ETL + XGBoost example to predict taxi trip fare amount with [NYC taxi trips data set](https://www1.nyc.gov/site/tlc/about/tlc-trip-record-data.page) | 4 | ML/DL | PCA End-to-End | Spark MLlib based PCA example to train and transform with a synthetic dataset | 5 | UDF | cuSpatial - Point in Polygon | Spark cuSpatial example for Point in Polygon function using NYC Taxi pickup location dataset -| 6 | UDF | URL Decode | Decodes URL-encoded strings using the [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable/) -| 7 | UDF | URL Encode | URL-encodes strings using the [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable/) +| 6 | UDF | URL Decode | Decodes URL-encoded strings using the [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy/) +| 7 | UDF | URL Encode | URL-encodes strings using the [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy/) | 8 | UDF | [CosineSimilarity](./examples/UDF-Examples/RAPIDS-accelerated-UDFs/src/main/java/com/nvidia/spark/rapids/udf/java/CosineSimilarity.java) | Computes the cosine similarity between two float vectors using [native code](./examples/UDF-Examples/RAPIDS-accelerated-UDFs/src/main/cpp/src) | 9 | UDF | [StringWordCount](./examples/UDF-Examples/RAPIDS-accelerated-UDFs/src/main/java/com/nvidia/spark/rapids/udf/hive/StringWordCount.java) | Implements a Hive simple UDF using [native code](./examples/UDF-Examples/RAPIDS-accelerated-UDFs/src/main/cpp/src) to count words in strings diff --git a/examples/UDF-Examples/RAPIDS-accelerated-UDFs/README.md b/examples/UDF-Examples/RAPIDS-accelerated-UDFs/README.md index c3794cb06..32b6d9a53 100644 --- a/examples/UDF-Examples/RAPIDS-accelerated-UDFs/README.md +++ b/examples/UDF-Examples/RAPIDS-accelerated-UDFs/README.md @@ -18,7 +18,7 @@ which provides a single method we need to override called evaluateColumnar returns a cudf ColumnVector, because the GPU get its speed by performing operations on many rows at a time. In the `evaluateColumnar` function, there is a cudf implementation of URL decode that we're leveraging, so we don't need to write any native C++ code. This is all done -through the [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable). The benefit to +through the [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy). The benefit to implement via the Java API is ease of development, but the memory model is not friendly for doing GPU operations because the JVM makes the assumption that everything we're trying to do is in heap memory. We need to free the GPU resources in a timely manner with try-finally blocks. Note that we @@ -27,10 +27,10 @@ involving the RAPIDS accelerated UDF falls back to the CPU. - [URLDecode](src/main/scala/com/nvidia/spark/rapids/udf/scala/URLDecode.scala) decodes URL-encoded strings using the - [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable) + [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy) - [URLEncode](src/main/scala/com/nvidia/spark/rapids/udf/scala/URLEncode.scala) URL-encodes strings using the - [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable) + [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy) ## Spark Java UDF Examples @@ -53,10 +53,10 @@ significant effort. - [URLDecode](src/main/java/com/nvidia/spark/rapids/udf/java/URLDecode.java) decodes URL-encoded strings using the - [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable) + [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy) - [URLEncode](src/main/java/com/nvidia/spark/rapids/udf/java/URLEncode.java) URL-encodes strings using the - [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable) + [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy) - [CosineSimilarity](src/main/java/com/nvidia/spark/rapids/udf/java/CosineSimilarity.java) computes the [cosine similarity](https://en.wikipedia.org/wiki/Cosine_similarity) between two float vectors using [native code](src/main/cpp/src) @@ -67,11 +67,11 @@ Below are some examples for implementing RAPIDS accelerated Hive UDF via JNI and - [URLDecode](src/main/java/com/nvidia/spark/rapids/udf/hive/URLDecode.java) implements a Hive simple UDF using the - [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable) + [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy) to decode URL-encoded strings - [URLEncode](src/main/java/com/nvidia/spark/rapids/udf/hive/URLEncode.java) implements a Hive generic UDF using the - [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/stable) + [Java APIs of RAPIDS cudf](https://docs.rapids.ai/api/cudf-java/legacy) to URL-encode strings - [StringWordCount](src/main/java/com/nvidia/spark/rapids/udf/hive/StringWordCount.java) implements a Hive simple UDF using From 9f2f1538b56571c221e3cfbdea653476a78f363d Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 15:35:05 +0800 Subject: [PATCH 3/8] fix deadlink in udf readme Signed-off-by: liyuan --- examples/UDF-Examples/Spark-cuSpatial/README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/UDF-Examples/Spark-cuSpatial/README.md b/examples/UDF-Examples/Spark-cuSpatial/README.md index 0daf9fe0f..ce1407f61 100644 --- a/examples/UDF-Examples/Spark-cuSpatial/README.md +++ b/examples/UDF-Examples/Spark-cuSpatial/README.md @@ -105,7 +105,7 @@ Note: The docker env is just for building the jar, not for running the applicati docker push : ``` -2. Follow the [Spark-rapids get-started document](https://nvidia.github.io/spark-rapids/docs/get-started/getting-started-databricks.html#start-a-databricks-cluster) to create a GPU cluster on AWS Databricks. +2. Follow the [Spark-rapids get-started document](https://docs.nvidia.com/spark-rapids/user-guide/latest/getting-started/databricks.html) to create a GPU cluster on AWS Databricks. Below are some different steps since a custom docker image is used with Databricks: * Databricks Runtime Version Choose a non-ML Databricks Runtime such as `Runtime: 9.1 LTS(Scala 2.12, Spark 3.1.2)` and From 744c64a4724b734b4cbca206e275e6c921cb4593 Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 15:49:57 +0800 Subject: [PATCH 4/8] update the markdown link checker version Signed-off-by: liyuan --- .github/workflows/markdown-links-check.yml | 2 +- examples/UDF-Examples/Spark-cuSpatial/README.md | 3 ++- 2 files changed, 3 insertions(+), 2 deletions(-) diff --git a/.github/workflows/markdown-links-check.yml b/.github/workflows/markdown-links-check.yml index 8cad61d76..7e6de1349 100644 --- a/.github/workflows/markdown-links-check.yml +++ b/.github/workflows/markdown-links-check.yml @@ -26,7 +26,7 @@ jobs: - name: work around permission issue run: git config --global --add safe.directory /github/workspace - uses: actions/checkout@master - - uses: gaurav-nelson/github-action-markdown-link-check@v1 + - uses: gaurav-nelson/github-action-markdown-link-check@v1.0.14 with: max-depth: -1 use-verbose-mode: 'yes' diff --git a/examples/UDF-Examples/Spark-cuSpatial/README.md b/examples/UDF-Examples/Spark-cuSpatial/README.md index ce1407f61..6ba27ae2e 100644 --- a/examples/UDF-Examples/Spark-cuSpatial/README.md +++ b/examples/UDF-Examples/Spark-cuSpatial/README.md @@ -82,7 +82,8 @@ Note: The docker env is just for building the jar, not for running the applicati ## Run ### GPU Demo on Spark Standalone on-premises cluster -1. Set up [a standalone cluster](/docs/get-started/xgboost-examples/on-prem-cluster/standalone-scala.md) of Spark. Make sure the conda/lib is included in LD_LIBRARY_PATH, so that spark executors can load libcuspatial.so. +1. Set up [a standalone cluster](../../../docs/get-started/xgboost-examples/on-prem-cluster/standalone-scala.md) of Spark. + Make sure the conda/lib is included in LD_LIBRARY_PATH, so that spark executors can load libcuspatial.so. 2. Download Spark RAPIDS JAR * [Spark RAPIDS JAR v23.02.0](https://repo1.maven.org/maven2/com/nvidia/rapids-4-spark_2.12/23.02.0/rapids-4-spark_2.12-23.02.0.jar) or above From 6a8f9eae0d18ff81ab6a3975f88c251a9a25c85d Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 15:53:32 +0800 Subject: [PATCH 5/8] update the markdown link checker version Signed-off-by: liyuan --- .github/workflows/markdown-links-check.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/markdown-links-check.yml b/.github/workflows/markdown-links-check.yml index 7e6de1349..b22ce6a6e 100644 --- a/.github/workflows/markdown-links-check.yml +++ b/.github/workflows/markdown-links-check.yml @@ -26,7 +26,7 @@ jobs: - name: work around permission issue run: git config --global --add safe.directory /github/workspace - uses: actions/checkout@master - - uses: gaurav-nelson/github-action-markdown-link-check@v1.0.14 + - uses: gaurav-nelson/github-action-markdown-link-check@v1.0.15 with: max-depth: -1 use-verbose-mode: 'yes' From 10e85548312924260bba2500e0bd4da1699c2f3a Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 15:56:14 +0800 Subject: [PATCH 6/8] update the markdown link checker version Signed-off-by: liyuan --- .github/workflows/markdown-links-check.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/markdown-links-check.yml b/.github/workflows/markdown-links-check.yml index b22ce6a6e..785b956b0 100644 --- a/.github/workflows/markdown-links-check.yml +++ b/.github/workflows/markdown-links-check.yml @@ -26,7 +26,7 @@ jobs: - name: work around permission issue run: git config --global --add safe.directory /github/workspace - uses: actions/checkout@master - - uses: gaurav-nelson/github-action-markdown-link-check@v1.0.15 + - uses: gaurav-nelson/github-action-markdown-link-check@0f074c8562c5a8fed38282b7c741d1970bb1512d with: max-depth: -1 use-verbose-mode: 'yes' From 01be7fa008e6a27587814eca949feae83ff8c0a6 Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 16:13:21 +0800 Subject: [PATCH 7/8] comments the anchor links to workarround the issue Signed-off-by: liyuan --- .github/workflows/markdown-links-check.yml | 2 +- .../markdown-links-check/markdown-links-check-config.json | 8 ++++++++ examples/ML+DL-Examples/Spark-DL/criteo_train/README.md | 2 +- .../Spark-DL/criteo_train/notebooks/Criteo-Training.ipynb | 2 +- 4 files changed, 11 insertions(+), 3 deletions(-) diff --git a/.github/workflows/markdown-links-check.yml b/.github/workflows/markdown-links-check.yml index 785b956b0..8cad61d76 100644 --- a/.github/workflows/markdown-links-check.yml +++ b/.github/workflows/markdown-links-check.yml @@ -26,7 +26,7 @@ jobs: - name: work around permission issue run: git config --global --add safe.directory /github/workspace - uses: actions/checkout@master - - uses: gaurav-nelson/github-action-markdown-link-check@0f074c8562c5a8fed38282b7c741d1970bb1512d + - uses: gaurav-nelson/github-action-markdown-link-check@v1 with: max-depth: -1 use-verbose-mode: 'yes' diff --git a/.github/workflows/markdown-links-check/markdown-links-check-config.json b/.github/workflows/markdown-links-check/markdown-links-check-config.json index 32fba2b72..360a9f642 100644 --- a/.github/workflows/markdown-links-check/markdown-links-check-config.json +++ b/.github/workflows/markdown-links-check/markdown-links-check-config.json @@ -1,4 +1,12 @@ { + "ignorePatterns": [ + { + "pattern": "/docs" + }, + { + "pattern": "/datasets" + } + ], "timeout": "15s", "retryOn429": true, "retryCount":30, diff --git a/examples/ML+DL-Examples/Spark-DL/criteo_train/README.md b/examples/ML+DL-Examples/Spark-DL/criteo_train/README.md index a743e45cf..083894055 100644 --- a/examples/ML+DL-Examples/Spark-DL/criteo_train/README.md +++ b/examples/ML+DL-Examples/Spark-DL/criteo_train/README.md @@ -7,7 +7,7 @@ _Please note: The following demo is dedicated for DGX-2 machine(with V100 GPUs). ## Dataset The dataset used here is from Criteo clicklog dataset. -It's preprocessed by [DLRM](https://github.com/NVIDIA/DeepLearningExamples/tree/master/TensorFlow2/Recommendation/DLRM/preproc) +It's preprocessed by [DLRM](https://github.com/NVIDIA/DeepLearningExamples/tree/master/TensorFlow2/Recommendation/DLRM_and_DCNv2/preproc) ETL job on Spark. We also provide a small size sample data in sample_data folder. All 40 columns(1 label + 39 features) are already numeric. diff --git a/examples/ML+DL-Examples/Spark-DL/criteo_train/notebooks/Criteo-Training.ipynb b/examples/ML+DL-Examples/Spark-DL/criteo_train/notebooks/Criteo-Training.ipynb index 93dcd16f6..d760106b4 100644 --- a/examples/ML+DL-Examples/Spark-DL/criteo_train/notebooks/Criteo-Training.ipynb +++ b/examples/ML+DL-Examples/Spark-DL/criteo_train/notebooks/Criteo-Training.ipynb @@ -9,7 +9,7 @@ "\n", "This notebook contains the same content as \"criteo_keras.py\" but in a notebook(interactive) form.\n", "\n", - "The dataset used here is from Criteo clicklog dataset. It's preprocessed by DLRM(https://github.com/NVIDIA/DeepLearningExamples/tree/master/TensorFlow2/Recommendation/DLRM/preproc) ETL job on Spark.\n", + "The dataset used here is from Criteo clicklog dataset. It's preprocessed by DLRM(https://github.com/NVIDIA/DeepLearningExamples/tree/master/TensorFlow2/Recommendation/DLRM_and_DCNv2/preproc) ETL job on Spark.\n", "\n", "We provide a small size sample data in `sample_data` folder.\n", "\n", From 3885d5b0f1a7e0baa7ee4d023036e82970e30995 Mon Sep 17 00:00:00 2001 From: liyuan Date: Mon, 19 Feb 2024 16:15:50 +0800 Subject: [PATCH 8/8] comments the anchor links to workarround the issue Signed-off-by: liyuan --- .../markdown-links-check/markdown-links-check-config.json | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/.github/workflows/markdown-links-check/markdown-links-check-config.json b/.github/workflows/markdown-links-check/markdown-links-check-config.json index 360a9f642..de3af9914 100644 --- a/.github/workflows/markdown-links-check/markdown-links-check-config.json +++ b/.github/workflows/markdown-links-check/markdown-links-check-config.json @@ -5,6 +5,12 @@ }, { "pattern": "/datasets" + }, + { + "pattern": "/dockerfile" + }, + { + "pattern": "/examples" } ], "timeout": "15s",