diff --git a/.gitignore b/.gitignore index 1bdc6b279..93baf27cd 100644 --- a/.gitignore +++ b/.gitignore @@ -11,4 +11,6 @@ hub/charts **/__pycache__ **/*.pyc +# random stuff **/.DS_Store +**/*.bak diff --git a/deployments/a11y/hubploy.yaml b/deployments/a11y/hubploy.yaml index 0c9d33c63..bfe84d961 100644 --- a/deployments/a11y/hubploy.yaml +++ b/deployments/a11y/hubploy.yaml @@ -1,6 +1,6 @@ images: images: - - name: us-central1-docker.pkg.dev/ucb-datahub-2018/user-images/a11y-user-image:9cb7ae26ffbe + - name: us-central1-docker.pkg.dev/ucb-datahub-2018/user-images/a11y-user-image:710a5e6cc90c cluster: provider: gcloud diff --git a/hub/Chart.yaml b/hub/Chart.yaml index 5559675a6..2dfb4abd2 100644 --- a/hub/Chart.yaml +++ b/hub/Chart.yaml @@ -2,4 +2,4 @@ apiVersion: v1 appVersion: '1.0' description: Deployment Chart for JupyterHub name: hub -version: 20240731-224556.git.8607.hf7abb041 +version: 20240731-224556.git.8735.h3e417283 diff --git a/hub/values.yaml b/hub/values.yaml index 1883d69db..fd5ae1a74 100644 --- a/hub/values.yaml +++ b/hub/values.yaml @@ -39,6 +39,16 @@ jupyterhub: # https://github.com/Jimbly/http-proxy-node16/commit/56283e33edfc7aad8c2605dd493da8a196b4371d # https://github.com/consideRatio/configurable-http-proxy/commits/main/ # https://jira-secure.berkeley.edu/browse/DH-382 timeouts break stuff + # + # bump the default ip_local_port_range from "32768 60999" to "10000 65000" + # https://z2jh.jupyter.org/en/latest/resources/reference.html#proxy-chp-extrapodspec + # https://kubernetes.io/docs/reference/kubernetes-api/workload-resources/pod-v1/#PodSpec + # + extraPodSpec: + securityContext: + sysctls: + - name: net.ipv4.ip_local_port_range + value: "10000 65000" image: tag: 4.6.2 # extraCommandLineFlags: diff --git a/node-placeholder/Chart.yaml b/node-placeholder/Chart.yaml index 912579940..e9e55d215 100644 --- a/node-placeholder/Chart.yaml +++ b/node-placeholder/Chart.yaml @@ -15,7 +15,7 @@ type: application # This is the chart version. This version number should be incremented each time you make changes # to the chart and its templates, including the app version. # Versions are expected to follow Semantic Versioning (https://semver.org/) -version: 20240731-224556.git.8610.hedc17750 +version: 20240731-224556.git.8735.h3e417283 # This is the version number of the application being deployed. This version number should be # incremented each time you make changes to the application. Versions are not expected to diff --git a/node-placeholder/values.yaml b/node-placeholder/values.yaml index 6dada1924..0d3589c06 100644 --- a/node-placeholder/values.yaml +++ b/node-placeholder/values.yaml @@ -4,7 +4,7 @@ image: repository: us-central1-docker.pkg.dev/ucb-datahub-2018/core/node-placeholder-scaler pullPolicy: IfNotPresent # Overrides the image tag whose default is the chart appVersion. - tag: "20240731-224556.git.8610.hedc17750" + tag: "20240731-224556.git.8652.h618d172c" imagePullSecrets: [] nameOverride: "" diff --git a/vendor/google/gke/node-pool/config/core-pool-sysctl.yaml b/vendor/google/gke/node-pool/config/core-pool-sysctl.yaml index 793ae82e3..c81b79680 100644 --- a/vendor/google/gke/node-pool/config/core-pool-sysctl.yaml +++ b/vendor/google/gke/node-pool/config/core-pool-sysctl.yaml @@ -4,48 +4,48 @@ kubeletConfig: # allowedUnsafeSysctls: 'net.core.*,net.ipv4.*' linuxConfig: sysctl: - # tune the ipv4 settings to not cause nginx to use all of the tcp memory - # addresses: https://jira-secure.berkeley.edu/browse/DH-3 - # - # following this process: - # https://cloud.google.com/kubernetes-engine/docs/how-to/node-system-config - # - # man page: - # https://man7.org/linux/man-pages/man7/tcp.7.html - # - # figures below are measured in units of system page size (4096B), - # and gleaned from the following articles: - # https://cromwell-intl.com/open-source/performance-tuning/tcp.html - # https://www.ibm.com/docs/en/linux-on-systems?topic=tuning-tcpip-ipv4-settings - # https://www.ibm.com/docs/en/linux-on-systems?topic=tuning-network-stack-settings - # - # net.ipv4.tcp_mem seems to be automagically generated from the supplied tcp_rmem - # and tcp_wmem settings. i believe? - # - # here be dragons. - # - # original values (as of 2023-04-19): - # net.core.netdev_max_backlog=1000 - # net.core.rmem_max=212992 - # net.core.wmem_max=212992 - # net.ipv4.tcp_rmem=4096 87380 6291456 - # net.ipv4.tcp_wmem=4096 16384 4194304 - # - # changes and additional tweaks (2024-04-11): - # net.ipv4.tcp_max_syn_backlog=4096 - # net.core.rmem_max=3276800 - # net.core.wmem_max=3276800 - # net.ipv4.tcp_rmem=4096 87380 16777216 - # net.ipv4.tcp_wmem=4096 87380 16777216 - # net.core.somaxconn=1024 - # - # https://fasterdata.es.net/host-tuning/linux/#toc-anchor-2 - net.core.netdev_max_backlog: '30000' - net.core.somaxconn: '4096' - # net.ipv4.tcp_max_syn_backlog: '8192' + # tune the ipv4 settings to not cause nginx to use all of the tcp memory + # addresses: https://jira-secure.berkeley.edu/browse/DH-3 + # + # following this process: + # https://cloud.google.com/kubernetes-engine/docs/how-to/node-system-config + # + # man page: + # https://man7.org/linux/man-pages/man7/tcp.7.html + # + # figures below are measured in units of system page size (4096B), + # and gleaned from the following articles: + # https://cromwell-intl.com/open-source/performance-tuning/tcp.html + # https://www.ibm.com/docs/en/linux-on-systems?topic=tuning-tcpip-ipv4-settings + # https://www.ibm.com/docs/en/linux-on-systems?topic=tuning-network-stack-settings + # + # net.ipv4.tcp_mem seems to be automagically generated from the supplied tcp_rmem + # and tcp_wmem settings. i believe? + # + # here be dragons. + # + # original values (as of 2023-04-19): + # net.core.netdev_max_backlog=1000 + # net.core.rmem_max=212992 + # net.core.wmem_max=212992 + # net.ipv4.tcp_rmem=4096 87380 6291456 + # net.ipv4.tcp_wmem=4096 16384 4194304 + # + # changes and additional tweaks (2024-04-11): + # net.ipv4.tcp_max_syn_backlog=4096 + # net.core.rmem_max=3276800 + # net.core.wmem_max=3276800 + # net.ipv4.tcp_rmem=4096 87380 16777216 + # net.ipv4.tcp_wmem=4096 87380 16777216 + # net.core.somaxconn=1024 + # + # https://fasterdata.es.net/host-tuning/linux/#toc-anchor-2 + net.core.netdev_max_backlog: '30000' + net.core.somaxconn: '4096' + # net.ipv4.tcp_max_syn_backlog: '8192' - # these values are in bytes - net.core.rmem_max: '67108864' - net.core.wmem_max: '67108864' - net.ipv4.tcp_rmem: '4096 87380 33554432' - net.ipv4.tcp_wmem: '4096 87380 33554432' + # these values are in bytes + net.core.rmem_max: '67108864' + net.core.wmem_max: '67108864' + net.ipv4.tcp_rmem: '4096 87380 33554432' + net.ipv4.tcp_wmem: '4096 87380 33554432' diff --git a/vendor/google/gke/node-pool/config/core-pool-sysctl.yaml.bak b/vendor/google/gke/node-pool/config/core-pool-sysctl.yaml.bak deleted file mode 100644 index 1e22a09b9..000000000 --- a/vendor/google/gke/node-pool/config/core-pool-sysctl.yaml.bak +++ /dev/null @@ -1,43 +0,0 @@ -kubeletConfig: - cpuManagerPolicy: static -# kubeletExtraConfig: -# allowedUnsafeSysctls: 'net.core.*,net.ipv4.*' -linuxConfig: - sysctl: - # tune the ipv4 settings to not cause nginx to use all of the tcp memory - # addresses: https://jira-secure.berkeley.edu/browse/DH-3 - # - # following this process: - # https://cloud.google.com/kubernetes-engine/docs/how-to/node-system-config - # - # man page: - # https://man7.org/linux/man-pages/man7/tcp.7.html - # - # figures below are measured in units of system page size (4096B), - # and gleaned from the following articles: - # https://cromwell-intl.com/open-source/performance-tuning/tcp.html - # https://www.ibm.com/docs/en/linux-on-systems?topic=tuning-tcpip-ipv4-settings - # https://www.ibm.com/docs/en/linux-on-systems?topic=tuning-network-stack-settings - # - # net.ipv4.tcp_mem seems to be automagically generated from the supplied tcp_rmem - # and tcp_wmem settings. i believe? - # - # here be dragons. - # - # original values (as of 2023-19-04): - # net.core.netdev_max_backlog=1000 - # net.core.rmem_max=212992 - # net.core.wmem_max=212992 - # net.ipv4.tcp_rmem=4096 87380 6291456 - # net.ipv4.tcp_wmem=4096 16384 4194304 - # - # https://fasterdata.es.net/host-tuning/linux/#toc-anchor-2 - net.core.netdev_max_backlog: '30000' - net.ipv4.tcp_max_syn_backlog: '8192' - net.core.rmem_default: - net.core.rmem_max: '67108864' - net.core.wmem_max: '67108864' - net.ipv4.tcp_rmem: '4096 87380 33554432' - net.ipv4.tcp_wmem: '4096 87380 33554432' - # http://simonhf.wordpress.com/2010/10/01/node-js-versus-sxe-hello-world-complexity-speed-and-memory-usage/ - net.core.somaxconn: '65535'