-
Notifications
You must be signed in to change notification settings - Fork 7
/
Copy pathvalues.yaml
101 lines (88 loc) · 2.12 KB
/
values.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
# Default values for netchecker.
# This is a YAML-formatted file.
# Declare variables to be passed into your templates.
image:
repository:
agent: mirantis/k8s-netchecker-agent
server: mirantis/k8s-netchecker-server
etcd: quay.io/coreos/etcd
pullPolicy: IfNotPresent
# Overrides the image tag whose default is the chart appVersion.
tag:
agent: v1.2.2
server: v1.2.2
etcd: v3.4.17
imagePullSecrets: []
nameOverride: ""
fullnameOverride: ""
serviceAccount:
# Specifies whether a service account should be created
create: true
# Annotations to add to the service account
annotations: {}
# The name of the service account to use.
# If not set and create is true, a name is generated using the fullname template
name: ""
podAnnotations: {}
podSecurityContext:
fsGroup: 1000
securityContext:
# capabilities:
# drop:
# - ALL
readOnlyRootFilesystem: true
runAsNonRoot: true
runAsUser: 1000
runAsGroup: 1000
service:
type: ClusterIP
port: 8081
ingress:
enabled: false
className: ""
annotations: {}
# kubernetes.io/ingress.class: nginx
# kubernetes.io/tls-acme: "true"
hosts:
- host: netchecker.local
paths:
- path: /
pathType: ImplementationSpecific
tls: []
# - secretName: netchecker-tls
# hosts:
# - netchecker.local
resources:
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
server:
limits:
cpu: 100m
memory: 256Mi
requests:
cpu: 50m
memory: 64Mi
agent:
limits:
cpu: 30m
memory: 100Mi
requests:
cpu: 15m
memory: 64Mi
etcd:
limits:
cpu: 200m
memory: 256Mi
requests:
cpu: 100m
memory: 100Mi
nodeSelector: {}
tolerations:
- effect: NoSchedule
operator: Exists
affinity: {}
config:
debug: 5
reportInterval: 60