-
Notifications
You must be signed in to change notification settings - Fork 3
/
values.yaml
124 lines (103 loc) · 5.08 KB
/
values.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
# Some default values for the Kamino VMSS-Prototype Pattern Image Generator
kamino:
name: kamino-gen
labels:
app: kamino-vmss-prototype
container:
# TODO: Point these to our public container registry once we have it setup
imageRegistry: ghcr.io
imageRepository: jackfrancis/kamino/vmss-prototype
imageTag: v0.1.0
# Pulling by hash has stronger assurance that the container is unchanged
imageHash: "a383124095bea6de2b90e4e2e81812f1796b63e641200f897cda1e0fcef0e621"
pullByHash: true
# include the name of the image pull secret in your cluster if you
# need one. (Note that they are local to the namespace you deploy in)
#pullSecret: skyman-acr
# Number of images to keep in the history
# Minimum is 2.
imageHistory: 3
# Number of seconds after the job completes before it is cleaned up
# see https://kubernetes.io/docs/concepts/workloads/controllers/job/#ttl-mechanism-for-finished-jobs
# This has it clean up the pod from the cluster within an hour, just to
# help reduce left over state in the cluster.
jobTtl: 3600
drain:
# Drain grace period is the maximum time to allow pods to drain load
# and leave the node. The default of 300 seconds is relatively long
# but extra safe.
gracePeriod: 300
# Drain Force, if set to true, will force drain the node if the normal
# drain fails to drain the node. This may be needed if the drain operation
# is not respected by the workloads in the cluster.
force: false
auto:
# This is the minimum amount of time a node needs to be ready before it
# is considered "good"
# Number of seconds or with single character suffix for seconds, minutes,
# hours, or days
# 3s == 3 seconds
# 3m == 3 minutes
# 3h == 3 hours
# 3d == 3 days
# We pick a default of 1h (1 hour)
minimumReadyTime: 1h
# The minimum number of valide candidates before we consider that we
# have enough signal to pick one of them. The default is 1 as you must
# have 1 candidate but large critical clusters you may wish to wait
# until you have a number of nodes that have taken the updates and are
# in ready state before you commit to picking one of them as a good
# candidate.
minimumCandidates: 1
# The maximum age of the current VMSS Prototype image before the auto
# node selection logic will try to find a new candate even if there
# has been no OS update/patch. If set to 0, this no age is defined
# and thus the age is never considered. Otherwise it is the number
# of days in age of the current image. For example, if set to 7 days,
# once the current VMSS Prototype image is over 7 days old, the auto
# node selection process will not limit its node selection only to nodes
# that have an OS update/patch that is newer than the prior image.
# OS Updates/patches are always preferred but if the age is over the
# limit, an OS update/patch is not required.
maximumImageAge: 0
# The annotation on a node that holds the last time a patch was
# "applied"
#lastPatchAnnotation: weave.works/kured-most-recent-reboot-needed
# If this annotation exists, we do not want the node as it is
# pending a servicing reboot. We don't want pending reboot nodes
#pendingRebootAnnotation: weave.works/kured-reboot-in-progress
# Set this to true such that auto-updates will only show what they
# would have done without actually doing it
dryRun: false
# If you are running in "automatic mode" you likely will want to have this
# run periodically (a cron job). This only takes effect if you have set
# targetVMSS and kamino.auto.cronjob.enabled is true
cronjob:
enabled: false
# This is the schedule in standard cron syntax. Note that the job will
# only be running one at a time. In some of my clusters, I run this every
# hour.
#schedule: "42 * * * *"
# A reasonable default is to do it every day. (Obviously, only if enalbed)
schedule: "42 0 * * *"
# This is the log-level for the process via python logging.
# Each level includes all below it, so DEBUG includes INFO, WARNING, etc...
# Best keep this at INFO
# DEBUG - Very verbose (albeit not that much more than INFO)
# INFO - Show all process executions and actions as progress
# WARNING - Show only the warnings
# ERROR - Show only the errors
# CRITICAL - Show only complete critical faults
logLevel: INFO
# The the target node are not things we can know
# automatically
#targetNode: k8s-agentpool1-12345678-vmss0001ct
# The target VMSS to use, or "ALL" for automatically doing all of them
# Only useful if you don't have a targetNode
#targetVMSS: k8s-agentpool1-12345678-vmss
#targetVMSS: ALL
# Set newUpdatedNodes to a higher value to immediately add more nodes built from the prototype after VMSS update is complete
newUpdatedNodes: 0
# Require scheduling the pod onto a control plane VM using the AKS Engine control plane VM identifier
# Default to false, which allows scheduling on any node other than the targetNode
scheduleOnControlPlane: false