blob: 2d403e5ca2354eac4a8454549af231a790a7d1cc [file] [log] [blame]
Victor Ryzhenkinef34a022018-06-22 19:36:13 +04001/**
2 * Update kuberentes cluster
3 *
4 * Expected parameters:
Victor Ryzhenkin3401ee62019-01-18 06:34:26 +04005 * SALT_MASTER_CREDENTIALS Credentials to the Salt API.
6 * SALT_MASTER_URL Full Salt API address [https://10.10.10.1:8000].
7 * KUBERNETES_HYPERKUBE_SOURCE Versioned hyperkube binary to update control plane from. Should be null if update rolling via reclass-system level
8 * KUBERNETES_HYPERKUBE_SOURCE_HASH Versioned hyperkube binary to update control plane from. Should be null if update rolling via reclass-system level
9 * KUBERNETES_PAUSE_IMAGE Kubernetes pause image should have same version as hyperkube. May be null in case of reclass-system rollout
10 * TARGET_UPDATES Comma separated list of nodes to update (Valid values are ctl,cmp)
11 * CTL_TARGET Salt targeted kubernetes CTL nodes (ex. I@kubernetes:master). Kubernetes control plane
12 * CMP_TARGET Salt targeted compute nodes (ex. cmp* and 'I@kubernetes:pool') Kubernetes computes
13 * PER_NODE Target nodes will be managed one by one (bool)
14 * SIMPLE_UPGRADE Use previous version of upgrade without conron/drain abilities
15 * CONFORMANCE_RUN_AFTER Run Kubernetes conformance tests after update
16 * CONFORMANCE_RUN_BEFORE Run Kubernetes conformance tests before update
17 * TEST_K8S_API_SERVER Kubernetes API server address for test execution
18 * ARTIFACTORY_URL Artifactory URL where docker images located. Needed to correctly fetch conformance images.
19 * UPGRADE_CALICO_V2_TO_V3 Perform Calico upgrade from v2 to v3.
20 * KUBERNETES_CALICO_IMAGE Target calico/node image. May be null in case of reclass-system rollout.
21 * KUBERNETES_CALICO_CALICOCTL_SOURCE Versioned calico/ctl binary. Should be null if update rolling via reclass-system level
22 * KUBERNETES_CALICO_CALICOCTL_SOURCE_HASH Calico/ctl binary md5 hash. Should be null if update rolling via reclass-system level
23 * KUBERNETES_CALICO_CNI_SOURCE Versioned calico/cni binary. Should be null if update rolling via reclass-system level
24 * KUBERNETES_CALICO_CNI_SOURCE_HASH Сalico/cni binary hash. Should be null if update rolling via reclass-system level
25 * KUBERNETES_CALICO_BIRDCL_SOURCE Versioned calico/bird binary. Should be null if update rolling via reclass-system level
26 * KUBERNETES_CALICO_BIRDCL_SOURCE_HASH Сalico/bird binary hash. Should be null if update rolling via reclass-system level
27 * KUBERNETES_CALICO_CNI_IPAM_SOURCE Versioned calico/ipam binary. Should be null if update rolling via reclass-system level
28 * KUBERNETES_CALICO_CNI_IPAM_SOURCE_HASH Сalico/ipam binary hash. Should be null if update rolling via reclass-system level
29 * KUBERNETES_CALICO_KUBE_CONTROLLERS_IMAGE Target calico/kube-controllers image. May be null in case of reclass-system rollout.
30 * CALICO_UPGRADE_VERSION Version of "calico-upgrade" utility to be used ("v1.0.5" for Calico v3.1.3 target).
ashestakov138b0a52019-01-11 11:34:54 +020031 * KUBERNETES_ETCD_SOURCE Target etcd binary. May be null in case of reclass-system rollout.
32 * KUBERNETES_ETCD_SOURCE_HASH Target etcd binary checksum. May be null in case of reclass-system rollout.
Victor Ryzhenkinef34a022018-06-22 19:36:13 +040033 *
34**/
Aleksei Kasatkin9ce11842018-11-23 14:27:33 +010035import groovy.json.JsonSlurper
36
Victor Ryzhenkinef34a022018-06-22 19:36:13 +040037def common = new com.mirantis.mk.Common()
38def salt = new com.mirantis.mk.Salt()
39def python = new com.mirantis.mk.Python()
Victor Ryzhenkin723bd062018-12-11 17:09:06 +040040def test = new com.mirantis.mk.Test()
Victor Ryzhenkinef34a022018-06-22 19:36:13 +040041
42def updates = TARGET_UPDATES.tokenize(",").collect{it -> it.trim()}
43def pepperEnv = "pepperEnv"
44
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +020045def POOL = "I@kubernetes:pool"
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +020046
Aleksei Kasatkin1f4f5ba2018-11-20 18:30:36 +010047ETCD_ENDPOINTS = ""
48
Victor Ryzhenkinef34a022018-06-22 19:36:13 +040049def overrideKubernetesImage(pepperEnv) {
50 def salt = new com.mirantis.mk.Salt()
51
52 def k8sSaltOverrides = """
Victor Ryzhenkin3401ee62019-01-18 06:34:26 +040053 kubernetes_hyperkube_source: ${KUBERNETES_HYPERKUBE_SOURCE}
54 kubernetes_hyperkube_source_hash: ${KUBERNETES_HYPERKUBE_SOURCE_HASH}
Victor Ryzhenkinef34a022018-06-22 19:36:13 +040055 kubernetes_pause_image: ${KUBERNETES_PAUSE_IMAGE}
56 """
57 stage("Override kubernetes images to target version") {
58 salt.setSaltOverrides(pepperEnv, k8sSaltOverrides)
59 }
60}
61
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +020062def overrideCalicoImages(pepperEnv) {
63 def salt = new com.mirantis.mk.Salt()
64
65 def calicoSaltOverrides = """
66 kubernetes_calico_image: ${KUBERNETES_CALICO_IMAGE}
Victor Ryzhenkin3401ee62019-01-18 06:34:26 +040067 kubernetes_calico_calicoctl_source: ${KUBERNETES_CALICO_CALICOCTL_SOURCE}
68 kubernetes_calico_calicoctl_source_hash: ${KUBERNETES_CALICO_CALICOCTL_SOURCE_HASH}
69 kubernetes_calico_birdcl_source: ${KUBERNETES_CALICO_BIRDCL_SOURCE}
70 kubernetes_calico_birdcl_source_hash: ${KUBERNETES_CALICO_BIRDCL_SOURCE_HASH}
71 kubernetes_calico_cni_source: ${KUBERNETES_CALICO_CNI_SOURCE}
72 kubernetes_calico_cni_source_hash: ${KUBERNETES_CALICO_CNI_SOURCE_HASH}
73 kubernetes_calico_cni_ipam_source: ${KUBERNETES_CALICO_CNI_IPAM_SOURCE}
74 kubernetes_calico_cni_ipam_source_hash: ${KUBERNETES_CALICO_CNI_IPAM_SOURCE_HASH}
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +020075 kubernetes_calico_kube_controllers_image: ${KUBERNETES_CALICO_KUBE_CONTROLLERS_IMAGE}
76 """
77 stage("Override calico images to target version") {
78 salt.setSaltOverrides(pepperEnv, calicoSaltOverrides)
79 }
80}
81
82def downloadCalicoUpgrader(pepperEnv, target) {
83 def salt = new com.mirantis.mk.Salt()
84
85 stage("Downloading calico-upgrade utility") {
86 salt.cmdRun(pepperEnv, target, "rm -f ./calico-upgrade")
87 salt.cmdRun(pepperEnv, target, "wget https://github.com/projectcalico/calico-upgrade/releases/download/${CALICO_UPGRADE_VERSION}/calico-upgrade")
88 salt.cmdRun(pepperEnv, target, "chmod +x ./calico-upgrade")
89 }
90}
91
ashestakov138b0a52019-01-11 11:34:54 +020092def overrideEtcdSource(pepperEnv) {
93 def salt = new com.mirantis.mk.Salt()
94
95 def k8sSaltOverrides = """
96 kubernetes_etcd_source: ${KUBERNETES_ETCD_SOURCE}
97 kubernetes_etcd_source_hash: ${KUBERNETES_ETCD_SOURCE_HASH}
98 """
99 stage("Override etcd binaries to target version") {
100 salt.setSaltOverrides(pepperEnv, k8sSaltOverrides)
101 }
102}
103
104def performEtcdUpdateAndServicesRestart(pepperEnv, target) {
105 def salt = new com.mirantis.mk.Salt()
106
107 stage("Performing etcd update and services restart on ${target}") {
108 salt.enforceState(pepperEnv, target, "etcd.server.service")
109 salt.cmdRun(pepperEnv, target, ". /var/lib/etcd/configenv && etcdctl cluster-health")
110 }
111}
112
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400113def performKubernetesComputeUpdate(pepperEnv, target) {
114 def salt = new com.mirantis.mk.Salt()
115
116 stage("Execute Kubernetes compute update on ${target}") {
117 salt.enforceState(pepperEnv, target, 'kubernetes.pool')
118 salt.runSaltProcessStep(pepperEnv, target, 'service.restart', ['kubelet'])
119 }
120}
121
122def performKubernetesControlUpdate(pepperEnv, target) {
123 def salt = new com.mirantis.mk.Salt()
124
125 stage("Execute Kubernetes control plane update on ${target}") {
Victor Ryzhenkinc2024132019-01-23 05:39:34 +0400126 salt.enforceStateWithExclude(pepperEnv, target, "kubernetes", "kubernetes.master.setup,kubernetes.master.kube-addons")
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400127 // Restart kubelet
128 salt.runSaltProcessStep(pepperEnv, target, 'service.restart', ['kubelet'])
129 }
130}
131
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200132def startCalicoUpgrade(pepperEnv, target) {
133 def salt = new com.mirantis.mk.Salt()
134
135 stage("Starting upgrade using calico-upgrade: migrate etcd schema and lock Calico") {
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200136 def cmd = "export APIV1_ETCD_ENDPOINTS=${ETCD_ENDPOINTS} && " +
137 "export APIV1_ETCD_CA_CERT_FILE=/var/lib/etcd/ca.pem && " +
138 "export APIV1_ETCD_CERT_FILE=/var/lib/etcd/etcd-client.crt && " +
139 "export APIV1_ETCD_KEY_FILE=/var/lib/etcd/etcd-client.key && " +
140 "export ETCD_ENDPOINTS=${ETCD_ENDPOINTS} && " +
141 "export ETCD_CA_CERT_FILE=/var/lib/etcd/ca.pem && " +
142 "export ETCD_CERT_FILE=/var/lib/etcd/etcd-client.crt && " +
143 "export ETCD_KEY_FILE=/var/lib/etcd/etcd-client.key && " +
144 "rm /root/upg_complete -f && " +
145 "./calico-upgrade start --no-prompts --ignore-v3-data > upgrade-start.log && " +
146 "until [ -f /root/upg_complete ]; do sleep 0.1; done && " +
147 "./calico-upgrade complete --no-prompts > upgrade-complete.log && " +
148 "rm /root/upg_complete -f"
149 // "saltArgs = ['async']" doesn't work, so we have to run "cmd.run --async"
150 salt.cmdRun(pepperEnv, "I@salt:master", "salt -C '${target}' cmd.run '${cmd}' --async")
151 salt.cmdRun(pepperEnv, target, "until [ -f /root/upgrade-start.log ]; do sleep 0.1; done")
152 }
153}
154
155def completeCalicoUpgrade(pepperEnv, target) {
156 def salt = new com.mirantis.mk.Salt()
157
158 stage("Complete upgrade using calico-upgrade: unlock Calico") {
159 salt.cmdRun(pepperEnv, target, "echo 'true' > /root/upg_complete")
160 salt.cmdRun(pepperEnv, target, "while [ -f /root/upg_complete ]; do sleep 0.1; done")
161 salt.cmdRun(pepperEnv, target, "cat /root/upgrade-start.log")
162 salt.cmdRun(pepperEnv, target, "cat /root/upgrade-complete.log")
163 }
164}
165
Aleksei Kasatkind9d682e2018-12-12 14:51:59 +0100166def performCalicoConfigurationUpdateAndServicesRestart(pepperEnv, target, ctl_node) {
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200167 def salt = new com.mirantis.mk.Salt()
168
169 stage("Performing Calico configuration update and services restart") {
Aleksei Kasatkind9d682e2018-12-12 14:51:59 +0100170 if (containerDenabled(pepperEnv, ctl_node)) {
171 salt.enforceState(pepperEnv, target, "kubernetes.pool")
172 } else {
173 salt.enforceState(pepperEnv, target, "kubernetes.pool.calico")
174 }
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200175 salt.runSaltProcessStep(pepperEnv, target, 'service.restart', ['kubelet'])
176 }
177}
178
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400179def cordonNode(pepperEnv, target) {
180 def salt = new com.mirantis.mk.Salt()
181 def originalTarget = "I@kubernetes:master and not ${target}"
182
183 stage("Cordoning ${target} kubernetes node") {
184 def nodeShortName = target.tokenize(".")[0]
185 salt.cmdRun(pepperEnv, originalTarget, "kubectl cordon ${nodeShortName}", true, 1)
186 }
187}
188
189def uncordonNode(pepperEnv, target) {
190 def salt = new com.mirantis.mk.Salt()
191 def originalTarget = "I@kubernetes:master and not ${target}"
192
193 stage("Uncordoning ${target} kubernetes node") {
194 def nodeShortName = target.tokenize(".")[0]
195 salt.cmdRun(pepperEnv, originalTarget, "kubectl uncordon ${nodeShortName}", true, 1)
196 }
197}
198
199def drainNode(pepperEnv, target) {
200 def salt = new com.mirantis.mk.Salt()
201 def originalTarget = "I@kubernetes:master and not ${target}"
202
203 stage("Draining ${target} kubernetes node") {
204 def nodeShortName = target.tokenize(".")[0]
205 salt.cmdRun(pepperEnv, originalTarget, "kubectl drain --force --ignore-daemonsets --grace-period 100 --timeout 300s --delete-local-data ${nodeShortName}", true, 1)
206 }
207}
208
209def regenerateCerts(pepperEnv, target) {
210 def salt = new com.mirantis.mk.Salt()
211
212 stage("Regenerate certs for ${target}") {
213 salt.enforceState(pepperEnv, target, 'salt.minion.cert')
214 }
215}
216
Victor Ryzhenkinae909182018-10-02 17:49:18 +0400217def updateAddons(pepperEnv, target) {
218 def salt = new com.mirantis.mk.Salt()
219
220 stage("Upgrading Addons at ${target}") {
Victor Ryzhenkin40625bc2018-10-04 16:15:27 +0400221 salt.enforceState(pepperEnv, target, "kubernetes.master.kube-addons")
Victor Ryzhenkinfd9677f2018-10-16 16:14:40 +0400222 }
223}
224
225def updateAddonManager(pepperEnv, target) {
226 def salt = new com.mirantis.mk.Salt()
227
228 stage("Upgrading AddonManager at ${target}") {
Victor Ryzhenkinae909182018-10-02 17:49:18 +0400229 salt.enforceState(pepperEnv, target, "kubernetes.master.setup")
230 }
231}
232
Victor Ryzhenkinc2024132019-01-23 05:39:34 +0400233def buildDaemonsetMap(pepperEnv, target) {
234 def salt = new com.mirantis.mk.Salt()
235 def daemonset_lists
236 daemonset_lists = salt.cmdRun(pepperEnv, target, "kubectl get ds --all-namespaces | tail -n+2 | awk '{print \$2, \$1}'"
237 )['return'][0].values()[0].replaceAll('Salt command execution success','').tokenize("\n")
238 def daemonset_map = []
239 for (ds in daemonset_lists) {
240 a = ds.tokenize(" ")
241 daemonset_map << a
242 }
243 print("Built daemonset map")
244 print(daemonset_map)
245 return daemonset_map
246}
247
248def purgeDaemonsetPods(pepperEnv, target, daemonSetMap) {
249 def salt = new com.mirantis.mk.Salt()
250 def originalTarget = "I@kubernetes:master and not ${target}"
251 def nodeShortName = target.tokenize(".")[0]
252 firstTarget = salt.getFirstMinion(pepperEnv, originalTarget)
253
254 if (daemonSetMap) {
255 stage("Purging daemonset-managed pods on ${target}") {
256 for (ds in daemonSetMap) {
257 print("Purging "+ ds[0] +" inside "+ ds[1] +" namespace")
258 salt.cmdRun(pepperEnv, firstTarget, "kubectl get po -n ${ds[1]} -o wide | grep ${nodeShortName}" +
259 " | grep ${ds[0]} | awk '{print \$1}' | xargs --no-run-if-empty kubectl delete po -n ${ds[1]} --grace-period=0 --force")
260 }
261 }
262 }
263}
264
265def isNodeReady(pepperEnv, target) {
266 def salt = new com.mirantis.mk.Salt()
267 def originalTarget = "I@kubernetes:master and not ${target}"
268 def nodeShortName = target.tokenize(".")[0]
269 firstTarget = salt.getFirstMinion(pepperEnv, originalTarget)
270
Victor Ryzhenkineb7e7482019-03-28 02:11:28 +0400271 status = salt.cmdRun(pepperEnv, firstTarget, "kubectl get no ${nodeShortName} | tail -n+2 | awk '{print \$2}'"
Victor Ryzhenkinc2024132019-01-23 05:39:34 +0400272 )['return'][0].values()[0].replaceAll('Salt command execution success',''
273 ).replaceAll(',SchedulingDisabled','').trim()
274
275 if (status == "Ready") {
276 return true
277 } else {
278 return false
279 }
280}
281
282def rebootKubernetesNode(pepperEnv, target, times=15, delay=10) {
283 def common = new com.mirantis.mk.Common()
284 def debian = new com.mirantis.mk.Debian()
285
286 stage("Rebooting ${target}") {
287 debian.osReboot(pepperEnv, target)
288 common.retry(times, delay) {
289 if(!isNodeReady(pepperEnv, target)) {
290 error("Node still not in Ready state...")
291 }
292 }
293 }
294}
295
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400296def upgradeDocker(pepperEnv, target) {
297 def salt = new com.mirantis.mk.Salt()
298
299 stage("Upgrading docker at ${target}") {
300 salt.enforceState(pepperEnv, target, 'docker.host')
301 }
302}
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400303
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400304def runConformance(pepperEnv, target, k8s_api, image) {
305 def salt = new com.mirantis.mk.Salt()
306 def containerName = 'conformance_tests'
307 output_file = image.replaceAll('/', '-') + '.output'
308 def output_file_full_path = "/tmp/" + image.replaceAll('/', '-') + '.output'
309 def artifacts_dir = '_artifacts/'
310 salt.cmdRun(pepperEnv, target, "docker rm -f ${containerName}", false)
311 salt.cmdRun(pepperEnv, target, "docker run -d --name ${containerName} --net=host -e API_SERVER=${k8s_api} ${image}")
312 sleep(10)
313
314 print("Waiting for tests to run...")
315 salt.runSaltProcessStep(pepperEnv, target, 'cmd.run', ["docker wait ${containerName}"], null, false)
316
317 print("Writing test results to output file...")
318 salt.runSaltProcessStep(pepperEnv, target, 'cmd.run', ["docker logs -t ${containerName} > ${output_file_full_path}"])
319 print("Conformance test output saved in " + output_file_full_path)
320
321 // collect output
322 sh "mkdir -p ${artifacts_dir}"
323 file_content = salt.getFileContent(pepperEnv, target, '/tmp/' + output_file)
324 writeFile file: "${artifacts_dir}${output_file}", text: file_content
325 sh "cat ${artifacts_dir}${output_file}"
326 try {
327 sh "cat ${artifacts_dir}${output_file} | grep 'Test Suite Failed' && exit 1 || exit 0"
328 } catch (Throwable e) {
329 print("Conformance tests failed. Please check output")
330 currentBuild.result = "FAILURE"
331 currentBuild.description = currentBuild.description ? e.message + " " + currentBuild.description : e.message
332 throw e
333 }
334}
335
336def buildImageURL(pepperEnv, target, mcp_repo) {
337 def salt = new com.mirantis.mk.Salt()
338 def raw_version = salt.cmdRun(pepperEnv, target, "kubectl version --short -o json")['return'][0].values()[0].replaceAll('Salt command execution success','')
339 print("Kubernetes version: " + raw_version)
340 def serialized_version = readJSON text: raw_version
341 def short_version = (serialized_version.serverVersion.gitVersion =~ /([v])(\d+\.)(\d+\.)(\d+\-)(\d+)/)[0][0]
342 print("Kubernetes short version: " + short_version)
343 def conformance_image = mcp_repo + "/mirantis/kubernetes/k8s-conformance:" + short_version
344 return conformance_image
345}
346
347def executeConformance(pepperEnv, target, k8s_api, mcp_repo) {
Victor Ryzhenkin7e28c152019-03-28 02:23:42 +0400348 def image = buildImageURL(pepperEnv, target, mcp_repo)
349 print("Using image: " + image)
350 runConformance(pepperEnv, target, k8s_api, image)
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400351}
352
Victor Ryzhenkin723bd062018-12-11 17:09:06 +0400353def containerDinstalled(pepperEnv, target) {
354 def salt = new com.mirantis.mk.Salt()
355 return salt.cmdRun(pepperEnv, target, "containerd --version 2>1 1>/dev/null && echo 'true' || echo 'false'"
356 )['return'][0].values()[0].replaceAll('Salt command execution success','').trim().toBoolean()
357}
358
359def containerDenabled(pepperEnv, target) {
360 def salt = new com.mirantis.mk.Salt()
Victor Ryzhenkin71ecdf42018-12-11 22:22:50 +0400361 return salt.getPillar(pepperEnv, target, "kubernetes:common:containerd:enabled"
Victor Ryzhenkin723bd062018-12-11 17:09:06 +0400362 )["return"][0].values()[0].toBoolean()
363}
364
365def conformancePodDefExists(pepperEnv, target) {
366 def salt = new com.mirantis.mk.Salt()
367 return salt.cmdRun(pepperEnv, target, "test -e /srv/kubernetes/conformance.yml && echo 'true' || echo 'false'"
368 )['return'][0].values()[0].replaceAll('Salt command execution success','').trim().toBoolean()
369}
370
Victor Ryzhenkin3401ee62019-01-18 06:34:26 +0400371def printVersionInfo(pepperEnv, target) {
372 def salt = new com.mirantis.mk.Salt()
373 def common = new com.mirantis.mk.Common()
374
375 stage("Gather version and runtime information") {
376 common.infoMsg("Version and runtime info:")
377 salt.cmdRun(pepperEnv, target, "kubectl get no -o wide")
378 common.infoMsg("Cluster health info:")
379 salt.cmdRun(pepperEnv, target, "kubectl get cs")
380 common.infoMsg("ETCD health info:")
Victor Ryzhenkin3029c8b2019-01-18 22:17:57 +0400381 salt.cmdRun(pepperEnv, target, ". /var/lib/etcd/configenv && etcdctl cluster-health")
Victor Ryzhenkin3401ee62019-01-18 06:34:26 +0400382 common.infoMsg("Calico peers info:")
383 salt.cmdRun(pepperEnv, target, "calicoctl node status")
384 }
385}
386
Aleksei Kasatkin5ccea272018-12-06 17:34:58 +0100387def calicoEnabled(pepperEnv, target) {
388 def salt = new com.mirantis.mk.Salt()
389 return salt.getPillar(pepperEnv, target, "kubernetes:pool:network:calico:enabled"
390 )["return"][0].values()[0].toBoolean()
391}
392
393def checkCalicoClusterState(pepperEnv, target) {
394 def common = new com.mirantis.mk.Common()
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200395 def salt = new com.mirantis.mk.Salt()
396
Aleksei Kasatkin5ccea272018-12-06 17:34:58 +0100397 stage("Checking Calico cluster state after upgrade") {
398 // check Calico cluster and cli clients versions
399 def checkVer = [
400 "Client Version:": [verStr: "", dif: false, wrong: false],
401 "Cluster Version:": [verStr: "", dif: false, wrong: false]
402 ]
403 def checkVerPassed = true
404 def versionResults = salt.cmdRun(pepperEnv, target, "calicoctl version | grep -i version")['return'][0]
405 versionResults.each { k, v ->
406 // println("Node:\n${k}\nResult:\n${v}")
407 for (verLine in v.split("\n")) {
408 for (verType in checkVer.keySet()) {
409 if (verLine.contains(verType)) {
410 def verRec = checkVer[verType]
411 ver = (verLine - verType).trim()
412 if (!verRec.verStr) {
413 verRec.verStr = ver
414 }
415 if (verRec.verStr != ver) {
416 verRec.dif = true
417 checkVerPassed = false
418 }
419 version = ver.tokenize(".")
420 if ((version.size() < 3) || (version[0] != "v3")) {
421 verRec.wrong = true
422 checkVerPassed = false
423 }
424 checkVer[verType] = verRec
425 }
426 }
427 }
428 }
429 if (checkVerPassed) {
430 common.infoMsg("Calico version verification passed")
431 }
432 else {
433 def warningMsg = "Calico version verification failed.\n"
434 checkVer.each { k, rec ->
435 if (rec.dif) {
436 warningMsg += "${k} versions are different across nodes.\n"
437 }
438 if (rec.wrong) {
439 warningMsg += "${k} (some) versions are wrong - should be v3.x.\n"
440 }
441 }
442 common.warningMsg(warningMsg)
443 currentBuild.description += "<br><b>${warningMsg}</b><br><br>"
444 }
445
446 // check Calico nodes' statuses
447 def nodeStatusResults = salt.cmdRun(pepperEnv, target, "calicoctl node status")['return'][0]
448 def nodesRunning = true
449 def peersNotFound = []
450 def peersNotOnline = []
451 nodeStatusResults.each { k, v ->
452 // println("Node:\n${k}\nResult:\n${v}")
453 if (!v.contains("Calico process is running")) {
454 nodesRunning = false
455 def warningMsg = "Node ${k}: Calico node is not running."
456 common.warningMsg(warningMsg)
457 currentBuild.description += "<br><b>${warningMsg}</b><br><br>"
458 }
459 def nodePeersFound = false
460 def nodePeersOnline = true
461 for (nodeLine in v.split("\n")) {
462 if (nodeLine.contains("|") && (!nodeLine.contains("STATE"))) {
463 def col = nodeLine.tokenize("|").collect{it.trim()}
464 if (col.size() == 5) {
465 nodePeersFound = true
466 if ((col[2] != "up") || (col[4] != "Established")) {
467 def warningMsg = "Node ${k}: BGP peer '${col[0]}' is out of reach. Peer state: '${col[2]}', connection info: '${col[4]}'."
468 common.warningMsg(warningMsg)
469 currentBuild.description += "<br><b>${warningMsg}</b><br><br>"
470 nodePeersOnline = false
471 }
472 }
473 }
474 }
475 if (!nodePeersFound) {
476 peersNotFound += k
477 }
478 if (!nodePeersOnline) {
479 peersNotOnline += k
480 }
481 }
482 if (nodesRunning) {
483 common.infoMsg("All the Calico nodes are running")
484 }
485 if (peersNotFound) {
486 def warningMsg = "BGP peers not found for the node(s): " + peersNotFound.join(', ') + "."
487 common.warningMsg(warningMsg)
488 currentBuild.description += "<br><b>${warningMsg}</b><br><br>"
489 } else {
490 common.infoMsg("BGP peers were found for all the nodes")
491 }
492 if (!peersNotOnline) {
493 common.infoMsg("All reported BGP peers are reachable")
494 }
495
496 // check that 'calico-kube-controllers' is running
497 // one CTL node will be used to get pod's state using kubectl
498 def ctl_node = salt.getMinionsSorted(pepperEnv, CTL_TARGET)[0]
499 def kubeCtrlResult = salt.cmdRun(
500 pepperEnv, ctl_node, "kubectl get pod -n kube-system --selector=k8s-app=calico-kube-controllers"
501 )['return'][0].values()[0].toString()
502 if (kubeCtrlResult.contains("calico-kube-controllers")) {
503 for (line in kubeCtrlResult.split("\n")) {
504 if (line.contains("calico-kube-controllers")) {
505 col = line.tokenize(" ")
506 if ((col[1] != "1/1") || (col[2] != "Running")) {
507 def warningMsg = "Calico kube-controllers pod is not running properly."
508 common.warningMsg(warningMsg)
509 currentBuild.description += "<br><b>${warningMsg}</b><br><br>"
510 }
511 else {
512 common.infoMsg("Calico kube-controllers pod is running.")
513 }
514 break
515 }
516 }
517 } else {
518 def warningMsg = "Calico kube-controllers pod was not scheduled."
519 common.warningMsg(warningMsg)
520 currentBuild.description += "<br><b>${warningMsg}</b><br><br>"
521 }
Aleksei Kasatkin1f4f5ba2018-11-20 18:30:36 +0100522 }
523}
524
525def checkCalicoUpgradePossibility(pepperEnv, target) {
526 def salt = new com.mirantis.mk.Salt()
527
528 stage("Verification of Calico upgrade possibility") {
529 // check Calico version
530 def versionResult = salt.cmdRun(
531 pepperEnv, target, "calicoctl version | grep 'Cluster Version'"
532 )['return'][0].values()[0].split("\n")[0].trim()
533 versionStr = (versionResult - "Cluster Version:").trim()
534 version = versionStr.tokenize(".")
535 if ((version.size() < 3) || (version[0] != "v2") || (version[1] != "6") || (version[2].toInteger() < 5)) {
536 error(
537 "Current Calico ${versionStr} cannot be upgraded to v3.x. " +
538 "Calico v2.6.x starting from v2.6.5 can be upgraded. " +
539 "For earlier versions, please update to v2.6.5 first."
540 )
541 }
542 print("Calico version was determined: ${versionStr}")
543
544 // check Calico is switched on
545 def readinessResult = salt.cmdRun(
546 pepperEnv, target, ". /var/lib/etcd/configenv && etcdctl get /calico/v1/Ready"
547 )['return'][0].values()[0].split("\n")[0].trim()
548 print("Calico readiness check result: ${readinessResult}")
549 if (readinessResult != "true") {
550 // try set it to true
551 readinessResult = salt.cmdRun(
552 pepperEnv, target, ". /var/lib/etcd/configenv && etcdctl set /calico/v1/Ready true"
553 )['return'][0].values()[0].split("\n")[0].trim()
554 print("Calico readiness result 2nd attempt: ${readinessResult}")
555 if (readinessResult != "true") {
556 error("Calico is not ready. '/calico/v1/Ready': '${readinessResult}'")
557 }
558 }
559
560 // Calico data upgrade dry-run
561 def cmd = "export APIV1_ETCD_ENDPOINTS=${ETCD_ENDPOINTS} && " +
562 "export APIV1_ETCD_CA_CERT_FILE=/var/lib/etcd/ca.pem && " +
563 "export APIV1_ETCD_CERT_FILE=/var/lib/etcd/etcd-client.crt && " +
564 "export APIV1_ETCD_KEY_FILE=/var/lib/etcd/etcd-client.key && " +
565 "export ETCD_ENDPOINTS=${ETCD_ENDPOINTS} && " +
566 "export ETCD_CA_CERT_FILE=/var/lib/etcd/ca.pem && " +
567 "export ETCD_CERT_FILE=/var/lib/etcd/etcd-client.crt && " +
568 "export ETCD_KEY_FILE=/var/lib/etcd/etcd-client.key && " +
569 "./calico-upgrade dry-run --ignore-v3-data"
570 def dryRunResult = salt.cmdRun(pepperEnv, target, cmd)['return'][0].values()[0]
571 // check dry-run result
572 def validationSuccessStr = "Successfully validated v1 to v3 conversion"
573 if (!dryRunResult.contains(validationSuccessStr)) {
574 error("Calico data upgrade dry-run has failed")
575 }
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200576 }
577}
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400578
Aleksei Kasatkin9ce11842018-11-23 14:27:33 +0100579def checkCalicoPolicySetting(pepperEnv, target) {
580 def common = new com.mirantis.mk.Common()
581 def salt = new com.mirantis.mk.Salt()
582
583 stage("Checking of Calico network policy setting") {
584 // check Calico policy enabled
585 def cniPolicy = false
586 def addonsPolicy = false
587 def kubeCtrlRunning = false
588
589 // check CNI config
590 def cniCfgResult = salt.cmdRun(
591 pepperEnv, target, "cat /etc/cni/net.d/10-calico.conf"
592 )['return'][0].values()[0].toString()
593 def cniCfg = new JsonSlurper().parseText(cniCfgResult)
594 if (cniCfg.get("policy") != null) {
595 if (cniCfg["policy"].get("type") == "k8s") {
596 cniPolicy = true
597 } else {
598 common.warningMsg("Calico policy type is unknown or not set.")
599 }
600 }
601
602 // check k8s addons
603 def addonsResult = salt.cmdRun(
604 pepperEnv, target, "ls /etc/kubernetes/addons"
605 )['return'][0].values()[0].toString()
606 if (addonsResult.contains("calico_policy")) {
607 addonsPolicy = true
608 }
609
610 // check kube-controllers is running
611 def kubeCtrlResult = salt.cmdRun(
612 pepperEnv, target, "kubectl get pod -n kube-system --selector=k8s-app=calico-kube-controllers"
613 )['return'][0].values()[0].toString()
614 if (kubeCtrlResult.contains("Running")) {
615 kubeCtrlRunning = true
616 }
617
618 // It's safe to enable Calico policy any time, but it may be unsafe to disable it.
619 // So, no need to disable Calico policy for v3.x if it's not in use currently.
620 // But if Calico policy is in use already, it should be enabled after upgrade as well.
621
622 // check for consistency
623 if ((cniPolicy != addonsPolicy) || (addonsPolicy != kubeCtrlRunning)) {
624 caution = "ATTENTION. Calico policy setting cannot be determined reliably (enabled in CNI config: ${cniPolicy}, " +
625 "presence in k8s addons: ${addonsPolicy}, kube-controllers is running: ${kubeCtrlRunning})."
626 currentBuild.description += "<br><b>${caution}</b><br><br>"
627 common.warningMsg(caution)
628 } else {
629 common.infoMsg("Current Calico policy state is detected as: ${cniPolicy}")
630 if (cniPolicy) {
631 // Calico policy is in use. Check policy setting for v3.x.
632 common.infoMsg("Calico policy is in use. It should be enabled for v3.x as well.")
633 def saltPolicyResult = salt.getPillar(
634 pepperEnv, target, "kubernetes:pool:network:calico:policy"
635 )["return"][0].values()[0].toString()
636
637 common.infoMsg("kubernetes.pool.network.calico.policy: ${saltPolicyResult}")
638 if (saltPolicyResult.toLowerCase().contains("true")) {
639 common.infoMsg("Calico policy setting for v3.x is detected as: true")
640 } else {
641 caution = "ATTENTION. Currently, Calico is running with policy switched on. " +
642 "Calico policy setting for v3.x is not set to true. " +
643 "After upgrade is completed, Calico policy will be switched off. " +
644 "You will need to switch it on manually if required."
645 currentBuild.description += "<br><b>${caution}</b><br><br>"
646 common.warningMsg(caution)
647 }
648 }
649 }
650
651 if (addonsPolicy) {
652 // Remove v2.6.x policy-related addons on masters to not interfere with v3.x kube-controllers
653 salt.cmdRun(pepperEnv, CTL_TARGET, "rm -rf /etc/kubernetes/addons/calico_policy")
654 }
655 }
656}
657
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400658timeout(time: 12, unit: 'HOURS') {
659 node() {
660 try {
661
662 stage("Setup virtualenv for Pepper") {
663 python.setupPepperVirtualenv(pepperEnv, SALT_MASTER_URL, SALT_MASTER_CREDENTIALS)
664 }
665
Victor Ryzhenkinc2024132019-01-23 05:39:34 +0400666 def ctl_node = salt.getMinionsSorted(pepperEnv, CTL_TARGET)[0]
667 def daemonsetMap = buildDaemonsetMap(pepperEnv, ctl_node)
668
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400669 if (CONFORMANCE_RUN_BEFORE.toBoolean()) {
Victor Ryzhenkin7e28c152019-03-28 02:23:42 +0400670 stage("Perform conformance run before upgrade") {
671 def target = CTL_TARGET
672 def mcp_repo = ARTIFACTORY_URL
673 def k8s_api = TEST_K8S_API_SERVER
674 firstTarget = salt.getFirstMinion(pepperEnv, target)
675 def containerd_enabled = containerDenabled(pepperEnv, firstTarget)
676 def containerd_installed = containerDinstalled(pepperEnv, firstTarget)
677 def conformance_pod_ready = conformancePodDefExists(pepperEnv, firstTarget)
678 if (containerd_enabled && containerd_installed && conformance_pod_ready) {
679 def config = ['master': pepperEnv,
680 'target': firstTarget,
681 'junitResults': false,
682 'autodetect': true]
683 test.executeConformance(config)
684 } else {
685 executeConformance(pepperEnv, firstTarget, k8s_api, mcp_repo)
686 }
Victor Ryzhenkin723bd062018-12-11 17:09:06 +0400687 }
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400688 }
689
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400690 if ((common.validInputParam('KUBERNETES_HYPERKUBE_IMAGE')) && (common.validInputParam('KUBERNETES_PAUSE_IMAGE'))) {
691 overrideKubernetesImage(pepperEnv)
692 }
693
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200694 if ((common.validInputParam('KUBERNETES_CALICO_IMAGE'))
Victor Ryzhenkin3401ee62019-01-18 06:34:26 +0400695 && (common.validInputParam('KUBERNETES_CALICO_CALICOCTL_SOURCE'))
696 && (common.validInputParam('KUBERNETES_CALICO_CALICOCTL_SOURCE_HASH'))
697 && (common.validInputParam('KUBERNETES_CALICO_CNI_SOURCE'))
698 && (common.validInputParam('KUBERNETES_CALICO_CNI_SOURCE_HASH'))
699 && (common.validInputParam('KUBERNETES_CALICO_BIRDCL_SOURCE'))
700 && (common.validInputParam('KUBERNETES_CALICO_BIRDCL_SOURCE_HASH'))
701 && (common.validInputParam('KUBERNETES_CALICO_CNI_IPAM_SOURCE'))
702 && (common.validInputParam('KUBERNETES_CALICO_CNI_IPAM_SOURCE_HASH'))
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200703 && (common.validInputParam('KUBERNETES_CALICO_KUBE_CONTROLLERS_IMAGE'))
704 ) {
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200705 overrideCalicoImages(pepperEnv)
706 }
707
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400708 /*
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200709 * Execute Calico upgrade if needed (only for v2 to v3 upgrade).
710 * This part causes workloads operations downtime.
711 * It is only required for Calico v2.x to v3.x upgrade when etcd is in use for Calico
712 * as Calico etcd schema has different formats for Calico v2.x and Calico v3.x.
713 */
714 if (UPGRADE_CALICO_V2_TO_V3.toBoolean()) {
Aleksei Kasatkin1f4f5ba2018-11-20 18:30:36 +0100715 // get ETCD_ENDPOINTS in use by Calico
716 def ep_str = salt.cmdRun(pepperEnv, ctl_node, "cat /etc/calico/calicoctl.cfg | grep etcdEndpoints")['return'][0].values()[0]
717 ETCD_ENDPOINTS = ep_str.split("\n")[0].tokenize(' ')[1]
718 print("ETCD_ENDPOINTS in use by Calico: '${ETCD_ENDPOINTS}'")
719
720 // download calico-upgrade utility
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200721 downloadCalicoUpgrader(pepperEnv, ctl_node)
Aleksei Kasatkin1f4f5ba2018-11-20 18:30:36 +0100722
723 // check the possibility of upgrading of Calico
724 checkCalicoUpgradePossibility(pepperEnv, ctl_node)
725
Aleksei Kasatkin9ce11842018-11-23 14:27:33 +0100726 // check and adjust Calico policy setting
727 checkCalicoPolicySetting(pepperEnv, ctl_node)
728
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200729 // this sequence implies workloads operations downtime
730 startCalicoUpgrade(pepperEnv, ctl_node)
Aleksei Kasatkind9d682e2018-12-12 14:51:59 +0100731 performCalicoConfigurationUpdateAndServicesRestart(pepperEnv, POOL, ctl_node)
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200732 completeCalicoUpgrade(pepperEnv, ctl_node)
Aleksei Kasatkin5ccea272018-12-06 17:34:58 +0100733 // no downtime is expected after this point
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200734 }
735
736 /*
ashestakov138b0a52019-01-11 11:34:54 +0200737 * Execute etcd update
738 */
739 if ((common.validInputParam('KUBERNETES_ETCD_SOURCE')) && (common.validInputParam('KUBERNETES_ETCD_SOURCE_HASH'))) {
740 overrideEtcdSource(pepperEnv)
741 }
Victor Ryzhenkin2c77cda2019-01-24 13:52:21 +0400742 def targetHostsEtcd = salt.getMinionsSorted(pepperEnv, "I@etcd:server")
743 for (t in targetHostsEtcd) {
ashestakov138b0a52019-01-11 11:34:54 +0200744 performEtcdUpdateAndServicesRestart(pepperEnv, t)
745 }
746
747 /*
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200748 * Execute k8s update
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400749 */
750 if (updates.contains("ctl")) {
751 def target = CTL_TARGET
752
753 if (PER_NODE.toBoolean()) {
754 def targetHosts = salt.getMinionsSorted(pepperEnv, target)
755
756 for (t in targetHosts) {
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400757 if (SIMPLE_UPGRADE.toBoolean()) {
758 performKubernetesControlUpdate(pepperEnv, t)
759 } else {
760 cordonNode(pepperEnv, t)
761 drainNode(pepperEnv, t)
762 regenerateCerts(pepperEnv, t)
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400763 performKubernetesControlUpdate(pepperEnv, t)
Victor Ryzhenkinfd9677f2018-10-16 16:14:40 +0400764 updateAddonManager(pepperEnv, t)
Victor Ryzhenkinc2024132019-01-23 05:39:34 +0400765 if (daemonsetMap) {
766 purgeDaemonsetPods(pepperEnv, t, daemonsetMap)
767 rebootKubernetesNode(pepperEnv, t)
768 }
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400769 uncordonNode(pepperEnv, t)
770 }
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400771 }
772 } else {
773 performKubernetesControlUpdate(pepperEnv, target)
774 }
Victor Ryzhenkinfd9677f2018-10-16 16:14:40 +0400775 if (!SIMPLE_UPGRADE.toBoolean()) {
Aleksei Kasatkinff9d5b52018-10-26 11:47:46 +0200776 // Addons upgrade should be performed after all nodes will be upgraded
Victor Ryzhenkinfd9677f2018-10-16 16:14:40 +0400777 updateAddons(pepperEnv, target)
778 // Wait for 90 sec for addons reconciling
779 sleep(90)
780 }
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400781 }
782
783 if (updates.contains("cmp")) {
784 def target = CMP_TARGET
785
786 if (PER_NODE.toBoolean()) {
787 def targetHosts = salt.getMinionsSorted(pepperEnv, target)
788
789 for (t in targetHosts) {
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400790 if (SIMPLE_UPGRADE.toBoolean()) {
791 performKubernetesComputeUpdate(pepperEnv, t)
792 } else {
793 cordonNode(pepperEnv, t)
794 drainNode(pepperEnv, t)
795 regenerateCerts(pepperEnv, t)
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400796 performKubernetesComputeUpdate(pepperEnv, t)
Victor Ryzhenkinc2024132019-01-23 05:39:34 +0400797 if (daemonsetMap) {
798 purgeDaemonsetPods(pepperEnv, t, daemonsetMap)
799 rebootKubernetesNode(pepperEnv, t)
800 }
Victor Ryzhenkin42e4b382018-09-11 17:57:56 +0400801 uncordonNode(pepperEnv, t)
802 }
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400803 }
804 } else {
805 performKubernetesComputeUpdate(pepperEnv, target)
806 }
807 }
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400808
Aleksei Kasatkin5ccea272018-12-06 17:34:58 +0100809 if (calicoEnabled(pepperEnv, ctl_node)) {
810 checkCalicoClusterState(pepperEnv, POOL)
811 }
Victor Ryzhenkin3401ee62019-01-18 06:34:26 +0400812 printVersionInfo(pepperEnv, ctl_node)
Aleksei Kasatkin5ccea272018-12-06 17:34:58 +0100813
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400814 if (CONFORMANCE_RUN_AFTER.toBoolean()) {
Victor Ryzhenkin7e28c152019-03-28 02:23:42 +0400815 stage("Perform conformance run after upgrade") {
816 def target = CTL_TARGET
817 def mcp_repo = ARTIFACTORY_URL
818 def k8s_api = TEST_K8S_API_SERVER
819 firstTarget = salt.getFirstMinion(pepperEnv, target)
820 def containerd_enabled = containerDenabled(pepperEnv, firstTarget)
821 def containerd_installed = containerDinstalled(pepperEnv, firstTarget)
822 def conformance_pod_ready = conformancePodDefExists(pepperEnv, firstTarget)
823 if (containerd_enabled && containerd_installed && conformance_pod_ready) {
824 def config = ['master': pepperEnv,
825 'target': firstTarget,
826 'junitResults': false,
827 'autodetect': true]
828 test.executeConformance(config)
829 } else {
830 executeConformance(pepperEnv, firstTarget, k8s_api, mcp_repo)
831 }
Victor Ryzhenkin723bd062018-12-11 17:09:06 +0400832 }
Victor Ryzhenkinae22a5a2018-10-12 15:52:27 +0400833 }
Victor Ryzhenkinef34a022018-06-22 19:36:13 +0400834 } catch (Throwable e) {
835 // If there was an error or exception thrown, the build failed
836 currentBuild.result = "FAILURE"
837 currentBuild.description = currentBuild.description ? e.message + " " + currentBuild.description : e.message
838 throw e
839 }
840 }
Victor Ryzhenkinc2024132019-01-23 05:39:34 +0400841}