OLD | NEW |
1 // Copyright 2016 The LUCI Authors. All rights reserved. | 1 // Copyright 2016 The LUCI Authors. All rights reserved. |
2 // Use of this source code is governed under the Apache License, Version 2.0 | 2 // Use of this source code is governed under the Apache License, Version 2.0 |
3 // that can be found in the LICENSE file. | 3 // that can be found in the LICENSE file. |
4 | 4 |
5 package main | 5 package main |
6 | 6 |
7 import ( | 7 import ( |
8 "fmt" | 8 "fmt" |
9 "sort" | 9 "sort" |
10 "strconv" | 10 "strconv" |
(...skipping 101 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
112 } | 112 } |
113 | 113 |
114 func (d *containerEngineDeployment) stage(w *work, root *managedfs.Dir, params *
deployParams) error { | 114 func (d *containerEngineDeployment) stage(w *work, root *managedfs.Dir, params *
deployParams) error { |
115 d.ignoreCurrentVersion = params.ignoreCurrentVersion | 115 d.ignoreCurrentVersion = params.ignoreCurrentVersion |
116 | 116 |
117 // Build a common timestamp suffix for our Docker images. | 117 // Build a common timestamp suffix for our Docker images. |
118 d.timestampSuffix = strconv.FormatInt(clock.Now(w).Unix(), 10) | 118 d.timestampSuffix = strconv.FormatInt(clock.Now(w).Unix(), 10) |
119 | 119 |
120 podRoot, err := root.EnsureDirectory("pods") | 120 podRoot, err := root.EnsureDirectory("pods") |
121 if err != nil { | 121 if err != nil { |
122 » » return errors.Annotate(err).Reason("failed to create pods direct
ory").Err() | 122 » » return errors.Annotate(err, "failed to create pods directory").E
rr() |
123 } | 123 } |
124 | 124 |
125 // Stage in parallel. We will stage all pods before we stage any contain
ers, | 125 // Stage in parallel. We will stage all pods before we stage any contain
ers, |
126 // as container staging requires some pod staging values to be populated
. | 126 // as container staging requires some pod staging values to be populated
. |
127 err = w.RunMulti(func(workC chan<- func() error) { | 127 err = w.RunMulti(func(workC chan<- func() error) { |
128 // Check and get all Kubernetes contexts in series. | 128 // Check and get all Kubernetes contexts in series. |
129 // | 129 // |
130 // These all share the same Kubernetes configuration file, so we
don't want | 130 // These all share the same Kubernetes configuration file, so we
don't want |
131 // them to stomp each other if we did them in parallel. | 131 // them to stomp each other if we did them in parallel. |
132 workC <- func() error { | 132 workC <- func() error { |
133 for _, name := range d.clusterNames { | 133 for _, name := range d.clusterNames { |
134 cluster := d.clusters[name] | 134 cluster := d.clusters[name] |
135 | 135 |
136 var err error | 136 var err error |
137 if cluster.kubeCtx, err = getContainerEngineKube
rnetesContext(w, cluster.cluster); err != nil { | 137 if cluster.kubeCtx, err = getContainerEngineKube
rnetesContext(w, cluster.cluster); err != nil { |
138 » » » » » return errors.Annotate(err).Reason("fail
ed to get Kubernetes context for %(cluster)q"). | 138 » » » » » return errors.Annotate(err, "failed to g
et Kubernetes context for %q", cluster.cluster.Name).Err() |
139 » » » » » » D("cluster", cluster.cluster.Nam
e).Err() | |
140 } | 139 } |
141 } | 140 } |
142 return nil | 141 return nil |
143 } | 142 } |
144 | 143 |
145 for _, pod := range d.pods { | 144 for _, pod := range d.pods { |
146 pod := pod | 145 pod := pod |
147 workC <- func() error { | 146 workC <- func() error { |
148 // Use the name of this Pod's Component for stag
ing directory. | 147 // Use the name of this Pod's Component for stag
ing directory. |
149 name := pod.pod.comp.comp.Name | 148 name := pod.pod.comp.comp.Name |
150 podDir, err := podRoot.EnsureDirectory(name) | 149 podDir, err := podRoot.EnsureDirectory(name) |
151 if err != nil { | 150 if err != nil { |
152 » » » » » return errors.Annotate(err).Reason("fail
ed to create pod directory for %(pod)q"). | 151 » » » » » return errors.Annotate(err, "failed to c
reate pod directory for %q", name).Err() |
153 » » » » » » D("pod", name).Err() | |
154 } | 152 } |
155 | 153 |
156 return pod.stage(w, podDir, params) | 154 return pod.stage(w, podDir, params) |
157 } | 155 } |
158 } | 156 } |
159 }) | 157 }) |
160 if err != nil { | 158 if err != nil { |
161 return err | 159 return err |
162 } | 160 } |
163 | 161 |
164 // Now that pods are deployed, deploy our clusters. | 162 // Now that pods are deployed, deploy our clusters. |
165 clusterRoot, err := root.EnsureDirectory("clusters") | 163 clusterRoot, err := root.EnsureDirectory("clusters") |
166 if err != nil { | 164 if err != nil { |
167 » » return errors.Annotate(err).Reason("failed to create clusters di
rectory").Err() | 165 » » return errors.Annotate(err, "failed to create clusters directory
").Err() |
168 } | 166 } |
169 | 167 |
170 return w.RunMulti(func(workC chan<- func() error) { | 168 return w.RunMulti(func(workC chan<- func() error) { |
171 // Stage each cluster and pod in parallel. | 169 // Stage each cluster and pod in parallel. |
172 for _, name := range d.clusterNames { | 170 for _, name := range d.clusterNames { |
173 cluster := d.clusters[name] | 171 cluster := d.clusters[name] |
174 | 172 |
175 workC <- func() error { | 173 workC <- func() error { |
176 clusterDir, err := clusterRoot.EnsureDirectory(c
luster.cluster.Name) | 174 clusterDir, err := clusterRoot.EnsureDirectory(c
luster.cluster.Name) |
177 if err != nil { | 175 if err != nil { |
178 » » » » » return errors.Annotate(err).Reason("fail
ed to create cluster directory for %(cluster)q"). | 176 » » » » » return errors.Annotate(err, "failed to c
reate cluster directory for %q", cluster.cluster.Name).Err() |
179 » » » » » » D("cluster", cluster.cluster.Nam
e).Err() | |
180 } | 177 } |
181 | 178 |
182 return cluster.stage(w, clusterDir) | 179 return cluster.stage(w, clusterDir) |
183 } | 180 } |
184 } | 181 } |
185 }) | 182 }) |
186 } | 183 } |
187 | 184 |
188 func (d *containerEngineDeployment) localBuild(w *work) error { | 185 func (d *containerEngineDeployment) localBuild(w *work) error { |
189 return w.RunMulti(func(workC chan<- func() error) { | 186 return w.RunMulti(func(workC chan<- func() error) { |
(...skipping 75 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
265 } | 262 } |
266 sort.Strings(c.scopes) | 263 sort.Strings(c.scopes) |
267 | 264 |
268 // Stage for each deploymend pod. | 265 // Stage for each deploymend pod. |
269 return w.RunMulti(func(workC chan<- func() error) { | 266 return w.RunMulti(func(workC chan<- func() error) { |
270 for _, bp := range c.pods { | 267 for _, bp := range c.pods { |
271 bp := bp | 268 bp := bp |
272 workC <- func() error { | 269 workC <- func() error { |
273 stageDir, err := root.EnsureDirectory(string(bp.
sp.pod.comp.comp.title)) | 270 stageDir, err := root.EnsureDirectory(string(bp.
sp.pod.comp.comp.title)) |
274 if err != nil { | 271 if err != nil { |
275 » » » » » return errors.Annotate(err).Reason("fail
ed to create staging directory").Err() | 272 » » » » » return errors.Annotate(err, "failed to c
reate staging directory").Err() |
276 } | 273 } |
277 | 274 |
278 return bp.stage(w, stageDir) | 275 return bp.stage(w, stageDir) |
279 } | 276 } |
280 } | 277 } |
281 }) | 278 }) |
282 } | 279 } |
283 | 280 |
284 func (c *containerEngineDeploymentCluster) commit(w *work) error { | 281 func (c *containerEngineDeploymentCluster) commit(w *work) error { |
285 // Push all pods in parallel. | 282 // Push all pods in parallel. |
(...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
320 // Generate our deployment YAML. | 317 // Generate our deployment YAML. |
321 depYAML := kubeBuildDeploymentYAML(bp.binding, bp.sp.deploymentName, bp.
sp.imageMap) | 318 depYAML := kubeBuildDeploymentYAML(bp.binding, bp.sp.deploymentName, bp.
sp.imageMap) |
322 depYAML.Metadata.addAnnotation(kubeManagedByKey, kubeManagedByMe) | 319 depYAML.Metadata.addAnnotation(kubeManagedByKey, kubeManagedByMe) |
323 depYAML.Metadata.addAnnotation(kubeVersionKey, bp.sp.version.String()) | 320 depYAML.Metadata.addAnnotation(kubeVersionKey, bp.sp.version.String()) |
324 depYAML.Metadata.addAnnotation(kubeSourceVersionKey, comp.source().Revis
ion) | 321 depYAML.Metadata.addAnnotation(kubeSourceVersionKey, comp.source().Revis
ion) |
325 depYAML.Spec.Template.Metadata.addLabel("luci/project", string(comp.comp
.proj.title)) | 322 depYAML.Spec.Template.Metadata.addLabel("luci/project", string(comp.comp
.proj.title)) |
326 depYAML.Spec.Template.Metadata.addLabel("luci/component", string(comp.co
mp.title)) | 323 depYAML.Spec.Template.Metadata.addLabel("luci/component", string(comp.co
mp.title)) |
327 | 324 |
328 deploymentYAML := root.File("deployment.yaml") | 325 deploymentYAML := root.File("deployment.yaml") |
329 if err := deploymentYAML.GenerateYAML(w, depYAML); err != nil { | 326 if err := deploymentYAML.GenerateYAML(w, depYAML); err != nil { |
330 » » return errors.Annotate(err).Reason("failed to generate deploymen
t YAML").Err() | 327 » » return errors.Annotate(err, "failed to generate deployment YAML"
).Err() |
331 } | 328 } |
332 bp.deploymentYAMLPath = deploymentYAML.String() | 329 bp.deploymentYAMLPath = deploymentYAML.String() |
333 return nil | 330 return nil |
334 } | 331 } |
335 | 332 |
336 func (bp *containerEngineBoundPod) commit(w *work) error { | 333 func (bp *containerEngineBoundPod) commit(w *work) error { |
337 kubectl, err := bp.c.kubectl(w) | 334 kubectl, err := bp.c.kubectl(w) |
338 if err != nil { | 335 if err != nil { |
339 » » return errors.Annotate(err).Err() | 336 » » return errors.Annotate(err, "").Err() |
340 } | 337 } |
341 | 338 |
342 // Get the current deployment status for this pod. | 339 // Get the current deployment status for this pod. |
343 var ( | 340 var ( |
344 kd kubeDeployment | 341 kd kubeDeployment |
345 currentVersion string | 342 currentVersion string |
346 ) | 343 ) |
347 switch err := kubectl.getResource(w, fmt.Sprintf("deployments/%s", bp.sp
.deploymentName), &kd); err { | 344 switch err := kubectl.getResource(w, fmt.Sprintf("deployments/%s", bp.sp
.deploymentName), &kd); err { |
348 case nil: | 345 case nil: |
349 // Got deployment status. | 346 // Got deployment status. |
350 md := kd.Metadata | 347 md := kd.Metadata |
351 if md == nil { | 348 if md == nil { |
352 return errors.Reason("current deployment has no metadata
").Err() | 349 return errors.Reason("current deployment has no metadata
").Err() |
353 } | 350 } |
354 | 351 |
355 // Make sure the current deployment is managed by this tool. | 352 // Make sure the current deployment is managed by this tool. |
356 v, ok := md.Annotations[kubeManagedByKey].(string) | 353 v, ok := md.Annotations[kubeManagedByKey].(string) |
357 if !ok { | 354 if !ok { |
358 return errors.Reason("missing '" + kubeManagedByKey + "'
annotation").Err() | 355 return errors.Reason("missing '" + kubeManagedByKey + "'
annotation").Err() |
359 } | 356 } |
360 if v != kubeManagedByMe { | 357 if v != kubeManagedByMe { |
361 log.Fields{ | 358 log.Fields{ |
362 "managedBy": v, | 359 "managedBy": v, |
363 "deployment": bp.sp.deploymentName, | 360 "deployment": bp.sp.deploymentName, |
364 }.Errorf(w, "Current deployment is not managed.") | 361 }.Errorf(w, "Current deployment is not managed.") |
365 » » » return errors.Reason("unknown manager %(managedBy)q").D(
"managedBy", v).Err() | 362 » » » return errors.Reason("unknown manager %q", v).Err() |
366 } | 363 } |
367 | 364 |
368 // Is the current deployment tagged at the current version? | 365 // Is the current deployment tagged at the current version? |
369 currentVersion, ok = md.Annotations[kubeVersionKey].(string) | 366 currentVersion, ok = md.Annotations[kubeVersionKey].(string) |
370 if !ok { | 367 if !ok { |
371 return errors.Reason("missing '" + kubeVersionKey + "' a
nnotation").Err() | 368 return errors.Reason("missing '" + kubeVersionKey + "' a
nnotation").Err() |
372 } | 369 } |
373 cloudVersion, err := parseCloudProjectVersion(bp.c.gke.project.V
ersionScheme, currentVersion) | 370 cloudVersion, err := parseCloudProjectVersion(bp.c.gke.project.V
ersionScheme, currentVersion) |
374 switch { | 371 switch { |
375 case err != nil: | 372 case err != nil: |
376 if !bp.c.gke.ignoreCurrentVersion { | 373 if !bp.c.gke.ignoreCurrentVersion { |
377 » » » » return errors.Annotate(err).Reason("failed to pa
rse current version %(version)q"). | 374 » » » » return errors.Annotate(err, "failed to parse cur
rent version %q", currentVersion).Err() |
378 » » » » » D("version", currentVersion).Err() | |
379 } | 375 } |
380 | 376 |
381 log.Fields{ | 377 log.Fields{ |
382 log.ErrorKey: err, | 378 log.ErrorKey: err, |
383 "currentVersion": currentVersion, | 379 "currentVersion": currentVersion, |
384 }.Warningf(w, "Could not parse current version, but conf
igured to ignore this failure.") | 380 }.Warningf(w, "Could not parse current version, but conf
igured to ignore this failure.") |
385 | 381 |
386 case cloudVersion.String() == bp.sp.version.String(): | 382 case cloudVersion.String() == bp.sp.version.String(): |
387 if !bp.c.gke.ignoreCurrentVersion { | 383 if !bp.c.gke.ignoreCurrentVersion { |
388 log.Fields{ | 384 log.Fields{ |
(...skipping 10 matching lines...) Expand all Loading... |
399 // fallthrough to "kubectl apply" the new configuration. | 395 // fallthrough to "kubectl apply" the new configuration. |
400 fallthrough | 396 fallthrough |
401 | 397 |
402 case errKubeResourceNotFound: | 398 case errKubeResourceNotFound: |
403 // No current deployment, create a new one. | 399 // No current deployment, create a new one. |
404 log.Fields{ | 400 log.Fields{ |
405 "currentVersion": currentVersion, | 401 "currentVersion": currentVersion, |
406 "deployVersion": bp.sp.version, | 402 "deployVersion": bp.sp.version, |
407 }.Infof(w, "Deploying new pod configuration.") | 403 }.Infof(w, "Deploying new pod configuration.") |
408 if err := kubectl.exec("apply", "-f", bp.deploymentYAMLPath).che
ck(w); err != nil { | 404 if err := kubectl.exec("apply", "-f", bp.deploymentYAMLPath).che
ck(w); err != nil { |
409 » » » return errors.Annotate(err).Reason("failed to create new
deployment configuration").Err() | 405 » » » return errors.Annotate(err, "failed to create new deploy
ment configuration").Err() |
410 } | 406 } |
411 return nil | 407 return nil |
412 | 408 |
413 default: | 409 default: |
414 » » return errors.Annotate(err).Reason("failed to get status for dep
loyment %(deployment)q"). | 410 » » return errors.Annotate(err, "failed to get status for deployment
%q", bp.sp.deploymentName).Err() |
415 » » » D("deployment", bp.sp.deploymentName).Err() | |
416 } | 411 } |
417 } | 412 } |
418 | 413 |
419 // stagedGKEPod is staging information for a Google Container Engine deployed | 414 // stagedGKEPod is staging information for a Google Container Engine deployed |
420 // Kubernetes Pod. | 415 // Kubernetes Pod. |
421 type stagedGKEPod struct { | 416 type stagedGKEPod struct { |
422 *deploy.ContainerEnginePod | 417 *deploy.ContainerEnginePod |
423 | 418 |
424 // gke is the container engine deployment that owns this pod. | 419 // gke is the container engine deployment that owns this pod. |
425 gke *containerEngineDeployment | 420 gke *containerEngineDeployment |
(...skipping 16 matching lines...) Expand all Loading... |
442 func (sp *stagedGKEPod) cloudProject() *layoutDeploymentCloudProject { | 437 func (sp *stagedGKEPod) cloudProject() *layoutDeploymentCloudProject { |
443 return sp.pod.comp.dep.cloudProject | 438 return sp.pod.comp.dep.cloudProject |
444 } | 439 } |
445 | 440 |
446 func (sp *stagedGKEPod) stage(w *work, root *managedfs.Dir, params *deployParams
) error { | 441 func (sp *stagedGKEPod) stage(w *work, root *managedfs.Dir, params *deployParams
) error { |
447 // Calculate the cloud project version for this pod. | 442 // Calculate the cloud project version for this pod. |
448 if sp.version = params.forceVersion; sp.version == nil { | 443 if sp.version = params.forceVersion; sp.version == nil { |
449 var err error | 444 var err error |
450 sp.version, err = makeCloudProjectVersion(sp.cloudProject(), sp.
pod.comp.source()) | 445 sp.version, err = makeCloudProjectVersion(sp.cloudProject(), sp.
pod.comp.source()) |
451 if err != nil { | 446 if err != nil { |
452 » » » return errors.Annotate(err).Reason("failed to get cloud
version").Err() | 447 » » » return errors.Annotate(err, "failed to get cloud version
").Err() |
453 } | 448 } |
454 } | 449 } |
455 | 450 |
456 comp := sp.pod.comp | 451 comp := sp.pod.comp |
457 sp.deploymentName = fmt.Sprintf("%s--%s", comp.comp.proj.title, comp.com
p.title) | 452 sp.deploymentName = fmt.Sprintf("%s--%s", comp.comp.proj.title, comp.com
p.title) |
458 | 453 |
459 sp.imageMap = make(map[string]string, len(sp.KubePod.Container)) | 454 sp.imageMap = make(map[string]string, len(sp.KubePod.Container)) |
460 sp.containers = make([]*stagedKubernetesContainer, len(sp.KubePod.Contai
ner)) | 455 sp.containers = make([]*stagedKubernetesContainer, len(sp.KubePod.Contai
ner)) |
461 for i, kc := range sp.KubePod.Container { | 456 for i, kc := range sp.KubePod.Container { |
462 skc := stagedKubernetesContainer{ | 457 skc := stagedKubernetesContainer{ |
(...skipping 14 matching lines...) Expand all Loading... |
477 if skc.needsGoPath() { | 472 if skc.needsGoPath() { |
478 needsGoPath = true | 473 needsGoPath = true |
479 break | 474 break |
480 } | 475 } |
481 } | 476 } |
482 if needsGoPath { | 477 if needsGoPath { |
483 // Build a GOPATH from our sources. | 478 // Build a GOPATH from our sources. |
484 // Construct a GOPATH for this module. | 479 // Construct a GOPATH for this module. |
485 goPath, err := root.EnsureDirectory("gopath") | 480 goPath, err := root.EnsureDirectory("gopath") |
486 if err != nil { | 481 if err != nil { |
487 » » » return errors.Annotate(err).Reason("failed to create GOP
ATH base").Err() | 482 » » » return errors.Annotate(err, "failed to create GOPATH bas
e").Err() |
488 } | 483 } |
489 if err := stageGoPath(w, comp, goPath); err != nil { | 484 if err := stageGoPath(w, comp, goPath); err != nil { |
490 » » » return errors.Annotate(err).Reason("failed to stage GOPA
TH").Err() | 485 » » » return errors.Annotate(err, "failed to stage GOPATH").Er
r() |
491 } | 486 } |
492 sp.goPath = []string{goPath.String()} | 487 sp.goPath = []string{goPath.String()} |
493 } | 488 } |
494 | 489 |
495 // Stage each of our containers. | 490 // Stage each of our containers. |
496 containersDir, err := root.EnsureDirectory("containers") | 491 containersDir, err := root.EnsureDirectory("containers") |
497 if err != nil { | 492 if err != nil { |
498 » » return errors.Annotate(err).Err() | 493 » » return errors.Annotate(err, "").Err() |
499 } | 494 } |
500 err = w.RunMulti(func(workC chan<- func() error) { | 495 err = w.RunMulti(func(workC chan<- func() error) { |
501 // Stage each component. | 496 // Stage each component. |
502 for _, skc := range sp.containers { | 497 for _, skc := range sp.containers { |
503 skc := skc | 498 skc := skc |
504 workC <- func() error { | 499 workC <- func() error { |
505 containerDir, err := containersDir.EnsureDirecto
ry(skc.Name) | 500 containerDir, err := containersDir.EnsureDirecto
ry(skc.Name) |
506 if err != nil { | 501 if err != nil { |
507 » » » » » return errors.Annotate(err).Err() | 502 » » » » » return errors.Annotate(err, "").Err() |
508 } | 503 } |
509 | 504 |
510 if err := skc.stage(w, containerDir); err != nil
{ | 505 if err := skc.stage(w, containerDir); err != nil
{ |
511 » » » » » return errors.Annotate(err).Reason("fail
ed to stage container %(container)q"). | 506 » » » » » return errors.Annotate(err, "failed to s
tage container %q", skc.Name).Err() |
512 » » » » » » D("container", skc.Name).Err() | |
513 } | 507 } |
514 return nil | 508 return nil |
515 } | 509 } |
516 } | 510 } |
517 }) | 511 }) |
518 if err != nil { | 512 if err != nil { |
519 return err | 513 return err |
520 } | 514 } |
521 | 515 |
522 if err := root.CleanUp(); err != nil { | 516 if err := root.CleanUp(); err != nil { |
523 » » return errors.Annotate(err).Reason("failed to cleanup staging ar
ea").Err() | 517 » » return errors.Annotate(err, "failed to cleanup staging area").Er
r() |
524 } | 518 } |
525 return nil | 519 return nil |
526 } | 520 } |
527 | 521 |
528 func (sp *stagedGKEPod) build(w *work) error { | 522 func (sp *stagedGKEPod) build(w *work) error { |
529 // Build any containers within this pod. | 523 // Build any containers within this pod. |
530 return w.RunMulti(func(workC chan<- func() error) { | 524 return w.RunMulti(func(workC chan<- func() error) { |
531 for _, cont := range sp.containers { | 525 for _, cont := range sp.containers { |
532 workC <- func() error { | 526 workC <- func() error { |
533 return cont.build(w) | 527 return cont.build(w) |
(...skipping 36 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
570 return true | 564 return true |
571 default: | 565 default: |
572 return false | 566 return false |
573 } | 567 } |
574 } | 568 } |
575 | 569 |
576 func (skc *stagedKubernetesContainer) stage(w *work, root *managedfs.Dir) error
{ | 570 func (skc *stagedKubernetesContainer) stage(w *work, root *managedfs.Dir) error
{ |
577 // Build each Component. | 571 // Build each Component. |
578 buildDir, err := root.EnsureDirectory("build") | 572 buildDir, err := root.EnsureDirectory("build") |
579 if err != nil { | 573 if err != nil { |
580 » » return errors.Annotate(err).Reason("failed to create build direc
tory").Err() | 574 » » return errors.Annotate(err, "failed to create build directory").
Err() |
581 } | 575 } |
582 if err := buildComponent(w, skc.pod.pod.comp, buildDir); err != nil { | 576 if err := buildComponent(w, skc.pod.pod.comp, buildDir); err != nil { |
583 » » return errors.Annotate(err).Reason("failed to build component").
Err() | 577 » » return errors.Annotate(err, "failed to build component").Err() |
584 } | 578 } |
585 | 579 |
586 switch skc.Type { | 580 switch skc.Type { |
587 case deploy.KubernetesPod_Container_GO: | 581 case deploy.KubernetesPod_Container_GO: |
588 // Specify how we are to be built. | 582 // Specify how we are to be built. |
589 skc.buildFn = func(w *work) error { | 583 skc.buildFn = func(w *work) error { |
590 path, err := skc.pod.pod.comp.buildPath(skc.GetBuild()) | 584 path, err := skc.pod.pod.comp.buildPath(skc.GetBuild()) |
591 if err != nil { | 585 if err != nil { |
592 » » » » return errors.Annotate(err).Err() | 586 » » » » return errors.Annotate(err, "").Err() |
593 } | 587 } |
594 return skc.buildGo(w, path) | 588 return skc.buildGo(w, path) |
595 } | 589 } |
596 | 590 |
597 default: | 591 default: |
598 » » return errors.Reason("unknown Kubernetes pod type %(type)T").D("
type", skc.Type).Err() | 592 » » return errors.Reason("unknown Kubernetes pod type %T", skc.Type)
.Err() |
599 } | 593 } |
600 return nil | 594 return nil |
601 } | 595 } |
602 | 596 |
603 func (skc *stagedKubernetesContainer) build(w *work) error { | 597 func (skc *stagedKubernetesContainer) build(w *work) error { |
604 if f := skc.buildFn; f != nil { | 598 if f := skc.buildFn; f != nil { |
605 return f(w) | 599 return f(w) |
606 } | 600 } |
607 return nil | 601 return nil |
608 } | 602 } |
609 | 603 |
610 // build builds the image associated with this container. | 604 // build builds the image associated with this container. |
611 func (skc *stagedKubernetesContainer) buildGo(w *work, entryPath string) error { | 605 func (skc *stagedKubernetesContainer) buildGo(w *work, entryPath string) error { |
612 gcloud, err := w.tools.gcloud(skc.pod.cloudProject().Name) | 606 gcloud, err := w.tools.gcloud(skc.pod.cloudProject().Name) |
613 if err != nil { | 607 if err != nil { |
614 » » return errors.Annotate(err).Reason("could not get gcloud tool").
Err() | 608 » » return errors.Annotate(err, "could not get gcloud tool").Err() |
615 } | 609 } |
616 | 610 |
617 // Use "aedeploy" to gather GOPATH and build against our root. | 611 // Use "aedeploy" to gather GOPATH and build against our root. |
618 aedeploy, err := w.tools.aedeploy(skc.pod.goPath) | 612 aedeploy, err := w.tools.aedeploy(skc.pod.goPath) |
619 if err != nil { | 613 if err != nil { |
620 » » return errors.Annotate(err).Err() | 614 » » return errors.Annotate(err, "").Err() |
621 } | 615 } |
622 | 616 |
623 x := gcloud.exec("docker", "--", "build", "-t", skc.image, ".") | 617 x := gcloud.exec("docker", "--", "build", "-t", skc.image, ".") |
624 return aedeploy.bootstrap(x).cwd(entryPath).check(w) | 618 return aedeploy.bootstrap(x).cwd(entryPath).check(w) |
625 } | 619 } |
626 | 620 |
627 func (skc *stagedKubernetesContainer) push(w *work) error { | 621 func (skc *stagedKubernetesContainer) push(w *work) error { |
628 gcloud, err := w.tools.gcloud(skc.pod.cloudProject().Name) | 622 gcloud, err := w.tools.gcloud(skc.pod.cloudProject().Name) |
629 if err != nil { | 623 if err != nil { |
630 » » return errors.Annotate(err).Reason("could not get gcloud tool").
Err() | 624 » » return errors.Annotate(err, "could not get gcloud tool").Err() |
631 } | 625 } |
632 | 626 |
633 if err := gcloud.exec("docker", "--", "push", skc.image).check(w); err !
= nil { | 627 if err := gcloud.exec("docker", "--", "push", skc.image).check(w); err !
= nil { |
634 » » return errors.Annotate(err).Reason("failed to push Docker image
%(image)q"). | 628 » » return errors.Annotate(err, "failed to push Docker image %q", sk
c.image).Err() |
635 » » » D("image", skc.image).Err() | |
636 } | 629 } |
637 return nil | 630 return nil |
638 } | 631 } |
639 | 632 |
640 func getContainerEngineKubernetesContext(w *work, cluster *layoutDeploymentGKECl
uster) ( | 633 func getContainerEngineKubernetesContext(w *work, cluster *layoutDeploymentGKECl
uster) ( |
641 string, error) { | 634 string, error) { |
642 // Generate our Kubernetes context name. This is derived from the Google | 635 // Generate our Kubernetes context name. This is derived from the Google |
643 // Container Engine cluster parameters. | 636 // Container Engine cluster parameters. |
644 kubeCtx := fmt.Sprintf("gke_%s_%s_%s", cluster.cloudProject.Name, cluste
r.Zone, cluster.Name) | 637 kubeCtx := fmt.Sprintf("gke_%s_%s_%s", cluster.cloudProject.Name, cluste
r.Zone, cluster.Name) |
645 | 638 |
646 kubectl, err := w.tools.kubectl(kubeCtx) | 639 kubectl, err := w.tools.kubectl(kubeCtx) |
647 if err != nil { | 640 if err != nil { |
648 » » return "", errors.Annotate(err).Err() | 641 » » return "", errors.Annotate(err, "").Err() |
649 } | 642 } |
650 | 643 |
651 // Check if the context is already installed in our Kubernetes configura
tion. | 644 // Check if the context is already installed in our Kubernetes configura
tion. |
652 switch has, err := kubectl.hasContext(w); { | 645 switch has, err := kubectl.hasContext(w); { |
653 case err != nil: | 646 case err != nil: |
654 » » return "", errors.Annotate(err).Reason("failed to check for Kube
rnetes context").Err() | 647 » » return "", errors.Annotate(err, "failed to check for Kubernetes
context").Err() |
655 | 648 |
656 case !has: | 649 case !has: |
657 gcloud, err := w.tools.gcloud(cluster.cloudProject.Name) | 650 gcloud, err := w.tools.gcloud(cluster.cloudProject.Name) |
658 if err != nil { | 651 if err != nil { |
659 » » » return "", errors.Annotate(err).Err() | 652 » » » return "", errors.Annotate(err, "").Err() |
660 } | 653 } |
661 | 654 |
662 // The context isn't cached, we will fetch it via: | 655 // The context isn't cached, we will fetch it via: |
663 // $ gcloud container clusters get-credentials | 656 // $ gcloud container clusters get-credentials |
664 x := gcloud.exec( | 657 x := gcloud.exec( |
665 "container", "clusters", | 658 "container", "clusters", |
666 "get-credentials", cluster.Name, | 659 "get-credentials", cluster.Name, |
667 "--zone", cluster.Zone) | 660 "--zone", cluster.Zone) |
668 if err := x.check(w); err != nil { | 661 if err := x.check(w); err != nil { |
669 » » » return "", errors.Annotate(err).Reason("failed to get cl
uster credentials").Err() | 662 » » » return "", errors.Annotate(err, "failed to get cluster c
redentials").Err() |
670 } | 663 } |
671 switch has, err = kubectl.hasContext(w); { | 664 switch has, err = kubectl.hasContext(w); { |
672 case err != nil: | 665 case err != nil: |
673 » » » return "", errors.Annotate(err).Reason("failed to confir
m Kubernetes context").Err() | 666 » » » return "", errors.Annotate(err, "failed to confirm Kuber
netes context").Err() |
674 case !has: | 667 case !has: |
675 » » » return "", errors.Reason("context %(context)q missing af
ter fetching credentials").D("context", kubeCtx).Err() | 668 » » » return "", errors.Reason("context %q missing after fetch
ing credentials", kubeCtx).Err() |
676 } | 669 } |
677 } | 670 } |
678 return kubeCtx, nil | 671 return kubeCtx, nil |
679 } | 672 } |
OLD | NEW |