func (d *Dispatcher) createAppliance(conf *config.VirtualContainerHostConfigSpec, settings *data.InstallerData) error { defer trace.End(trace.Begin("")) log.Infof("Creating appliance on target") spec, err := d.createApplianceSpec(conf, settings) if err != nil { log.Errorf("Unable to create appliance spec: %s", err) return err } var info *types.TaskInfo // create appliance VM if d.isVC && d.vchVapp != nil { info, err = tasks.WaitForResult(d.ctx, func(ctx context.Context) (tasks.ResultWaiter, error) { return d.vchVapp.CreateChildVM_Task(ctx, *spec, d.session.Host) }) } else { // if vapp is not created, fall back to create VM under default resource pool info, err = tasks.WaitForResult(d.ctx, func(ctx context.Context) (tasks.ResultWaiter, error) { return d.session.Folders(ctx).VmFolder.CreateVM(ctx, *spec, d.vchPool, d.session.Host) }) } if err != nil { log.Errorf("Unable to create appliance VM: %s", err) return err } if info.Error != nil || info.State != types.TaskInfoStateSuccess { log.Errorf("Create appliance reported: %s", info.Error.LocalizedMessage) } // get VM reference and save it moref := info.Result.(types.ManagedObjectReference) conf.SetMoref(&moref) obj, err := d.session.Finder.ObjectReference(d.ctx, moref) if err != nil { log.Errorf("Failed to reacquire reference to appliance VM after creation: %s", err) return err } gvm, ok := obj.(*object.VirtualMachine) if !ok { return fmt.Errorf("Required reference after appliance creation was not for a VM: %T", obj) } vm2 := vm.NewVirtualMachineFromVM(d.ctx, d.session, gvm) // update the displayname to the actual folder name used if d.vmPathName, err = vm2.FolderName(d.ctx); err != nil { log.Errorf("Failed to get canonical name for appliance: %s", err) return err } log.Debugf("vm folder name: %q", d.vmPathName) log.Debugf("vm inventory path: %q", vm2.InventoryPath) // create an extension to register the appliance as if err = d.GenerateExtensionName(conf, vm2); err != nil { return errors.Errorf("Could not generate extension name during appliance creation due to error: %s", err) } settings.Extension = types.Extension{ Description: &types.Description{ Label: "VIC", Summary: "vSphere Integrated Containers Virtual Container Host", }, Company: "VMware, Inc.", Version: "0.0", Key: conf.ExtensionName, } conf.AddComponent("vicadmin", &executor.SessionConfig{ User: "******", Group: "vicadmin", Cmd: executor.Cmd{ Path: "/sbin/vicadmin", Args: []string{ "/sbin/vicadmin", "-docker-host=unix:///var/run/docker.sock", // FIXME: hack during config migration "-insecure", "-ds=" + conf.ImageStores[0].Host, "-cluster=" + settings.ClusterPath, "-pool=" + settings.ResourcePoolPath, "-vm-path=" + vm2.InventoryPath, }, Env: []string{ "PATH=/sbin:/bin", }, Dir: "/home/vicadmin", }, Restart: true, }, ) if conf.HostCertificate != nil { d.VICAdminProto = "https" d.DockerPort = fmt.Sprintf("%d", opts.DefaultTLSHTTPPort) } else { d.VICAdminProto = "http" d.DockerPort = fmt.Sprintf("%d", opts.DefaultHTTPPort) } conf.AddComponent("docker-personality", &executor.SessionConfig{ Cmd: executor.Cmd{ Path: "/sbin/docker-engine-server", Args: []string{ "/sbin/docker-engine-server", //FIXME: hack during config migration "-serveraddr=0.0.0.0", "-port=" + d.DockerPort, "-port-layer-port=8080", }, Env: []string{ "PATH=/sbin", }, }, Restart: true, }, ) conf.AddComponent("port-layer", &executor.SessionConfig{ Cmd: executor.Cmd{ Path: "/sbin/port-layer-server", Args: []string{ "/sbin/port-layer-server", //FIXME: hack during config migration "--host=localhost", "--port=8080", "--insecure", "--sdk=" + conf.Target.String(), "--datacenter=" + settings.DatacenterName, "--cluster=" + settings.ClusterPath, "--pool=" + settings.ResourcePoolPath, "--datastore=" + conf.ImageStores[0].Host, "--vch=" + conf.ExecutorConfig.Name, }, }, Restart: true, }, ) conf.BootstrapImagePath = fmt.Sprintf("[%s] %s/%s", conf.ImageStores[0].Host, d.vmPathName, settings.BootstrapISO) spec, err = d.reconfigureApplianceSpec(vm2, conf, settings) if err != nil { log.Errorf("Error while getting appliance reconfig spec: %s", err) return err } // reconfig info, err = tasks.WaitForResult(d.ctx, func(ctx context.Context) (tasks.ResultWaiter, error) { return vm2.Reconfigure(ctx, *spec) }) if err != nil { log.Errorf("Error while setting component parameters to appliance: %s", err) return err } if info.State != types.TaskInfoStateSuccess { log.Errorf("Setting parameters to appliance reported: %s", info.Error.LocalizedMessage) return err } d.appliance = vm2 return nil }
// Upgrade will try to upgrade vch appliance to new version. If failed will try to roll back to original status. func (d *Dispatcher) Upgrade(vch *vm.VirtualMachine, conf *config.VirtualContainerHostConfigSpec, settings *data.InstallerData) (err error) { defer trace.End(trace.Begin(conf.Name)) d.appliance = vch // update the displayname to the actual folder name used if d.vmPathName, err = d.appliance.FolderName(d.ctx); err != nil { log.Errorf("Failed to get canonical name for appliance: %s", err) return err } ds, err := d.session.Finder.Datastore(d.ctx, conf.ImageStores[0].Host) if err != nil { err = errors.Errorf("Failed to find image datastore %q", conf.ImageStores[0].Host) return err } d.session.Datastore = ds if !conf.HostCertificate.IsNil() { d.VICAdminProto = "https" d.DockerPort = fmt.Sprintf("%d", opts.DefaultTLSHTTPPort) } else { d.VICAdminProto = "http" d.DockerPort = fmt.Sprintf("%d", opts.DefaultHTTPPort) } if err = d.uploadImages(settings.ImageFiles); err != nil { return errors.Errorf("Uploading images failed with %s. Exiting...", err) } conf.BootstrapImagePath = fmt.Sprintf("[%s] %s/%s", conf.ImageStores[0].Host, d.vmPathName, settings.BootstrapISO) // ensure that we wait for components to come up for _, s := range conf.ExecutorConfig.Sessions { s.Started = "" } snapshotName := fmt.Sprintf("%s %s", UpgradePrefix, conf.Version.BuildNumber) snapshotName = strings.TrimSpace(snapshotName) snapshotRefID, err := d.createSnapshot(snapshotName, "upgrade snapshot") if err != nil { d.deleteUpgradeImages(ds, settings) return err } defer func() { if err == nil { // do clean up aggressively, even the previous operation failed with context deadline excceeded. d.deleteSnapshot(*snapshotRefID, snapshotName, conf.Name) } }() if err = d.update(conf, settings); err == nil { return nil } log.Errorf("Failed to upgrade: %s", err) log.Infof("Rolling back upgrade") // reset timeout, to make sure rollback still happens in case of deadline exceeded error in previous step var cancel context.CancelFunc d.ctx, cancel = context.WithTimeout(context.Background(), settings.RollbackTimeout) defer cancel() if rerr := d.rollback(conf, snapshotName); rerr != nil { log.Errorf("Failed to revert appliance to snapshot: %s", rerr) // return the error message for upgrade, instead of rollback return err } d.deleteUpgradeImages(ds, settings) log.Infof("Appliance is rollback to old version") return err }