func (j *containerLog) Execute(resp jobs.Response) { if _, err := os.Stat(j.Id.UnitPathFor()); err != nil { resp.Failure(ErrContainerNotFound) return } w := resp.SuccessWithWrite(jobs.ResponseOk, true, false) err := systemd.WriteLogsTo(w, j.Id.UnitNameFor(), 30, time.After(30*time.Second)) if err != nil { log.Printf("job_container_log: Unable to fetch journal logs: %s\n", err.Error()) } }
func (j *getEnvironment) Execute(resp jobs.Response) { id := j.Id file, erro := os.Open(id.EnvironmentPathFor()) if erro != nil { resp.Failure(ErrEnvironmentNotFound) return } defer file.Close() w := resp.SuccessWithWrite(jobs.ResponseOk, false, false) if _, err := io.Copy(w, file); err != nil { log.Printf("job_content: Unable to write environment file: %+v", err) return } }
func (j *startContainer) Execute(resp jobs.Response) { unitName := j.Id.UnitNameFor() unitPath := j.Id.UnitPathFor() inState, tooSoon := inStateOrTooSoon(j.Id, unitName, true, false, rateLimitChanges) if inState { w := resp.SuccessWithWrite(jobs.ResponseAccepted, true, false) fmt.Fprintf(w, "Container %s starting\n", j.Id) return } if tooSoon { resp.Failure(ErrStartRequestThrottled) return } if errs := csystemd.SetUnitStartOnBoot(j.Id, true); errs != nil { log.Print("alter_container_state: Unable to persist whether the unit is started on boot: ", errs) resp.Failure(ErrContainerStartFailed) return } if err := systemd.EnableAndReloadUnit(systemd.Connection(), unitName, unitPath); err != nil { if systemd.IsNoSuchUnit(err) || systemd.IsFileNotFound(err) { resp.Failure(ErrContainerNotFound) return } log.Printf("alter_container_state: Could not enable container %s: %v", unitName, err) resp.Failure(ErrContainerStartFailed) return } if err := systemd.Connection().StartUnitJob(unitName, "replace"); err != nil { log.Printf("alter_container_state: Could not start container %s: %v", unitName, err) resp.Failure(ErrContainerStartFailed) return } w := resp.SuccessWithWrite(jobs.ResponseAccepted, true, false) fmt.Fprintf(w, "Container %s starting\n", j.Id) }
func (j *listImages) Execute(resp jobs.Response) { // TODO: config item for docker port dockerClient, err := docker.NewClient(j.DockerSocket) if err != nil { log.Printf("job_list_images: Couldn't connect to docker: %+v", err) resp.Failure(ErrListImagesFailed) return } imgs, err := dockerClient.ListImages(false) if err != nil { log.Printf("job_list_images: Couldn't connect to docker: %+v", err) resp.Failure(ErrListImagesFailed) return } w := resp.SuccessWithWrite(jobs.ResponseAccepted, true, false) for _, img := range imgs { fmt.Fprintf(w, "%+v\n", img.RepoTags[0]) } }
func (j *buildImage) Execute(resp jobs.Response) { w := resp.SuccessWithWrite(jobs.ResponseAccepted, true, false) fmt.Fprintf(w, "Processing build-image request:\n") // TODO: download source, add bind-mount unitName := containers.JobIdentifier(j.Name).UnitNameForBuild() unitDescription := fmt.Sprintf("Builder for %s", j.Tag) stdout, err := systemd.ProcessLogsForUnit(unitName) if err != nil { stdout = utils.EmptyReader log.Printf("job_build_image: Unable to fetch build logs: %s, %+v", err.Error(), err) } defer stdout.Close() conn, errc := systemd.NewConnection() if errc != nil { log.Print("job_build_image:", errc) fmt.Fprintf(w, "Unable to watch start status", errc) return } slice := DefaultSlice systemd.IsUnitProperty(conn, j.BaseImage+".service", func(properties map[string]interface{}) bool { value, ok := properties["Slice"] if ok { slice = value.(string) } return ok }) if err := conn.Subscribe(); err != nil { log.Print("job_build_image:", err) fmt.Fprintf(w, "Unable to watch start status", errc) return } defer conn.Unsubscribe() // make subscription global for efficiency var ( changes <-chan map[string]*dbus.UnitStatus errch <-chan error ) if resp.StreamResult() { changes, errch = conn.SubscribeUnitsCustom(1*time.Second, 2, func(s1 *dbus.UnitStatus, s2 *dbus.UnitStatus) bool { return true }, func(unit string) bool { return unit != unitName }) } fmt.Fprintf(w, "Running sti build unit: %s\n", unitName) log.Printf("build_image: Running build %s", unitName) var startCmd []string if _, err := os.Stat(gearBinaryPath); err != nil { log.Println("gear executable is not installed on system; using sti builder image") startCmd = []string{ "/usr/bin/docker", "run", "-rm", "-v", "/run/docker.sock:/run/docker.sock", "-t", builderImage, "sti", "build", j.Source, j.BaseImage, j.Tag, "-U", "unix:///run/docker.sock", } } else { startCmd = []string{ gearBinaryPath, "build", j.Source, j.BaseImage, j.Tag, } } if j.RuntimeImage != "" { startCmd = append(startCmd, "--runtime-image") startCmd = append(startCmd, j.RuntimeImage) } if j.Clean { startCmd = append(startCmd, "--clean") } if j.Verbose { startCmd = append(startCmd, "--debug") } if j.CallbackUrl != "" { startCmd = append(startCmd, "--callbackUrl="+j.CallbackUrl) } log.Printf("build_image: Will execute %v", startCmd) status, err := systemd.Connection().StartTransientUnit( unitName, "fail", dbus.PropExecStart(startCmd, true), dbus.PropDescription(unitDescription), dbus.PropRemainAfterExit(true), dbus.PropSlice(slice), ) if err != nil { errType := reflect.TypeOf(err) fmt.Fprintf(w, "Unable to start build container for this image due to (%s): %s\n", errType, err.Error()) return } else if status != "done" { fmt.Fprintf(w, "Build did not complete successfully: %s\n", status) } else { fmt.Fprintf(w, "Sti build is running\n") } if resp.StreamResult() { go io.Copy(w, stdout) wait: for { select { case c := <-changes: if changed, ok := c[unitName]; ok { if changed.SubState != "running" { fmt.Fprintf(w, "Build completed\n") break wait } } case err := <-errch: fmt.Fprintf(w, "Error %+v\n", err) case <-time.After(25 * time.Second): log.Print("job_build_image:", "timeout") break wait } } } }
func (j *stopContainer) Execute(resp jobs.Response) { unitName := j.Id.UnitNameFor() inState, tooSoon := inStateOrTooSoon(j.Id, unitName, false, false, rateLimitChanges) if inState { w := resp.SuccessWithWrite(jobs.ResponseAccepted, true, false) fmt.Fprintf(w, "Container %s is stopped\n", j.Id) return } if tooSoon { resp.Failure(ErrStopRequestThrottled) return } if errs := csystemd.SetUnitStartOnBoot(j.Id, false); errs != nil { log.Print("alter_container_state: Unable to persist whether the unit is started on boot: ", errs) resp.Failure(ErrContainerStopFailed) return } if !j.Wait { err := systemd.Connection().StopUnitJob(unitName, "replace") if err != nil { log.Print("alter_container_state: Failed to queue a restart ", err) resp.Failure(ErrContainerStopFailed) } w := resp.SuccessWithWrite(jobs.ResponseAccepted, false, false) fmt.Fprintf(w, "Container %s is stopping\n", j.Id) return } w := resp.SuccessWithWrite(jobs.ResponseAccepted, true, false) done := make(chan time.Time) ioerr := make(chan error) go func() { ioerr <- systemd.WriteLogsTo(w, unitName, 0, done) }() joberr := make(chan error) go func() { status, err := systemd.Connection().StopUnit(unitName, "replace") if err == nil && status != "done" { err = errors.New(fmt.Sprintf("Job status 'done' != %s", status)) } joberr <- err }() var err error select { case err = <-ioerr: log.Printf("alter_container_state: Client hung up") close(ioerr) case err = <-joberr: log.Printf("alter_container_state: Stop job done") case <-time.After(15 * time.Second): log.Printf("alter_container_state: Timeout waiting for stop completion") } close(done) select { case <-ioerr: } switch { case systemd.IsNoSuchUnit(err): if _, err := os.Stat(j.Id.UnitPathFor()); err == nil { fmt.Fprintf(w, "Container %s is stopped\n", j.Id) } else { fmt.Fprintf(w, "No such container %s\n", j.Id) } case err != nil: fmt.Fprintf(w, "Could not stop container: %s\n", err.Error()) default: fmt.Fprintf(w, "Container %s is stopped\n", j.Id) } }
func (req *installContainer) Execute(resp jobs.Response) { id := req.Id unitName := id.UnitNameFor() unitPath := id.UnitPathFor() unitVersionPath := id.VersionedUnitPathFor(req.RequestIdentifier.String()) socketUnitName := id.SocketUnitNameFor() socketUnitPath := id.SocketUnitPathFor() var socketActivationType string if req.SocketActivation { socketActivationType = "enabled" if !req.SkipSocketProxy { socketActivationType = "proxied" } } // attempt to download the environment if it is remote env := req.Environment if env != nil { if err := env.Fetch(100 * 1024); err != nil { resp.Failure(ErrContainerCreateFailed) return } if env.Empty() { env = nil } } // open and lock the base path (to prevent simultaneous updates) state, exists, err := utils.OpenFileExclusive(unitPath, 0664) if err != nil { log.Print("install_container: Unable to lock unit file: ", err) resp.Failure(ErrContainerCreateFailed) } defer state.Close() // write a new file to disk that describes the new service unit, err := utils.CreateFileExclusive(unitVersionPath, 0664) if err != nil { log.Print("install_container: Unable to open unit file definition: ", err) resp.Failure(ErrContainerCreateFailed) return } defer unit.Close() // if this is an existing container, read the currently reserved ports existingPorts := port.PortPairs{} if exists { existingPorts, err = containers.GetExistingPorts(id) if err != nil { if _, ok := err.(*os.PathError); !ok { log.Print("install_container: Unable to read existing ports from file: ", err) resp.Failure(ErrContainerCreateFailed) return } } } // allocate and reserve ports for this container reserved, erra := portReserver.AtomicReserveExternalPorts(unitVersionPath, req.Ports, existingPorts) if erra != nil { log.Printf("install_container: Unable to reserve external ports: %+v", erra) resp.Failure(ErrContainerCreateFailedPortsReserved) return } if len(reserved) > 0 { resp.WritePendingSuccess(PendingPortMappingName, reserved) } var portSpec string if req.Simple && len(reserved) == 0 { portSpec = "-P" } else { portSpec = dockerPortSpec(reserved) } // write the environment to disk var environmentPath string if env != nil { if errw := env.Write(false); errw != nil { resp.Failure(ErrContainerCreateFailed) return } environmentPath = env.Id.EnvironmentPathFor() } // write the network links (if any) to disk if req.NetworkLinks != nil { if errw := req.NetworkLinks.Write(id.NetworkLinksPathFor(), false); errw != nil { resp.Failure(ErrContainerCreateFailed) return } } var sliceName string if "" == req.SystemdSlice { sliceName = DefaultSlice } else { sliceNames := []string{} sliceFound := false for _, name := range ListSliceNames() { sliceNames = append(sliceNames, name) if req.SystemdSlice == name { sliceFound = true break } } if sliceFound { sliceName = req.SystemdSlice } else { log.Printf("'%s' is not a valid systemd slice. Must be one of [%s]", req.SystemdSlice, strings.Join(sliceNames, ", ")) resp.Failure(ErrContainerCreateFailedInvalidSlice) return } } // write the definition unit file args := csystemd.ContainerUnit{ Id: id, Image: req.Image, PortSpec: portSpec, Slice: sliceName + ".slice", Isolate: req.Isolate, ReqId: req.RequestIdentifier.String(), HomeDir: id.HomePath(), RunDir: id.RunPathFor(), EnvironmentPath: environmentPath, ExecutablePath: filepath.Join("/", "usr", "bin", "gear"), IncludePath: "", PortPairs: reserved, SocketUnitName: socketUnitName, SocketActivationType: socketActivationType, DockerFeatures: config.SystemDockerFeatures, } var templateName string switch { case req.SocketActivation: templateName = "SOCKETACTIVATED" case config.SystemDockerFeatures.ForegroundRun: templateName = "FOREGROUND" default: templateName = "SIMPLE" } if erre := csystemd.ContainerUnitTemplate.ExecuteTemplate(unit, templateName, args); erre != nil { log.Printf("install_container: Unable to output template: %+v", erre) resp.Failure(ErrContainerCreateFailed) defer os.Remove(unitVersionPath) return } if err := unit.Close(); err != nil { log.Printf("install_container: Unable to finish writing unit: %+v", err) resp.Failure(ErrContainerCreateFailed) defer os.Remove(unitVersionPath) return } // swap the new definition with the old one if err := utils.AtomicReplaceLink(unitVersionPath, unitPath); err != nil { log.Printf("install_container: Failed to activate new unit: %+v", err) resp.Failure(ErrContainerCreateFailed) return } state.Close() // write whether this container should be started on next boot if req.Started { if errs := csystemd.SetUnitStartOnBoot(id, true); errs != nil { log.Print("install_container: Unable to write container boot link: ", err) resp.Failure(ErrContainerCreateFailed) return } } // Generate the socket file and ignore failures paths := []string{unitPath} if req.SocketActivation { if err := writeSocketUnit(socketUnitPath, &args); err == nil { paths = []string{unitPath, socketUnitPath} } } if err := systemd.EnableAndReloadUnit(systemd.Connection(), unitName, paths...); err != nil { log.Printf("install_container: Could not enable container %s (%v): %v", unitName, paths, err) resp.Failure(ErrContainerCreateFailed) return } if req.Started { if req.SocketActivation { // Start the socket file, not the service and ignore failures if err := systemd.Connection().StartUnitJob(socketUnitName, "replace"); err != nil { log.Printf("install_container: Could not start container socket %s: %v", socketUnitName, err) resp.Failure(ErrContainerCreateFailed) return } } else { if err := systemd.Connection().StartUnitJob(unitName, "replace"); err != nil { log.Printf("install_container: Could not start container %s: %v", unitName, err) resp.Failure(ErrContainerCreateFailed) return } } } w := resp.SuccessWithWrite(jobs.ResponseAccepted, true, false) if req.Started { fmt.Fprintf(w, "Container %s is starting\n", id) } else { fmt.Fprintf(w, "Container %s is installed\n", id) } }