// NewDeployer returns a Worker that deploys and recalls unit agents // via ctx, taking a machine id to operate on. func NewDeployer(st *apideployer.State, ctx Context) worker.Worker { d := &Deployer{ st: st, ctx: ctx, } return worker.NewStringsWorker(d) }
func (s *stringsWorkerSuite) SetUpTest(c *gc.C) { s.LoggingSuite.SetUpTest(c) s.actor = &stringsHandler{ actions: nil, handled: make(chan []string, 1), watcher: &testStringsWatcher{ changes: make(chan []string), }, } s.worker = worker.NewStringsWorker(s.actor) }
func (s *ContainerSetupSuite) setupContainerWorker(c *gc.C, tag string) { runner := worker.NewRunner(allFatal, noImportance) pr := s.st.Provisioner() machine, err := pr.Machine(tag) c.Assert(err, gc.IsNil) err = machine.SetSupportedContainers(instance.ContainerTypes...) c.Assert(err, gc.IsNil) cfg := s.AgentConfigForTag(c, tag) watcherName := fmt.Sprintf("%s-container-watcher", machine.Id()) handler := provisioner.NewContainerSetupHandler(runner, watcherName, instance.ContainerTypes, machine, pr, cfg) runner.StartWorker(watcherName, func() (worker.Worker, error) { return worker.NewStringsWorker(handler), nil }) go func() { runner.Wait() }() }
func (s *stringsWorkerSuite) TestHandleErrorStopsWorkerAndWatcher(c *gc.C) { s.stopWorker(c) actor := &stringsHandler{ actions: nil, handled: make(chan []string, 1), handlerError: fmt.Errorf("my handling error"), watcher: &testStringsWatcher{ changes: make(chan []string), }, } w := worker.NewStringsWorker(actor) actor.watcher.TriggerChange(c, []string{"aa", "bb"}) waitForHandledStrings(c, actor.handled, []string{"aa", "bb"}) err := waitShort(c, w) c.Check(err, gc.ErrorMatches, "my handling error") actor.CheckActions(c, "setup", "handler", "teardown") c.Check(actor.watcher.stopped, jc.IsTrue) }
func (s *stringsWorkerSuite) TestSetUpFailureStopsWithTearDown(c *gc.C) { // Stop the worker and SetUp again, this time with an error s.stopWorker(c) actor := &stringsHandler{ actions: nil, handled: make(chan []string, 1), setupError: fmt.Errorf("my special error"), watcher: &testStringsWatcher{ changes: make(chan []string), }, } w := worker.NewStringsWorker(actor) err := waitShort(c, w) c.Check(err, gc.ErrorMatches, "my special error") // TearDown is not called on SetUp error. actor.CheckActions(c, "setup") c.Check(actor.watcher.stopped, jc.IsTrue) }
// updateSupportedContainers records in state that a machine can run the specified containers. // It starts a watcher and when a container of a given type is first added to the machine, // the watcher is killed, the machine is set up to be able to start containers of the given type, // and a suitable provisioner is started. func (a *MachineAgent) updateSupportedContainers( runner worker.Runner, st *api.State, tag string, containers []instance.ContainerType, agentConfig agent.Config, ) error { pr := st.Provisioner() machine, err := pr.Machine(tag) if err != nil { return fmt.Errorf("%s is not in state: %v", tag, err) } if len(containers) == 0 { if err := machine.SupportsNoContainers(); err != nil { return fmt.Errorf("clearing supported containers for %s: %v", tag, err) } return nil } if err := machine.SetSupportedContainers(containers...); err != nil { return fmt.Errorf("setting supported containers for %s: %v", tag, err) } // Start the watcher to fire when a container is first requested on the machine. watcherName := fmt.Sprintf("%s-container-watcher", machine.Id()) handler := provisioner.NewContainerSetupHandler( runner, watcherName, containers, machine, pr, agentConfig, ) a.startWorkerAfterUpgrade(runner, watcherName, func() (worker.Worker, error) { return worker.NewStringsWorker(handler), nil }) return nil }
// NewMinUnitsWorker returns a Worker that runs service.EnsureMinUnits() // if the number of alive units belonging to a service decreases, or if the // minimum required number of units for a service is increased. func NewMinUnitsWorker(st *state.State) worker.Worker { mu := &MinUnitsWorker{st: st} return worker.NewStringsWorker(mu) }