func (*ManifoldSuite) TestOutputBadTarget(c *gc.C) { manifold := migrationflag.Manifold(migrationflag.ManifoldConfig{}) in := &migrationflag.Worker{} var out bool err := manifold.Output(in, &out) c.Check(err, gc.ErrorMatches, "expected out to be a \\*Flag; got a .*") }
// checkManifoldNotValid checks that the supplied ManifoldConfig creates // a manifold that cannot be started. func checkManifoldNotValid(c *gc.C, config migrationflag.ManifoldConfig, expect string) { manifold := migrationflag.Manifold(config) worker, err := manifold.Start(dt.StubContext(nil, nil)) c.Check(worker, gc.IsNil) c.Check(err, gc.ErrorMatches, expect) c.Check(err, jc.Satisfies, errors.IsNotValid) }
func (*ManifoldSuite) TestOutputBadWorker(c *gc.C) { manifold := migrationflag.Manifold(migrationflag.ManifoldConfig{}) in := &struct{ worker.Worker }{} var out util.Flag err := manifold.Output(in, &out) c.Check(err, gc.ErrorMatches, "expected in to implement Flag; got a .*") }
func (*ManifoldSuite) TestOutputBadInput(c *gc.C) { manifold := migrationflag.Manifold(migrationflag.ManifoldConfig{}) in := &migrationflag.Worker{} var out util.Flag err := manifold.Output(in, &out) c.Check(err, jc.ErrorIsNil) c.Check(out, gc.Equals, in) }
func (*ManifoldSuite) TestStartMissingAPICaller(c *gc.C) { context := dt.StubContext(nil, map[string]interface{}{ "api-caller": dependency.ErrMissing, }) manifold := migrationflag.Manifold(validManifoldConfig()) worker, err := manifold.Start(context) c.Check(worker, gc.IsNil) c.Check(errors.Cause(err), gc.Equals, dependency.ErrMissing) }
func (*ManifoldSuite) TestStartNewFacadeError(c *gc.C) { expectCaller := &stubCaller{} context := dt.StubContext(nil, map[string]interface{}{ "api-caller": expectCaller, }) config := validManifoldConfig() config.NewFacade = func(caller base.APICaller) (migrationflag.Facade, error) { c.Check(caller, gc.Equals, expectCaller) return nil, errors.New("bort") } manifold := migrationflag.Manifold(config) worker, err := manifold.Start(context) c.Check(worker, gc.IsNil) c.Check(err, gc.ErrorMatches, "bort") }
func (*ManifoldSuite) TestStartSuccess(c *gc.C) { context := dt.StubContext(nil, map[string]interface{}{ "api-caller": &stubCaller{}, }) expectWorker := &struct{ worker.Worker }{} config := validManifoldConfig() config.NewFacade = func(base.APICaller) (migrationflag.Facade, error) { return &struct{ migrationflag.Facade }{}, nil } config.NewWorker = func(migrationflag.Config) (worker.Worker, error) { return expectWorker, nil } manifold := migrationflag.Manifold(config) worker, err := manifold.Start(context) c.Check(err, jc.ErrorIsNil) c.Check(worker, gc.Equals, expectWorker) }
func (*ManifoldSuite) TestStartNewWorkerError(c *gc.C) { context := dt.StubContext(nil, map[string]interface{}{ "api-caller": &stubCaller{}, }) expectFacade := &struct{ migrationflag.Facade }{} config := validManifoldConfig() config.NewFacade = func(base.APICaller) (migrationflag.Facade, error) { return expectFacade, nil } config.NewWorker = func(workerConfig migrationflag.Config) (worker.Worker, error) { c.Check(workerConfig.Facade, gc.Equals, expectFacade) c.Check(workerConfig.Model, gc.Equals, validUUID) c.Check(workerConfig.Check, gc.NotNil) // uncomparable return nil, errors.New("snerk") } manifold := migrationflag.Manifold(config) worker, err := manifold.Start(context) c.Check(worker, gc.IsNil) c.Check(err, gc.ErrorMatches, "snerk") }
func (*ManifoldSuite) TestFilterOther(c *gc.C) { manifold := migrationflag.Manifold(migrationflag.ManifoldConfig{}) expect := errors.New("whatever") actual := manifold.Filter(expect) c.Check(actual, gc.Equals, expect) }
func (*ManifoldSuite) TestFilterErrChanged(c *gc.C) { manifold := migrationflag.Manifold(migrationflag.ManifoldConfig{}) err := manifold.Filter(migrationflag.ErrChanged) c.Check(err, gc.Equals, dependency.ErrBounce) }
func (*ManifoldSuite) TestFilterNil(c *gc.C) { manifold := migrationflag.Manifold(migrationflag.ManifoldConfig{}) err := manifold.Filter(nil) c.Check(err, jc.ErrorIsNil) }
func (*ManifoldSuite) TestInputs(c *gc.C) { manifold := migrationflag.Manifold(validManifoldConfig()) c.Check(manifold.Inputs, jc.DeepEquals, []string{"api-caller"}) }
// Manifolds returns a set of co-configured manifolds covering the // various responsibilities of a machine agent. // // Thou Shalt Not Use String Literals In This Function. Or Else. func Manifolds(config ManifoldsConfig) dependency.Manifolds { // connectFilter exists: // 1) to let us retry api connections immediately on password change, // rather than causing the dependency engine to wait for a while; // 2) to ensure that certain connection failures correctly trigger // complete agent removal. (It's not safe to let any agent other // than the machine mess around with SetCanUninstall). connectFilter := func(err error) error { cause := errors.Cause(err) if cause == apicaller.ErrConnectImpossible { err2 := coreagent.SetCanUninstall(config.Agent) if err2 != nil { return errors.Trace(err2) } return worker.ErrTerminateAgent } else if cause == apicaller.ErrChangedPassword { return dependency.ErrBounce } return err } var externalUpdateProxyFunc func(proxy.Settings) error if runtime.GOOS == "linux" { externalUpdateProxyFunc = lxd.ConfigureLXDProxies } return dependency.Manifolds{ // The agent manifold references the enclosing agent, and is the // foundation stone on which most other manifolds ultimately depend. agentName: agent.Manifold(config.Agent), // The termination worker returns ErrTerminateAgent if a // termination signal is received by the process it's running // in. It has no inputs and its only output is the error it // returns. It depends on the uninstall file having been // written *by the manual provider* at install time; it would // be Very Wrong Indeed to use SetCanUninstall in conjunction // with this code. terminationName: terminationworker.Manifold(), // The stateconfigwatcher manifold watches the machine agent's // configuration and reports if state serving info is // present. It will bounce itself if state serving info is // added or removed. It is intended as a dependency just for // the state manifold. stateConfigWatcherName: stateconfigwatcher.Manifold(stateconfigwatcher.ManifoldConfig{ AgentName: agentName, AgentConfigChanged: config.AgentConfigChanged, }), // The state manifold creates a *state.State and makes it // available to other manifolds. It pings the mongodb session // regularly and will die if pings fail. stateName: workerstate.Manifold(workerstate.ManifoldConfig{ AgentName: agentName, StateConfigWatcherName: stateConfigWatcherName, OpenState: config.OpenState, }), // The stateworkers manifold starts workers which rely on a // *state.State but which haven't been converted to run // directly under the dependency engine yet. This manifold // will be removed once all such workers have been converted; // until then, the workers are expected to handle their own // checks for upgrades etc, rather than blocking this whole // worker on upgrade completion. stateWorkersName: StateWorkersManifold(StateWorkersConfig{ StateName: stateName, StartStateWorkers: config.StartStateWorkers, }), // The api-config-watcher manifold monitors the API server // addresses in the agent config and bounces when they // change. It's required as part of model migrations. apiConfigWatcherName: apiconfigwatcher.Manifold(apiconfigwatcher.ManifoldConfig{ AgentName: agentName, AgentConfigChanged: config.AgentConfigChanged, }), // The api caller is a thin concurrent wrapper around a connection // to some API server. It's used by many other manifolds, which all // select their own desired facades. It will be interesting to see // how this works when we consolidate the agents; might be best to // handle the auth changes server-side..? apiCallerName: apicaller.Manifold(apicaller.ManifoldConfig{ AgentName: agentName, APIConfigWatcherName: apiConfigWatcherName, APIOpen: api.Open, NewConnection: apicaller.ScaryConnect, Filter: connectFilter, }), // The upgrade steps gate is used to coordinate workers which // shouldn't do anything until the upgrade-steps worker has // finished running any required upgrade steps. The flag of // similar name is used to implement the isFullyUpgraded func // that keeps upgrade concerns out of unrelated manifolds. upgradeStepsGateName: gate.ManifoldEx(config.UpgradeStepsLock), upgradeStepsFlagName: gate.FlagManifold(gate.FlagManifoldConfig{ GateName: upgradeStepsGateName, NewWorker: gate.NewFlagWorker, }), // The upgrade check gate is used to coordinate workers which // shouldn't do anything until the upgrader worker has // completed its first check for a new tools version to // upgrade to. The flag of similar name is used to implement // the isFullyUpgraded func that keeps upgrade concerns out of // unrelated manifolds. upgradeCheckGateName: gate.ManifoldEx(config.UpgradeCheckLock), upgradeCheckFlagName: gate.FlagManifold(gate.FlagManifoldConfig{ GateName: upgradeCheckGateName, NewWorker: gate.NewFlagWorker, }), // The upgrader is a leaf worker that returns a specific error // type recognised by the machine agent, causing other workers // to be stopped and the agent to be restarted running the new // tools. We should only need one of these in a consolidated // agent, but we'll need to be careful about behavioural // differences, and interactions with the upgrade-steps // worker. upgraderName: upgrader.Manifold(upgrader.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, UpgradeStepsGateName: upgradeStepsGateName, UpgradeCheckGateName: upgradeCheckGateName, PreviousAgentVersion: config.PreviousAgentVersion, }), // The upgradesteps worker runs soon after the machine agent // starts and runs any steps required to upgrade to the // running jujud version. Once upgrade steps have run, the // upgradesteps gate is unlocked and the worker exits. upgradeStepsName: upgradesteps.Manifold(upgradesteps.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, UpgradeStepsGateName: upgradeStepsGateName, OpenStateForUpgrade: config.OpenStateForUpgrade, PreUpgradeSteps: config.PreUpgradeSteps, }), // The migration workers collaborate to run migrations; // and to create a mechanism for running other workers // so they can't accidentally interfere with a migration // in progress. Such a manifold should (1) depend on the // migration-inactive flag, to know when to start or die; // and (2) occupy the migration-fortress, so as to avoid // possible interference with the minion (which will not // take action until it's gained sole control of the // fortress). // // Note that the fortress itself will not be created // until the upgrade process is complete; this frees all // its dependencies from upgrade concerns. migrationFortressName: ifFullyUpgraded(fortress.Manifold()), migrationInactiveFlagName: migrationflag.Manifold(migrationflag.ManifoldConfig{ APICallerName: apiCallerName, Check: migrationflag.IsTerminal, NewFacade: migrationflag.NewFacade, NewWorker: migrationflag.NewWorker, }), migrationMinionName: migrationminion.Manifold(migrationminion.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, FortressName: migrationFortressName, APIOpen: api.Open, ValidateMigration: config.ValidateMigration, NewFacade: migrationminion.NewFacade, NewWorker: migrationminion.NewWorker, }), // The serving-info-setter manifold sets grabs the state // serving info from the API connection and writes it to the // agent config. servingInfoSetterName: ifNotMigrating(ServingInfoSetterManifold(ServingInfoSetterConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The apiworkers manifold starts workers which rely on the // machine agent's API connection but have not been converted // to work directly under the dependency engine. It waits for // upgrades to be finished before starting these workers. apiWorkersName: ifNotMigrating(APIWorkersManifold(APIWorkersConfig{ APICallerName: apiCallerName, StartAPIWorkers: config.StartAPIWorkers, })), // The reboot manifold manages a worker which will reboot the // machine when requested. It needs an API connection and // waits for upgrades to be complete. rebootName: ifNotMigrating(reboot.Manifold(reboot.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, MachineLockName: coreagent.MachineLockName, Clock: config.Clock, })), // The logging config updater is a leaf worker that indirectly // controls the messages sent via the log sender or rsyslog, // according to changes in environment config. We should only need // one of these in a consolidated agent. loggingConfigUpdaterName: ifNotMigrating(logger.Manifold(logger.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The diskmanager worker periodically lists block devices on the // machine it runs on. This worker will be run on all Juju-managed // machines (one per machine agent). diskManagerName: ifNotMigrating(diskmanager.Manifold(diskmanager.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The proxy config updater is a leaf worker that sets http/https/apt/etc // proxy settings. proxyConfigUpdater: ifNotMigrating(proxyupdater.Manifold(proxyupdater.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, WorkerFunc: proxyupdater.NewWorker, ExternalUpdate: externalUpdateProxyFunc, })), // The api address updater is a leaf worker that rewrites agent config // as the state server addresses change. We should only need one of // these in a consolidated agent. apiAddressUpdaterName: ifNotMigrating(apiaddressupdater.Manifold(apiaddressupdater.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The machiner Worker will wait for the identified machine to become // Dying and make it Dead; or until the machine becomes Dead by other // means. machinerName: ifNotMigrating(machiner.Manifold(machiner.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The log sender is a leaf worker that sends log messages to some // API server, when configured so to do. We should only need one of // these in a consolidated agent. // // NOTE: the LogSource will buffer a large number of messages as an upgrade // runs; it currently seems better to fill the buffer and send when stable, // optimising for stable controller upgrades rather than up-to-the-moment // observable normal-machine upgrades. logSenderName: ifNotMigrating(logsender.Manifold(logsender.ManifoldConfig{ APICallerName: apiCallerName, LogSource: config.LogSource, })), // The deployer worker is responsible for deploying and recalling unit // agents, according to changes in a set of state units; and for the // final removal of its agents' units from state when they are no // longer needed. deployerName: ifNotMigrating(deployer.Manifold(deployer.ManifoldConfig{ NewDeployContext: config.NewDeployContext, AgentName: agentName, APICallerName: apiCallerName, })), authenticationWorkerName: ifNotMigrating(authenticationworker.Manifold(authenticationworker.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The storageProvisioner worker manages provisioning // (deprovisioning), and attachment (detachment) of first-class // volumes and filesystems. storageProvisionerName: ifNotMigrating(storageprovisioner.MachineManifold(storageprovisioner.MachineManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, Clock: config.Clock, })), resumerName: ifNotMigrating(resumer.Manifold(resumer.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, Clock: config.Clock, Interval: time.Minute, NewFacade: resumer.NewFacade, NewWorker: resumer.NewWorker, })), identityFileWriterName: ifNotMigrating(identityfilewriter.Manifold(identityfilewriter.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), toolsVersionCheckerName: ifNotMigrating(toolsversionchecker.Manifold(toolsversionchecker.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), machineActionName: ifNotMigrating(machineactions.Manifold(machineactions.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, NewFacade: machineactions.NewFacade, NewWorker: machineactions.NewMachineActionsWorker, })), hostKeyReporterName: ifNotMigrating(hostkeyreporter.Manifold(hostkeyreporter.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, RootDir: config.RootDir, NewFacade: hostkeyreporter.NewFacade, NewWorker: hostkeyreporter.NewWorker, })), logForwarderName: ifFullyUpgraded(logforwarder.Manifold(logforwarder.ManifoldConfig{ StateName: stateName, APICallerName: apiCallerName, Sinks: []logforwarder.LogSinkSpec{{ Name: "juju-log-forward", OpenFn: sinks.OpenSyslog, }}, })), } }
// Manifolds returns a set of co-configured manifolds covering the various // responsibilities of a standalone unit agent. It also accepts the logSource // argument because we haven't figured out how to thread all the logging bits // through a dependency engine yet. // // Thou Shalt Not Use String Literals In This Function. Or Else. func Manifolds(config ManifoldsConfig) dependency.Manifolds { // connectFilter exists to let us retry api connections immediately // on password change, rather than causing the dependency engine to // wait for a while. connectFilter := func(err error) error { cause := errors.Cause(err) if cause == apicaller.ErrChangedPassword { return dependency.ErrBounce } else if cause == apicaller.ErrConnectImpossible { return worker.ErrTerminateAgent } return err } return dependency.Manifolds{ // The agent manifold references the enclosing agent, and is the // foundation stone on which most other manifolds ultimately depend. // (Currently, that is "all manifolds", but consider a shared clock.) agentName: agent.Manifold(config.Agent), // The api-config-watcher manifold monitors the API server // addresses in the agent config and bounces when they // change. It's required as part of model migrations. apiConfigWatcherName: apiconfigwatcher.Manifold(apiconfigwatcher.ManifoldConfig{ AgentName: agentName, AgentConfigChanged: config.AgentConfigChanged, }), // The api caller is a thin concurrent wrapper around a connection // to some API server. It's used by many other manifolds, which all // select their own desired facades. It will be interesting to see // how this works when we consolidate the agents; might be best to // handle the auth changes server-side..? apiCallerName: apicaller.Manifold(apicaller.ManifoldConfig{ AgentName: agentName, APIConfigWatcherName: apiConfigWatcherName, APIOpen: api.Open, NewConnection: apicaller.ScaryConnect, Filter: connectFilter, }), // The log sender is a leaf worker that sends log messages to some // API server, when configured so to do. We should only need one of // these in a consolidated agent. logSenderName: logsender.Manifold(logsender.ManifoldConfig{ APICallerName: apiCallerName, LogSource: config.LogSource, }), // The upgrader is a leaf worker that returns a specific error type // recognised by the unit agent, causing other workers to be stopped // and the agent to be restarted running the new tools. We should only // need one of these in a consolidated agent, but we'll need to be // careful about behavioural differences, and interactions with the // upgradesteps worker. upgraderName: upgrader.Manifold(upgrader.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, }), // The migration workers collaborate to run migrations; // and to create a mechanism for running other workers // so they can't accidentally interfere with a migration // in progress. Such a manifold should (1) depend on the // migration-inactive flag, to know when to start or die; // and (2) occupy the migration-fortress, so as to avoid // possible interference with the minion (which will not // take action until it's gained sole control of the // fortress). migrationFortressName: fortress.Manifold(), migrationInactiveFlagName: migrationflag.Manifold(migrationflag.ManifoldConfig{ APICallerName: apiCallerName, Check: migrationflag.IsTerminal, NewFacade: migrationflag.NewFacade, NewWorker: migrationflag.NewWorker, }), migrationMinionName: migrationminion.Manifold(migrationminion.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, FortressName: migrationFortressName, APIOpen: api.Open, ValidateMigration: config.ValidateMigration, NewFacade: migrationminion.NewFacade, NewWorker: migrationminion.NewWorker, }), // The logging config updater is a leaf worker that indirectly // controls the messages sent via the log sender according to // changes in environment config. We should only need one of // these in a consolidated agent. loggingConfigUpdaterName: ifNotMigrating(logger.Manifold(logger.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The api address updater is a leaf worker that rewrites agent config // as the controller addresses change. We should only need one of // these in a consolidated agent. apiAddressUpdaterName: ifNotMigrating(apiaddressupdater.Manifold(apiaddressupdater.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The proxy config updater is a leaf worker that sets http/https/apt/etc // proxy settings. // TODO(fwereade): timing of this is suspicious. There was superstitious // code trying to run this early; if that ever helped, it was only by // coincidence. Probably we ought to be making components that might // need proxy config into explicit dependencies of the proxy updater... proxyConfigUpdaterName: ifNotMigrating(proxyupdater.Manifold(proxyupdater.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, WorkerFunc: proxyupdater.NewWorker, })), // The charmdir resource coordinates whether the charm directory is // available or not; after 'start' hook and before 'stop' hook // executes, and not during upgrades. charmDirName: ifNotMigrating(fortress.Manifold()), // The leadership tracker attempts to secure and retain leadership of // the unit's service, and is consulted on such matters by the // uniter. As it stannds today, we'll need one per unit in a // consolidated agent. leadershipTrackerName: ifNotMigrating(leadership.Manifold(leadership.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, Clock: clock.WallClock, LeadershipGuarantee: config.LeadershipGuarantee, })), // HookRetryStrategy uses a retrystrategy worker to get a // retry strategy that will be used by the uniter to run its hooks. hookRetryStrategyName: ifNotMigrating(retrystrategy.Manifold(retrystrategy.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, NewFacade: retrystrategy.NewFacade, NewWorker: retrystrategy.NewRetryStrategyWorker, })), // The uniter installs charms; manages the unit's presence in its // relations; creates suboordinate units; runs all the hooks; sends // metrics; etc etc etc. We expect to break it up further in the // coming weeks, and to need one per unit in a consolidated agent // (and probably one for each component broken out). uniterName: ifNotMigrating(uniter.Manifold(uniter.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, MachineLockName: coreagent.MachineLockName, Clock: clock.WallClock, LeadershipTrackerName: leadershipTrackerName, CharmDirName: charmDirName, HookRetryStrategyName: hookRetryStrategyName, })), // TODO (mattyw) should be added to machine agent. metricSpoolName: ifNotMigrating(spool.Manifold(spool.ManifoldConfig{ AgentName: agentName, })), // The metric collect worker executes the collect-metrics hook in a // restricted context that can safely run concurrently with other hooks. metricCollectName: ifNotMigrating(collect.Manifold(collect.ManifoldConfig{ AgentName: agentName, MetricSpoolName: metricSpoolName, CharmDirName: charmDirName, })), // The meter status worker executes the meter-status-changed hook when it detects // that the meter status has changed. meterStatusName: ifNotMigrating(meterstatus.Manifold(meterstatus.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, MachineLockName: coreagent.MachineLockName, Clock: clock.WallClock, NewHookRunner: meterstatus.NewHookRunner, NewMeterStatusAPIClient: msapi.NewClient, NewConnectedStatusWorker: meterstatus.NewConnectedStatusWorker, NewIsolatedStatusWorker: meterstatus.NewIsolatedStatusWorker, })), // The metric sender worker periodically sends accumulated metrics to the controller. metricSenderName: ifNotMigrating(sender.Manifold(sender.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, MetricSpoolName: metricSpoolName, })), } }
// Manifolds returns a set of interdependent dependency manifolds that will // run together to administer a model, as configured. func Manifolds(config ManifoldsConfig) dependency.Manifolds { modelTag := config.Agent.CurrentConfig().Model() return dependency.Manifolds{ // The first group are foundational; the agent and clock // which wrap those supplied in config, and the api-caller // through which everything else communicates with the // controller. agentName: agent.Manifold(config.Agent), clockName: clockManifold(config.Clock), apiConfigWatcherName: apiconfigwatcher.Manifold(apiconfigwatcher.ManifoldConfig{ AgentName: agentName, AgentConfigChanged: config.AgentConfigChanged, }), apiCallerName: apicaller.Manifold(apicaller.ManifoldConfig{ AgentName: agentName, APIOpen: api.Open, NewConnection: apicaller.OnlyConnect, Filter: apiConnectFilter, }), // The spaces-imported gate will be unlocked when space // discovery is known to be complete. Various manifolds // should also come to depend upon it (or rather, on a // Flag depending on it) in the future. spacesImportedGateName: gate.ManifoldEx(config.SpacesImportedGate), // All other manifolds should depend on at least one of these // three, which handle all the tasks that are safe and sane // to run in *all* controller machines. notDeadFlagName: lifeflag.Manifold(lifeflag.ManifoldConfig{ APICallerName: apiCallerName, Entity: modelTag, Result: life.IsNotDead, Filter: LifeFilter, NewFacade: lifeflag.NewFacade, NewWorker: lifeflag.NewWorker, }), notAliveFlagName: lifeflag.Manifold(lifeflag.ManifoldConfig{ APICallerName: apiCallerName, Entity: modelTag, Result: life.IsNotAlive, Filter: LifeFilter, NewFacade: lifeflag.NewFacade, NewWorker: lifeflag.NewWorker, }), isResponsibleFlagName: singular.Manifold(singular.ManifoldConfig{ ClockName: clockName, AgentName: agentName, APICallerName: apiCallerName, Duration: config.RunFlagDuration, NewFacade: singular.NewFacade, NewWorker: singular.NewWorker, }), // The migration workers collaborate to run migrations; // and to create a mechanism for running other workers // so they can't accidentally interfere with a migration // in progress. Such a manifold should (1) depend on the // migration-inactive flag, to know when to start or die; // and (2) occupy the migration-fortress, so as to avoid // possible interference with the minion (which will not // take action until it's gained sole control of the // fortress). // // Note that the fortress and flag will only exist while // the model is not dead; this frees their dependencies // from model-lifetime concerns. migrationFortressName: ifNotDead(fortress.Manifold()), migrationInactiveFlagName: ifNotDead(migrationflag.Manifold(migrationflag.ManifoldConfig{ APICallerName: apiCallerName, Check: migrationflag.IsTerminal, NewFacade: migrationflag.NewFacade, NewWorker: migrationflag.NewWorker, })), migrationMasterName: ifNotDead(migrationmaster.Manifold(migrationmaster.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, FortressName: migrationFortressName, Clock: config.Clock, NewFacade: migrationmaster.NewFacade, NewWorker: config.NewMigrationMaster, })), // Everything else should be wrapped in ifResponsible, // ifNotAlive, ifNotDead, or ifNotMigrating (which also // implies NotDead), to ensure that only a single // controller is attempting to administer this model at // any one time. // // NOTE: not perfectly reliable at this stage? i.e. a // worker that ignores its stop signal for "too long" // might continue to take admin actions after the window // of responsibility closes. This *is* a pre-existing // problem, but demands some thought/care: e.g. should // we make sure the apiserver also closes any // connections that lose responsibility..? can we make // sure all possible environ operations are either time- // bounded or interruptible? etc // // On the other hand, all workers *should* be written in // the expectation of dealing with sucky infrastructure // running things in parallel unexpectedly, just because // the universe hates us and will engineer matters such // that it happens sometimes, even when we try to avoid // it. // The environ tracker could/should be used by several other // workers (firewaller, provisioners, address-cleaner?). environTrackerName: ifResponsible(environ.Manifold(environ.ManifoldConfig{ APICallerName: apiCallerName, NewEnvironFunc: config.NewEnvironFunc, })), // The undertaker is currently the only ifNotAlive worker. undertakerName: ifNotAlive(undertaker.Manifold(undertaker.ManifoldConfig{ APICallerName: apiCallerName, EnvironName: environTrackerName, NewFacade: undertaker.NewFacade, NewWorker: undertaker.NewWorker, })), // All the rest depend on ifNotMigrating. spaceImporterName: ifNotMigrating(discoverspaces.Manifold(discoverspaces.ManifoldConfig{ EnvironName: environTrackerName, APICallerName: apiCallerName, UnlockerName: spacesImportedGateName, NewFacade: discoverspaces.NewFacade, NewWorker: discoverspaces.NewWorker, })), computeProvisionerName: ifNotMigrating(provisioner.Manifold(provisioner.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, EnvironName: environTrackerName, NewProvisionerFunc: provisioner.NewEnvironProvisioner, })), storageProvisionerName: ifNotMigrating(storageprovisioner.ModelManifold(storageprovisioner.ModelManifoldConfig{ APICallerName: apiCallerName, ClockName: clockName, EnvironName: environTrackerName, Scope: modelTag, })), firewallerName: ifNotMigrating(firewaller.Manifold(firewaller.ManifoldConfig{ APICallerName: apiCallerName, })), unitAssignerName: ifNotMigrating(unitassigner.Manifold(unitassigner.ManifoldConfig{ APICallerName: apiCallerName, })), applicationScalerName: ifNotMigrating(applicationscaler.Manifold(applicationscaler.ManifoldConfig{ APICallerName: apiCallerName, NewFacade: applicationscaler.NewFacade, NewWorker: applicationscaler.New, })), instancePollerName: ifNotMigrating(instancepoller.Manifold(instancepoller.ManifoldConfig{ APICallerName: apiCallerName, EnvironName: environTrackerName, ClockName: clockName, Delay: config.InstPollerAggregationDelay, })), charmRevisionUpdaterName: ifNotMigrating(charmrevisionmanifold.Manifold(charmrevisionmanifold.ManifoldConfig{ APICallerName: apiCallerName, ClockName: clockName, Period: config.CharmRevisionUpdateInterval, NewFacade: charmrevisionmanifold.NewAPIFacade, NewWorker: charmrevision.NewWorker, })), metricWorkerName: ifNotMigrating(metricworker.Manifold(metricworker.ManifoldConfig{ APICallerName: apiCallerName, })), stateCleanerName: ifNotMigrating(cleaner.Manifold(cleaner.ManifoldConfig{ APICallerName: apiCallerName, })), statusHistoryPrunerName: ifNotMigrating(statushistorypruner.Manifold(statushistorypruner.ManifoldConfig{ APICallerName: apiCallerName, MaxHistoryTime: config.StatusHistoryPrunerMaxHistoryTime, MaxHistoryMB: config.StatusHistoryPrunerMaxHistoryMB, PruneInterval: config.StatusHistoryPrunerInterval, // TODO(fwereade): 2016-03-17 lp:1558657 NewTimer: worker.NewTimer, })), machineUndertakerName: ifNotMigrating(machineundertaker.Manifold(machineundertaker.ManifoldConfig{ APICallerName: apiCallerName, EnvironName: environTrackerName, NewWorker: machineundertaker.NewWorker, })), } }