// Manifolds returns a set of co-configured manifolds covering the // various responsibilities of a machine agent. // // Thou Shalt Not Use String Literals In This Function. Or Else. func Manifolds(config ManifoldsConfig) dependency.Manifolds { // connectFilter exists: // 1) to let us retry api connections immediately on password change, // rather than causing the dependency engine to wait for a while; // 2) to ensure that certain connection failures correctly trigger // complete agent removal. (It's not safe to let any agent other // than the machine mess around with SetCanUninstall). connectFilter := func(err error) error { cause := errors.Cause(err) if cause == apicaller.ErrConnectImpossible { err2 := coreagent.SetCanUninstall(config.Agent) if err2 != nil { return errors.Trace(err2) } return worker.ErrTerminateAgent } else if cause == apicaller.ErrChangedPassword { return dependency.ErrBounce } return err } return dependency.Manifolds{ // The agent manifold references the enclosing agent, and is the // foundation stone on which most other manifolds ultimately depend. agentName: agent.Manifold(config.Agent), // The termination worker returns ErrTerminateAgent if a // termination signal is received by the process it's running // in. It has no inputs and its only output is the error it // returns. It depends on the uninstall file having been // written *by the manual provider* at install time; it would // be Very Wrong Indeed to use SetCanUninstall in conjunction // with this code. terminationName: terminationworker.Manifold(), // The stateconfigwatcher manifold watches the machine agent's // configuration and reports if state serving info is // present. It will bounce itself if state serving info is // added or removed. It is intended as a dependency just for // the state manifold. stateConfigWatcherName: stateconfigwatcher.Manifold(stateconfigwatcher.ManifoldConfig{ AgentName: agentName, AgentConfigChanged: config.AgentConfigChanged, }), // The state manifold creates a *state.State and makes it // available to other manifolds. It pings the mongodb session // regularly and will die if pings fail. stateName: workerstate.Manifold(workerstate.ManifoldConfig{ AgentName: agentName, StateConfigWatcherName: stateConfigWatcherName, OpenState: config.OpenState, }), // The stateworkers manifold starts workers which rely on a // *state.State but which haven't been converted to run // directly under the dependency engine yet. This manifold // will be removed once all such workers have been converted; // until then, the workers are expected to handle their own // checks for upgrades etc, rather than blocking this whole // worker on upgrade completion. stateWorkersName: StateWorkersManifold(StateWorkersConfig{ StateName: stateName, StartStateWorkers: config.StartStateWorkers, }), // The api-config-watcher manifold monitors the API server // addresses in the agent config and bounces when they // change. It's required as part of model migrations. apiConfigWatcherName: apiconfigwatcher.Manifold(apiconfigwatcher.ManifoldConfig{ AgentName: agentName, AgentConfigChanged: config.AgentConfigChanged, }), // The api caller is a thin concurrent wrapper around a connection // to some API server. It's used by many other manifolds, which all // select their own desired facades. It will be interesting to see // how this works when we consolidate the agents; might be best to // handle the auth changes server-side..? apiCallerName: apicaller.Manifold(apicaller.ManifoldConfig{ AgentName: agentName, APIConfigWatcherName: apiConfigWatcherName, APIOpen: apicaller.APIOpen, NewConnection: apicaller.ScaryConnect, Filter: connectFilter, }), // The upgrade steps gate is used to coordinate workers which // shouldn't do anything until the upgrade-steps worker has // finished running any required upgrade steps. The flag of // similar name is used to implement the isFullyUpgraded func // that keeps upgrade concerns out of unrelated manifolds. upgradeStepsGateName: gate.ManifoldEx(config.UpgradeStepsLock), upgradeStepsFlagName: gate.FlagManifold(gate.FlagManifoldConfig{ GateName: upgradeStepsGateName, NewWorker: gate.NewFlagWorker, }), // The upgrade check gate is used to coordinate workers which // shouldn't do anything until the upgrader worker has // completed its first check for a new tools version to // upgrade to. The flag of similar name is used to implement // the isFullyUpgraded func that keeps upgrade concerns out of // unrelated manifolds. upgradeCheckGateName: gate.ManifoldEx(config.UpgradeCheckLock), upgradeCheckFlagName: gate.FlagManifold(gate.FlagManifoldConfig{ GateName: upgradeCheckGateName, NewWorker: gate.NewFlagWorker, }), // The upgrader is a leaf worker that returns a specific error // type recognised by the machine agent, causing other workers // to be stopped and the agent to be restarted running the new // tools. We should only need one of these in a consolidated // agent, but we'll need to be careful about behavioural // differences, and interactions with the upgrade-steps // worker. upgraderName: upgrader.Manifold(upgrader.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, UpgradeStepsGateName: upgradeStepsGateName, UpgradeCheckGateName: upgradeCheckGateName, PreviousAgentVersion: config.PreviousAgentVersion, }), // The upgradesteps worker runs soon after the machine agent // starts and runs any steps required to upgrade to the // running jujud version. Once upgrade steps have run, the // upgradesteps gate is unlocked and the worker exits. upgradeStepsName: upgradesteps.Manifold(upgradesteps.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, UpgradeStepsGateName: upgradeStepsGateName, OpenStateForUpgrade: config.OpenStateForUpgrade, PreUpgradeSteps: config.PreUpgradeSteps, }), // The migration minion handles the agent side aspects of model migrations. migrationFortressName: ifFullyUpgraded(fortress.Manifold()), migrationMinionName: ifFullyUpgraded(migrationminion.Manifold(migrationminion.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, FortressName: migrationFortressName, NewFacade: migrationminion.NewFacade, NewWorker: migrationminion.NewWorker, })), // The serving-info-setter manifold sets grabs the state // serving info from the API connection and writes it to the // agent config. servingInfoSetterName: ifFullyUpgraded(ServingInfoSetterManifold(ServingInfoSetterConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The apiworkers manifold starts workers which rely on the // machine agent's API connection but have not been converted // to work directly under the dependency engine. It waits for // upgrades to be finished before starting these workers. apiWorkersName: ifFullyUpgraded(APIWorkersManifold(APIWorkersConfig{ APICallerName: apiCallerName, StartAPIWorkers: config.StartAPIWorkers, })), // The reboot manifold manages a worker which will reboot the // machine when requested. It needs an API connection and // waits for upgrades to be complete. rebootName: ifFullyUpgraded(reboot.Manifold(reboot.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The logging config updater is a leaf worker that indirectly // controls the messages sent via the log sender or rsyslog, // according to changes in environment config. We should only need // one of these in a consolidated agent. loggingConfigUpdaterName: ifFullyUpgraded(logger.Manifold(logger.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The diskmanager worker periodically lists block devices on the // machine it runs on. This worker will be run on all Juju-managed // machines (one per machine agent). diskManagerName: ifFullyUpgraded(diskmanager.Manifold(diskmanager.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The proxy config updater is a leaf worker that sets http/https/apt/etc // proxy settings. proxyConfigUpdater: ifFullyUpgraded(proxyupdater.Manifold(proxyupdater.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The api address updater is a leaf worker that rewrites agent config // as the state server addresses change. We should only need one of // these in a consolidated agent. apiAddressUpdaterName: ifFullyUpgraded(apiaddressupdater.Manifold(apiaddressupdater.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The machiner Worker will wait for the identified machine to become // Dying and make it Dead; or until the machine becomes Dead by other // means. machinerName: ifFullyUpgraded(machiner.Manifold(machiner.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The log sender is a leaf worker that sends log messages to some // API server, when configured so to do. We should only need one of // these in a consolidated agent. // // NOTE: the LogSource will buffer a large number of messages as an upgrade // runs; it currently seems better to fill the buffer and send when stable, // optimising for stable controller upgrades rather than up-to-the-moment // observable normal-machine upgrades. logSenderName: ifFullyUpgraded(logsender.Manifold(logsender.ManifoldConfig{ APICallerName: apiCallerName, LogSource: config.LogSource, })), // The deployer worker is responsible for deploying and recalling unit // agents, according to changes in a set of state units; and for the // final removal of its agents' units from state when they are no // longer needed. deployerName: ifFullyUpgraded(deployer.Manifold(deployer.ManifoldConfig{ NewDeployContext: config.NewDeployContext, AgentName: agentName, APICallerName: apiCallerName, })), authenticationWorkerName: ifFullyUpgraded(authenticationworker.Manifold(authenticationworker.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), // The storageProvisioner worker manages provisioning // (deprovisioning), and attachment (detachment) of first-class // volumes and filesystems. storageProvisionerName: ifFullyUpgraded(storageprovisioner.MachineManifold(storageprovisioner.MachineManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, Clock: config.Clock, })), resumerName: ifFullyUpgraded(resumer.Manifold(resumer.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), identityFileWriterName: ifFullyUpgraded(identityfilewriter.Manifold(identityfilewriter.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), toolsVersionCheckerName: ifFullyUpgraded(toolsversionchecker.Manifold(toolsversionchecker.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, })), machineActionName: ifFullyUpgraded(machineactions.Manifold(machineactions.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, NewFacade: machineactions.NewFacade, NewWorker: machineactions.NewMachineActionsWorker, })), hostKeyReporterName: ifFullyUpgraded(hostkeyreporter.Manifold(hostkeyreporter.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, RootDir: config.RootDir, NewFacade: hostkeyreporter.NewFacade, NewWorker: hostkeyreporter.NewWorker, })), } }
// Manifolds returns a set of co-configured manifolds covering the various // responsibilities of a standalone unit agent. It also accepts the logSource // argument because we haven't figured out how to thread all the logging bits // through a dependency engine yet. // // Thou Shalt Not Use String Literals In This Function. Or Else. func Manifolds(config ManifoldsConfig) dependency.Manifolds { // connectFilter exists to let us retry api connections immediately // on password change, rather than causing the dependency engine to // wait for a while. connectFilter := func(err error) error { cause := errors.Cause(err) if cause == apicaller.ErrChangedPassword { return dependency.ErrBounce } else if cause == apicaller.ErrConnectImpossible { return worker.ErrTerminateAgent } return err } return dependency.Manifolds{ // The agent manifold references the enclosing agent, and is the // foundation stone on which most other manifolds ultimately depend. // (Currently, that is "all manifolds", but consider a shared clock.) agentName: agent.Manifold(config.Agent), // The machine lock manifold is a thin concurrent wrapper around an // FSLock in an agreed location. We expect it to be replaced with an // in-memory lock when the unit agent moves into the machine agent. machineLockName: machinelock.Manifold(machinelock.ManifoldConfig{ AgentName: agentName, }), // The api-config-watcher manifold monitors the API server // addresses in the agent config and bounces when they // change. It's required as part of model migrations. apiConfigWatcherName: apiconfigwatcher.Manifold(apiconfigwatcher.ManifoldConfig{ AgentName: agentName, AgentConfigChanged: config.AgentConfigChanged, }), // The api caller is a thin concurrent wrapper around a connection // to some API server. It's used by many other manifolds, which all // select their own desired facades. It will be interesting to see // how this works when we consolidate the agents; might be best to // handle the auth changes server-side..? apiCallerName: apicaller.Manifold(apicaller.ManifoldConfig{ AgentName: agentName, APIConfigWatcherName: apiConfigWatcherName, APIOpen: apicaller.APIOpen, NewConnection: apicaller.ScaryConnect, Filter: connectFilter, }), // The log sender is a leaf worker that sends log messages to some // API server, when configured so to do. We should only need one of // these in a consolidated agent. logSenderName: logsender.Manifold(logsender.ManifoldConfig{ APICallerName: apiCallerName, LogSource: config.LogSource, }), // The upgrader is a leaf worker that returns a specific error type // recognised by the unit agent, causing other workers to be stopped // and the agent to be restarted running the new tools. We should only // need one of these in a consolidated agent, but we'll need to be // careful about behavioural differences, and interactions with the // upgradesteps worker. upgraderName: upgrader.Manifold(upgrader.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, }), migrationFortressName: fortress.Manifold(), // The migration minion handles the agent side aspects of model migrations. migrationMinionName: migrationminion.Manifold(migrationminion.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, FortressName: migrationFortressName, NewFacade: migrationminion.NewFacade, NewWorker: migrationminion.NewWorker, }), // The logging config updater is a leaf worker that indirectly // controls the messages sent via the log sender according to // changes in environment config. We should only need one of // these in a consolidated agent. loggingConfigUpdaterName: logger.Manifold(logger.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, }), // The api address updater is a leaf worker that rewrites agent config // as the controller addresses change. We should only need one of // these in a consolidated agent. apiAddressUpdaterName: apiaddressupdater.Manifold(apiaddressupdater.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, }), // The proxy config updater is a leaf worker that sets http/https/apt/etc // proxy settings. // TODO(fwereade): timing of this is suspicious. There was superstitious // code trying to run this early; if that ever helped, it was only by // coincidence. Probably we ought to be making components that might // need proxy config into explicit dependencies of the proxy updater... proxyConfigUpdaterName: proxyupdater.Manifold(proxyupdater.ManifoldConfig{ APICallerName: apiCallerName, }), // The charmdir resource coordinates whether the charm directory is // available or not; after 'start' hook and before 'stop' hook // executes, and not during upgrades. charmDirName: fortress.Manifold(), // The leadership tracker attempts to secure and retain leadership of // the unit's service, and is consulted on such matters by the // uniter. As it stannds today, we'll need one per unit in a // consolidated agent. leadershipTrackerName: leadership.Manifold(leadership.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, LeadershipGuarantee: config.LeadershipGuarantee, }), // HookRetryStrategy uses a retrystrategy worker to get a // retry strategy that will be used by the uniter to run its hooks. hookRetryStrategyName: retrystrategy.Manifold(retrystrategy.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, NewFacade: retrystrategy.NewFacade, NewWorker: retrystrategy.NewRetryStrategyWorker, }), // The uniter installs charms; manages the unit's presence in its // relations; creates suboordinate units; runs all the hooks; sends // metrics; etc etc etc. We expect to break it up further in the // coming weeks, and to need one per unit in a consolidated agent // (and probably one for each component broken out). uniterName: uniter.Manifold(uniter.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, LeadershipTrackerName: leadershipTrackerName, MachineLockName: machineLockName, CharmDirName: charmDirName, HookRetryStrategyName: hookRetryStrategyName, }), // TODO (mattyw) should be added to machine agent. metricSpoolName: spool.Manifold(spool.ManifoldConfig{ AgentName: agentName, }), // The metric collect worker executes the collect-metrics hook in a // restricted context that can safely run concurrently with other hooks. metricCollectName: collect.Manifold(collect.ManifoldConfig{ AgentName: agentName, MetricSpoolName: metricSpoolName, CharmDirName: charmDirName, }), // The meter status worker executes the meter-status-changed hook when it detects // that the meter status has changed. meterStatusName: meterstatus.Manifold(meterstatus.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, MachineLockName: machineLockName, NewHookRunner: meterstatus.NewHookRunner, NewMeterStatusAPIClient: msapi.NewClient, NewConnectedStatusWorker: meterstatus.NewConnectedStatusWorker, NewIsolatedStatusWorker: meterstatus.NewIsolatedStatusWorker, }), // The metric sender worker periodically sends accumulated metrics to the controller. metricSenderName: sender.Manifold(sender.ManifoldConfig{ AgentName: agentName, APICallerName: apiCallerName, MetricSpoolName: metricSpoolName, }), } }