// Return the SharedResources from our base stack that are needed for pool // stacks. We need the IDs for subnets and security groups, since they cannot // be referenced by name in a VPC. We also lookup the IAM instance profile // created by the base stack for use in pool's launch configs. This could be // cached to disk so that we don't need to lookup the base stack to build a // pool template. func GetSharedResources(stackName string) (SharedResources, error) { shared := SharedResources{ SecurityGroups: make(map[string]string), Roles: make(map[string]string), Parameters: make(map[string]string), ServerCerts: make(map[string]string), } // we need to use DescribeStacks to get any parameters that were used in // the base stack, such as KeyName descResp, err := DescribeStacks(stackName) if err != nil { return shared, err } // load all parameters from the base stack into the shared values for _, stack := range descResp.Stacks { if stack.Name == stackName { for _, param := range stack.Parameters { shared.Parameters[param.Key] = param.Value } } } res, err := ListStackResources(stackName) if err != nil { return shared, err } for _, resource := range res.Resources { switch resource.Type { case "AWS::EC2::SecurityGroup": shared.SecurityGroups[resource.LogicalId] = resource.PhysicalId case "AWS::IAM::InstanceProfile": shared.Roles[resource.LogicalId] = resource.PhysicalId case "AWS::EC2::VPC": shared.VPCID = resource.PhysicalId } } // NOTE: using default AZ snResp, err := DescribeSubnets(shared.VPCID, "") if err != nil { return shared, err } shared.Subnets = snResp.Subnets // now we need to find any server certs we may have certResp, err := ListServerCertificates() if err != nil { // we've made it this far, just log this error so we can at least get the CF data log.Error("error listing server certificates:", err) } for _, cert := range certResp.Certs { shared.ServerCerts[cert.ServerCertificateName] = cert.Arn } return shared, nil }
func SSHCmd(host string, command string, background bool, debug bool) { port := "22" hostPort := strings.SplitN(host, ":", 2) if len(hostPort) > 1 { host, port = hostPort[0], hostPort[1] } cmd := exec.Command("/usr/bin/ssh", "-o", "RequestTTY=yes", host, "-p", port, "-C", "/bin/sh", "-i", "-l", "-c", "'"+command+"'") cmd.Stdin = os.Stdin cmd.Stdout = os.Stdout var b []byte buf := bytes.NewBuffer(b) cmd.Stderr = buf err := cmd.Start() if err != nil { log.Fatal(err) } if err := cmd.Wait(); err != nil { log.Error(buf.String()) if exiterr, ok := err.(*exec.ExitError); ok { // The program has exited with an exit code != 0 // This works on both Unix and Windows. Although package // syscall is generally platform dependent, WaitStatus is // defined for both Unix and Windows and in both cases has // an ExitStatus() method with the same signature. if status, ok := exiterr.Sys().(syscall.WaitStatus); ok { fmt.Fprintf(os.Stderr, "Command finished with error: %v\n", err) os.Exit(status.ExitStatus()) } } else { fmt.Fprintf(os.Stderr, "Command finished with error: %v\n", err) os.Exit(1) } } }
func main() { flag.Int64Var(&stopCutoff, "cutoff", 10, "Seconds to wait before stopping old containers") flag.StringVar(®istryURL, "registry", utils.GetEnv("GALAXY_REGISTRY_URL", "redis://127.0.0.1:6379"), "registry URL") flag.StringVar(&env, "env", utils.GetEnv("GALAXY_ENV", ""), "Environment namespace") flag.StringVar(&pool, "pool", utils.GetEnv("GALAXY_POOL", ""), "Pool namespace") flag.StringVar(&hostIP, "host-ip", "127.0.0.1", "Host IP") flag.StringVar(&shuttleAddr, "shuttle-addr", "", "Shuttle API addr (127.0.0.1:9090)") flag.StringVar(&dns, "dns", "", "DNS addr to use for containers") flag.BoolVar(&debug, "debug", false, "verbose logging") flag.BoolVar(&version, "v", false, "display version info") flag.Usage = func() { println("Usage: commander [options] <command> [<args>]\n") println("Available commands are:") println(" agent Runs commander agent") println(" app List all apps") println(" app:assign Assign an app to a pool") println(" app:create Create an app") println(" app:deploy Deploy an app") println(" app:delete Delete an app") println(" app:restart Restart an app") println(" app:run Run a command within an app on this host") println(" app:shell Run a shell within an app on this host") println(" app:start Starts one or more apps") println(" app:stop Stops one or more apps") println(" app:unassign Unassign an app from a pool") println(" config List config for an app") println(" config:get Get config values for an app") println(" config:set Set config values for an app") println(" config:unset Unset config values for an app") println(" runtime List container runtime policies") println(" runtime:set Set container runtime policies") println(" hosts List hosts in an env and pool") println("\nOptions:\n") flag.PrintDefaults() } flag.Parse() if version { fmt.Println(buildVersion) return } log.DefaultLogger = log.New(os.Stdout, "", log.INFO) log.DefaultLogger.SetFlags(0) if debug { log.DefaultLogger.Level = log.DEBUG } if flag.NArg() < 1 { fmt.Println("Need a command") flag.Usage() os.Exit(1) } initOrDie() switch flag.Args()[0] { case "dump": if flag.NArg() < 2 { fmt.Println("Usage: commander dump ENV") os.Exit(1) } dump(flag.Arg(1)) return case "restore": if flag.NArg() < 2 { fmt.Println("Usage: commander dump ENV FILE") os.Exit(1) } restore(flag.Arg(1)) return case "agent": log.DefaultLogger.SetFlags(golog.LstdFlags) loop = true agentFs := flag.NewFlagSet("agent", flag.ExitOnError) agentFs.Usage = func() { println("Usage: commander agent [options]\n") println(" Runs commander continuously\n\n") println("Options:\n\n") agentFs.PrintDefaults() } agentFs.Parse(flag.Args()[1:]) ensureEnv() ensurePool() case "app": appFs := flag.NewFlagSet("app", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app\n") println(" List all apps or apps in an environment\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) err := commander.AppList(configStore, env) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:assign": appFs := flag.NewFlagSet("app:assign", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:assign <app>\n") println(" Assign an app to a pool\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() ensurePool() if appFs.NArg() != 1 { appFs.Usage() os.Exit(1) } err := commander.AppAssign(configStore, appFs.Args()[0], env, pool) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:create": appFs := flag.NewFlagSet("app:create", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:create <app>\n") println(" Create an app in an environment\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() if appFs.NArg() == 0 { appFs.Usage() os.Exit(1) } err := commander.AppCreate(configStore, appFs.Args()[0], env) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:delete": appFs := flag.NewFlagSet("app:delete", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:delete <app>\n") println(" Delete an app in an environment\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() if appFs.NArg() == 0 { appFs.Usage() os.Exit(1) } err := commander.AppDelete(configStore, appFs.Args()[0], env) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:deploy": appFs := flag.NewFlagSet("app:delete", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:deploy [-force] <app> <version>\n") println(" Deploy an app in an environment\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() if appFs.NArg() != 2 { appFs.Usage() os.Exit(1) } err := commander.AppDeploy(configStore, serviceRuntime, appFs.Args()[0], env, appFs.Args()[1]) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:restart": appFs := flag.NewFlagSet("app:restart", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:restart <app>\n") println(" Restart an app in an environment\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() if appFs.NArg() == 0 { appFs.Usage() os.Exit(1) } err := commander.AppRestart(configStore, appFs.Args()[0], env) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:run": appFs := flag.NewFlagSet("app:run", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:run <app> <cmd>\n") println(" Restart an app in an environment\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() if appFs.NArg() < 2 { appFs.Usage() os.Exit(1) } err := commander.AppRun(configStore, serviceRuntime, appFs.Args()[0], env, appFs.Args()[1:]) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:shell": appFs := flag.NewFlagSet("app:shell", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:shell <app>\n") println(" Run a shell for an app\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() ensurePool() if appFs.NArg() != 1 { appFs.Usage() os.Exit(1) } err := commander.AppShell(configStore, serviceRuntime, appFs.Args()[0], env, pool) if err != nil { log.Fatalf("ERROR: %s", err) } return case "app:start": startFs := flag.NewFlagSet("app:start", flag.ExitOnError) startFs.Usage = func() { println("Usage: commander app:start [options] [<app>]*\n") println(" Starts one or more apps. If no apps are specified, starts all apps.\n") println("Options:\n") startFs.PrintDefaults() } startFs.Parse(flag.Args()[1:]) apps = startFs.Args() if len(apps) == 0 { acs, err := configStore.ListApps(env) if err != nil { log.Fatalf("ERROR: Unable to list apps: %s", err) } for _, ac := range acs { apps = append(apps, ac.Name()) } } break case "app:status": // FIXME: undocumented statusFs := flag.NewFlagSet("app:status", flag.ExitOnError) statusFs.Usage = func() { println("Usage: commander app:status [options] [<app>]*\n") println(" Lists status of running apps.\n") println("Options:\n") statusFs.PrintDefaults() } statusFs.Parse(flag.Args()[1:]) ensureEnv() ensurePool() err := discovery.Status(serviceRuntime, configStore, env, pool, hostIP) if err != nil { log.Fatalf("ERROR: Unable to list app status: %s", err) } return case "app:stop": stopFs := flag.NewFlagSet("app:stop", flag.ExitOnError) stopFs.Usage = func() { println("Usage: commander app:stop [options] [<app>]*\n") println(" Stops one or more apps. If no apps are specified, stops all apps.\n") println("Options:\n") stopFs.PrintDefaults() } stopFs.Parse(flag.Args()[1:]) apps = stopFs.Args() for _, app := range apps { err := serviceRuntime.StopAllMatching(app) if err != nil { log.Fatalf("ERROR: Unable able to stop all containers: %s", err) } } if len(apps) > 0 { return } err := serviceRuntime.StopAll(env) if err != nil { log.Fatalf("ERROR: Unable able to stop all containers: %s", err) } return case "app:unassign": appFs := flag.NewFlagSet("app:unassign", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander app:unassign <app>\n") println(" Unassign an app to a pool\n") println("Options:\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() ensurePool() if appFs.NArg() != 1 { appFs.Usage() os.Exit(1) } err := commander.AppUnassign(configStore, appFs.Args()[0], env, pool) if err != nil { log.Fatalf("ERROR: %s", err) } return case "hosts": hostFs := flag.NewFlagSet("hosts", flag.ExitOnError) hostFs.Usage = func() { println("Usage: commander hosts\n") println(" List hosts in an env and pool\n") println("Options:\n") hostFs.PrintDefaults() } err := hostFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } ensureEnv() ensurePool() err = commander.HostsList(configStore, env, pool) if err != nil { log.Fatalf("ERROR: %s", err) } return case "config": configFs := flag.NewFlagSet("config", flag.ExitOnError) usage := "Usage: commander config <app>" configFs.Usage = func() { println(usage) println(" List config values for an app\n") println("Options:\n") configFs.PrintDefaults() } err := configFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } ensureEnv() if configFs.NArg() != 1 { log.Error("ERROR: Missing app name argument") log.Printf("Usage: %s", usage) os.Exit(1) } app := configFs.Args()[0] err = commander.ConfigList(configStore, app, env) if err != nil { log.Fatalf("ERROR: %s", err) } return case "config:get": configFs := flag.NewFlagSet("config:get", flag.ExitOnError) configFs.Usage = func() { println("Usage: commander config <app> KEY [KEY]*\n") println(" Get config values for an app\n") println("Options:\n") configFs.PrintDefaults() } err := configFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } ensureEnv() if configFs.NArg() == 0 { log.Errorf("ERROR: Missing app name") configFs.Usage() os.Exit(1) } app := configFs.Args()[0] err = commander.ConfigGet(configStore, app, env, configFs.Args()[1:]) if err != nil { log.Fatalf("ERROR: %s", err) } return case "config:set": configFs := flag.NewFlagSet("config:set", flag.ExitOnError) configFs.Usage = func() { println("Usage: commander config <app> KEY=VALUE [KEY=VALUE]*\n") println(" Set config values for an app\n") println("Options:\n") configFs.PrintDefaults() } err := configFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } ensureEnv() if configFs.NArg() == 0 { log.Errorf("ERROR: Missing app name") configFs.Usage() os.Exit(1) } app := configFs.Args()[0] err = commander.ConfigSet(configStore, app, env, configFs.Args()[1:]) if err != nil { log.Fatalf("ERROR: %s", err) } return case "config:unset": configFs := flag.NewFlagSet("config:unset", flag.ExitOnError) configFs.Usage = func() { println("Usage: commander config <app> KEY [KEY]*\n") println(" Unset config values for an app\n") println("Options:\n") configFs.PrintDefaults() } err := configFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } ensureEnv() if configFs.NArg() == 0 { log.Errorf("ERROR: Missing app name") configFs.Usage() os.Exit(1) } app := configFs.Args()[0] err = commander.ConfigUnset(configStore, app, env, configFs.Args()[1:]) if err != nil { log.Fatalf("ERROR: %s", err) } return case "runtime": runtimeFs := flag.NewFlagSet("runtime", flag.ExitOnError) runtimeFs.Usage = func() { println("Usage: commander runtime\n") println(" List container runtime policies\n") println("Options:\n") runtimeFs.PrintDefaults() } err := runtimeFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } app := "" if runtimeFs.NArg() > 0 { app = runtimeFs.Args()[0] } err = commander.RuntimeList(configStore, app, env, pool) if err != nil { log.Fatalf("ERROR: %s", err) } return case "runtime:set": var ps int var m string var c string var vhost string var port string var maint string runtimeFs := flag.NewFlagSet("runtime:set", flag.ExitOnError) runtimeFs.IntVar(&ps, "ps", 0, "Number of instances to run across all hosts") runtimeFs.StringVar(&m, "m", "", "Memory limit (format: <number><optional unit>, where unit = b, k, m or g)") runtimeFs.StringVar(&c, "c", "", "CPU shares (relative weight)") runtimeFs.StringVar(&vhost, "vhost", "", "Virtual host for HTTP routing") runtimeFs.StringVar(&port, "port", "", "Service port for service discovery") runtimeFs.StringVar(&maint, "maint", "", "Enable or disable maintenance mode") runtimeFs.Usage = func() { println("Usage: commander runtime:set [-ps 1] [-m 100m] [-c 512] [-vhost x.y.z] [-port 8000] [-maint false] <app>\n") println(" Set container runtime policies\n") println("Options:\n") runtimeFs.PrintDefaults() } err := runtimeFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } ensureEnv() if ps != 0 || m != "" || c != "" || maint != "" { ensurePool() } if runtimeFs.NArg() != 1 { runtimeFs.Usage() os.Exit(1) } app := runtimeFs.Args()[0] _, err = utils.ParseMemory(m) if err != nil { log.Fatalf("ERROR: Bad memory option %s: %s", m, err) } updated, err := commander.RuntimeSet(configStore, app, env, pool, commander.RuntimeOptions{ Ps: ps, Memory: m, CPUShares: c, VirtualHost: vhost, Port: port, MaintenanceMode: maint, }) if err != nil { log.Fatalf("ERROR: %s", err) } if !updated { log.Fatalf("ERROR: Failed to set runtime options.") } if pool != "" { log.Printf("Runtime options updated for %s in %s running on %s", app, env, pool) } else { log.Printf("Runtime options updated for %s in %s", app, env) } return case "runtime:unset": var ps, m, c, port bool var vhost string runtimeFs := flag.NewFlagSet("runtime:unset", flag.ExitOnError) runtimeFs.BoolVar(&ps, "ps", false, "Number of instances to run across all hosts") runtimeFs.BoolVar(&m, "m", false, "Memory limit") runtimeFs.BoolVar(&c, "c", false, "CPU shares (relative weight)") runtimeFs.StringVar(&vhost, "vhost", "", "Virtual host for HTTP routing") runtimeFs.BoolVar(&port, "port", false, "Service port for service discovery") runtimeFs.Usage = func() { println("Usage: commander runtime:unset [-ps] [-m] [-c] [-vhost x.y.z] [-port] <app>\n") println(" Reset and removes container runtime policies to defaults\n") println("Options:\n") runtimeFs.PrintDefaults() } err := runtimeFs.Parse(flag.Args()[1:]) if err != nil { log.Fatalf("ERROR: Bad command line options: %s", err) } ensureEnv() if ps || m || c { ensurePool() } if runtimeFs.NArg() != 1 { runtimeFs.Usage() os.Exit(1) } app := runtimeFs.Args()[0] options := commander.RuntimeOptions{ VirtualHost: vhost, } if ps { options.Ps = -1 } if m { options.Memory = "-" } if c { options.CPUShares = "-" } if port { options.Port = "-" } updated, err := commander.RuntimeUnset(configStore, app, env, pool, options) if err != nil { log.Fatalf("ERROR: %s", err) } if !updated { log.Fatalf("ERROR: Failed to set runtime options.") } if pool != "" { log.Printf("Runtime options updated for %s in %s running on %s", app, env, pool) } else { log.Printf("Runtime options updated for %s in %s", app, env) } return case "pool": err := commander.ListPools(configStore, env) if err != nil { log.Fatal(err) } return case "pool:create": appFs := flag.NewFlagSet("pool:create", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander -env <env> pool:create <pool>\n") println(" Create a pool in <env>\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() if pool == "" && appFs.NArg() > 0 { pool = appFs.Arg(0) } else { ensurePool() } err := commander.PoolCreate(configStore, env, pool) if err != nil { log.Fatalf("ERROR: Could not create pool: %s", err) } fmt.Println("created pool:", pool) return case "pool:delete": appFs := flag.NewFlagSet("pool:delete", flag.ExitOnError) appFs.Usage = func() { println("Usage: commander -env <env> pool:delete <pool>\n") println(" Delete a pool from <env>\n") appFs.PrintDefaults() } appFs.Parse(flag.Args()[1:]) ensureEnv() if pool == "" && flag.NArg() > 1 { pool = flag.Arg(1) } else { ensurePool() } err := commander.PoolDelete(configStore, env, pool) if err != nil { log.Fatalf("ERROR: Could not delete pool: %s", err) return } fmt.Println("deleted pool:", pool) return default: fmt.Println("Unknown command") flag.Usage() os.Exit(1) } ensureEnv() ensurePool() log.Printf("Starting commander %s", buildVersion) log.Printf("env=%s pool=%s host-ip=%s registry=%s shuttle-addr=%s dns=%s cutoff=%ds", env, pool, hostIP, registryURL, shuttleAddr, dns, stopCutoff) defer func() { configStore.DeleteHost(env, pool, config.HostInfo{ HostIP: hostIP, }) }() for app, ch := range workerChans { if len(apps) == 0 || utils.StringInSlice(app, apps) { wg.Add(1) go restartContainers(app, ch) ch <- "deploy" } } if loop { wg.Add(1) go heartbeatHost() go discovery.Register(serviceRuntime, configStore, env, pool, hostIP, shuttleAddr) cancelChan := make(chan struct{}) // do we need to cancel ever? restartChan := configStore.Watch(env, cancelChan) monitorService(restartChan) } // TODO: do we still need a WaitGroup? wg.Wait() }
func stackCreatePool(c *cli.Context) { var err error ensureEnvArg(c) ensurePoolArg(c) if c.String("region") != "" { stack.Region = c.String("region") } poolName := utils.GalaxyPool(c) baseStack := getBase(c) poolEnv := utils.GalaxyEnv(c) stackName := fmt.Sprintf("%s-%s-%s", baseStack, poolEnv, poolName) pool := stack.NewPool() // get the resources we need from the base stack // TODO: this may search for the base stack a second time resources := sharedResources(c, baseStack) desiredCap := c.Int("desired-size") if desiredCap == 0 { desiredCap = 1 } numZones := c.Int("availability-zones") if numZones == 0 { // default to running one host per zone numZones = desiredCap } minSize := c.Int("min-size") maxSize := c.Int("max-size") httpPort := c.Int("http-port") if httpPort == 0 { httpPort = 80 } sslCert := "" if cert := c.String("ssl-cert"); cert != "" { sslCert = resources.ServerCerts[cert] if sslCert == "" { log.Fatalf("Could not find certificate '%s'", cert) } } // Create our Launch Config lc := pool.LCTemplate lcName := "lc" + poolEnv + poolName if amiID := c.String("ami"); amiID != "" { lc.Properties.ImageId = amiID } else { lc.Properties.ImageId = resources.Parameters["PoolImageId"] } if insType := c.String("instance-type"); insType != "" { lc.Properties.InstanceType = insType } else { lc.Properties.InstanceType = resources.Parameters["PoolInstanceType"] } if keyName := c.String("keyname"); keyName != "" { lc.Properties.KeyName = keyName } else { lc.Properties.KeyName = resources.Parameters["KeyName"] } lc.Properties.IamInstanceProfile = resources.Roles["galaxyInstanceProfile"] lc.Properties.SecurityGroups = []string{ resources.SecurityGroups["sshSG"], resources.SecurityGroups["defaultSG"], } lc.SetVolumeSize(c.Int("volume-size")) pool.Resources[lcName] = lc // Create the Auto Scaling Group asg := pool.ASGTemplate asgName := "asg" + poolEnv + poolName asg.AddTag("Name", fmt.Sprintf("%s-%s-%s", baseStack, poolEnv, poolName), true) asg.AddTag("env", poolEnv, true) asg.AddTag("pool", poolName, true) asg.AddTag("galaxy", "pool", true) asg.Properties.DesiredCapacity = desiredCap // Don't always run in all zones subnets := resources.Subnets if numZones <= len(subnets) { subnets = subnets[:numZones] } else { log.Fatal("ERROR: cannot run in %d zones, only %d available.", numZones, len(subnets)) } // break the subnets info into separate subnet and AZ slices for the template subnetIDs := []string{} azIDs := []string{} for _, sn := range subnets { subnetIDs = append(subnetIDs, sn.ID) azIDs = append(azIDs, sn.AvailabilityZone) } asg.SetLaunchConfiguration(lcName) asg.Properties.AvailabilityZones = azIDs asg.Properties.VPCZoneIdentifier = subnetIDs if maxSize > 0 { asg.Properties.MaxSize = maxSize } if minSize > 0 { asg.Properties.MinSize = minSize } if c.Bool("auto-update") { asg.SetASGUpdatePolicy(c.Int("update-min"), c.Int("update-batch"), c.Duration("update-pause")) } pool.Resources[asgName] = asg // Optionally create the Elastic Load Balancer if c.Bool("elb") { elb := pool.ELBTemplate elbName := "elb" + poolEnv + poolName // make sure to add this to the ASG asg.AddLoadBalancer(elbName) elb.Properties.Subnets = subnetIDs elb.Properties.SecurityGroups = []string{ resources.SecurityGroups["webSG"], resources.SecurityGroups["defaultSG"], } elb.Properties.HealthCheck.Target = c.String("http-health-check") elb.AddListener(80, "HTTP", httpPort, "HTTP", "", nil) if sslCert != "" { elb.AddListener(443, "HTTPS", httpPort, "HTTP", sslCert, nil) } pool.Resources[elbName] = elb } // add autoscaling if it's required setCPUAutoScale(c, pool) poolTmpl, err := json.MarshalIndent(pool, "", " ") if err != nil { log.Fatal(err) } if c.Bool("print") { fmt.Println(string(poolTmpl)) return } opts := make(map[string]string) opts["tag.env"] = poolEnv opts["tag.pool"] = poolName opts["tag.galaxy"] = "pool" _, err = stack.Create(stackName, poolTmpl, opts) if err != nil { log.Fatal(err) } log.Println("Creating stack:", stackName) // do we want to wait on this by default? if err := stack.Wait(stackName, 5*time.Minute); err != nil { log.Error(err) log.Error("CreateStack Failed, attempting to delete") waitAndDelete(stackName) return } log.Println("CreateStack complete") }