func TestStop(t *testing.T) { th.SetupHTTP() defer th.TeardownHTTP() mockStopServerResponse(t, serverID) err := startstop.Stop(client.ServiceClient(), serverID).ExtractErr() th.AssertNoErr(t, err) }
func (s *StepStopServer) Run(state multistep.StateBag) multistep.StepAction { ui := state.Get("ui").(packer.Ui) config := state.Get("config").(Config) extensions := state.Get("extensions").(map[string]struct{}) server := state.Get("server").(*servers.Server) // Verify we have the extension if _, ok := extensions["os-server-start-stop"]; !ok { ui.Say("OpenStack cluster doesn't support stop, skipping...") return multistep.ActionContinue } // We need the v2 compute client client, err := config.computeV2Client() if err != nil { err = fmt.Errorf("Error initializing compute client: %s", err) state.Put("error", err) return multistep.ActionHalt } ui.Say(fmt.Sprintf("Stopping server: %s ...", server.ID)) if err := startstop.Stop(client, server.ID).ExtractErr(); err != nil { err = fmt.Errorf("Error stopping server: %s", err) state.Put("error", err) return multistep.ActionHalt } ui.Message(fmt.Sprintf("Waiting for server to stop: %s ...", server.ID)) stateChange := StateChangeConf{ Pending: []string{"ACTIVE"}, Target: []string{"SHUTOFF", "STOPPED"}, Refresh: ServerStateRefreshFunc(client, server), StepState: state, } if _, err := WaitForState(&stateChange); err != nil { err := fmt.Errorf("Error waiting for server (%s) to stop: %s", server.ID, err) state.Put("error", err) ui.Error(err.Error()) return multistep.ActionHalt } return multistep.ActionContinue }
func resourceComputeInstanceV2Delete(d *schema.ResourceData, meta interface{}) error { config := meta.(*Config) computeClient, err := config.computeV2Client(GetRegion(d)) if err != nil { return fmt.Errorf("Error creating OpenStack compute client: %s", err) } // Make sure all volumes are detached before deleting volumes := d.Get("volume") if volumeSet, ok := volumes.(*schema.Set); ok { volumeList := volumeSet.List() if len(volumeList) > 0 { log.Printf("[DEBUG] Attempting to detach the following volumes: %#v", volumeList) if blockClient, err := config.blockStorageV1Client(GetRegion(d)); err != nil { return err } else { if err := detachVolumesFromInstance(computeClient, blockClient, d.Id(), volumeList); err != nil { return err } } } } if d.Get("stop_before_destroy").(bool) { err = startstop.Stop(computeClient, d.Id()).ExtractErr() if err != nil { log.Printf("[WARN] Error stopping OpenStack instance: %s", err) } else { stopStateConf := &resource.StateChangeConf{ Pending: []string{"ACTIVE"}, Target: []string{"SHUTOFF"}, Refresh: ServerV2StateRefreshFunc(computeClient, d.Id()), Timeout: 3 * time.Minute, Delay: 10 * time.Second, MinTimeout: 3 * time.Second, } log.Printf("[DEBUG] Waiting for instance (%s) to stop", d.Id()) _, err = stopStateConf.WaitForState() if err != nil { log.Printf("[WARN] Error waiting for instance (%s) to stop: %s, proceeding to delete", d.Id(), err) } } } err = servers.Delete(computeClient, d.Id()).ExtractErr() if err != nil { return fmt.Errorf("Error deleting OpenStack server: %s", err) } // Wait for the instance to delete before moving on. log.Printf("[DEBUG] Waiting for instance (%s) to delete", d.Id()) stateConf := &resource.StateChangeConf{ Pending: []string{"ACTIVE", "SHUTOFF"}, Target: []string{"DELETED"}, Refresh: ServerV2StateRefreshFunc(computeClient, d.Id()), Timeout: 30 * time.Minute, Delay: 10 * time.Second, MinTimeout: 3 * time.Second, } _, err = stateConf.WaitForState() if err != nil { return fmt.Errorf( "Error waiting for instance (%s) to delete: %s", d.Id(), err) } d.SetId("") return nil }