func resourceAwsEfsMountTargetCreate(d *schema.ResourceData, meta interface{}) error { conn := meta.(*AWSClient).efsconn input := efs.CreateMountTargetInput{ FileSystemId: aws.String(d.Get("file_system_id").(string)), SubnetId: aws.String(d.Get("subnet_id").(string)), } if v, ok := d.GetOk("ip_address"); ok { input.IpAddress = aws.String(v.(string)) } if v, ok := d.GetOk("security_groups"); ok { input.SecurityGroups = expandStringList(v.(*schema.Set).List()) } log.Printf("[DEBUG] Creating EFS mount target: %#v", input) mt, err := conn.CreateMountTarget(&input) if err != nil { return err } d.SetId(*mt.MountTargetId) stateConf := &resource.StateChangeConf{ Pending: []string{"creating"}, Target: "available", Refresh: func() (interface{}, string, error) { resp, err := conn.DescribeMountTargets(&efs.DescribeMountTargetsInput{ MountTargetId: aws.String(d.Id()), }) if err != nil { return nil, "error", err } if len(resp.MountTargets) < 1 { return nil, "error", fmt.Errorf("EFS mount target %q not found", d.Id()) } mt := resp.MountTargets[0] log.Printf("[DEBUG] Current status of %q: %q", *mt.MountTargetId, *mt.LifeCycleState) return mt, *mt.LifeCycleState, nil }, Timeout: 10 * time.Minute, Delay: 2 * time.Second, MinTimeout: 3 * time.Second, } _, err = stateConf.WaitForState() if err != nil { return fmt.Errorf("Error waiting for EFS mount target (%s) to create: %s", d.Id(), err) } log.Printf("[DEBUG] EFS mount target created: %s", *mt.MountTargetId) return resourceAwsEfsMountTargetRead(d, meta) }
func resourceAwsEfsMountTargetCreate(d *schema.ResourceData, meta interface{}) error { conn := meta.(*AWSClient).efsconn fsId := d.Get("file_system_id").(string) subnetId := d.Get("subnet_id").(string) // CreateMountTarget would return the same Mount Target ID // to parallel requests if they both include the same AZ // and we would end up managing the same MT as 2 resources. // So we make it fail by calling 1 request per AZ at a time. az, err := getAzFromSubnetId(subnetId, meta.(*AWSClient).ec2conn) if err != nil { return fmt.Errorf("Failed getting Availability Zone from subnet ID (%s): %s", subnetId, err) } mtKey := "efs-mt-" + fsId + "-" + az awsMutexKV.Lock(mtKey) defer awsMutexKV.Unlock(mtKey) input := efs.CreateMountTargetInput{ FileSystemId: aws.String(fsId), SubnetId: aws.String(subnetId), } if v, ok := d.GetOk("ip_address"); ok { input.IpAddress = aws.String(v.(string)) } if v, ok := d.GetOk("security_groups"); ok { input.SecurityGroups = expandStringList(v.(*schema.Set).List()) } log.Printf("[DEBUG] Creating EFS mount target: %#v", input) mt, err := conn.CreateMountTarget(&input) if err != nil { return err } d.SetId(*mt.MountTargetId) log.Printf("[INFO] EFS mount target ID: %s", d.Id()) stateConf := &resource.StateChangeConf{ Pending: []string{"creating"}, Target: []string{"available"}, Refresh: func() (interface{}, string, error) { resp, err := conn.DescribeMountTargets(&efs.DescribeMountTargetsInput{ MountTargetId: aws.String(d.Id()), }) if err != nil { return nil, "error", err } if hasEmptyMountTargets(resp) { return nil, "error", fmt.Errorf("EFS mount target %q could not be found.", d.Id()) } mt := resp.MountTargets[0] log.Printf("[DEBUG] Current status of %q: %q", *mt.MountTargetId, *mt.LifeCycleState) return mt, *mt.LifeCycleState, nil }, Timeout: 10 * time.Minute, Delay: 2 * time.Second, MinTimeout: 3 * time.Second, } _, err = stateConf.WaitForState() if err != nil { return fmt.Errorf("Error waiting for EFS mount target (%s) to create: %s", d.Id(), err) } log.Printf("[DEBUG] EFS mount target created: %s", *mt.MountTargetId) return resourceAwsEfsMountTargetRead(d, meta) }