func ExampleClient_Annotate_oneImage() { ctx := context.Background() client, err := vision.NewClient(ctx) if err != nil { // TODO: handle error. } annsSlice, err := client.Annotate(ctx, &vision.AnnotateRequest{ Image: vision.NewImageFromGCS("gs://my-bucket/my-image.png"), MaxLogos: 100, MaxTexts: 100, SafeSearch: true, }) if err != nil { // TODO: handle error. } anns := annsSlice[0] if anns.Logos != nil { fmt.Println(anns.Logos) } if anns.Texts != nil { fmt.Println(anns.Texts) } if anns.SafeSearch != nil { fmt.Println(anns.SafeSearch) } if anns.Error != nil { fmt.Printf("at least one of the features failed: %v", anns.Error) } }
func ExampleNewClient() { ctx := context.Background() client, err := vision.NewClient(ctx) if err != nil { // TODO: handle error. } // Use the client. // Close the client when finished. if err := client.Close(); err != nil { // TODO: handle error. } }
func ExampleClient_DetectFaces() { ctx := context.Background() client, err := vision.NewClient(ctx) if err != nil { // TODO: handle error. } img := vision.NewImageFromGCS("gs://my-bucket/my-image.png") faces, err := client.DetectFaces(ctx, img, 10) if err != nil { // TODO: handle error. } fmt.Println(faces[0].Face.Nose.Tip) fmt.Println(faces[0].Face.Eyes.Left.Pupil) }
// findLabels gets labels from the Vision API for an image at the given file path. func findLabels(file string) ([]string, error) { // [START init] ctx := context.Background() // Create the client. client, err := vision.NewClient(ctx) if err != nil { return nil, err } // [END init] // [START request] // Open the file. f, err := os.Open(file) if err != nil { return nil, err } image, err := vision.NewImageFromReader(f) if err != nil { return nil, err } // Perform the request. annotations, err := client.DetectLabels(ctx, image, 10) if err != nil { return nil, err } // [END request] // [START transform] var labels []string for _, annotation := range annotations { labels = append(labels, annotation.Description) } return labels, nil // [END transform] }