// Code generated by private/model/cli/gen-api/main.go. DO NOT EDIT. package rekognition import ( "github.com/aws/aws-sdk-go/aws" "github.com/aws/aws-sdk-go/aws/awsutil" "github.com/aws/aws-sdk-go/aws/request" ) const opCompareFaces = "CompareFaces" // CompareFacesRequest generates a "aws/request.Request" representing the // client's request for the CompareFaces operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See CompareFaces for more information on using the CompareFaces // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the CompareFacesRequest method. // req, resp := client.CompareFacesRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) CompareFacesRequest(input *CompareFacesInput) (req *request.Request, output *CompareFacesOutput) { op := &request.Operation{ Name: opCompareFaces, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &CompareFacesInput{} } output = &CompareFacesOutput{} req = c.newRequest(op, input, output) return } // CompareFaces API operation for Amazon Rekognition. // // Compares a face in the source input image with each of the 100 largest faces // detected in the target input image. // // If the source image contains multiple faces, the service detects the largest // face and compares it with each face detected in the target image. // // You pass the input and target images either as base64-encoded image bytes // or as a references to images in an Amazon S3 bucket. If you use the Amazon // CLI to call Amazon Rekognition operations, passing image bytes is not supported. // The image must be either a PNG or JPEG formatted file. // // In response, the operation returns an array of face matches ordered by similarity // score in descending order. For each face match, the response provides a bounding // box of the face, facial landmarks, pose details (pitch, role, and yaw), quality // (brightness and sharpness), and confidence value (indicating the level of // confidence that the bounding box contains a face). The response also provides // a similarity score, which indicates how closely the faces match. // // By default, only faces with a similarity score of greater than or equal to // 80% are returned in the response. You can change this value by specifying // the SimilarityThreshold parameter. // // CompareFaces also returns an array of faces that don't match the source image. // For each face, it returns a bounding box, confidence value, landmarks, pose // details, and quality. The response also returns information about the face // in the source image, including the bounding box of the face and confidence // value. // // If the image doesn't contain Exif metadata, CompareFaces returns orientation // information for the source and target images. Use these values to display // the images with the correct image orientation. // // If no faces are detected in the source or target images, CompareFaces returns // an InvalidParameterException error. // // This is a stateless API operation. That is, data returned by this operation // doesn't persist. // // For an example, see get-started-exercise-compare-faces. // // This operation requires permissions to perform the rekognition:CompareFaces // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation CompareFaces for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) CompareFaces(input *CompareFacesInput) (*CompareFacesOutput, error) { req, out := c.CompareFacesRequest(input) return out, req.Send() } // CompareFacesWithContext is the same as CompareFaces with the addition of // the ability to pass a context and additional request options. // // See CompareFaces for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) CompareFacesWithContext(ctx aws.Context, input *CompareFacesInput, opts ...request.Option) (*CompareFacesOutput, error) { req, out := c.CompareFacesRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opCreateCollection = "CreateCollection" // CreateCollectionRequest generates a "aws/request.Request" representing the // client's request for the CreateCollection operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See CreateCollection for more information on using the CreateCollection // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the CreateCollectionRequest method. // req, resp := client.CreateCollectionRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) CreateCollectionRequest(input *CreateCollectionInput) (req *request.Request, output *CreateCollectionOutput) { op := &request.Operation{ Name: opCreateCollection, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &CreateCollectionInput{} } output = &CreateCollectionOutput{} req = c.newRequest(op, input, output) return } // CreateCollection API operation for Amazon Rekognition. // // Creates a collection in an AWS Region. You can add faces to the collection // using the operation. // // For example, you might create collections, one for each of your application // users. A user can then index faces using the IndexFaces operation and persist // results in a specific collection. Then, a user can search the collection // for faces in the user-specific container. // // Collection names are case-sensitive. // // For an example, see example1. // // This operation requires permissions to perform the rekognition:CreateCollection // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation CreateCollection for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeResourceAlreadyExistsException "ResourceAlreadyExistsException" // A collection with the specified ID already exists. // func (c *Rekognition) CreateCollection(input *CreateCollectionInput) (*CreateCollectionOutput, error) { req, out := c.CreateCollectionRequest(input) return out, req.Send() } // CreateCollectionWithContext is the same as CreateCollection with the addition of // the ability to pass a context and additional request options. // // See CreateCollection for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) CreateCollectionWithContext(ctx aws.Context, input *CreateCollectionInput, opts ...request.Option) (*CreateCollectionOutput, error) { req, out := c.CreateCollectionRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opDeleteCollection = "DeleteCollection" // DeleteCollectionRequest generates a "aws/request.Request" representing the // client's request for the DeleteCollection operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See DeleteCollection for more information on using the DeleteCollection // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the DeleteCollectionRequest method. // req, resp := client.DeleteCollectionRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) DeleteCollectionRequest(input *DeleteCollectionInput) (req *request.Request, output *DeleteCollectionOutput) { op := &request.Operation{ Name: opDeleteCollection, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &DeleteCollectionInput{} } output = &DeleteCollectionOutput{} req = c.newRequest(op, input, output) return } // DeleteCollection API operation for Amazon Rekognition. // // Deletes the specified collection. Note that this operation removes all faces // in the collection. For an example, see example1. // // This operation requires permissions to perform the rekognition:DeleteCollection // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation DeleteCollection for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // func (c *Rekognition) DeleteCollection(input *DeleteCollectionInput) (*DeleteCollectionOutput, error) { req, out := c.DeleteCollectionRequest(input) return out, req.Send() } // DeleteCollectionWithContext is the same as DeleteCollection with the addition of // the ability to pass a context and additional request options. // // See DeleteCollection for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) DeleteCollectionWithContext(ctx aws.Context, input *DeleteCollectionInput, opts ...request.Option) (*DeleteCollectionOutput, error) { req, out := c.DeleteCollectionRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opDeleteFaces = "DeleteFaces" // DeleteFacesRequest generates a "aws/request.Request" representing the // client's request for the DeleteFaces operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See DeleteFaces for more information on using the DeleteFaces // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the DeleteFacesRequest method. // req, resp := client.DeleteFacesRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) DeleteFacesRequest(input *DeleteFacesInput) (req *request.Request, output *DeleteFacesOutput) { op := &request.Operation{ Name: opDeleteFaces, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &DeleteFacesInput{} } output = &DeleteFacesOutput{} req = c.newRequest(op, input, output) return } // DeleteFaces API operation for Amazon Rekognition. // // Deletes faces from a collection. You specify a collection ID and an array // of face IDs to remove from the collection. // // This operation requires permissions to perform the rekognition:DeleteFaces // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation DeleteFaces for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // func (c *Rekognition) DeleteFaces(input *DeleteFacesInput) (*DeleteFacesOutput, error) { req, out := c.DeleteFacesRequest(input) return out, req.Send() } // DeleteFacesWithContext is the same as DeleteFaces with the addition of // the ability to pass a context and additional request options. // // See DeleteFaces for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) DeleteFacesWithContext(ctx aws.Context, input *DeleteFacesInput, opts ...request.Option) (*DeleteFacesOutput, error) { req, out := c.DeleteFacesRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opDetectFaces = "DetectFaces" // DetectFacesRequest generates a "aws/request.Request" representing the // client's request for the DetectFaces operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See DetectFaces for more information on using the DetectFaces // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the DetectFacesRequest method. // req, resp := client.DetectFacesRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) DetectFacesRequest(input *DetectFacesInput) (req *request.Request, output *DetectFacesOutput) { op := &request.Operation{ Name: opDetectFaces, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &DetectFacesInput{} } output = &DetectFacesOutput{} req = c.newRequest(op, input, output) return } // DetectFaces API operation for Amazon Rekognition. // // Detects faces within an image that is provided as input. // // DetectFaces detects the 100 largest faces in the image. For each face detected, // the operation returns face details including a bounding box of the face, // a confidence value (that the bounding box contains a face), and a fixed set // of attributes such as facial landmarks (for example, coordinates of eye and // mouth), gender, presence of beard, sunglasses, etc. // // The face-detection algorithm is most effective on frontal faces. For non-frontal // or obscured faces, the algorithm may not detect the faces or might detect // faces with lower confidence. // // You pass the input image either as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon // Rekognition operations, passing image bytes is not supported. The image must // be either a PNG or JPEG formatted file. // // This is a stateless API operation. That is, the operation does not persist // any data. // // For an example, see get-started-exercise-detect-faces. // // This operation requires permissions to perform the rekognition:DetectFaces // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation DetectFaces for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) DetectFaces(input *DetectFacesInput) (*DetectFacesOutput, error) { req, out := c.DetectFacesRequest(input) return out, req.Send() } // DetectFacesWithContext is the same as DetectFaces with the addition of // the ability to pass a context and additional request options. // // See DetectFaces for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) DetectFacesWithContext(ctx aws.Context, input *DetectFacesInput, opts ...request.Option) (*DetectFacesOutput, error) { req, out := c.DetectFacesRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opDetectLabels = "DetectLabels" // DetectLabelsRequest generates a "aws/request.Request" representing the // client's request for the DetectLabels operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See DetectLabels for more information on using the DetectLabels // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the DetectLabelsRequest method. // req, resp := client.DetectLabelsRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) DetectLabelsRequest(input *DetectLabelsInput) (req *request.Request, output *DetectLabelsOutput) { op := &request.Operation{ Name: opDetectLabels, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &DetectLabelsInput{} } output = &DetectLabelsOutput{} req = c.newRequest(op, input, output) return } // DetectLabels API operation for Amazon Rekognition. // // Detects instances of real-world labels within an image (JPEG or PNG) provided // as input. This includes objects like flower, tree, and table; events like // wedding, graduation, and birthday party; and concepts like landscape, evening, // and nature. For an example, see get-started-exercise-detect-labels. // // You pass the input image as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon // Rekognition operations, passing image bytes is not supported. The image must // be either a PNG or JPEG formatted file. // // For each object, scene, and concept the API returns one or more labels. Each // label provides the object name, and the level of confidence that the image // contains the object. For example, suppose the input image has a lighthouse, // the sea, and a rock. The response will include all three labels, one for // each object. // // {Name: lighthouse, Confidence: 98.4629} // // {Name: rock,Confidence: 79.2097} // // {Name: sea,Confidence: 75.061} // // In the preceding example, the operation returns one label for each of the // three objects. The operation can also return multiple labels for the same // object in the image. For example, if the input image shows a flower (for // example, a tulip), the operation might return the following three labels. // // {Name: flower,Confidence: 99.0562} // // {Name: plant,Confidence: 99.0562} // // {Name: tulip,Confidence: 99.0562} // // In this example, the detection algorithm more precisely identifies the flower // as a tulip. // // In response, the API returns an array of labels. In addition, the response // also includes the orientation correction. Optionally, you can specify MinConfidence // to control the confidence threshold for the labels returned. The default // is 50%. You can also add the MaxLabels parameter to limit the number of labels // returned. // // If the object detected is a person, the operation doesn't provide the same // facial details that the DetectFaces operation provides. // // This is a stateless API operation. That is, the operation does not persist // any data. // // This operation requires permissions to perform the rekognition:DetectLabels // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation DetectLabels for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) DetectLabels(input *DetectLabelsInput) (*DetectLabelsOutput, error) { req, out := c.DetectLabelsRequest(input) return out, req.Send() } // DetectLabelsWithContext is the same as DetectLabels with the addition of // the ability to pass a context and additional request options. // // See DetectLabels for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) DetectLabelsWithContext(ctx aws.Context, input *DetectLabelsInput, opts ...request.Option) (*DetectLabelsOutput, error) { req, out := c.DetectLabelsRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opDetectModerationLabels = "DetectModerationLabels" // DetectModerationLabelsRequest generates a "aws/request.Request" representing the // client's request for the DetectModerationLabels operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See DetectModerationLabels for more information on using the DetectModerationLabels // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the DetectModerationLabelsRequest method. // req, resp := client.DetectModerationLabelsRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) DetectModerationLabelsRequest(input *DetectModerationLabelsInput) (req *request.Request, output *DetectModerationLabelsOutput) { op := &request.Operation{ Name: opDetectModerationLabels, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &DetectModerationLabelsInput{} } output = &DetectModerationLabelsOutput{} req = c.newRequest(op, input, output) return } // DetectModerationLabels API operation for Amazon Rekognition. // // Detects explicit or suggestive adult content in a specified JPEG or PNG format // image. Use DetectModerationLabels to moderate images depending on your requirements. // For example, you might want to filter images that contain nudity, but not // images containing suggestive content. // // To filter images, use the labels returned by DetectModerationLabels to determine // which types of content are appropriate. For information about moderation // labels, see image-moderation. // // You pass the input image either as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon // Rekognition operations, passing image bytes is not supported. The image must // be either a PNG or JPEG formatted file. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation DetectModerationLabels for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) DetectModerationLabels(input *DetectModerationLabelsInput) (*DetectModerationLabelsOutput, error) { req, out := c.DetectModerationLabelsRequest(input) return out, req.Send() } // DetectModerationLabelsWithContext is the same as DetectModerationLabels with the addition of // the ability to pass a context and additional request options. // // See DetectModerationLabels for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) DetectModerationLabelsWithContext(ctx aws.Context, input *DetectModerationLabelsInput, opts ...request.Option) (*DetectModerationLabelsOutput, error) { req, out := c.DetectModerationLabelsRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opDetectText = "DetectText" // DetectTextRequest generates a "aws/request.Request" representing the // client's request for the DetectText operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See DetectText for more information on using the DetectText // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the DetectTextRequest method. // req, resp := client.DetectTextRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) DetectTextRequest(input *DetectTextInput) (req *request.Request, output *DetectTextOutput) { op := &request.Operation{ Name: opDetectText, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &DetectTextInput{} } output = &DetectTextOutput{} req = c.newRequest(op, input, output) return } // DetectText API operation for Amazon Rekognition. // // Detects text in the input image and converts it into machine-readable text. // // Pass the input image as base64-encoded image bytes or as a reference to an // image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition // operations, you must pass it as a reference to an image in an Amazon S3 bucket. // For the AWS CLI, passing image bytes is not supported. The image must be // either a .png or .jpeg formatted file. // // The DetectText operation returns text in an array of elements, TextDetections. // Each TextDetection element provides information about a single word or line // of text that was detected in the image. // // A word is one or more ISO basic latin script characters that are not separated // by spaces. DetectText can detect up to 50 words in an image. // // A line is a string of equally spaced words. A line isn't necessarily a complete // sentence. For example, a driver's license number is detected as a line. A // line ends when there is no aligned text after it. Also, a line ends when // there is a large gap between words, relative to the length of the words. // This means, depending on the gap between words, Amazon Rekognition may detect // multiple lines in text aligned in the same direction. Periods don't represent // the end of a line. If a sentence spans multiple lines, the DetectText operation // returns multiple lines. // // To determine whether a TextDetection element is a line of text or a word, // use the TextDetection object Type field. // // To be detected, text must be within +/- 30 degrees orientation of the horizontal // axis. // // For more information, see text-detection. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation DetectText for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) DetectText(input *DetectTextInput) (*DetectTextOutput, error) { req, out := c.DetectTextRequest(input) return out, req.Send() } // DetectTextWithContext is the same as DetectText with the addition of // the ability to pass a context and additional request options. // // See DetectText for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) DetectTextWithContext(ctx aws.Context, input *DetectTextInput, opts ...request.Option) (*DetectTextOutput, error) { req, out := c.DetectTextRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opGetCelebrityInfo = "GetCelebrityInfo" // GetCelebrityInfoRequest generates a "aws/request.Request" representing the // client's request for the GetCelebrityInfo operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See GetCelebrityInfo for more information on using the GetCelebrityInfo // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the GetCelebrityInfoRequest method. // req, resp := client.GetCelebrityInfoRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) GetCelebrityInfoRequest(input *GetCelebrityInfoInput) (req *request.Request, output *GetCelebrityInfoOutput) { op := &request.Operation{ Name: opGetCelebrityInfo, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &GetCelebrityInfoInput{} } output = &GetCelebrityInfoOutput{} req = c.newRequest(op, input, output) return } // GetCelebrityInfo API operation for Amazon Rekognition. // // Gets the name and additional information about a celebrity based on his or // her Rekognition ID. The additional information is returned as an array of // URLs. If there is no additional information about the celebrity, this list // is empty. For more information, see celebrity-recognition. // // This operation requires permissions to perform the rekognition:GetCelebrityInfo // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation GetCelebrityInfo for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // func (c *Rekognition) GetCelebrityInfo(input *GetCelebrityInfoInput) (*GetCelebrityInfoOutput, error) { req, out := c.GetCelebrityInfoRequest(input) return out, req.Send() } // GetCelebrityInfoWithContext is the same as GetCelebrityInfo with the addition of // the ability to pass a context and additional request options. // // See GetCelebrityInfo for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) GetCelebrityInfoWithContext(ctx aws.Context, input *GetCelebrityInfoInput, opts ...request.Option) (*GetCelebrityInfoOutput, error) { req, out := c.GetCelebrityInfoRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opIndexFaces = "IndexFaces" // IndexFacesRequest generates a "aws/request.Request" representing the // client's request for the IndexFaces operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See IndexFaces for more information on using the IndexFaces // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the IndexFacesRequest method. // req, resp := client.IndexFacesRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) IndexFacesRequest(input *IndexFacesInput) (req *request.Request, output *IndexFacesOutput) { op := &request.Operation{ Name: opIndexFaces, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &IndexFacesInput{} } output = &IndexFacesOutput{} req = c.newRequest(op, input, output) return } // IndexFaces API operation for Amazon Rekognition. // // Detects faces in the input image and adds them to the specified collection. // // Amazon Rekognition does not save the actual faces detected. Instead, the // underlying detection algorithm first detects the faces in the input image, // and for each face extracts facial features into a feature vector, and stores // it in the back-end database. Amazon Rekognition uses feature vectors when // performing face match and search operations using the and operations. // // If you are using version 1.0 of the face detection model, IndexFaces indexes // the 15 largest faces in the input image. Later versions of the face detection // model index the 100 largest faces in the input image. To determine which // version of the model you are using, check the the value of FaceModelVersion // in the response from IndexFaces. For more information, see face-detection-model. // // If you provide the optional ExternalImageID for the input image you provided, // Amazon Rekognition associates this ID with all faces that it detects. When // you call the operation, the response returns the external ID. You can use // this external image ID to create a client-side index to associate the faces // with each image. You can then use the index to find all faces in an image. // // In response, the operation returns an array of metadata for all detected // faces. This includes, the bounding box of the detected face, confidence value // (indicating the bounding box contains a face), a face ID assigned by the // service for each face that is detected and stored, and an image ID assigned // by the service for the input image. If you request all facial attributes // (using the detectionAttributes parameter, Amazon Rekognition returns detailed // facial attributes such as facial landmarks (for example, location of eye // and mount) and other facial attributes such gender. If you provide the same // image, specify the same collection, and use the same external ID in the IndexFaces // operation, Amazon Rekognition doesn't save duplicate face metadata. // // The input image is passed either as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon // Rekognition operations, passing image bytes is not supported. The image must // be either a PNG or JPEG formatted file. // // For an example, see example2. // // This operation requires permissions to perform the rekognition:IndexFaces // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation IndexFaces for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) IndexFaces(input *IndexFacesInput) (*IndexFacesOutput, error) { req, out := c.IndexFacesRequest(input) return out, req.Send() } // IndexFacesWithContext is the same as IndexFaces with the addition of // the ability to pass a context and additional request options. // // See IndexFaces for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) IndexFacesWithContext(ctx aws.Context, input *IndexFacesInput, opts ...request.Option) (*IndexFacesOutput, error) { req, out := c.IndexFacesRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opListCollections = "ListCollections" // ListCollectionsRequest generates a "aws/request.Request" representing the // client's request for the ListCollections operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See ListCollections for more information on using the ListCollections // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the ListCollectionsRequest method. // req, resp := client.ListCollectionsRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) ListCollectionsRequest(input *ListCollectionsInput) (req *request.Request, output *ListCollectionsOutput) { op := &request.Operation{ Name: opListCollections, HTTPMethod: "POST", HTTPPath: "/", Paginator: &request.Paginator{ InputTokens: []string{"NextToken"}, OutputTokens: []string{"NextToken"}, LimitToken: "MaxResults", TruncationToken: "", }, } if input == nil { input = &ListCollectionsInput{} } output = &ListCollectionsOutput{} req = c.newRequest(op, input, output) return } // ListCollections API operation for Amazon Rekognition. // // Returns list of collection IDs in your account. If the result is truncated, // the response also provides a NextToken that you can use in the subsequent // request to fetch the next set of collection IDs. // // For an example, see example1. // // This operation requires permissions to perform the rekognition:ListCollections // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation ListCollections for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidPaginationTokenException "InvalidPaginationTokenException" // Pagination token in the request is not valid. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // func (c *Rekognition) ListCollections(input *ListCollectionsInput) (*ListCollectionsOutput, error) { req, out := c.ListCollectionsRequest(input) return out, req.Send() } // ListCollectionsWithContext is the same as ListCollections with the addition of // the ability to pass a context and additional request options. // // See ListCollections for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) ListCollectionsWithContext(ctx aws.Context, input *ListCollectionsInput, opts ...request.Option) (*ListCollectionsOutput, error) { req, out := c.ListCollectionsRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } // ListCollectionsPages iterates over the pages of a ListCollections operation, // calling the "fn" function with the response data for each page. To stop // iterating, return false from the fn function. // // See ListCollections method for more information on how to use this operation. // // Note: This operation can generate multiple requests to a service. // // // Example iterating over at most 3 pages of a ListCollections operation. // pageNum := 0 // err := client.ListCollectionsPages(params, // func(page *ListCollectionsOutput, lastPage bool) bool { // pageNum++ // fmt.Println(page) // return pageNum <= 3 // }) // func (c *Rekognition) ListCollectionsPages(input *ListCollectionsInput, fn func(*ListCollectionsOutput, bool) bool) error { return c.ListCollectionsPagesWithContext(aws.BackgroundContext(), input, fn) } // ListCollectionsPagesWithContext same as ListCollectionsPages except // it takes a Context and allows setting request options on the pages. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) ListCollectionsPagesWithContext(ctx aws.Context, input *ListCollectionsInput, fn func(*ListCollectionsOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ NewRequest: func() (*request.Request, error) { var inCpy *ListCollectionsInput if input != nil { tmp := *input inCpy = &tmp } req, _ := c.ListCollectionsRequest(inCpy) req.SetContext(ctx) req.ApplyOptions(opts...) return req, nil }, } cont := true for p.Next() && cont { cont = fn(p.Page().(*ListCollectionsOutput), !p.HasNextPage()) } return p.Err() } const opListFaces = "ListFaces" // ListFacesRequest generates a "aws/request.Request" representing the // client's request for the ListFaces operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See ListFaces for more information on using the ListFaces // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the ListFacesRequest method. // req, resp := client.ListFacesRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) ListFacesRequest(input *ListFacesInput) (req *request.Request, output *ListFacesOutput) { op := &request.Operation{ Name: opListFaces, HTTPMethod: "POST", HTTPPath: "/", Paginator: &request.Paginator{ InputTokens: []string{"NextToken"}, OutputTokens: []string{"NextToken"}, LimitToken: "MaxResults", TruncationToken: "", }, } if input == nil { input = &ListFacesInput{} } output = &ListFacesOutput{} req = c.newRequest(op, input, output) return } // ListFaces API operation for Amazon Rekognition. // // Returns metadata for faces in the specified collection. This metadata includes // information such as the bounding box coordinates, the confidence (that the // bounding box contains a face), and face ID. For an example, see example3. // // This operation requires permissions to perform the rekognition:ListFaces // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation ListFaces for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidPaginationTokenException "InvalidPaginationTokenException" // Pagination token in the request is not valid. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // func (c *Rekognition) ListFaces(input *ListFacesInput) (*ListFacesOutput, error) { req, out := c.ListFacesRequest(input) return out, req.Send() } // ListFacesWithContext is the same as ListFaces with the addition of // the ability to pass a context and additional request options. // // See ListFaces for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) ListFacesWithContext(ctx aws.Context, input *ListFacesInput, opts ...request.Option) (*ListFacesOutput, error) { req, out := c.ListFacesRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } // ListFacesPages iterates over the pages of a ListFaces operation, // calling the "fn" function with the response data for each page. To stop // iterating, return false from the fn function. // // See ListFaces method for more information on how to use this operation. // // Note: This operation can generate multiple requests to a service. // // // Example iterating over at most 3 pages of a ListFaces operation. // pageNum := 0 // err := client.ListFacesPages(params, // func(page *ListFacesOutput, lastPage bool) bool { // pageNum++ // fmt.Println(page) // return pageNum <= 3 // }) // func (c *Rekognition) ListFacesPages(input *ListFacesInput, fn func(*ListFacesOutput, bool) bool) error { return c.ListFacesPagesWithContext(aws.BackgroundContext(), input, fn) } // ListFacesPagesWithContext same as ListFacesPages except // it takes a Context and allows setting request options on the pages. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) ListFacesPagesWithContext(ctx aws.Context, input *ListFacesInput, fn func(*ListFacesOutput, bool) bool, opts ...request.Option) error { p := request.Pagination{ NewRequest: func() (*request.Request, error) { var inCpy *ListFacesInput if input != nil { tmp := *input inCpy = &tmp } req, _ := c.ListFacesRequest(inCpy) req.SetContext(ctx) req.ApplyOptions(opts...) return req, nil }, } cont := true for p.Next() && cont { cont = fn(p.Page().(*ListFacesOutput), !p.HasNextPage()) } return p.Err() } const opRecognizeCelebrities = "RecognizeCelebrities" // RecognizeCelebritiesRequest generates a "aws/request.Request" representing the // client's request for the RecognizeCelebrities operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See RecognizeCelebrities for more information on using the RecognizeCelebrities // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the RecognizeCelebritiesRequest method. // req, resp := client.RecognizeCelebritiesRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) RecognizeCelebritiesRequest(input *RecognizeCelebritiesInput) (req *request.Request, output *RecognizeCelebritiesOutput) { op := &request.Operation{ Name: opRecognizeCelebrities, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &RecognizeCelebritiesInput{} } output = &RecognizeCelebritiesOutput{} req = c.newRequest(op, input, output) return } // RecognizeCelebrities API operation for Amazon Rekognition. // // Returns an array of celebrities recognized in the input image. For more information, // see celebrity-recognition. // // RecognizeCelebrities returns the 100 largest faces in the image. It lists // recognized celebrities in the CelebrityFaces array and unrecognized faces // in the UnrecognizedFaces array. RecognizeCelebrities doesn't return celebrities // whose faces are not amongst the largest 100 faces in the image. // // For each celebrity recognized, the RecognizeCelebrities returns a Celebrity // object. The Celebrity object contains the celebrity name, ID, URL links to // additional information, match confidence, and a ComparedFace object that // you can use to locate the celebrity's face on the image. // // Rekognition does not retain information about which images a celebrity has // been recognized in. Your application must store this information and use // the Celebrity ID property as a unique identifier for the celebrity. If you // don't store the celebrity name or additional information URLs returned by // RecognizeCelebrities, you will need the ID to identify the celebrity in a // call to the operation. // // You pass the imput image either as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon // Rekognition operations, passing image bytes is not supported. The image must // be either a PNG or JPEG formatted file. // // For an example, see recognize-celebrities-tutorial. // // This operation requires permissions to perform the rekognition:RecognizeCelebrities // operation. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation RecognizeCelebrities for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) RecognizeCelebrities(input *RecognizeCelebritiesInput) (*RecognizeCelebritiesOutput, error) { req, out := c.RecognizeCelebritiesRequest(input) return out, req.Send() } // RecognizeCelebritiesWithContext is the same as RecognizeCelebrities with the addition of // the ability to pass a context and additional request options. // // See RecognizeCelebrities for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) RecognizeCelebritiesWithContext(ctx aws.Context, input *RecognizeCelebritiesInput, opts ...request.Option) (*RecognizeCelebritiesOutput, error) { req, out := c.RecognizeCelebritiesRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opSearchFaces = "SearchFaces" // SearchFacesRequest generates a "aws/request.Request" representing the // client's request for the SearchFaces operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See SearchFaces for more information on using the SearchFaces // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the SearchFacesRequest method. // req, resp := client.SearchFacesRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) SearchFacesRequest(input *SearchFacesInput) (req *request.Request, output *SearchFacesOutput) { op := &request.Operation{ Name: opSearchFaces, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &SearchFacesInput{} } output = &SearchFacesOutput{} req = c.newRequest(op, input, output) return } // SearchFaces API operation for Amazon Rekognition. // // For a given input face ID, searches for matching faces in the collection // the face belongs to. You get a face ID when you add a face to the collection // using the IndexFaces operation. The operation compares the features of the // input face with faces in the specified collection. // // You can also search faces without indexing faces by using the SearchFacesByImage // operation. // // The operation response returns an array of faces that match, ordered by similarity // score with the highest similarity first. More specifically, it is an array // of metadata for each face match that is found. Along with the metadata, the // response also includes a confidence value for each face match, indicating // the confidence that the specific face matches the input face. // // For an example, see example3. // // This operation requires permissions to perform the rekognition:SearchFaces // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation SearchFaces for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // func (c *Rekognition) SearchFaces(input *SearchFacesInput) (*SearchFacesOutput, error) { req, out := c.SearchFacesRequest(input) return out, req.Send() } // SearchFacesWithContext is the same as SearchFaces with the addition of // the ability to pass a context and additional request options. // // See SearchFaces for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) SearchFacesWithContext(ctx aws.Context, input *SearchFacesInput, opts ...request.Option) (*SearchFacesOutput, error) { req, out := c.SearchFacesRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } const opSearchFacesByImage = "SearchFacesByImage" // SearchFacesByImageRequest generates a "aws/request.Request" representing the // client's request for the SearchFacesByImage operation. The "output" return // value will be populated with the request's response once the request complets // successfuly. // // Use "Send" method on the returned Request to send the API call to the service. // the "output" return value is not valid until after Send returns without error. // // See SearchFacesByImage for more information on using the SearchFacesByImage // API call, and error handling. // // This method is useful when you want to inject custom logic or configuration // into the SDK's request lifecycle. Such as custom headers, or retry logic. // // // // Example sending a request using the SearchFacesByImageRequest method. // req, resp := client.SearchFacesByImageRequest(params) // // err := req.Send() // if err == nil { // resp is now filled // fmt.Println(resp) // } func (c *Rekognition) SearchFacesByImageRequest(input *SearchFacesByImageInput) (req *request.Request, output *SearchFacesByImageOutput) { op := &request.Operation{ Name: opSearchFacesByImage, HTTPMethod: "POST", HTTPPath: "/", } if input == nil { input = &SearchFacesByImageInput{} } output = &SearchFacesByImageOutput{} req = c.newRequest(op, input, output) return } // SearchFacesByImage API operation for Amazon Rekognition. // // For a given input image, first detects the largest face in the image, and // then searches the specified collection for matching faces. The operation // compares the features of the input face with faces in the specified collection. // // To search for all faces in an input image, you might first call the operation, // and then use the face IDs returned in subsequent calls to the operation. // // You can also call the DetectFaces operation and use the bounding boxes in // the response to make face crops, which then you can pass in to the SearchFacesByImage // operation. // // You pass the input image either as base64-encoded image bytes or as a reference // to an image in an Amazon S3 bucket. If you use the Amazon CLI to call Amazon // Rekognition operations, passing image bytes is not supported. The image must // be either a PNG or JPEG formatted file. // // The response returns an array of faces that match, ordered by similarity // score with the highest similarity first. More specifically, it is an array // of metadata for each face match found. Along with the metadata, the response // also includes a similarity indicating how similar the face is to the input // face. In the response, the operation also returns the bounding box (and a // confidence level that the bounding box contains a face) of the face that // Amazon Rekognition used for the input image. // // For an example, see example3. // // This operation requires permissions to perform the rekognition:SearchFacesByImage // action. // // Returns awserr.Error for service API and SDK errors. Use runtime type assertions // with awserr.Error's Code and Message methods to get detailed information about // the error. // // See the AWS API reference guide for Amazon Rekognition's // API operation SearchFacesByImage for usage and error information. // // Returned Error Codes: // * ErrCodeInvalidS3ObjectException "InvalidS3ObjectException" // Amazon Rekognition is unable to access the S3 object specified in the request. // // * ErrCodeInvalidParameterException "InvalidParameterException" // Input parameter violated a constraint. Validate your parameter before calling // the API operation again. // // * ErrCodeImageTooLargeException "ImageTooLargeException" // The input image size exceeds the allowed limit. For more information, see // limits. // // * ErrCodeAccessDeniedException "AccessDeniedException" // You are not authorized to perform the action. // // * ErrCodeInternalServerError "InternalServerError" // Amazon Rekognition experienced a service issue. Try your call again. // // * ErrCodeThrottlingException "ThrottlingException" // Amazon Rekognition is temporarily unable to process the request. Try your // call again. // // * ErrCodeProvisionedThroughputExceededException "ProvisionedThroughputExceededException" // The number of requests exceeded your throughput limit. If you want to increase // this limit, contact Amazon Rekognition. // // * ErrCodeResourceNotFoundException "ResourceNotFoundException" // The collection specified in the request cannot be found. // // * ErrCodeInvalidImageFormatException "InvalidImageFormatException" // The provided image format is not supported. // func (c *Rekognition) SearchFacesByImage(input *SearchFacesByImageInput) (*SearchFacesByImageOutput, error) { req, out := c.SearchFacesByImageRequest(input) return out, req.Send() } // SearchFacesByImageWithContext is the same as SearchFacesByImage with the addition of // the ability to pass a context and additional request options. // // See SearchFacesByImage for details on how to use this API operation. // // The context must be non-nil and will be used for request cancellation. If // the context is nil a panic will occur. In the future the SDK may create // sub-contexts for http.Requests. See https://golang.org/pkg/context/ // for more information on using Contexts. func (c *Rekognition) SearchFacesByImageWithContext(ctx aws.Context, input *SearchFacesByImageInput, opts ...request.Option) (*SearchFacesByImageOutput, error) { req, out := c.SearchFacesByImageRequest(input) req.SetContext(ctx) req.ApplyOptions(opts...) return out, req.Send() } // Structure containing the estimated age range, in years, for a face. // // Rekognition estimates an age-range for faces detected in the input image. // Estimated age ranges can overlap; a face of a 5 year old may have an estimated // range of 4-6 whilst the face of a 6 year old may have an estimated range // of 4-8. type AgeRange struct { _ struct{} `type:"structure"` // The highest estimated age. High *int64 `type:"integer"` // The lowest estimated age. Low *int64 `type:"integer"` } // String returns the string representation func (s AgeRange) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s AgeRange) GoString() string { return s.String() } // SetHigh sets the High field's value. func (s *AgeRange) SetHigh(v int64) *AgeRange { s.High = &v return s } // SetLow sets the Low field's value. func (s *AgeRange) SetLow(v int64) *AgeRange { s.Low = &v return s } // Indicates whether or not the face has a beard, and the confidence level in // the determination. type Beard struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Boolean value that indicates whether the face has beard or not. Value *bool `type:"boolean"` } // String returns the string representation func (s Beard) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Beard) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Beard) SetConfidence(v float64) *Beard { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *Beard) SetValue(v bool) *Beard { s.Value = &v return s } // Identifies the bounding box around the object, face or text. The left (x-coordinate) // and top (y-coordinate) are coordinates representing the top and left sides // of the bounding box. Note that the upper-left corner of the image is the // origin (0,0). // // The top and left values returned are ratios of the overall image size. For // example, if the input image is 700x200 pixels, and the top-left coordinate // of the bounding box is 350x50 pixels, the API returns a left value of 0.5 // (350/700) and a top value of 0.25 (50/200). // // The width and height values represent the dimensions of the bounding box // as a ratio of the overall image dimension. For example, if the input image // is 700x200 pixels, and the bounding box width is 70 pixels, the width returned // is 0.1. // // The bounding box coordinates can have negative values. For example, if Amazon // Rekognition is able to detect a face that is at the image edge and is only // partially visible, the service can return coordinates that are outside the // image bounds and, depending on the image edge, you might get negative values // or values greater than 1 for the left or top values. type BoundingBox struct { _ struct{} `type:"structure"` // Height of the bounding box as a ratio of the overall image height. Height *float64 `type:"float"` // Left coordinate of the bounding box as a ratio of overall image width. Left *float64 `type:"float"` // Top coordinate of the bounding box as a ratio of overall image height. Top *float64 `type:"float"` // Width of the bounding box as a ratio of the overall image width. Width *float64 `type:"float"` } // String returns the string representation func (s BoundingBox) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s BoundingBox) GoString() string { return s.String() } // SetHeight sets the Height field's value. func (s *BoundingBox) SetHeight(v float64) *BoundingBox { s.Height = &v return s } // SetLeft sets the Left field's value. func (s *BoundingBox) SetLeft(v float64) *BoundingBox { s.Left = &v return s } // SetTop sets the Top field's value. func (s *BoundingBox) SetTop(v float64) *BoundingBox { s.Top = &v return s } // SetWidth sets the Width field's value. func (s *BoundingBox) SetWidth(v float64) *BoundingBox { s.Width = &v return s } // Provides information about a celebrity recognized by the operation. type Celebrity struct { _ struct{} `type:"structure"` // Provides information about the celebrity's face, such as its location on // the image. Face *ComparedFace `type:"structure"` // A unique identifier for the celebrity. Id *string `type:"string"` // The confidence, in percentage, that Rekognition has that the recognized face // is the celebrity. MatchConfidence *float64 `type:"float"` // The name of the celebrity. Name *string `type:"string"` // An array of URLs pointing to additional information about the celebrity. // If there is no additional information about the celebrity, this list is empty. Urls []*string `type:"list"` } // String returns the string representation func (s Celebrity) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Celebrity) GoString() string { return s.String() } // SetFace sets the Face field's value. func (s *Celebrity) SetFace(v *ComparedFace) *Celebrity { s.Face = v return s } // SetId sets the Id field's value. func (s *Celebrity) SetId(v string) *Celebrity { s.Id = &v return s } // SetMatchConfidence sets the MatchConfidence field's value. func (s *Celebrity) SetMatchConfidence(v float64) *Celebrity { s.MatchConfidence = &v return s } // SetName sets the Name field's value. func (s *Celebrity) SetName(v string) *Celebrity { s.Name = &v return s } // SetUrls sets the Urls field's value. func (s *Celebrity) SetUrls(v []*string) *Celebrity { s.Urls = v return s } type CompareFacesInput struct { _ struct{} `type:"structure"` // The minimum level of confidence in the face matches that a match must meet // to be included in the FaceMatches array. SimilarityThreshold *float64 `type:"float"` // The input image as base64-encoded bytes or an S3 object. If you use the AWS // CLI to call Amazon Rekognition operations, passing base64-encoded image bytes // is not supported. // // SourceImage is a required field SourceImage *Image `type:"structure" required:"true"` // The target image as base64-encoded bytes or an S3 object. If you use the // AWS CLI to call Amazon Rekognition operations, passing base64-encoded image // bytes is not supported. // // TargetImage is a required field TargetImage *Image `type:"structure" required:"true"` } // String returns the string representation func (s CompareFacesInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s CompareFacesInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *CompareFacesInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "CompareFacesInput"} if s.SourceImage == nil { invalidParams.Add(request.NewErrParamRequired("SourceImage")) } if s.TargetImage == nil { invalidParams.Add(request.NewErrParamRequired("TargetImage")) } if s.SourceImage != nil { if err := s.SourceImage.Validate(); err != nil { invalidParams.AddNested("SourceImage", err.(request.ErrInvalidParams)) } } if s.TargetImage != nil { if err := s.TargetImage.Validate(); err != nil { invalidParams.AddNested("TargetImage", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetSimilarityThreshold sets the SimilarityThreshold field's value. func (s *CompareFacesInput) SetSimilarityThreshold(v float64) *CompareFacesInput { s.SimilarityThreshold = &v return s } // SetSourceImage sets the SourceImage field's value. func (s *CompareFacesInput) SetSourceImage(v *Image) *CompareFacesInput { s.SourceImage = v return s } // SetTargetImage sets the TargetImage field's value. func (s *CompareFacesInput) SetTargetImage(v *Image) *CompareFacesInput { s.TargetImage = v return s } // Provides information about a face in a target image that matches the source // image face analysed by CompareFaces. The Face property contains the bounding // box of the face in the target image. The Similarity property is the confidence // that the source image face matches the face in the bounding box. type CompareFacesMatch struct { _ struct{} `type:"structure"` // Provides face metadata (bounding box and confidence that the bounding box // actually contains a face). Face *ComparedFace `type:"structure"` // Level of confidence that the faces match. Similarity *float64 `type:"float"` } // String returns the string representation func (s CompareFacesMatch) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s CompareFacesMatch) GoString() string { return s.String() } // SetFace sets the Face field's value. func (s *CompareFacesMatch) SetFace(v *ComparedFace) *CompareFacesMatch { s.Face = v return s } // SetSimilarity sets the Similarity field's value. func (s *CompareFacesMatch) SetSimilarity(v float64) *CompareFacesMatch { s.Similarity = &v return s } type CompareFacesOutput struct { _ struct{} `type:"structure"` // An array of faces in the target image that match the source image face. Each // CompareFacesMatch object provides the bounding box, the confidence level // that the bounding box contains a face, and the similarity score for the face // in the bounding box and the face in the source image. FaceMatches []*CompareFacesMatch `type:"list"` // The face in the source image that was used for comparison. SourceImageFace *ComparedSourceImageFace `type:"structure"` // The orientation of the source image (counterclockwise direction). If your // application displays the source image, you can use this value to correct // image orientation. The bounding box coordinates returned in SourceImageFace // represent the location of the face before the image orientation is corrected. // // If the source image is in .jpeg format, it might contain exchangeable image // (Exif) metadata that includes the image's orientation. If the Exif metadata // for the source image populates the orientation field, the value of OrientationCorrection // is null and the SourceImageFace bounding box coordinates represent the location // of the face after Exif metadata is used to correct the orientation. Images // in .png format don't contain Exif metadata. SourceImageOrientationCorrection *string `type:"string" enum:"OrientationCorrection"` // The orientation of the target image (in counterclockwise direction). If your // application displays the target image, you can use this value to correct // the orientation of the image. The bounding box coordinates returned in FaceMatches // and UnmatchedFaces represent face locations before the image orientation // is corrected. // // If the target image is in .jpg format, it might contain Exif metadata that // includes the orientation of the image. If the Exif metadata for the target // image populates the orientation field, the value of OrientationCorrection // is null and the bounding box coordinates in FaceMatches and UnmatchedFaces // represent the location of the face after Exif metadata is used to correct // the orientation. Images in .png format don't contain Exif metadata. TargetImageOrientationCorrection *string `type:"string" enum:"OrientationCorrection"` // An array of faces in the target image that did not match the source image // face. UnmatchedFaces []*ComparedFace `type:"list"` } // String returns the string representation func (s CompareFacesOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s CompareFacesOutput) GoString() string { return s.String() } // SetFaceMatches sets the FaceMatches field's value. func (s *CompareFacesOutput) SetFaceMatches(v []*CompareFacesMatch) *CompareFacesOutput { s.FaceMatches = v return s } // SetSourceImageFace sets the SourceImageFace field's value. func (s *CompareFacesOutput) SetSourceImageFace(v *ComparedSourceImageFace) *CompareFacesOutput { s.SourceImageFace = v return s } // SetSourceImageOrientationCorrection sets the SourceImageOrientationCorrection field's value. func (s *CompareFacesOutput) SetSourceImageOrientationCorrection(v string) *CompareFacesOutput { s.SourceImageOrientationCorrection = &v return s } // SetTargetImageOrientationCorrection sets the TargetImageOrientationCorrection field's value. func (s *CompareFacesOutput) SetTargetImageOrientationCorrection(v string) *CompareFacesOutput { s.TargetImageOrientationCorrection = &v return s } // SetUnmatchedFaces sets the UnmatchedFaces field's value. func (s *CompareFacesOutput) SetUnmatchedFaces(v []*ComparedFace) *CompareFacesOutput { s.UnmatchedFaces = v return s } // Provides face metadata for target image faces that are analysed by CompareFaces // and RecognizeCelebrities. type ComparedFace struct { _ struct{} `type:"structure"` // Bounding box of the face. BoundingBox *BoundingBox `type:"structure"` // Level of confidence that what the bounding box contains is a face. Confidence *float64 `type:"float"` // An array of facial landmarks. Landmarks []*Landmark `type:"list"` // Indicates the pose of the face as determined by its pitch, roll, and yaw. Pose *Pose `type:"structure"` // Identifies face image brightness and sharpness. Quality *ImageQuality `type:"structure"` } // String returns the string representation func (s ComparedFace) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ComparedFace) GoString() string { return s.String() } // SetBoundingBox sets the BoundingBox field's value. func (s *ComparedFace) SetBoundingBox(v *BoundingBox) *ComparedFace { s.BoundingBox = v return s } // SetConfidence sets the Confidence field's value. func (s *ComparedFace) SetConfidence(v float64) *ComparedFace { s.Confidence = &v return s } // SetLandmarks sets the Landmarks field's value. func (s *ComparedFace) SetLandmarks(v []*Landmark) *ComparedFace { s.Landmarks = v return s } // SetPose sets the Pose field's value. func (s *ComparedFace) SetPose(v *Pose) *ComparedFace { s.Pose = v return s } // SetQuality sets the Quality field's value. func (s *ComparedFace) SetQuality(v *ImageQuality) *ComparedFace { s.Quality = v return s } // Type that describes the face Amazon Rekognition chose to compare with the // faces in the target. This contains a bounding box for the selected face and // confidence level that the bounding box contains a face. Note that Amazon // Rekognition selects the largest face in the source image for this comparison. type ComparedSourceImageFace struct { _ struct{} `type:"structure"` // Bounding box of the face. BoundingBox *BoundingBox `type:"structure"` // Confidence level that the selected bounding box contains a face. Confidence *float64 `type:"float"` } // String returns the string representation func (s ComparedSourceImageFace) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ComparedSourceImageFace) GoString() string { return s.String() } // SetBoundingBox sets the BoundingBox field's value. func (s *ComparedSourceImageFace) SetBoundingBox(v *BoundingBox) *ComparedSourceImageFace { s.BoundingBox = v return s } // SetConfidence sets the Confidence field's value. func (s *ComparedSourceImageFace) SetConfidence(v float64) *ComparedSourceImageFace { s.Confidence = &v return s } type CreateCollectionInput struct { _ struct{} `type:"structure"` // ID for the collection that you are creating. // // CollectionId is a required field CollectionId *string `min:"1" type:"string" required:"true"` } // String returns the string representation func (s CreateCollectionInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s CreateCollectionInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *CreateCollectionInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "CreateCollectionInput"} if s.CollectionId == nil { invalidParams.Add(request.NewErrParamRequired("CollectionId")) } if s.CollectionId != nil && len(*s.CollectionId) < 1 { invalidParams.Add(request.NewErrParamMinLen("CollectionId", 1)) } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetCollectionId sets the CollectionId field's value. func (s *CreateCollectionInput) SetCollectionId(v string) *CreateCollectionInput { s.CollectionId = &v return s } type CreateCollectionOutput struct { _ struct{} `type:"structure"` // Amazon Resource Name (ARN) of the collection. You can use this to manage // permissions on your resources. CollectionArn *string `type:"string"` // Version number of the face detection model associated with the collection // you are creating. FaceModelVersion *string `type:"string"` // HTTP status code indicating the result of the operation. StatusCode *int64 `type:"integer"` } // String returns the string representation func (s CreateCollectionOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s CreateCollectionOutput) GoString() string { return s.String() } // SetCollectionArn sets the CollectionArn field's value. func (s *CreateCollectionOutput) SetCollectionArn(v string) *CreateCollectionOutput { s.CollectionArn = &v return s } // SetFaceModelVersion sets the FaceModelVersion field's value. func (s *CreateCollectionOutput) SetFaceModelVersion(v string) *CreateCollectionOutput { s.FaceModelVersion = &v return s } // SetStatusCode sets the StatusCode field's value. func (s *CreateCollectionOutput) SetStatusCode(v int64) *CreateCollectionOutput { s.StatusCode = &v return s } type DeleteCollectionInput struct { _ struct{} `type:"structure"` // ID of the collection to delete. // // CollectionId is a required field CollectionId *string `min:"1" type:"string" required:"true"` } // String returns the string representation func (s DeleteCollectionInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DeleteCollectionInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *DeleteCollectionInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "DeleteCollectionInput"} if s.CollectionId == nil { invalidParams.Add(request.NewErrParamRequired("CollectionId")) } if s.CollectionId != nil && len(*s.CollectionId) < 1 { invalidParams.Add(request.NewErrParamMinLen("CollectionId", 1)) } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetCollectionId sets the CollectionId field's value. func (s *DeleteCollectionInput) SetCollectionId(v string) *DeleteCollectionInput { s.CollectionId = &v return s } type DeleteCollectionOutput struct { _ struct{} `type:"structure"` // HTTP status code that indicates the result of the operation. StatusCode *int64 `type:"integer"` } // String returns the string representation func (s DeleteCollectionOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DeleteCollectionOutput) GoString() string { return s.String() } // SetStatusCode sets the StatusCode field's value. func (s *DeleteCollectionOutput) SetStatusCode(v int64) *DeleteCollectionOutput { s.StatusCode = &v return s } type DeleteFacesInput struct { _ struct{} `type:"structure"` // Collection from which to remove the specific faces. // // CollectionId is a required field CollectionId *string `min:"1" type:"string" required:"true"` // An array of face IDs to delete. // // FaceIds is a required field FaceIds []*string `min:"1" type:"list" required:"true"` } // String returns the string representation func (s DeleteFacesInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DeleteFacesInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *DeleteFacesInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "DeleteFacesInput"} if s.CollectionId == nil { invalidParams.Add(request.NewErrParamRequired("CollectionId")) } if s.CollectionId != nil && len(*s.CollectionId) < 1 { invalidParams.Add(request.NewErrParamMinLen("CollectionId", 1)) } if s.FaceIds == nil { invalidParams.Add(request.NewErrParamRequired("FaceIds")) } if s.FaceIds != nil && len(s.FaceIds) < 1 { invalidParams.Add(request.NewErrParamMinLen("FaceIds", 1)) } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetCollectionId sets the CollectionId field's value. func (s *DeleteFacesInput) SetCollectionId(v string) *DeleteFacesInput { s.CollectionId = &v return s } // SetFaceIds sets the FaceIds field's value. func (s *DeleteFacesInput) SetFaceIds(v []*string) *DeleteFacesInput { s.FaceIds = v return s } type DeleteFacesOutput struct { _ struct{} `type:"structure"` // An array of strings (face IDs) of the faces that were deleted. DeletedFaces []*string `min:"1" type:"list"` } // String returns the string representation func (s DeleteFacesOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DeleteFacesOutput) GoString() string { return s.String() } // SetDeletedFaces sets the DeletedFaces field's value. func (s *DeleteFacesOutput) SetDeletedFaces(v []*string) *DeleteFacesOutput { s.DeletedFaces = v return s } type DetectFacesInput struct { _ struct{} `type:"structure"` // An array of facial attributes you want to be returned. This can be the default // list of attributes or all attributes. If you don't specify a value for Attributes // or if you specify ["DEFAULT"], the API returns the following subset of facial // attributes: BoundingBox, Confidence, Pose, Quality and Landmarks. If you // provide ["ALL"], all facial attributes are returned but the operation will // take longer to complete. // // If you provide both, ["ALL", "DEFAULT"], the service uses a logical AND operator // to determine which attributes to return (in this case, all attributes). Attributes []*string `type:"list"` // The input image as base64-encoded bytes or an S3 object. If you use the AWS // CLI to call Amazon Rekognition operations, passing base64-encoded image bytes // is not supported. // // Image is a required field Image *Image `type:"structure" required:"true"` } // String returns the string representation func (s DetectFacesInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectFacesInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *DetectFacesInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "DetectFacesInput"} if s.Image == nil { invalidParams.Add(request.NewErrParamRequired("Image")) } if s.Image != nil { if err := s.Image.Validate(); err != nil { invalidParams.AddNested("Image", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetAttributes sets the Attributes field's value. func (s *DetectFacesInput) SetAttributes(v []*string) *DetectFacesInput { s.Attributes = v return s } // SetImage sets the Image field's value. func (s *DetectFacesInput) SetImage(v *Image) *DetectFacesInput { s.Image = v return s } type DetectFacesOutput struct { _ struct{} `type:"structure"` // Details of each face found in the image. FaceDetails []*FaceDetail `type:"list"` // The orientation of the input image (counter-clockwise direction). If your // application displays the image, you can use this value to correct image orientation. // The bounding box coordinates returned in FaceDetails represent face locations // before the image orientation is corrected. // // If the input image is in .jpeg format, it might contain exchangeable image // (Exif) metadata that includes the image's orientation. If so, and the Exif // metadata for the input image populates the orientation field, the value of // OrientationCorrection is null and the FaceDetails bounding box coordinates // represent face locations after Exif metadata is used to correct the image // orientation. Images in .png format don't contain Exif metadata. OrientationCorrection *string `type:"string" enum:"OrientationCorrection"` } // String returns the string representation func (s DetectFacesOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectFacesOutput) GoString() string { return s.String() } // SetFaceDetails sets the FaceDetails field's value. func (s *DetectFacesOutput) SetFaceDetails(v []*FaceDetail) *DetectFacesOutput { s.FaceDetails = v return s } // SetOrientationCorrection sets the OrientationCorrection field's value. func (s *DetectFacesOutput) SetOrientationCorrection(v string) *DetectFacesOutput { s.OrientationCorrection = &v return s } type DetectLabelsInput struct { _ struct{} `type:"structure"` // The input image as base64-encoded bytes or an S3 object. If you use the AWS // CLI to call Amazon Rekognition operations, passing base64-encoded image bytes // is not supported. // // Image is a required field Image *Image `type:"structure" required:"true"` // Maximum number of labels you want the service to return in the response. // The service returns the specified number of highest confidence labels. MaxLabels *int64 `type:"integer"` // Specifies the minimum confidence level for the labels to return. Amazon Rekognition // doesn't return any labels with confidence lower than this specified value. // // If MinConfidence is not specified, the operation returns labels with a confidence // values greater than or equal to 50 percent. MinConfidence *float64 `type:"float"` } // String returns the string representation func (s DetectLabelsInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectLabelsInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *DetectLabelsInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "DetectLabelsInput"} if s.Image == nil { invalidParams.Add(request.NewErrParamRequired("Image")) } if s.Image != nil { if err := s.Image.Validate(); err != nil { invalidParams.AddNested("Image", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetImage sets the Image field's value. func (s *DetectLabelsInput) SetImage(v *Image) *DetectLabelsInput { s.Image = v return s } // SetMaxLabels sets the MaxLabels field's value. func (s *DetectLabelsInput) SetMaxLabels(v int64) *DetectLabelsInput { s.MaxLabels = &v return s } // SetMinConfidence sets the MinConfidence field's value. func (s *DetectLabelsInput) SetMinConfidence(v float64) *DetectLabelsInput { s.MinConfidence = &v return s } type DetectLabelsOutput struct { _ struct{} `type:"structure"` // An array of labels for the real-world objects detected. Labels []*Label `type:"list"` // The orientation of the input image (counter-clockwise direction). If your // application displays the image, you can use this value to correct the orientation. // If Amazon Rekognition detects that the input image was rotated (for example, // by 90 degrees), it first corrects the orientation before detecting the labels. // // If the input image Exif metadata populates the orientation field, Amazon // Rekognition does not perform orientation correction and the value of OrientationCorrection // will be null. OrientationCorrection *string `type:"string" enum:"OrientationCorrection"` } // String returns the string representation func (s DetectLabelsOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectLabelsOutput) GoString() string { return s.String() } // SetLabels sets the Labels field's value. func (s *DetectLabelsOutput) SetLabels(v []*Label) *DetectLabelsOutput { s.Labels = v return s } // SetOrientationCorrection sets the OrientationCorrection field's value. func (s *DetectLabelsOutput) SetOrientationCorrection(v string) *DetectLabelsOutput { s.OrientationCorrection = &v return s } type DetectModerationLabelsInput struct { _ struct{} `type:"structure"` // The input image as base64-encoded bytes or an S3 object. If you use the AWS // CLI to call Amazon Rekognition operations, passing base64-encoded image bytes // is not supported. // // Image is a required field Image *Image `type:"structure" required:"true"` // Specifies the minimum confidence level for the labels to return. Amazon Rekognition // doesn't return any labels with a confidence level lower than this specified // value. // // If you don't specify MinConfidence, the operation returns labels with confidence // values greater than or equal to 50 percent. MinConfidence *float64 `type:"float"` } // String returns the string representation func (s DetectModerationLabelsInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectModerationLabelsInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *DetectModerationLabelsInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "DetectModerationLabelsInput"} if s.Image == nil { invalidParams.Add(request.NewErrParamRequired("Image")) } if s.Image != nil { if err := s.Image.Validate(); err != nil { invalidParams.AddNested("Image", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetImage sets the Image field's value. func (s *DetectModerationLabelsInput) SetImage(v *Image) *DetectModerationLabelsInput { s.Image = v return s } // SetMinConfidence sets the MinConfidence field's value. func (s *DetectModerationLabelsInput) SetMinConfidence(v float64) *DetectModerationLabelsInput { s.MinConfidence = &v return s } type DetectModerationLabelsOutput struct { _ struct{} `type:"structure"` // An array of labels for explicit or suggestive adult content found in the // image. The list includes the top-level label and each second-level label // detected in the image. This is useful for filtering specific categories of // content. ModerationLabels []*ModerationLabel `type:"list"` } // String returns the string representation func (s DetectModerationLabelsOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectModerationLabelsOutput) GoString() string { return s.String() } // SetModerationLabels sets the ModerationLabels field's value. func (s *DetectModerationLabelsOutput) SetModerationLabels(v []*ModerationLabel) *DetectModerationLabelsOutput { s.ModerationLabels = v return s } type DetectTextInput struct { _ struct{} `type:"structure"` // The input image as base64-encoded bytes or an Amazon S3 object. If you use // the AWS CLI to call Amazon Rekognition operations, you can't pass image bytes. // // Image is a required field Image *Image `type:"structure" required:"true"` } // String returns the string representation func (s DetectTextInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectTextInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *DetectTextInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "DetectTextInput"} if s.Image == nil { invalidParams.Add(request.NewErrParamRequired("Image")) } if s.Image != nil { if err := s.Image.Validate(); err != nil { invalidParams.AddNested("Image", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetImage sets the Image field's value. func (s *DetectTextInput) SetImage(v *Image) *DetectTextInput { s.Image = v return s } type DetectTextOutput struct { _ struct{} `type:"structure"` // An array of text that was detected in the input image. TextDetections []*TextDetection `type:"list"` } // String returns the string representation func (s DetectTextOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s DetectTextOutput) GoString() string { return s.String() } // SetTextDetections sets the TextDetections field's value. func (s *DetectTextOutput) SetTextDetections(v []*TextDetection) *DetectTextOutput { s.TextDetections = v return s } // The emotions detected on the face, and the confidence level in the determination. // For example, HAPPY, SAD, and ANGRY. type Emotion struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Type of emotion detected. Type *string `type:"string" enum:"EmotionName"` } // String returns the string representation func (s Emotion) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Emotion) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Emotion) SetConfidence(v float64) *Emotion { s.Confidence = &v return s } // SetType sets the Type field's value. func (s *Emotion) SetType(v string) *Emotion { s.Type = &v return s } // Indicates whether or not the eyes on the face are open, and the confidence // level in the determination. type EyeOpen struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Boolean value that indicates whether the eyes on the face are open. Value *bool `type:"boolean"` } // String returns the string representation func (s EyeOpen) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s EyeOpen) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *EyeOpen) SetConfidence(v float64) *EyeOpen { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *EyeOpen) SetValue(v bool) *EyeOpen { s.Value = &v return s } // Indicates whether or not the face is wearing eye glasses, and the confidence // level in the determination. type Eyeglasses struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Boolean value that indicates whether the face is wearing eye glasses or not. Value *bool `type:"boolean"` } // String returns the string representation func (s Eyeglasses) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Eyeglasses) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Eyeglasses) SetConfidence(v float64) *Eyeglasses { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *Eyeglasses) SetValue(v bool) *Eyeglasses { s.Value = &v return s } // Describes the face properties such as the bounding box, face ID, image ID // of the input image, and external image ID that you assigned. type Face struct { _ struct{} `type:"structure"` // Bounding box of the face. BoundingBox *BoundingBox `type:"structure"` // Confidence level that the bounding box contains a face (and not a different // object such as a tree). Confidence *float64 `type:"float"` // Identifier that you assign to all the faces in the input image. ExternalImageId *string `min:"1" type:"string"` // Unique identifier that Amazon Rekognition assigns to the face. FaceId *string `type:"string"` // Unique identifier that Amazon Rekognition assigns to the input image. ImageId *string `type:"string"` } // String returns the string representation func (s Face) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Face) GoString() string { return s.String() } // SetBoundingBox sets the BoundingBox field's value. func (s *Face) SetBoundingBox(v *BoundingBox) *Face { s.BoundingBox = v return s } // SetConfidence sets the Confidence field's value. func (s *Face) SetConfidence(v float64) *Face { s.Confidence = &v return s } // SetExternalImageId sets the ExternalImageId field's value. func (s *Face) SetExternalImageId(v string) *Face { s.ExternalImageId = &v return s } // SetFaceId sets the FaceId field's value. func (s *Face) SetFaceId(v string) *Face { s.FaceId = &v return s } // SetImageId sets the ImageId field's value. func (s *Face) SetImageId(v string) *Face { s.ImageId = &v return s } // Structure containing attributes of the face that the algorithm detected. type FaceDetail struct { _ struct{} `type:"structure"` // The estimated age range, in years, for the face. Low represents the lowest // estimated age and High represents the highest estimated age. AgeRange *AgeRange `type:"structure"` // Indicates whether or not the face has a beard, and the confidence level in // the determination. Beard *Beard `type:"structure"` // Bounding box of the face. BoundingBox *BoundingBox `type:"structure"` // Confidence level that the bounding box contains a face (and not a different // object such as a tree). Confidence *float64 `type:"float"` // The emotions detected on the face, and the confidence level in the determination. // For example, HAPPY, SAD, and ANGRY. Emotions []*Emotion `type:"list"` // Indicates whether or not the face is wearing eye glasses, and the confidence // level in the determination. Eyeglasses *Eyeglasses `type:"structure"` // Indicates whether or not the eyes on the face are open, and the confidence // level in the determination. EyesOpen *EyeOpen `type:"structure"` // Gender of the face and the confidence level in the determination. Gender *Gender `type:"structure"` // Indicates the location of landmarks on the face. Landmarks []*Landmark `type:"list"` // Indicates whether or not the mouth on the face is open, and the confidence // level in the determination. MouthOpen *MouthOpen `type:"structure"` // Indicates whether or not the face has a mustache, and the confidence level // in the determination. Mustache *Mustache `type:"structure"` // Indicates the pose of the face as determined by its pitch, roll, and yaw. Pose *Pose `type:"structure"` // Identifies image brightness and sharpness. Quality *ImageQuality `type:"structure"` // Indicates whether or not the face is smiling, and the confidence level in // the determination. Smile *Smile `type:"structure"` // Indicates whether or not the face is wearing sunglasses, and the confidence // level in the determination. Sunglasses *Sunglasses `type:"structure"` } // String returns the string representation func (s FaceDetail) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s FaceDetail) GoString() string { return s.String() } // SetAgeRange sets the AgeRange field's value. func (s *FaceDetail) SetAgeRange(v *AgeRange) *FaceDetail { s.AgeRange = v return s } // SetBeard sets the Beard field's value. func (s *FaceDetail) SetBeard(v *Beard) *FaceDetail { s.Beard = v return s } // SetBoundingBox sets the BoundingBox field's value. func (s *FaceDetail) SetBoundingBox(v *BoundingBox) *FaceDetail { s.BoundingBox = v return s } // SetConfidence sets the Confidence field's value. func (s *FaceDetail) SetConfidence(v float64) *FaceDetail { s.Confidence = &v return s } // SetEmotions sets the Emotions field's value. func (s *FaceDetail) SetEmotions(v []*Emotion) *FaceDetail { s.Emotions = v return s } // SetEyeglasses sets the Eyeglasses field's value. func (s *FaceDetail) SetEyeglasses(v *Eyeglasses) *FaceDetail { s.Eyeglasses = v return s } // SetEyesOpen sets the EyesOpen field's value. func (s *FaceDetail) SetEyesOpen(v *EyeOpen) *FaceDetail { s.EyesOpen = v return s } // SetGender sets the Gender field's value. func (s *FaceDetail) SetGender(v *Gender) *FaceDetail { s.Gender = v return s } // SetLandmarks sets the Landmarks field's value. func (s *FaceDetail) SetLandmarks(v []*Landmark) *FaceDetail { s.Landmarks = v return s } // SetMouthOpen sets the MouthOpen field's value. func (s *FaceDetail) SetMouthOpen(v *MouthOpen) *FaceDetail { s.MouthOpen = v return s } // SetMustache sets the Mustache field's value. func (s *FaceDetail) SetMustache(v *Mustache) *FaceDetail { s.Mustache = v return s } // SetPose sets the Pose field's value. func (s *FaceDetail) SetPose(v *Pose) *FaceDetail { s.Pose = v return s } // SetQuality sets the Quality field's value. func (s *FaceDetail) SetQuality(v *ImageQuality) *FaceDetail { s.Quality = v return s } // SetSmile sets the Smile field's value. func (s *FaceDetail) SetSmile(v *Smile) *FaceDetail { s.Smile = v return s } // SetSunglasses sets the Sunglasses field's value. func (s *FaceDetail) SetSunglasses(v *Sunglasses) *FaceDetail { s.Sunglasses = v return s } // Provides face metadata. In addition, it also provides the confidence in the // match of this face with the input face. type FaceMatch struct { _ struct{} `type:"structure"` // Describes the face properties such as the bounding box, face ID, image ID // of the source image, and external image ID that you assigned. Face *Face `type:"structure"` // Confidence in the match of this face with the input face. Similarity *float64 `type:"float"` } // String returns the string representation func (s FaceMatch) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s FaceMatch) GoString() string { return s.String() } // SetFace sets the Face field's value. func (s *FaceMatch) SetFace(v *Face) *FaceMatch { s.Face = v return s } // SetSimilarity sets the Similarity field's value. func (s *FaceMatch) SetSimilarity(v float64) *FaceMatch { s.Similarity = &v return s } // Object containing both the face metadata (stored in the back-end database) // and facial attributes that are detected but aren't stored in the database. type FaceRecord struct { _ struct{} `type:"structure"` // Describes the face properties such as the bounding box, face ID, image ID // of the input image, and external image ID that you assigned. Face *Face `type:"structure"` // Structure containing attributes of the face that the algorithm detected. FaceDetail *FaceDetail `type:"structure"` } // String returns the string representation func (s FaceRecord) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s FaceRecord) GoString() string { return s.String() } // SetFace sets the Face field's value. func (s *FaceRecord) SetFace(v *Face) *FaceRecord { s.Face = v return s } // SetFaceDetail sets the FaceDetail field's value. func (s *FaceRecord) SetFaceDetail(v *FaceDetail) *FaceRecord { s.FaceDetail = v return s } // Gender of the face and the confidence level in the determination. type Gender struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Gender of the face. Value *string `type:"string" enum:"GenderType"` } // String returns the string representation func (s Gender) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Gender) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Gender) SetConfidence(v float64) *Gender { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *Gender) SetValue(v string) *Gender { s.Value = &v return s } // Information about where text detected by is located on an image. type Geometry struct { _ struct{} `type:"structure"` // An axis-aligned coarse representation of the detected text's location on // the image. BoundingBox *BoundingBox `type:"structure"` // Within the bounding box, a fine-grained polygon around the detected text. Polygon []*Point `type:"list"` } // String returns the string representation func (s Geometry) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Geometry) GoString() string { return s.String() } // SetBoundingBox sets the BoundingBox field's value. func (s *Geometry) SetBoundingBox(v *BoundingBox) *Geometry { s.BoundingBox = v return s } // SetPolygon sets the Polygon field's value. func (s *Geometry) SetPolygon(v []*Point) *Geometry { s.Polygon = v return s } type GetCelebrityInfoInput struct { _ struct{} `type:"structure"` // The ID for the celebrity. You get the celebrity ID from a call to the operation, // which recognizes celebrities in an image. // // Id is a required field Id *string `type:"string" required:"true"` } // String returns the string representation func (s GetCelebrityInfoInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s GetCelebrityInfoInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *GetCelebrityInfoInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "GetCelebrityInfoInput"} if s.Id == nil { invalidParams.Add(request.NewErrParamRequired("Id")) } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetId sets the Id field's value. func (s *GetCelebrityInfoInput) SetId(v string) *GetCelebrityInfoInput { s.Id = &v return s } type GetCelebrityInfoOutput struct { _ struct{} `type:"structure"` // The name of the celebrity. Name *string `type:"string"` // An array of URLs pointing to additional celebrity information. Urls []*string `type:"list"` } // String returns the string representation func (s GetCelebrityInfoOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s GetCelebrityInfoOutput) GoString() string { return s.String() } // SetName sets the Name field's value. func (s *GetCelebrityInfoOutput) SetName(v string) *GetCelebrityInfoOutput { s.Name = &v return s } // SetUrls sets the Urls field's value. func (s *GetCelebrityInfoOutput) SetUrls(v []*string) *GetCelebrityInfoOutput { s.Urls = v return s } // Provides the input image either as bytes or an S3 object. // // You pass image bytes to a Rekognition API operation by using the Bytes property. // For example, you would use the Bytes property to pass an image loaded from // a local file system. Image bytes passed by using the Bytes property must // be base64-encoded. Your code may not need to encode image bytes if you are // using an AWS SDK to call Rekognition API operations. For more information, // see example4. // // You pass images stored in an S3 bucket to a Rekognition API operation by // using the S3Object property. Images stored in an S3 bucket do not need to // be base64-encoded. // // The region for the S3 bucket containing the S3 object must match the region // you use for Amazon Rekognition operations. // // If you use the Amazon CLI to call Amazon Rekognition operations, passing // image bytes using the Bytes property is not supported. You must first upload // the image to an Amazon S3 bucket and then call the operation using the S3Object // property. // // For Amazon Rekognition to process an S3 object, the user must have permission // to access the S3 object. For more information, see manage-access-resource-policies. type Image struct { _ struct{} `type:"structure"` // Blob of image bytes up to 5 MBs. // // Bytes is automatically base64 encoded/decoded by the SDK. Bytes []byte `min:"1" type:"blob"` // Identifies an S3 object as the image source. S3Object *S3Object `type:"structure"` } // String returns the string representation func (s Image) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Image) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *Image) Validate() error { invalidParams := request.ErrInvalidParams{Context: "Image"} if s.Bytes != nil && len(s.Bytes) < 1 { invalidParams.Add(request.NewErrParamMinLen("Bytes", 1)) } if s.S3Object != nil { if err := s.S3Object.Validate(); err != nil { invalidParams.AddNested("S3Object", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetBytes sets the Bytes field's value. func (s *Image) SetBytes(v []byte) *Image { s.Bytes = v return s } // SetS3Object sets the S3Object field's value. func (s *Image) SetS3Object(v *S3Object) *Image { s.S3Object = v return s } // Identifies face image brightness and sharpness. type ImageQuality struct { _ struct{} `type:"structure"` // Value representing brightness of the face. The service returns a value between // 0 and 100 (inclusive). A higher value indicates a brighter face image. Brightness *float64 `type:"float"` // Value representing sharpness of the face. The service returns a value between // 0 and 100 (inclusive). A higher value indicates a sharper face image. Sharpness *float64 `type:"float"` } // String returns the string representation func (s ImageQuality) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ImageQuality) GoString() string { return s.String() } // SetBrightness sets the Brightness field's value. func (s *ImageQuality) SetBrightness(v float64) *ImageQuality { s.Brightness = &v return s } // SetSharpness sets the Sharpness field's value. func (s *ImageQuality) SetSharpness(v float64) *ImageQuality { s.Sharpness = &v return s } type IndexFacesInput struct { _ struct{} `type:"structure"` // The ID of an existing collection to which you want to add the faces that // are detected in the input images. // // CollectionId is a required field CollectionId *string `min:"1" type:"string" required:"true"` // An array of facial attributes that you want to be returned. This can be the // default list of attributes or all attributes. If you don't specify a value // for Attributes or if you specify ["DEFAULT"], the API returns the following // subset of facial attributes: BoundingBox, Confidence, Pose, Quality and Landmarks. // If you provide ["ALL"], all facial attributes are returned but the operation // will take longer to complete. // // If you provide both, ["ALL", "DEFAULT"], the service uses a logical AND operator // to determine which attributes to return (in this case, all attributes). DetectionAttributes []*string `type:"list"` // ID you want to assign to all the faces detected in the image. ExternalImageId *string `min:"1" type:"string"` // The input image as base64-encoded bytes or an S3 object. If you use the AWS // CLI to call Amazon Rekognition operations, passing base64-encoded image bytes // is not supported. // // Image is a required field Image *Image `type:"structure" required:"true"` } // String returns the string representation func (s IndexFacesInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s IndexFacesInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *IndexFacesInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "IndexFacesInput"} if s.CollectionId == nil { invalidParams.Add(request.NewErrParamRequired("CollectionId")) } if s.CollectionId != nil && len(*s.CollectionId) < 1 { invalidParams.Add(request.NewErrParamMinLen("CollectionId", 1)) } if s.ExternalImageId != nil && len(*s.ExternalImageId) < 1 { invalidParams.Add(request.NewErrParamMinLen("ExternalImageId", 1)) } if s.Image == nil { invalidParams.Add(request.NewErrParamRequired("Image")) } if s.Image != nil { if err := s.Image.Validate(); err != nil { invalidParams.AddNested("Image", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetCollectionId sets the CollectionId field's value. func (s *IndexFacesInput) SetCollectionId(v string) *IndexFacesInput { s.CollectionId = &v return s } // SetDetectionAttributes sets the DetectionAttributes field's value. func (s *IndexFacesInput) SetDetectionAttributes(v []*string) *IndexFacesInput { s.DetectionAttributes = v return s } // SetExternalImageId sets the ExternalImageId field's value. func (s *IndexFacesInput) SetExternalImageId(v string) *IndexFacesInput { s.ExternalImageId = &v return s } // SetImage sets the Image field's value. func (s *IndexFacesInput) SetImage(v *Image) *IndexFacesInput { s.Image = v return s } type IndexFacesOutput struct { _ struct{} `type:"structure"` // Version number of the face detection model associated with the input collection // (CollectionId). FaceModelVersion *string `type:"string"` // An array of faces detected and added to the collection. For more information, // see howitworks-index-faces. FaceRecords []*FaceRecord `type:"list"` // The orientation of the input image (counterclockwise direction). If your // application displays the image, you can use this value to correct image orientation. // The bounding box coordinates returned in FaceRecords represent face locations // before the image orientation is corrected. // // If the input image is in jpeg format, it might contain exchangeable image // (Exif) metadata. If so, and the Exif metadata populates the orientation field, // the value of OrientationCorrection is null and the bounding box coordinates // in FaceRecords represent face locations after Exif metadata is used to correct // the image orientation. Images in .png format don't contain Exif metadata. OrientationCorrection *string `type:"string" enum:"OrientationCorrection"` } // String returns the string representation func (s IndexFacesOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s IndexFacesOutput) GoString() string { return s.String() } // SetFaceModelVersion sets the FaceModelVersion field's value. func (s *IndexFacesOutput) SetFaceModelVersion(v string) *IndexFacesOutput { s.FaceModelVersion = &v return s } // SetFaceRecords sets the FaceRecords field's value. func (s *IndexFacesOutput) SetFaceRecords(v []*FaceRecord) *IndexFacesOutput { s.FaceRecords = v return s } // SetOrientationCorrection sets the OrientationCorrection field's value. func (s *IndexFacesOutput) SetOrientationCorrection(v string) *IndexFacesOutput { s.OrientationCorrection = &v return s } // Structure containing details about the detected label, including name, and // level of confidence. type Label struct { _ struct{} `type:"structure"` // Level of confidence. Confidence *float64 `type:"float"` // The name (label) of the object. Name *string `type:"string"` } // String returns the string representation func (s Label) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Label) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Label) SetConfidence(v float64) *Label { s.Confidence = &v return s } // SetName sets the Name field's value. func (s *Label) SetName(v string) *Label { s.Name = &v return s } // Indicates the location of the landmark on the face. type Landmark struct { _ struct{} `type:"structure"` // Type of the landmark. Type *string `type:"string" enum:"LandmarkType"` // x-coordinate from the top left of the landmark expressed as the ratio of // the width of the image. For example, if the images is 700x200 and the x-coordinate // of the landmark is at 350 pixels, this value is 0.5. X *float64 `type:"float"` // y-coordinate from the top left of the landmark expressed as the ratio of // the height of the image. For example, if the images is 700x200 and the y-coordinate // of the landmark is at 100 pixels, this value is 0.5. Y *float64 `type:"float"` } // String returns the string representation func (s Landmark) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Landmark) GoString() string { return s.String() } // SetType sets the Type field's value. func (s *Landmark) SetType(v string) *Landmark { s.Type = &v return s } // SetX sets the X field's value. func (s *Landmark) SetX(v float64) *Landmark { s.X = &v return s } // SetY sets the Y field's value. func (s *Landmark) SetY(v float64) *Landmark { s.Y = &v return s } type ListCollectionsInput struct { _ struct{} `type:"structure"` // Maximum number of collection IDs to return. MaxResults *int64 `type:"integer"` // Pagination token from the previous response. NextToken *string `type:"string"` } // String returns the string representation func (s ListCollectionsInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ListCollectionsInput) GoString() string { return s.String() } // SetMaxResults sets the MaxResults field's value. func (s *ListCollectionsInput) SetMaxResults(v int64) *ListCollectionsInput { s.MaxResults = &v return s } // SetNextToken sets the NextToken field's value. func (s *ListCollectionsInput) SetNextToken(v string) *ListCollectionsInput { s.NextToken = &v return s } type ListCollectionsOutput struct { _ struct{} `type:"structure"` // An array of collection IDs. CollectionIds []*string `type:"list"` // Version numbers of the face detection models associated with the collections // in the array CollectionIds. For example, the value of FaceModelVersions[2] // is the version number for the face detection model used by the collection // in CollectionId[2]. FaceModelVersions []*string `type:"list"` // If the result is truncated, the response provides a NextToken that you can // use in the subsequent request to fetch the next set of collection IDs. NextToken *string `type:"string"` } // String returns the string representation func (s ListCollectionsOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ListCollectionsOutput) GoString() string { return s.String() } // SetCollectionIds sets the CollectionIds field's value. func (s *ListCollectionsOutput) SetCollectionIds(v []*string) *ListCollectionsOutput { s.CollectionIds = v return s } // SetFaceModelVersions sets the FaceModelVersions field's value. func (s *ListCollectionsOutput) SetFaceModelVersions(v []*string) *ListCollectionsOutput { s.FaceModelVersions = v return s } // SetNextToken sets the NextToken field's value. func (s *ListCollectionsOutput) SetNextToken(v string) *ListCollectionsOutput { s.NextToken = &v return s } type ListFacesInput struct { _ struct{} `type:"structure"` // ID of the collection from which to list the faces. // // CollectionId is a required field CollectionId *string `min:"1" type:"string" required:"true"` // Maximum number of faces to return. MaxResults *int64 `type:"integer"` // If the previous response was incomplete (because there is more data to retrieve), // Amazon Rekognition returns a pagination token in the response. You can use // this pagination token to retrieve the next set of faces. NextToken *string `type:"string"` } // String returns the string representation func (s ListFacesInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ListFacesInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *ListFacesInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "ListFacesInput"} if s.CollectionId == nil { invalidParams.Add(request.NewErrParamRequired("CollectionId")) } if s.CollectionId != nil && len(*s.CollectionId) < 1 { invalidParams.Add(request.NewErrParamMinLen("CollectionId", 1)) } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetCollectionId sets the CollectionId field's value. func (s *ListFacesInput) SetCollectionId(v string) *ListFacesInput { s.CollectionId = &v return s } // SetMaxResults sets the MaxResults field's value. func (s *ListFacesInput) SetMaxResults(v int64) *ListFacesInput { s.MaxResults = &v return s } // SetNextToken sets the NextToken field's value. func (s *ListFacesInput) SetNextToken(v string) *ListFacesInput { s.NextToken = &v return s } type ListFacesOutput struct { _ struct{} `type:"structure"` // Version number of the face detection model associated with the input collection // (CollectionId). FaceModelVersion *string `type:"string"` // An array of Face objects. Faces []*Face `type:"list"` // If the response is truncated, Amazon Rekognition returns this token that // you can use in the subsequent request to retrieve the next set of faces. NextToken *string `type:"string"` } // String returns the string representation func (s ListFacesOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ListFacesOutput) GoString() string { return s.String() } // SetFaceModelVersion sets the FaceModelVersion field's value. func (s *ListFacesOutput) SetFaceModelVersion(v string) *ListFacesOutput { s.FaceModelVersion = &v return s } // SetFaces sets the Faces field's value. func (s *ListFacesOutput) SetFaces(v []*Face) *ListFacesOutput { s.Faces = v return s } // SetNextToken sets the NextToken field's value. func (s *ListFacesOutput) SetNextToken(v string) *ListFacesOutput { s.NextToken = &v return s } // Provides information about a single type of moderated content found in an // image. Each type of moderated content has a label within a hierarchical taxonomy. // For more information, see image-moderation. type ModerationLabel struct { _ struct{} `type:"structure"` // Specifies the confidence that Amazon Rekognition has that the label has been // correctly identified. // // If you don't specify the MinConfidence parameter in the call to DetectModerationLabels, // the operation returns labels with a confidence value greater than or equal // to 50 percent. Confidence *float64 `type:"float"` // The label name for the type of content detected in the image. Name *string `type:"string"` // The name for the parent label. Labels at the top-level of the hierarchy have // the parent label "". ParentName *string `type:"string"` } // String returns the string representation func (s ModerationLabel) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s ModerationLabel) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *ModerationLabel) SetConfidence(v float64) *ModerationLabel { s.Confidence = &v return s } // SetName sets the Name field's value. func (s *ModerationLabel) SetName(v string) *ModerationLabel { s.Name = &v return s } // SetParentName sets the ParentName field's value. func (s *ModerationLabel) SetParentName(v string) *ModerationLabel { s.ParentName = &v return s } // Indicates whether or not the mouth on the face is open, and the confidence // level in the determination. type MouthOpen struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Boolean value that indicates whether the mouth on the face is open or not. Value *bool `type:"boolean"` } // String returns the string representation func (s MouthOpen) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s MouthOpen) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *MouthOpen) SetConfidence(v float64) *MouthOpen { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *MouthOpen) SetValue(v bool) *MouthOpen { s.Value = &v return s } // Indicates whether or not the face has a mustache, and the confidence level // in the determination. type Mustache struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Boolean value that indicates whether the face has mustache or not. Value *bool `type:"boolean"` } // String returns the string representation func (s Mustache) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Mustache) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Mustache) SetConfidence(v float64) *Mustache { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *Mustache) SetValue(v bool) *Mustache { s.Value = &v return s } // The X and Y coordinates of a point on an image. The X and Y values returned // are ratios of the overall image size. For example, if the input image is // 700x200 and the operation returns X=0.5 and Y=0.25, then the point is at // the (350,50) pixel coordinate on the image. // // An array of Point objects, Polygon, is returned by . Polygon represents a // fine-grained polygon around detected text. For more information, see . type Point struct { _ struct{} `type:"structure"` // The value of the X coordinate for a point on a Polygon. X *float64 `type:"float"` // The value of the Y coordinate for a point on a Polygon. Y *float64 `type:"float"` } // String returns the string representation func (s Point) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Point) GoString() string { return s.String() } // SetX sets the X field's value. func (s *Point) SetX(v float64) *Point { s.X = &v return s } // SetY sets the Y field's value. func (s *Point) SetY(v float64) *Point { s.Y = &v return s } // Indicates the pose of the face as determined by its pitch, roll, and yaw. type Pose struct { _ struct{} `type:"structure"` // Value representing the face rotation on the pitch axis. Pitch *float64 `type:"float"` // Value representing the face rotation on the roll axis. Roll *float64 `type:"float"` // Value representing the face rotation on the yaw axis. Yaw *float64 `type:"float"` } // String returns the string representation func (s Pose) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Pose) GoString() string { return s.String() } // SetPitch sets the Pitch field's value. func (s *Pose) SetPitch(v float64) *Pose { s.Pitch = &v return s } // SetRoll sets the Roll field's value. func (s *Pose) SetRoll(v float64) *Pose { s.Roll = &v return s } // SetYaw sets the Yaw field's value. func (s *Pose) SetYaw(v float64) *Pose { s.Yaw = &v return s } type RecognizeCelebritiesInput struct { _ struct{} `type:"structure"` // The input image as base64-encoded bytes or an S3 object. If you use the AWS // CLI to call Amazon Rekognition operations, passing base64-encoded image bytes // is not supported. // // Image is a required field Image *Image `type:"structure" required:"true"` } // String returns the string representation func (s RecognizeCelebritiesInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s RecognizeCelebritiesInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *RecognizeCelebritiesInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "RecognizeCelebritiesInput"} if s.Image == nil { invalidParams.Add(request.NewErrParamRequired("Image")) } if s.Image != nil { if err := s.Image.Validate(); err != nil { invalidParams.AddNested("Image", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetImage sets the Image field's value. func (s *RecognizeCelebritiesInput) SetImage(v *Image) *RecognizeCelebritiesInput { s.Image = v return s } type RecognizeCelebritiesOutput struct { _ struct{} `type:"structure"` // Details about each celebrity found in the image. Amazon Rekognition can detect // a maximum of 15 celebrities in an image. CelebrityFaces []*Celebrity `type:"list"` // The orientation of the input image (counterclockwise direction). If your // application displays the image, you can use this value to correct the orientation. // The bounding box coordinates returned in CelebrityFaces and UnrecognizedFaces // represent face locations before the image orientation is corrected. // // If the input image is in .jpeg format, it might contain exchangeable image // (Exif) metadata that includes the image's orientation. If so, and the Exif // metadata for the input image populates the orientation field, the value of // OrientationCorrection is null and the CelebrityFaces and UnrecognizedFaces // bounding box coordinates represent face locations after Exif metadata is // used to correct the image orientation. Images in .png format don't contain // Exif metadata. OrientationCorrection *string `type:"string" enum:"OrientationCorrection"` // Details about each unrecognized face in the image. UnrecognizedFaces []*ComparedFace `type:"list"` } // String returns the string representation func (s RecognizeCelebritiesOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s RecognizeCelebritiesOutput) GoString() string { return s.String() } // SetCelebrityFaces sets the CelebrityFaces field's value. func (s *RecognizeCelebritiesOutput) SetCelebrityFaces(v []*Celebrity) *RecognizeCelebritiesOutput { s.CelebrityFaces = v return s } // SetOrientationCorrection sets the OrientationCorrection field's value. func (s *RecognizeCelebritiesOutput) SetOrientationCorrection(v string) *RecognizeCelebritiesOutput { s.OrientationCorrection = &v return s } // SetUnrecognizedFaces sets the UnrecognizedFaces field's value. func (s *RecognizeCelebritiesOutput) SetUnrecognizedFaces(v []*ComparedFace) *RecognizeCelebritiesOutput { s.UnrecognizedFaces = v return s } // Provides the S3 bucket name and object name. // // The region for the S3 bucket containing the S3 object must match the region // you use for Amazon Rekognition operations. // // For Amazon Rekognition to process an S3 object, the user must have permission // to access the S3 object. For more information, see manage-access-resource-policies. type S3Object struct { _ struct{} `type:"structure"` // Name of the S3 bucket. Bucket *string `min:"3" type:"string"` // S3 object key name. Name *string `min:"1" type:"string"` // If the bucket is versioning enabled, you can specify the object version. Version *string `min:"1" type:"string"` } // String returns the string representation func (s S3Object) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s S3Object) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *S3Object) Validate() error { invalidParams := request.ErrInvalidParams{Context: "S3Object"} if s.Bucket != nil && len(*s.Bucket) < 3 { invalidParams.Add(request.NewErrParamMinLen("Bucket", 3)) } if s.Name != nil && len(*s.Name) < 1 { invalidParams.Add(request.NewErrParamMinLen("Name", 1)) } if s.Version != nil && len(*s.Version) < 1 { invalidParams.Add(request.NewErrParamMinLen("Version", 1)) } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetBucket sets the Bucket field's value. func (s *S3Object) SetBucket(v string) *S3Object { s.Bucket = &v return s } // SetName sets the Name field's value. func (s *S3Object) SetName(v string) *S3Object { s.Name = &v return s } // SetVersion sets the Version field's value. func (s *S3Object) SetVersion(v string) *S3Object { s.Version = &v return s } type SearchFacesByImageInput struct { _ struct{} `type:"structure"` // ID of the collection to search. // // CollectionId is a required field CollectionId *string `min:"1" type:"string" required:"true"` // (Optional) Specifies the minimum confidence in the face match to return. // For example, don't return any matches where confidence in matches is less // than 70%. FaceMatchThreshold *float64 `type:"float"` // The input image as base64-encoded bytes or an S3 object. If you use the AWS // CLI to call Amazon Rekognition operations, passing base64-encoded image bytes // is not supported. // // Image is a required field Image *Image `type:"structure" required:"true"` // Maximum number of faces to return. The operation returns the maximum number // of faces with the highest confidence in the match. MaxFaces *int64 `min:"1" type:"integer"` } // String returns the string representation func (s SearchFacesByImageInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s SearchFacesByImageInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *SearchFacesByImageInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "SearchFacesByImageInput"} if s.CollectionId == nil { invalidParams.Add(request.NewErrParamRequired("CollectionId")) } if s.CollectionId != nil && len(*s.CollectionId) < 1 { invalidParams.Add(request.NewErrParamMinLen("CollectionId", 1)) } if s.Image == nil { invalidParams.Add(request.NewErrParamRequired("Image")) } if s.MaxFaces != nil && *s.MaxFaces < 1 { invalidParams.Add(request.NewErrParamMinValue("MaxFaces", 1)) } if s.Image != nil { if err := s.Image.Validate(); err != nil { invalidParams.AddNested("Image", err.(request.ErrInvalidParams)) } } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetCollectionId sets the CollectionId field's value. func (s *SearchFacesByImageInput) SetCollectionId(v string) *SearchFacesByImageInput { s.CollectionId = &v return s } // SetFaceMatchThreshold sets the FaceMatchThreshold field's value. func (s *SearchFacesByImageInput) SetFaceMatchThreshold(v float64) *SearchFacesByImageInput { s.FaceMatchThreshold = &v return s } // SetImage sets the Image field's value. func (s *SearchFacesByImageInput) SetImage(v *Image) *SearchFacesByImageInput { s.Image = v return s } // SetMaxFaces sets the MaxFaces field's value. func (s *SearchFacesByImageInput) SetMaxFaces(v int64) *SearchFacesByImageInput { s.MaxFaces = &v return s } type SearchFacesByImageOutput struct { _ struct{} `type:"structure"` // An array of faces that match the input face, along with the confidence in // the match. FaceMatches []*FaceMatch `type:"list"` // Version number of the face detection model associated with the input collection // (CollectionId). FaceModelVersion *string `type:"string"` // The bounding box around the face in the input image that Amazon Rekognition // used for the search. SearchedFaceBoundingBox *BoundingBox `type:"structure"` // The level of confidence that the searchedFaceBoundingBox, contains a face. SearchedFaceConfidence *float64 `type:"float"` } // String returns the string representation func (s SearchFacesByImageOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s SearchFacesByImageOutput) GoString() string { return s.String() } // SetFaceMatches sets the FaceMatches field's value. func (s *SearchFacesByImageOutput) SetFaceMatches(v []*FaceMatch) *SearchFacesByImageOutput { s.FaceMatches = v return s } // SetFaceModelVersion sets the FaceModelVersion field's value. func (s *SearchFacesByImageOutput) SetFaceModelVersion(v string) *SearchFacesByImageOutput { s.FaceModelVersion = &v return s } // SetSearchedFaceBoundingBox sets the SearchedFaceBoundingBox field's value. func (s *SearchFacesByImageOutput) SetSearchedFaceBoundingBox(v *BoundingBox) *SearchFacesByImageOutput { s.SearchedFaceBoundingBox = v return s } // SetSearchedFaceConfidence sets the SearchedFaceConfidence field's value. func (s *SearchFacesByImageOutput) SetSearchedFaceConfidence(v float64) *SearchFacesByImageOutput { s.SearchedFaceConfidence = &v return s } type SearchFacesInput struct { _ struct{} `type:"structure"` // ID of the collection the face belongs to. // // CollectionId is a required field CollectionId *string `min:"1" type:"string" required:"true"` // ID of a face to find matches for in the collection. // // FaceId is a required field FaceId *string `type:"string" required:"true"` // Optional value specifying the minimum confidence in the face match to return. // For example, don't return any matches where confidence in matches is less // than 70%. FaceMatchThreshold *float64 `type:"float"` // Maximum number of faces to return. The operation returns the maximum number // of faces with the highest confidence in the match. MaxFaces *int64 `min:"1" type:"integer"` } // String returns the string representation func (s SearchFacesInput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s SearchFacesInput) GoString() string { return s.String() } // Validate inspects the fields of the type to determine if they are valid. func (s *SearchFacesInput) Validate() error { invalidParams := request.ErrInvalidParams{Context: "SearchFacesInput"} if s.CollectionId == nil { invalidParams.Add(request.NewErrParamRequired("CollectionId")) } if s.CollectionId != nil && len(*s.CollectionId) < 1 { invalidParams.Add(request.NewErrParamMinLen("CollectionId", 1)) } if s.FaceId == nil { invalidParams.Add(request.NewErrParamRequired("FaceId")) } if s.MaxFaces != nil && *s.MaxFaces < 1 { invalidParams.Add(request.NewErrParamMinValue("MaxFaces", 1)) } if invalidParams.Len() > 0 { return invalidParams } return nil } // SetCollectionId sets the CollectionId field's value. func (s *SearchFacesInput) SetCollectionId(v string) *SearchFacesInput { s.CollectionId = &v return s } // SetFaceId sets the FaceId field's value. func (s *SearchFacesInput) SetFaceId(v string) *SearchFacesInput { s.FaceId = &v return s } // SetFaceMatchThreshold sets the FaceMatchThreshold field's value. func (s *SearchFacesInput) SetFaceMatchThreshold(v float64) *SearchFacesInput { s.FaceMatchThreshold = &v return s } // SetMaxFaces sets the MaxFaces field's value. func (s *SearchFacesInput) SetMaxFaces(v int64) *SearchFacesInput { s.MaxFaces = &v return s } type SearchFacesOutput struct { _ struct{} `type:"structure"` // An array of faces that matched the input face, along with the confidence // in the match. FaceMatches []*FaceMatch `type:"list"` // Version number of the face detection model associated with the input collection // (CollectionId). FaceModelVersion *string `type:"string"` // ID of the face that was searched for matches in a collection. SearchedFaceId *string `type:"string"` } // String returns the string representation func (s SearchFacesOutput) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s SearchFacesOutput) GoString() string { return s.String() } // SetFaceMatches sets the FaceMatches field's value. func (s *SearchFacesOutput) SetFaceMatches(v []*FaceMatch) *SearchFacesOutput { s.FaceMatches = v return s } // SetFaceModelVersion sets the FaceModelVersion field's value. func (s *SearchFacesOutput) SetFaceModelVersion(v string) *SearchFacesOutput { s.FaceModelVersion = &v return s } // SetSearchedFaceId sets the SearchedFaceId field's value. func (s *SearchFacesOutput) SetSearchedFaceId(v string) *SearchFacesOutput { s.SearchedFaceId = &v return s } // Indicates whether or not the face is smiling, and the confidence level in // the determination. type Smile struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Boolean value that indicates whether the face is smiling or not. Value *bool `type:"boolean"` } // String returns the string representation func (s Smile) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Smile) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Smile) SetConfidence(v float64) *Smile { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *Smile) SetValue(v bool) *Smile { s.Value = &v return s } // Indicates whether or not the face is wearing sunglasses, and the confidence // level in the determination. type Sunglasses struct { _ struct{} `type:"structure"` // Level of confidence in the determination. Confidence *float64 `type:"float"` // Boolean value that indicates whether the face is wearing sunglasses or not. Value *bool `type:"boolean"` } // String returns the string representation func (s Sunglasses) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s Sunglasses) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *Sunglasses) SetConfidence(v float64) *Sunglasses { s.Confidence = &v return s } // SetValue sets the Value field's value. func (s *Sunglasses) SetValue(v bool) *Sunglasses { s.Value = &v return s } // Information about a word or line of text detected by . // // The DetectedText field contains the text that Amazon Rekognition detected // in the image. // // Every word and line has an identifier (Id). Each word belongs to a line and // has a parent identifier (ParentId) that identifies the line of text in which // the word appears. The word Id is also an index for the word within a line // of words. // // For more information, see text-detection. type TextDetection struct { _ struct{} `type:"structure"` // The confidence that Amazon Rekognition has in the accuracy of the detected // text and the accuracy of the geometry points around the detected text. Confidence *float64 `type:"float"` // The word or line of text recognized by Amazon Rekognition. DetectedText *string `type:"string"` // The location of the detected text on the image. Includes an axis aligned // coarse bounding box surrounding the text and a finer grain polygon for more // accurate spatial information. Geometry *Geometry `type:"structure"` // The identifier for the detected text. The identifier is only unique for a // single call to DetectText. Id *int64 `type:"integer"` // The Parent identifier for the detected text identified by the value of ID. // If the type of detected text is LINE, the value of ParentId is Null. ParentId *int64 `type:"integer"` // The type of text that was detected. Type *string `type:"string" enum:"TextTypes"` } // String returns the string representation func (s TextDetection) String() string { return awsutil.Prettify(s) } // GoString returns the string representation func (s TextDetection) GoString() string { return s.String() } // SetConfidence sets the Confidence field's value. func (s *TextDetection) SetConfidence(v float64) *TextDetection { s.Confidence = &v return s } // SetDetectedText sets the DetectedText field's value. func (s *TextDetection) SetDetectedText(v string) *TextDetection { s.DetectedText = &v return s } // SetGeometry sets the Geometry field's value. func (s *TextDetection) SetGeometry(v *Geometry) *TextDetection { s.Geometry = v return s } // SetId sets the Id field's value. func (s *TextDetection) SetId(v int64) *TextDetection { s.Id = &v return s } // SetParentId sets the ParentId field's value. func (s *TextDetection) SetParentId(v int64) *TextDetection { s.ParentId = &v return s } // SetType sets the Type field's value. func (s *TextDetection) SetType(v string) *TextDetection { s.Type = &v return s } const ( // AttributeDefault is a Attribute enum value AttributeDefault = "DEFAULT" // AttributeAll is a Attribute enum value AttributeAll = "ALL" ) const ( // EmotionNameHappy is a EmotionName enum value EmotionNameHappy = "HAPPY" // EmotionNameSad is a EmotionName enum value EmotionNameSad = "SAD" // EmotionNameAngry is a EmotionName enum value EmotionNameAngry = "ANGRY" // EmotionNameConfused is a EmotionName enum value EmotionNameConfused = "CONFUSED" // EmotionNameDisgusted is a EmotionName enum value EmotionNameDisgusted = "DISGUSTED" // EmotionNameSurprised is a EmotionName enum value EmotionNameSurprised = "SURPRISED" // EmotionNameCalm is a EmotionName enum value EmotionNameCalm = "CALM" // EmotionNameUnknown is a EmotionName enum value EmotionNameUnknown = "UNKNOWN" ) const ( // GenderTypeMale is a GenderType enum value GenderTypeMale = "Male" // GenderTypeFemale is a GenderType enum value GenderTypeFemale = "Female" ) const ( // LandmarkTypeEyeLeft is a LandmarkType enum value LandmarkTypeEyeLeft = "eyeLeft" // LandmarkTypeEyeRight is a LandmarkType enum value LandmarkTypeEyeRight = "eyeRight" // LandmarkTypeNose is a LandmarkType enum value LandmarkTypeNose = "nose" // LandmarkTypeMouthLeft is a LandmarkType enum value LandmarkTypeMouthLeft = "mouthLeft" // LandmarkTypeMouthRight is a LandmarkType enum value LandmarkTypeMouthRight = "mouthRight" // LandmarkTypeLeftEyeBrowLeft is a LandmarkType enum value LandmarkTypeLeftEyeBrowLeft = "leftEyeBrowLeft" // LandmarkTypeLeftEyeBrowRight is a LandmarkType enum value LandmarkTypeLeftEyeBrowRight = "leftEyeBrowRight" // LandmarkTypeLeftEyeBrowUp is a LandmarkType enum value LandmarkTypeLeftEyeBrowUp = "leftEyeBrowUp" // LandmarkTypeRightEyeBrowLeft is a LandmarkType enum value LandmarkTypeRightEyeBrowLeft = "rightEyeBrowLeft" // LandmarkTypeRightEyeBrowRight is a LandmarkType enum value LandmarkTypeRightEyeBrowRight = "rightEyeBrowRight" // LandmarkTypeRightEyeBrowUp is a LandmarkType enum value LandmarkTypeRightEyeBrowUp = "rightEyeBrowUp" // LandmarkTypeLeftEyeLeft is a LandmarkType enum value LandmarkTypeLeftEyeLeft = "leftEyeLeft" // LandmarkTypeLeftEyeRight is a LandmarkType enum value LandmarkTypeLeftEyeRight = "leftEyeRight" // LandmarkTypeLeftEyeUp is a LandmarkType enum value LandmarkTypeLeftEyeUp = "leftEyeUp" // LandmarkTypeLeftEyeDown is a LandmarkType enum value LandmarkTypeLeftEyeDown = "leftEyeDown" // LandmarkTypeRightEyeLeft is a LandmarkType enum value LandmarkTypeRightEyeLeft = "rightEyeLeft" // LandmarkTypeRightEyeRight is a LandmarkType enum value LandmarkTypeRightEyeRight = "rightEyeRight" // LandmarkTypeRightEyeUp is a LandmarkType enum value LandmarkTypeRightEyeUp = "rightEyeUp" // LandmarkTypeRightEyeDown is a LandmarkType enum value LandmarkTypeRightEyeDown = "rightEyeDown" // LandmarkTypeNoseLeft is a LandmarkType enum value LandmarkTypeNoseLeft = "noseLeft" // LandmarkTypeNoseRight is a LandmarkType enum value LandmarkTypeNoseRight = "noseRight" // LandmarkTypeMouthUp is a LandmarkType enum value LandmarkTypeMouthUp = "mouthUp" // LandmarkTypeMouthDown is a LandmarkType enum value LandmarkTypeMouthDown = "mouthDown" // LandmarkTypeLeftPupil is a LandmarkType enum value LandmarkTypeLeftPupil = "leftPupil" // LandmarkTypeRightPupil is a LandmarkType enum value LandmarkTypeRightPupil = "rightPupil" ) const ( // OrientationCorrectionRotate0 is a OrientationCorrection enum value OrientationCorrectionRotate0 = "ROTATE_0" // OrientationCorrectionRotate90 is a OrientationCorrection enum value OrientationCorrectionRotate90 = "ROTATE_90" // OrientationCorrectionRotate180 is a OrientationCorrection enum value OrientationCorrectionRotate180 = "ROTATE_180" // OrientationCorrectionRotate270 is a OrientationCorrection enum value OrientationCorrectionRotate270 = "ROTATE_270" ) const ( // TextTypesLine is a TextTypes enum value TextTypesLine = "LINE" // TextTypesWord is a TextTypes enum value TextTypesWord = "WORD" )