diff --git a/codegen/sdk-codegen/aws-models/amplifyuibuilder.json b/codegen/sdk-codegen/aws-models/amplifyuibuilder.json index cd6968b1759..73dd9a65013 100644 --- a/codegen/sdk-codegen/aws-models/amplifyuibuilder.json +++ b/codegen/sdk-codegen/aws-models/amplifyuibuilder.json @@ -81,6 +81,9 @@ } ], "resources": [ + { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobResource" + }, { "target": "com.amazonaws.amplifyuibuilder#ComponentResource" }, @@ -489,8 +492,8 @@ }, "params": { "Region": "ap-northeast-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -502,8 +505,8 @@ }, "params": { "Region": "ap-northeast-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -515,8 +518,8 @@ }, "params": { "Region": "ap-south-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -528,8 +531,8 @@ }, "params": { "Region": "ap-southeast-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -541,8 +544,8 @@ }, "params": { "Region": "ap-southeast-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -554,8 +557,8 @@ }, "params": { "Region": "ca-central-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -567,8 +570,8 @@ }, "params": { "Region": "eu-central-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -580,8 +583,8 @@ }, "params": { "Region": "eu-north-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -593,8 +596,8 @@ }, "params": { "Region": "eu-west-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -606,8 +609,8 @@ }, "params": { "Region": "eu-west-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -619,8 +622,8 @@ }, "params": { "Region": "eu-west-3", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -632,8 +635,8 @@ }, "params": { "Region": "me-south-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -645,8 +648,8 @@ }, "params": { "Region": "sa-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -658,8 +661,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -671,8 +674,8 @@ }, "params": { "Region": "us-east-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -684,8 +687,8 @@ }, "params": { "Region": "us-west-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -697,8 +700,8 @@ }, "params": { "Region": "us-west-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -710,8 +713,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -723,8 +726,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -736,8 +739,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -749,8 +752,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -762,8 +765,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -775,8 +778,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -788,8 +791,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -801,8 +804,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -814,8 +817,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -827,8 +830,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -840,8 +843,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -851,8 +854,8 @@ }, "params": { "Region": "us-iso-east-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -864,8 +867,8 @@ }, "params": { "Region": "us-iso-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -875,8 +878,8 @@ }, "params": { "Region": "us-iso-east-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -888,8 +891,8 @@ }, "params": { "Region": "us-iso-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -899,8 +902,8 @@ }, "params": { "Region": "us-isob-east-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -912,8 +915,8 @@ }, "params": { "Region": "us-isob-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -923,8 +926,8 @@ }, "params": { "Region": "us-isob-east-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -936,8 +939,8 @@ }, "params": { "Region": "us-isob-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -949,8 +952,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, "UseFIPS": false, + "UseDualStack": false, "Endpoint": "https://example.com" } }, @@ -962,8 +965,8 @@ } }, "params": { - "UseDualStack": false, "UseFIPS": false, + "UseDualStack": false, "Endpoint": "https://example.com" } }, @@ -974,8 +977,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, "UseFIPS": true, + "UseDualStack": false, "Endpoint": "https://example.com" } }, @@ -986,8 +989,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": true, "UseFIPS": false, + "UseDualStack": true, "Endpoint": "https://example.com" } }, @@ -1002,6 +1005,666 @@ } } }, + "com.amazonaws.amplifyuibuilder#AppId": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 20 + }, + "smithy.api#pattern": "^d[a-z0-9]+$" + } + }, + "com.amazonaws.amplifyuibuilder#AssociatedFieldsList": { + "type": "list", + "member": { + "target": "smithy.api#String" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenFeatureFlags": { + "type": "structure", + "members": { + "isRelationshipSupported": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "
Specifes whether a code generation job supports data relationships.
" + } + }, + "isNonModelSupported": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether a code generation job supports non models.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Describes the feature flags that you can specify for a code generation job.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataEnum": { + "type": "structure", + "members": { + "values": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataEnumValuesList", + "traits": { + "smithy.api#documentation": "The list of enum values in the generic data schema.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#documentation": "Describes the enums in a generic data schema.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataEnumValuesList": { + "type": "list", + "member": { + "target": "smithy.api#String" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataEnums": { + "type": "map", + "key": { + "target": "smithy.api#String" + }, + "value": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataEnum" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataField": { + "type": "structure", + "members": { + "dataType": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataFieldDataType", + "traits": { + "smithy.api#documentation": "The data type for the generic data field.
", + "smithy.api#required": {} + } + }, + "dataTypeValue": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The value of the data type for the generic data field.
", + "smithy.api#required": {} + } + }, + "required": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether the generic data field is required.
", + "smithy.api#required": {} + } + }, + "readOnly": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether the generic data field is read-only.
", + "smithy.api#required": {} + } + }, + "isArray": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether the generic data field is an array.
", + "smithy.api#required": {} + } + }, + "relationship": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataRelationshipType", + "traits": { + "smithy.api#documentation": "The relationship of the generic data schema.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Describes a field in a generic data schema.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataFieldDataType": { + "type": "enum", + "members": { + "ID": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ID" + } + }, + "STRING": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "String" + } + }, + "INT": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Int" + } + }, + "FLOAT": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Float" + } + }, + "AWS_DATE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSDate" + } + }, + "AWS_TIME": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSTime" + } + }, + "AWS_DATE_TIME": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSDateTime" + } + }, + "AWS_TIMESTAMP": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSTimestamp" + } + }, + "AWS_EMAIL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSEmail" + } + }, + "AWS_URL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSURL" + } + }, + "AWS_IP_ADDRESS": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSIPAddress" + } + }, + "BOOLEAN": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Boolean" + } + }, + "AWS_JSON": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSJSON" + } + }, + "AWS_PHONE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "AWSPhone" + } + }, + "ENUM": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Enum" + } + }, + "MODEL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "Model" + } + }, + "NON_MODEL": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "NonModel" + } + } + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataFields": { + "type": "map", + "key": { + "target": "smithy.api#String" + }, + "value": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataField" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataModel": { + "type": "structure", + "members": { + "fields": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataFields", + "traits": { + "smithy.api#documentation": "The fields in the generic data model.
", + "smithy.api#required": {} + } + }, + "isJoinTable": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether the generic data model is a join table.
" + } + }, + "primaryKeys": { + "target": "com.amazonaws.amplifyuibuilder#CodegenPrimaryKeysList", + "traits": { + "smithy.api#documentation": "The primary keys of the generic data model.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#documentation": "Describes a model in a generic data schema.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataModels": { + "type": "map", + "key": { + "target": "smithy.api#String" + }, + "value": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataModel" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataNonModel": { + "type": "structure", + "members": { + "fields": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataNonModelFields", + "traits": { + "smithy.api#documentation": "The fields in a generic data schema non model.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#documentation": "Describes a non-model in a generic data schema.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataNonModelFields": { + "type": "map", + "key": { + "target": "smithy.api#String" + }, + "value": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataField" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataNonModels": { + "type": "map", + "key": { + "target": "smithy.api#String" + }, + "value": { + "target": "com.amazonaws.amplifyuibuilder#CodegenGenericDataNonModel" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenGenericDataRelationshipType": { + "type": "structure", + "members": { + "type": { + "target": "com.amazonaws.amplifyuibuilder#GenericDataRelationshipType", + "traits": { + "smithy.api#documentation": "The data relationship type.
", + "smithy.api#required": {} + } + }, + "relatedModelName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the related model in the data relationship.
", + "smithy.api#required": {} + } + }, + "relatedModelFields": { + "target": "com.amazonaws.amplifyuibuilder#RelatedModelFieldsList", + "traits": { + "smithy.api#documentation": "The related model fields in the data relationship.
" + } + }, + "canUnlinkAssociatedModel": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether the relationship can unlink the associated model.
" + } + }, + "relatedJoinFieldName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the related join field in the data relationship.
" + } + }, + "relatedJoinTableName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the related join table in the data relationship.
" + } + }, + "belongsToFieldOnRelatedModel": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The value of the belongsTo
field on the related data model.
The associated fields of the data relationship.
" + } + }, + "isHasManyIndex": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether the @index
directive is supported for a hasMany
data relationship.
Describes the relationship between generic data models.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJob": { + "type": "structure", + "members": { + "id": { + "target": "com.amazonaws.amplifyuibuilder#Uuid", + "traits": { + "smithy.api#documentation": "The unique ID for the code generation job.
", + "smithy.api#required": {} + } + }, + "appId": { + "target": "com.amazonaws.amplifyuibuilder#AppId", + "traits": { + "smithy.api#documentation": "The ID of the Amplify app associated with the code generation job.
", + "smithy.api#required": {} + } + }, + "environmentName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the backend environment associated with the code generation job.
", + "smithy.api#required": {} + } + }, + "renderConfig": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobRenderConfig", + "traits": { + "aws.cloudformation#cfnMutability": "create-and-read" + } + }, + "genericDataSchema": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobGenericDataSchema", + "traits": { + "aws.cloudformation#cfnMutability": "create-and-read" + } + }, + "autoGenerateForms": { + "target": "smithy.api#Boolean", + "traits": { + "aws.cloudformation#cfnMutability": "create-and-read", + "smithy.api#documentation": "Specifies whether to autogenerate forms in the code generation job.
" + } + }, + "features": { + "target": "com.amazonaws.amplifyuibuilder#CodegenFeatureFlags", + "traits": { + "aws.cloudformation#cfnMutability": "create-and-read" + } + }, + "status": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobStatus", + "traits": { + "aws.cloudformation#cfnMutability": "read", + "smithy.api#documentation": "The status of the code generation job.
" + } + }, + "statusMessage": { + "target": "smithy.api#String", + "traits": { + "aws.cloudformation#cfnMutability": "read", + "smithy.api#documentation": "The customized status message for the code generation job.
" + } + }, + "asset": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobAsset", + "traits": { + "aws.cloudformation#cfnMutability": "read", + "smithy.api#documentation": "The CodegenJobAsset
to use for the code generation job.
One or more key-value pairs to use when tagging the code generation job.
" + } + }, + "createdAt": { + "target": "smithy.api#Timestamp", + "traits": { + "aws.cloudformation#cfnExcludeProperty": {}, + "smithy.api#documentation": "The time that the code generation job was created.
", + "smithy.api#timestampFormat": "date-time" + } + }, + "modifiedAt": { + "target": "smithy.api#Timestamp", + "traits": { + "aws.cloudformation#cfnExcludeProperty": {}, + "smithy.api#documentation": "The time that the code generation job was modified.
", + "smithy.api#timestampFormat": "date-time" + } + } + }, + "traits": { + "smithy.api#documentation": "Describes the configuration for a code generation job that is associated with an Amplify app.
", + "smithy.api#references": [ + { + "resource": "com.amazonaws.amplifyuibuilder#CodegenJobResource" + } + ] + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobAsset": { + "type": "structure", + "members": { + "downloadUrl": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The URL to use to access the asset.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Describes an asset for a code generation job.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobGenericDataSchema": { + "type": "structure", + "members": { + "dataSourceType": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobGenericDataSourceType", + "traits": { + "smithy.api#documentation": "The type of the data source for the schema. Currently, the only valid value is an Amplify DataStore
.
The name of a CodegenGenericDataModel
.
The name of a CodegenGenericDataEnum
.
The name of a CodegenGenericDataNonModel
.
Describes the data schema for a code generation job.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobGenericDataSourceType": { + "type": "enum", + "members": { + "DATA_STORE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "DataStore" + } + } + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobRenderConfig": { + "type": "union", + "members": { + "react": { + "target": "com.amazonaws.amplifyuibuilder#ReactStartCodegenJobData", + "traits": { + "smithy.api#documentation": "The name of the ReactStartCodegenJobData
object.
Describes the configuration information for rendering the UI component associated the code generation job.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobResource": { + "type": "resource", + "identifiers": { + "id": { + "target": "com.amazonaws.amplifyuibuilder#Uuid" + }, + "appId": { + "target": "com.amazonaws.amplifyuibuilder#AppId" + }, + "environmentName": { + "target": "smithy.api#String" + } + }, + "create": { + "target": "com.amazonaws.amplifyuibuilder#StartCodegenJob" + }, + "read": { + "target": "com.amazonaws.amplifyuibuilder#GetCodegenJob" + }, + "list": { + "target": "com.amazonaws.amplifyuibuilder#ListCodegenJobs" + }, + "traits": { + "aws.api#arn": { + "template": "app/{appId}/environment/{environmentName}/codegen-jobs/{id}" + }, + "aws.cloudformation#cfnResource": { + "name": "CodegenJob", + "additionalSchemas": [ + "com.amazonaws.amplifyuibuilder#CodegenJob" + ] + }, + "aws.iam#conditionKeys": [ + "aws:ResourceTag/${TagKey}" + ] + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobStatus": { + "type": "enum", + "members": { + "IN_PROGRESS": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "in_progress" + } + }, + "FAILED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "failed" + } + }, + "SUCCEEDED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "succeeded" + } + } + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobSummary": { + "type": "structure", + "members": { + "appId": { + "target": "com.amazonaws.amplifyuibuilder#AppId", + "traits": { + "smithy.api#documentation": "The unique ID of the Amplify app associated with the code generation job.
", + "smithy.api#required": {} + } + }, + "environmentName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the backend environment associated with the code generation job.
", + "smithy.api#required": {} + } + }, + "id": { + "target": "com.amazonaws.amplifyuibuilder#Uuid", + "traits": { + "smithy.api#documentation": "The unique ID for the code generation job summary.
", + "smithy.api#required": {} + } + }, + "createdAt": { + "target": "smithy.api#Timestamp", + "traits": { + "aws.cloudformation#cfnExcludeProperty": {}, + "smithy.api#documentation": "The time that the code generation job summary was created.
", + "smithy.api#timestampFormat": "date-time" + } + }, + "modifiedAt": { + "target": "smithy.api#Timestamp", + "traits": { + "aws.cloudformation#cfnExcludeProperty": {}, + "smithy.api#documentation": "The time that the code generation job summary was modified.
", + "smithy.api#timestampFormat": "date-time" + } + } + }, + "traits": { + "smithy.api#documentation": "A summary of the basic information about the code generation job.
" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenJobSummaryList": { + "type": "list", + "member": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobSummary" + } + }, + "com.amazonaws.amplifyuibuilder#CodegenPrimaryKeysList": { + "type": "list", + "member": { + "target": "smithy.api#String" + } + }, "com.amazonaws.amplifyuibuilder#Component": { "type": "structure", "members": { @@ -1916,7 +2579,7 @@ "aws.iam#requiredActions": [ "amplify:GetApp" ], - "smithy.api#documentation": "Creates a new form for an Amplify app.
", + "smithy.api#documentation": "Creates a new form for an Amplify.
", "smithy.api#http": { "uri": "/app/{appId}/environment/{environmentName}/forms", "method": "POST" @@ -3479,6 +4142,111 @@ "target": "com.amazonaws.amplifyuibuilder#FormSummary" } }, + "com.amazonaws.amplifyuibuilder#GenericDataRelationshipType": { + "type": "enum", + "members": { + "HAS_MANY": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "HAS_MANY" + } + }, + "HAS_ONE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "HAS_ONE" + } + }, + "BELONGS_TO": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "BELONGS_TO" + } + } + } + }, + "com.amazonaws.amplifyuibuilder#GetCodegenJob": { + "type": "operation", + "input": { + "target": "com.amazonaws.amplifyuibuilder#GetCodegenJobRequest" + }, + "output": { + "target": "com.amazonaws.amplifyuibuilder#GetCodegenJobResponse" + }, + "errors": [ + { + "target": "com.amazonaws.amplifyuibuilder#InternalServerException" + }, + { + "target": "com.amazonaws.amplifyuibuilder#InvalidParameterException" + }, + { + "target": "com.amazonaws.amplifyuibuilder#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.amplifyuibuilder#ThrottlingException" + } + ], + "traits": { + "aws.iam#requiredActions": [ + "amplify:GetApp" + ], + "smithy.api#documentation": "Returns an existing code generation job.
", + "smithy.api#http": { + "uri": "/app/{appId}/environment/{environmentName}/codegen-jobs/{id}", + "method": "GET" + }, + "smithy.api#readonly": {} + } + }, + "com.amazonaws.amplifyuibuilder#GetCodegenJobRequest": { + "type": "structure", + "members": { + "appId": { + "target": "com.amazonaws.amplifyuibuilder#AppId", + "traits": { + "smithy.api#documentation": "The unique ID of the Amplify app associated with the code generation job.
", + "smithy.api#httpLabel": {}, + "smithy.api#required": {} + } + }, + "environmentName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the backend environment that is a part of the Amplify app associated with the code generation job.
", + "smithy.api#httpLabel": {}, + "smithy.api#required": {} + } + }, + "id": { + "target": "com.amazonaws.amplifyuibuilder#Uuid", + "traits": { + "smithy.api#documentation": "The unique ID of the code generation job.
", + "smithy.api#httpLabel": {}, + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.amplifyuibuilder#GetCodegenJobResponse": { + "type": "structure", + "members": { + "job": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJob", + "traits": { + "aws.cloudformation#cfnExcludeProperty": {}, + "smithy.api#documentation": "The configuration settings for the code generation job.
", + "smithy.api#httpPayload": {} + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, "com.amazonaws.amplifyuibuilder#GetComponent": { "type": "operation", "input": { @@ -3788,6 +4556,63 @@ "smithy.api#httpError": 400 } }, + "com.amazonaws.amplifyuibuilder#JSModule": { + "type": "enum", + "members": { + "ES2020": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "es2020" + } + }, + "ESNEXT": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "esnext" + } + } + } + }, + "com.amazonaws.amplifyuibuilder#JSScript": { + "type": "enum", + "members": { + "JSX": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "jsx" + } + }, + "TSX": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "tsx" + } + }, + "JS": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "js" + } + } + } + }, + "com.amazonaws.amplifyuibuilder#JSTarget": { + "type": "enum", + "members": { + "ES2015": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "es2015" + } + }, + "ES2020": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "es2020" + } + } + } + }, "com.amazonaws.amplifyuibuilder#LabelDecorator": { "type": "string", "traits": { @@ -3807,6 +4632,112 @@ ] } }, + "com.amazonaws.amplifyuibuilder#ListCodegenJobs": { + "type": "operation", + "input": { + "target": "com.amazonaws.amplifyuibuilder#ListCodegenJobsRequest" + }, + "output": { + "target": "com.amazonaws.amplifyuibuilder#ListCodegenJobsResponse" + }, + "errors": [ + { + "target": "com.amazonaws.amplifyuibuilder#InternalServerException" + }, + { + "target": "com.amazonaws.amplifyuibuilder#InvalidParameterException" + }, + { + "target": "com.amazonaws.amplifyuibuilder#ThrottlingException" + } + ], + "traits": { + "aws.iam#requiredActions": [ + "amplify:GetApp" + ], + "smithy.api#documentation": "Retrieves a list of code generation jobs for a specified Amplify app and backend environment.
", + "smithy.api#http": { + "uri": "/app/{appId}/environment/{environmentName}/codegen-jobs", + "method": "GET" + }, + "smithy.api#paginated": { + "inputToken": "nextToken", + "outputToken": "nextToken", + "pageSize": "maxResults", + "items": "entities" + }, + "smithy.api#readonly": {} + } + }, + "com.amazonaws.amplifyuibuilder#ListCodegenJobsLimit": { + "type": "integer", + "traits": { + "smithy.api#range": { + "min": 1, + "max": 100 + } + } + }, + "com.amazonaws.amplifyuibuilder#ListCodegenJobsRequest": { + "type": "structure", + "members": { + "appId": { + "target": "com.amazonaws.amplifyuibuilder#AppId", + "traits": { + "smithy.api#documentation": "The unique ID for the Amplify app.
", + "smithy.api#httpLabel": {}, + "smithy.api#required": {} + } + }, + "environmentName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the backend environment that is a part of the Amplify app.
", + "smithy.api#httpLabel": {}, + "smithy.api#required": {} + } + }, + "nextToken": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The token to request the next page of results.
", + "smithy.api#httpQuery": "nextToken" + } + }, + "maxResults": { + "target": "com.amazonaws.amplifyuibuilder#ListCodegenJobsLimit", + "traits": { + "smithy.api#default": 100, + "smithy.api#documentation": "The maximum number of jobs to retrieve.
", + "smithy.api#httpQuery": "maxResults" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.amplifyuibuilder#ListCodegenJobsResponse": { + "type": "structure", + "members": { + "entities": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobSummaryList", + "traits": { + "smithy.api#documentation": "The list of code generation jobs for the Amplify app.
", + "smithy.api#required": {} + } + }, + "nextToken": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The pagination token that's included if more results are available.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, "com.amazonaws.amplifyuibuilder#ListComponents": { "type": "operation", "input": { @@ -4268,6 +5199,46 @@ } } }, + "com.amazonaws.amplifyuibuilder#ReactStartCodegenJobData": { + "type": "structure", + "members": { + "module": { + "target": "com.amazonaws.amplifyuibuilder#JSModule", + "traits": { + "smithy.api#documentation": "The JavaScript module type.
" + } + }, + "target": { + "target": "com.amazonaws.amplifyuibuilder#JSTarget", + "traits": { + "smithy.api#documentation": "The ECMAScript specification to use.
" + } + }, + "script": { + "target": "com.amazonaws.amplifyuibuilder#JSScript", + "traits": { + "smithy.api#documentation": "The file type to use for a JavaScript project.
" + } + }, + "renderTypeDeclarations": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#default": false, + "smithy.api#documentation": "Specifies whether the code generation job should render type declaration files.
" + } + }, + "inlineSourceMap": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#default": false, + "smithy.api#documentation": "Specifies whether the code generation job should render inline source maps.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Describes the code generation job configuration for a React project.
" + } + }, "com.amazonaws.amplifyuibuilder#RefreshToken": { "type": "operation", "input": { @@ -4350,6 +5321,12 @@ } } }, + "com.amazonaws.amplifyuibuilder#RelatedModelFieldsList": { + "type": "list", + "member": { + "target": "smithy.api#String" + } + }, "com.amazonaws.amplifyuibuilder#ResourceConflictException": { "type": "structure", "members": { @@ -4490,6 +5467,132 @@ "target": "com.amazonaws.amplifyuibuilder#SortProperty" } }, + "com.amazonaws.amplifyuibuilder#StartCodegenJob": { + "type": "operation", + "input": { + "target": "com.amazonaws.amplifyuibuilder#StartCodegenJobRequest" + }, + "output": { + "target": "com.amazonaws.amplifyuibuilder#StartCodegenJobResponse" + }, + "errors": [ + { + "target": "com.amazonaws.amplifyuibuilder#InternalServerException" + }, + { + "target": "com.amazonaws.amplifyuibuilder#InvalidParameterException" + }, + { + "target": "com.amazonaws.amplifyuibuilder#ThrottlingException" + } + ], + "traits": { + "aws.iam#requiredActions": [ + "amplify:GetApp" + ], + "smithy.api#documentation": "Starts a code generation job for for a specified Amplify app and backend environment.
", + "smithy.api#http": { + "uri": "/app/{appId}/environment/{environmentName}/codegen-jobs", + "method": "POST" + } + } + }, + "com.amazonaws.amplifyuibuilder#StartCodegenJobData": { + "type": "structure", + "members": { + "renderConfig": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobRenderConfig", + "traits": { + "smithy.api#documentation": "The code generation configuration for the codegen job.
", + "smithy.api#required": {} + } + }, + "genericDataSchema": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJobGenericDataSchema", + "traits": { + "smithy.api#documentation": "The data schema to use for a code generation job.
" + } + }, + "autoGenerateForms": { + "target": "smithy.api#Boolean", + "traits": { + "smithy.api#documentation": "Specifies whether to autogenerate forms in the code generation job.
" + } + }, + "features": { + "target": "com.amazonaws.amplifyuibuilder#CodegenFeatureFlags", + "traits": { + "smithy.api#documentation": "The feature flags for a code generation job.
" + } + }, + "tags": { + "target": "com.amazonaws.amplifyuibuilder#Tags", + "traits": { + "smithy.api#documentation": "One or more key-value pairs to use when tagging the code generation job data.
" + } + } + }, + "traits": { + "smithy.api#documentation": "The code generation job resource configuration.
" + } + }, + "com.amazonaws.amplifyuibuilder#StartCodegenJobRequest": { + "type": "structure", + "members": { + "appId": { + "target": "com.amazonaws.amplifyuibuilder#AppId", + "traits": { + "smithy.api#documentation": "The unique ID for the Amplify app.
", + "smithy.api#httpLabel": {}, + "smithy.api#required": {} + } + }, + "environmentName": { + "target": "smithy.api#String", + "traits": { + "smithy.api#documentation": "The name of the backend environment that is a part of the Amplify app.
", + "smithy.api#httpLabel": {}, + "smithy.api#required": {} + } + }, + "clientToken": { + "target": "smithy.api#String", + "traits": { + "aws.cloudformation#cfnExcludeProperty": {}, + "smithy.api#documentation": "The idempotency token used to ensure that the code generation job request completes only once.
", + "smithy.api#httpQuery": "clientToken", + "smithy.api#idempotencyToken": {} + } + }, + "codegenJobToCreate": { + "target": "com.amazonaws.amplifyuibuilder#StartCodegenJobData", + "traits": { + "aws.cloudformation#cfnExcludeProperty": {}, + "smithy.api#documentation": "The code generation job resource configuration.
", + "smithy.api#httpPayload": {}, + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.amplifyuibuilder#StartCodegenJobResponse": { + "type": "structure", + "members": { + "entity": { + "target": "com.amazonaws.amplifyuibuilder#CodegenJob", + "traits": { + "smithy.api#documentation": "The code generation job for a UI component that is associated with an Amplify app.
", + "smithy.api#httpPayload": {} + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, "com.amazonaws.amplifyuibuilder#StorageAccessLevel": { "type": "string", "traits": { @@ -4772,6 +5875,19 @@ "target": "com.amazonaws.amplifyuibuilder#ThemeValues" } }, + "com.amazonaws.amplifyuibuilder#ThrottlingException": { + "type": "structure", + "members": { + "message": { + "target": "smithy.api#String" + } + }, + "traits": { + "smithy.api#documentation": "The request was denied due to request throttling.
", + "smithy.api#error": "client", + "smithy.api#httpError": 429 + } + }, "com.amazonaws.amplifyuibuilder#TokenProviders": { "type": "string", "traits": { diff --git a/codegen/sdk-codegen/aws-models/dynamodb-streams.json b/codegen/sdk-codegen/aws-models/dynamodb-streams.json index 16b2846c95a..aefb201e259 100644 --- a/codegen/sdk-codegen/aws-models/dynamodb-streams.json +++ b/codegen/sdk-codegen/aws-models/dynamodb-streams.json @@ -174,7 +174,8 @@ } }, "traits": { - "smithy.api#documentation": "Represents the input of a DescribeStream
operation.
Represents the input of a DescribeStream
operation.
Represents the output of a DescribeStream
operation.
Represents the output of a DescribeStream
operation.
Amazon DynamoDB Streams provides API actions for accessing streams and processing\n stream records. To learn more about application development with Streams, see Capturing\n Table Activity with DynamoDB Streams in the Amazon DynamoDB Developer\n Guide.
", + "smithy.api#documentation": "Amazon DynamoDB Streams provides API actions for accessing streams and processing\n stream records. To learn more about application development with Streams, see Capturing\n Table Activity with DynamoDB Streams in the Amazon DynamoDB Developer\n Guide.
", "smithy.api#title": "Amazon DynamoDB Streams", "smithy.api#xmlNamespace": { "uri": "http://dynamodb.amazonaws.com/doc/2012-08-10/" @@ -763,8 +765,8 @@ }, "params": { "Region": "af-south-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -776,8 +778,8 @@ }, "params": { "Region": "ap-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -789,8 +791,8 @@ }, "params": { "Region": "ap-northeast-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -802,8 +804,8 @@ }, "params": { "Region": "ap-northeast-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -815,8 +817,8 @@ }, "params": { "Region": "ap-northeast-3", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -828,8 +830,8 @@ }, "params": { "Region": "ap-south-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -841,8 +843,8 @@ }, "params": { "Region": "ap-southeast-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -854,8 +856,8 @@ }, "params": { "Region": "ap-southeast-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -867,8 +869,8 @@ }, "params": { "Region": "ap-southeast-3", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -880,8 +882,8 @@ }, "params": { "Region": "ca-central-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -893,8 +895,8 @@ }, "params": { "Region": "eu-central-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -906,8 +908,8 @@ }, "params": { "Region": "eu-north-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -919,8 +921,8 @@ }, "params": { "Region": "eu-south-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -932,8 +934,8 @@ }, "params": { "Region": "eu-west-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -945,8 +947,8 @@ }, "params": { "Region": "eu-west-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -958,8 +960,8 @@ }, "params": { "Region": "eu-west-3", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -980,8 +982,8 @@ }, "params": { "Region": "local", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -993,8 +995,8 @@ }, "params": { "Region": "me-south-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1006,8 +1008,8 @@ }, "params": { "Region": "sa-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1019,8 +1021,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1032,8 +1034,8 @@ }, "params": { "Region": "us-east-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1045,8 +1047,8 @@ }, "params": { "Region": "us-west-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1058,8 +1060,8 @@ }, "params": { "Region": "us-west-2", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1071,8 +1073,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -1084,8 +1086,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -1097,8 +1099,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -1110,8 +1112,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1123,8 +1125,8 @@ }, "params": { "Region": "cn-northwest-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1136,8 +1138,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -1149,8 +1151,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -1162,8 +1164,8 @@ }, "params": { "Region": "cn-north-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -1175,8 +1177,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1188,8 +1190,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -1201,8 +1203,8 @@ }, "params": { "Region": "us-gov-west-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false } }, { @@ -1214,8 +1216,8 @@ }, "params": { "Region": "us-gov-west-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false } }, { @@ -1227,8 +1229,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": true, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": true } }, { @@ -1240,8 +1242,8 @@ }, "params": { "Region": "us-gov-east-1", - "UseDualStack": true, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": true } }, { @@ -1253,8 +1255,19 @@ }, "params": { "Region": "us-iso-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + }, + "params": { + "Region": "us-iso-east-1", + "UseFIPS": true, + "UseDualStack": true } }, { @@ -1266,8 +1279,19 @@ }, "params": { "Region": "us-iso-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false + } + }, + { + "documentation": "For region us-iso-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "error": "DualStack is enabled but this partition does not support DualStack" + }, + "params": { + "Region": "us-iso-east-1", + "UseFIPS": false, + "UseDualStack": true } }, { @@ -1279,8 +1303,19 @@ }, "params": { "Region": "us-isob-east-1", - "UseDualStack": false, - "UseFIPS": false + "UseFIPS": false, + "UseDualStack": false + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS enabled and DualStack enabled", + "expect": { + "error": "FIPS and DualStack are enabled, but this partition does not support one or both" + }, + "params": { + "Region": "us-isob-east-1", + "UseFIPS": true, + "UseDualStack": true } }, { @@ -1292,8 +1327,19 @@ }, "params": { "Region": "us-isob-east-1", - "UseDualStack": false, - "UseFIPS": true + "UseFIPS": true, + "UseDualStack": false + } + }, + { + "documentation": "For region us-isob-east-1 with FIPS disabled and DualStack enabled", + "expect": { + "error": "DualStack is enabled but this partition does not support DualStack" + }, + "params": { + "Region": "us-isob-east-1", + "UseFIPS": false, + "UseDualStack": true } }, { @@ -1305,8 +1351,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, "UseFIPS": false, + "UseDualStack": false, "Endpoint": "https://example.com" } }, @@ -1318,8 +1364,8 @@ } }, "params": { - "UseDualStack": false, "UseFIPS": false, + "UseDualStack": false, "Endpoint": "https://example.com" } }, @@ -1330,8 +1376,8 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": false, "UseFIPS": true, + "UseDualStack": false, "Endpoint": "https://example.com" } }, @@ -1342,10 +1388,16 @@ }, "params": { "Region": "us-east-1", - "UseDualStack": true, "UseFIPS": false, + "UseDualStack": true, "Endpoint": "https://example.com" } + }, + { + "documentation": "Missing region", + "expect": { + "error": "Invalid Configuration: Missing Region" + } } ], "version": "1.0" @@ -1417,7 +1469,8 @@ } }, "traits": { - "smithy.api#documentation": "Represents the input of a GetRecords
operation.
Represents the input of a GetRecords
operation.
Represents the output of a GetRecords
operation.
Represents the output of a GetRecords
operation.
Represents the input of a GetShardIterator
operation.
Represents the input of a GetShardIterator
operation.
Represents the output of a GetShardIterator
operation.
Represents the output of a GetShardIterator
operation.
The role that this key attribute will assume:
\n\n HASH
- partition key
\n RANGE
- sort key
The partition key of an item is also known as its hash\n attribute. The term \"hash attribute\" derives from DynamoDB's usage of\n an internal hash function to evenly distribute data items across partitions, based\n on their partition key values.
\nThe sort key of an item is also known as its range\n attribute. The term \"range attribute\" derives from the way DynamoDB\n stores items with the same partition key physically close together, in sorted order\n by the sort key value.
\nThe role that this key attribute will assume:
\n\n HASH
- partition key
\n RANGE
- sort key
The partition key of an item is also known as its hash\n attribute. The term \"hash attribute\" derives from DynamoDB's usage of\n an internal hash function to evenly distribute data items across partitions, based\n on their partition key values.
\nThe sort key of an item is also known as its range\n attribute. The term \"range attribute\" derives from the way DynamoDB\n stores items with the same partition key physically close together, in sorted order\n by the sort key value.
\nThere is no limit to the number of daily on-demand backups that can be taken.
\nFor most purposes, up to 500 simultaneous table operations are allowed per account. These operations\n include CreateTable
, UpdateTable
,\n DeleteTable
,UpdateTimeToLive
,\n RestoreTableFromBackup
, and RestoreTableToPointInTime
.
When you are creating a table with one or more secondary\n indexes, you can have up to 250 such requests running at a time. However, if the table or\n index specifications are complex, then DynamoDB might temporarily reduce the number\n of concurrent operations.
\nWhen importing into DynamoDB, up to 50 simultaneous import table operations are allowed per account.
\nThere is a soft account quota of 2,500 tables.
", + "smithy.api#documentation": "There is no limit to the number of daily on-demand backups that can be taken.
\nFor most purposes, up to 500 simultaneous table operations are allowed per account. These operations\n include CreateTable
, UpdateTable
,\n DeleteTable
,UpdateTimeToLive
,\n RestoreTableFromBackup
, and RestoreTableToPointInTime
.
When you are creating a table with one or more secondary\n indexes, you can have up to 250 such requests running at a time. However, if the table or\n index specifications are complex, then DynamoDB might temporarily reduce the number\n of concurrent operations.
\nWhen importing into DynamoDB, up to 50 simultaneous import table operations are allowed per account.
\nThere is a soft account quota of 2,500 tables.
\nGetRecords was called with a value of more than 1000 for the limit request parameter.
\nMore than 2 processes are reading from the same streams shard at the same time. Exceeding\n this limit may result in request throttling.
", "smithy.api#error": "client" } }, @@ -1671,7 +1727,8 @@ } }, "traits": { - "smithy.api#documentation": "Represents the input of a ListStreams
operation.
Represents the input of a ListStreams
operation.
Represents the output of a ListStreams
operation.
Represents the output of a ListStreams
operation.
The AWS service from which the stream record originated. For DynamoDB Streams, this is aws:dynamodb
.
The Amazon Web Services service from which the stream record originated. For DynamoDB Streams, this is aws:dynamodb
.
A timestamp, in ISO 8601 format, for this stream.
\nNote that LatestStreamLabel
is not a unique identifier for the stream, because it is\n possible that a stream from another table might have the same timestamp. However, the\n combination of the following three elements is guaranteed to be unique:
the AWS customer ID.
\nthe table name
\nthe StreamLabel
\n
A timestamp, in ISO 8601 format, for this stream.
\nNote that LatestStreamLabel
is not a unique identifier for the stream, because it is\n possible that a stream from another table might have the same timestamp. However, the\n combination of the following three elements is guaranteed to be unique:
the Amazon Web Services customer ID.
\nthe table name
\nthe StreamLabel
\n
A timestamp, in ISO 8601 format, for this stream.
\nNote that LatestStreamLabel
is not a unique identifier for the stream, because it is\n possible that a stream from another table might have the same timestamp. However, the\n combination of the following three elements is guaranteed to be unique:
the AWS customer ID.
\nthe table name
\nthe StreamLabel
\n
A timestamp, in ISO 8601 format, for this stream.
\nNote that LatestStreamLabel
is not a unique identifier for the stream, because it is\n possible that a stream from another table might have the same timestamp. However, the\n combination of the following three elements is guaranteed to be unique:
the Amazon Web Services customer ID.
\nthe table name
\nthe StreamLabel
\n
The approximate date and time when the stream record was created, in UNIX epoch time format.
" + "smithy.api#documentation": "The approximate date and time when the stream record was created, in UNIX epoch time format and rounded down to the closest second.
" } }, "Keys": { diff --git a/codegen/sdk-codegen/aws-models/dynamodb.json b/codegen/sdk-codegen/aws-models/dynamodb.json index adb4cf00ebc..7c5a2d76787 100644 --- a/codegen/sdk-codegen/aws-models/dynamodb.json +++ b/codegen/sdk-codegen/aws-models/dynamodb.json @@ -833,7 +833,7 @@ "aws.api#clientDiscoveredEndpoint": { "required": false }, - "smithy.api#documentation": "The BatchGetItem
operation returns the attributes of one or more items\n from one or more tables. You identify requested items by primary key.
A single operation can retrieve up to 16 MB of data, which can contain as many as 100\n items. BatchGetItem
returns a partial result if the response size limit is\n exceeded, the table's provisioned throughput is exceeded, or an internal processing\n failure occurs. If a partial result is returned, the operation returns a value for\n UnprocessedKeys
. You can use this value to retry the operation starting\n with the next item to get.
If you request more than 100 items, BatchGetItem
returns a\n ValidationException
with the message \"Too many items requested for\n the BatchGetItem call.\"
For example, if you ask to retrieve 100 items, but each individual item is 300 KB in\n size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns\n an appropriate UnprocessedKeys
value so you can get the next page of\n results. If desired, your application can include its own logic to assemble the pages of\n results into one dataset.
If none of the items can be processed due to insufficient\n provisioned throughput on all of the tables in the request, then\n BatchGetItem
returns a\n ProvisionedThroughputExceededException
. If at least\n one of the items is successfully processed, then\n BatchGetItem
completes successfully, while returning the keys of the\n unread items in UnprocessedKeys
.
If DynamoDB returns any unprocessed items, you should retry the batch operation on\n those items. However, we strongly recommend that you use an exponential\n backoff algorithm. If you retry the batch operation immediately, the\n underlying read or write requests can still fail due to throttling on the individual\n tables. If you delay the batch operation using exponential backoff, the individual\n requests in the batch are much more likely to succeed.
\nFor more information, see Batch Operations and Error Handling in the Amazon DynamoDB\n Developer Guide.
\nBy default, BatchGetItem
performs eventually consistent reads on every\n table in the request. If you want strongly consistent reads instead, you can set\n ConsistentRead
to true
for any or all tables.
In order to minimize response latency, BatchGetItem
may retrieve items in\n parallel.
When designing your application, keep in mind that DynamoDB does not return items in\n any particular order. To help parse the response by item, include the primary key values\n for the items in your request in the ProjectionExpression
parameter.
If a requested item does not exist, it is not returned in the result. Requests for\n nonexistent items consume the minimum read capacity units according to the type of read.\n For more information, see Working with Tables in the Amazon DynamoDB Developer\n Guide.
" + "smithy.api#documentation": "The BatchGetItem
operation returns the attributes of one or more items\n from one or more tables. You identify requested items by primary key.
A single operation can retrieve up to 16 MB of data, which can contain as many as 100\n items. BatchGetItem
returns a partial result if the response size limit is\n exceeded, the table's provisioned throughput is exceeded, more than 1MB per partition is requested,\n or an internal processing failure occurs. If a partial result is returned, the operation returns a value for\n UnprocessedKeys
. You can use this value to retry the operation starting\n with the next item to get.
If you request more than 100 items, BatchGetItem
returns a\n ValidationException
with the message \"Too many items requested for\n the BatchGetItem call.\"
For example, if you ask to retrieve 100 items, but each individual item is 300 KB in\n size, the system returns 52 items (so as not to exceed the 16 MB limit). It also returns\n an appropriate UnprocessedKeys
value so you can get the next page of\n results. If desired, your application can include its own logic to assemble the pages of\n results into one dataset.
If none of the items can be processed due to insufficient\n provisioned throughput on all of the tables in the request, then\n BatchGetItem
returns a\n ProvisionedThroughputExceededException
. If at least\n one of the items is successfully processed, then\n BatchGetItem
completes successfully, while returning the keys of the\n unread items in UnprocessedKeys
.
If DynamoDB returns any unprocessed items, you should retry the batch operation on\n those items. However, we strongly recommend that you use an exponential\n backoff algorithm. If you retry the batch operation immediately, the\n underlying read or write requests can still fail due to throttling on the individual\n tables. If you delay the batch operation using exponential backoff, the individual\n requests in the batch are much more likely to succeed.
\nFor more information, see Batch Operations and Error Handling in the Amazon DynamoDB\n Developer Guide.
\nBy default, BatchGetItem
performs eventually consistent reads on every\n table in the request. If you want strongly consistent reads instead, you can set\n ConsistentRead
to true
for any or all tables.
In order to minimize response latency, BatchGetItem
may retrieve items in\n parallel.
When designing your application, keep in mind that DynamoDB does not return items in\n any particular order. To help parse the response by item, include the primary key values\n for the items in your request in the ProjectionExpression
parameter.
If a requested item does not exist, it is not returned in the result. Requests for\n nonexistent items consume the minimum read capacity units according to the type of read.\n For more information, see Working with Tables in the Amazon DynamoDB Developer\n Guide.
" } }, "com.amazonaws.dynamodb#BatchGetItemInput": { @@ -6732,7 +6732,7 @@ } }, "traits": { - "smithy.api#documentation": "There is no limit to the number of daily on-demand backups that can be taken.
\nFor most purposes, up to 500 simultaneous table operations are allowed per account. These operations\n include CreateTable
, UpdateTable
,\n DeleteTable
,UpdateTimeToLive
,\n RestoreTableFromBackup
, and RestoreTableToPointInTime
.
When you are creating a table with one or more secondary\n indexes, you can have up to 250 such requests running at a time. However, if the table or\n index specifications are complex, then DynamoDB might temporarily reduce the number\n of concurrent operations.
\nWhen importing into DynamoDB, up to 50 simultaneous import table operations are allowed per account.
\nThere is a soft account quota of 2,500 tables.
", + "smithy.api#documentation": "There is no limit to the number of daily on-demand backups that can be taken.
\nFor most purposes, up to 500 simultaneous table operations are allowed per account. These operations\n include CreateTable
, UpdateTable
,\n DeleteTable
,UpdateTimeToLive
,\n RestoreTableFromBackup
, and RestoreTableToPointInTime
.
When you are creating a table with one or more secondary\n indexes, you can have up to 250 such requests running at a time. However, if the table or\n index specifications are complex, then DynamoDB might temporarily reduce the number\n of concurrent operations.
\nWhen importing into DynamoDB, up to 50 simultaneous import table operations are allowed per account.
\nThere is a soft account quota of 2,500 tables.
\nGetRecords was called with a value of more than 1000 for the limit request parameter.
\nMore than 2 processes are reading from the same streams shard at the same time. Exceeding\n this limit may result in request throttling.
", "smithy.api#error": "client" } }, @@ -7689,14 +7689,14 @@ "ReadCapacityUnits": { "target": "com.amazonaws.dynamodb#PositiveLongObject", "traits": { - "smithy.api#documentation": "The maximum number of strongly consistent reads consumed per second before DynamoDB\n returns a ThrottlingException
. For more information, see Specifying Read and Write Requirements in the Amazon DynamoDB\n Developer Guide.
If read/write capacity mode is PAY_PER_REQUEST
the value is set to\n 0.
The maximum number of strongly consistent reads consumed per second before DynamoDB\n returns a ThrottlingException
. For more information, see Specifying Read and Write Requirements in the Amazon DynamoDB\n Developer Guide.
If read/write capacity mode is PAY_PER_REQUEST
the value is set to\n 0.
The maximum number of writes consumed per second before DynamoDB returns a\n ThrottlingException
. For more information, see Specifying Read and Write Requirements in the Amazon DynamoDB\n Developer Guide.
If read/write capacity mode is PAY_PER_REQUEST
the value is set to\n 0.
The maximum number of writes consumed per second before DynamoDB returns a\n ThrottlingException
. For more information, see Specifying Read and Write Requirements in the Amazon DynamoDB\n Developer Guide.
If read/write capacity mode is PAY_PER_REQUEST
the value is set to\n 0.
Creates an Amazon FSx for Lustre data repository association (DRA). A data\n repository association is a link between a directory on the file system and\n an Amazon S3 bucket or prefix. You can have a maximum of 8 data repository\n associations on a file system. Data repository associations are supported\n for all file systems except for Scratch_1
deployment type.
Each data repository association must have a unique Amazon FSx file\n system directory and a unique S3 bucket or prefix associated with it. You\n can configure a data repository association for automatic import only,\n for automatic export only, or for both. To learn more about linking a\n data repository to your file system, see \n Linking your file system to an S3 bucket.
\n\n CreateDataRepositoryAssociation
isn't supported\n on Amazon File Cache resources. To create a DRA on Amazon File Cache,\n use the CreateFileCache
operation.
Creates an Amazon FSx for Lustre data repository association (DRA). A data\n repository association is a link between a directory on the file system and\n an Amazon S3 bucket or prefix. You can have a maximum of 8 data repository\n associations on a file system. Data repository associations are supported\n on all FSx for Lustre 2.12 and newer file systems, excluding\n scratch_1
deployment type.
Each data repository association must have a unique Amazon FSx file\n system directory and a unique S3 bucket or prefix associated with it. You\n can configure a data repository association for automatic import only,\n for automatic export only, or for both. To learn more about linking a\n data repository to your file system, see \n Linking your file system to an S3 bucket.
\n\n CreateDataRepositoryAssociation
isn't supported\n on Amazon File Cache resources. To create a DRA on Amazon File Cache,\n use the CreateFileCache
operation.
Specifies the file system deployment type. Single AZ deployment types are configured\n for redundancy within a single Availability Zone in an Amazon Web Services Region .\n Valid values are the following:
\n\n SINGLE_AZ_1
- (Default) Creates file systems with throughput capacities of 64 - 4,096 MB/s.\n Single_AZ_1
is available in all Amazon Web Services Regions where Amazon FSx \n for OpenZFS is available, except US West (Oregon).
\n SINGLE_AZ_2
- Creates file systems with throughput capacities of 160 - 10,240 MB/s\n using an NVMe L2ARC cache. Single_AZ_2
is available only in the US East (N. Virginia), US East (Ohio), \n US West (Oregon), and Europe (Ireland) Amazon Web Services Regions.
For more information, see: Deployment type availability\n and File system performance\n in the Amazon FSx for OpenZFS User Guide.
", + "smithy.api#documentation": "Specifies the file system deployment type. Single AZ deployment types are configured\n for redundancy within a single Availability Zone in an Amazon Web Services Region .\n Valid values are the following:
\n\n SINGLE_AZ_1
- (Default) Creates file systems with throughput capacities of 64 - 4,096 MBps.\n Single_AZ_1
is available in all Amazon Web Services Regions where Amazon FSx \n for OpenZFS is available, except US West (Oregon).
\n SINGLE_AZ_2
- Creates file systems with throughput capacities of 160 - 10,240 MBps\n using an NVMe L2ARC cache. Single_AZ_2
is available only in the US East (N. Virginia), US East (Ohio), \n US West (Oregon), and Europe (Ireland) Amazon Web Services Regions.
For more information, see: Deployment type availability\n and File system performance\n in the Amazon FSx for OpenZFS User Guide.
", "smithy.api#required": {} } }, "ThroughputCapacity": { "target": "com.amazonaws.fsx#MegabytesPerSecond", "traits": { - "smithy.api#documentation": "Specifies the throughput of an Amazon FSx for OpenZFS file system, measured in megabytes per second (MB/s). Valid values depend on the DeploymentType you choose, as follows:
\nFor SINGLE_AZ_1
, valid values are 64, 128, 256, 512, 1024, 2048, 3072, or 4096 MB/s.
For SINGLE_AZ_2
, valid values are 160, 320, 640, 1280, 2560, 3840, 5120, 7680, or 10240 MB/s.
You pay for additional throughput capacity that you provision.
", + "smithy.api#documentation": "Specifies the throughput of an Amazon FSx for OpenZFS file system, measured in megabytes per second (MBps). Valid values depend on the DeploymentType you choose, as follows:
\nFor SINGLE_AZ_1
, valid values are 64, 128, 256, 512, 1024, 2048, 3072, or 4096 MBps.
For SINGLE_AZ_2
, valid values are 160, 320, 640, 1280, 2560, 3840, 5120, 7680, or 10240 MBps.
You pay for additional throughput capacity that you provision.
", "smithy.api#required": {} } }, @@ -3717,7 +3717,7 @@ "com.amazonaws.fsx#DNSName": { "type": "string", "traits": { - "smithy.api#documentation": "The Domain Name Service (DNS) name for the file system. You can mount your file\n system using its DNS name.
", + "smithy.api#documentation": "The file system's DNS name. You can mount your file\n system using its DNS name.
", "smithy.api#length": { "min": 16, "max": 275 @@ -3839,7 +3839,7 @@ } }, "traits": { - "smithy.api#documentation": "The configuration of a data repository association that links\n an Amazon FSx for Lustre file system to an Amazon S3 bucket\n or an Amazon File Cache resource to an Amazon S3 bucket or an NFS file system.\n The data repository association configuration object is returned\n in the response of the following operations:
\n\n CreateDataRepositoryAssociation
\n
\n UpdateDataRepositoryAssociation
\n
\n DescribeDataRepositoryAssociations
\n
Data repository associations are supported on Amazon File Cache resources and\n all Amazon FSx for Lustre file systems excluding Scratch_1
deployment\n types.
The configuration of a data repository association that links\n an Amazon FSx for Lustre file system to an Amazon S3 bucket\n or an Amazon File Cache resource to an Amazon S3 bucket or an NFS file system.\n The data repository association configuration object is returned\n in the response of the following operations:
\n\n CreateDataRepositoryAssociation
\n
\n UpdateDataRepositoryAssociation
\n
\n DescribeDataRepositoryAssociations
\n
Data repository associations are supported on Amazon File Cache resources and\n all FSx for Lustre 2.12 and newer file systems, excluding\n scratch_1
deployment type.
Deletes a data repository association on an Amazon FSx for Lustre\n file system. Deleting the data repository association unlinks the\n file system from the Amazon S3 bucket. When deleting a data repository\n association, you have the option of deleting the data in the file system\n that corresponds to the data repository association. Data repository\n associations are supported for all file systems except for Scratch_1
\n deployment type.
Deletes a data repository association on an Amazon FSx for Lustre\n file system. Deleting the data repository association unlinks the\n file system from the Amazon S3 bucket. When deleting a data repository\n association, you have the option of deleting the data in the file system\n that corresponds to the data repository association. Data repository\n associations are supported on all FSx for Lustre 2.12 and newer file\n systems, excluding scratch_1
deployment type.
Returns the description of specific Amazon FSx for Lustre or Amazon File Cache\n data repository associations, if one or more AssociationIds
values\n are provided in the request, or if filters are used in the request. Data repository\n associations are supported on Amazon File Cache resources and all Amazon FSx for\n Lustre file systems excluding Scratch_1
deployment types.
You can use filters to narrow the response to include just data repository\n associations for specific file systems (use the file-system-id
filter with\n the ID of the file system) or caches (use the file-cache-id
filter with\n the ID of the cache), or data repository associations for a specific repository type\n (use the data-repository-type
filter with a value of S3
\n or NFS
). If you don't use filters, the response returns all data\n repository associations owned by your Amazon Web Services account in the Amazon Web Services Region\n of the endpoint that you're calling.
When retrieving all data repository associations, you can paginate the response by using\n the optional MaxResults
parameter to limit the number of data repository associations\n returned in a response. If more data repository associations remain, a\n NextToken
value is returned in the response. In this case, send a later\n request with the NextToken
request parameter set to the value of\n NextToken
from the last response.
Returns the description of specific Amazon FSx for Lustre or Amazon File Cache\n data repository associations, if one or more AssociationIds
values\n are provided in the request, or if filters are used in the request. Data repository\n associations are supported on Amazon File Cache resources and all FSx for Lustre\n 2.12 and newer file systems, excluding scratch_1
deployment type.
You can use filters to narrow the response to include just data repository\n associations for specific file systems (use the file-system-id
filter with\n the ID of the file system) or caches (use the file-cache-id
filter with\n the ID of the cache), or data repository associations for a specific repository type\n (use the data-repository-type
filter with a value of S3
\n or NFS
). If you don't use filters, the response returns all data\n repository associations owned by your Amazon Web Services account in the Amazon Web Services Region\n of the endpoint that you're calling.
When retrieving all data repository associations, you can paginate the response by using\n the optional MaxResults
parameter to limit the number of data repository associations\n returned in a response. If more data repository associations remain, a\n NextToken
value is returned in the response. In this case, send a later\n request with the NextToken
request parameter set to the value of\n NextToken
from the last response.
Specifies whether the number of IOPS for the file system is\n using the system default (AUTOMATIC
) or was\n provisioned by the customer (USER_PROVISIONED
).
Specifies whether the file system is \n using the AUTOMATIC
setting of SSD IOPS of 3 IOPS per GB of storage capacity, , or \n if it using a USER_PROVISIONED
value.
The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for NetApp ONTAP or Amazon FSx for OpenZFS file system. The\n default is 3 IOPS per GB of storage capacity, but you can provision additional IOPS per\n GB of storage. The configuration consists of the total number of provisioned SSD IOPS\n and how the amount was provisioned (by the customer or by the system).
" + "smithy.api#documentation": "The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for NetApp ONTAP or FSx for OpenZFS file system. By default, Amazon FSx \n automatically provisions 3 IOPS per GB of storage capacity. You can provision additional IOPS per\n GB of storage. The configuration consists of the total number of provisioned SSD IOPS\n and how it is was provisioned, or the mode (by the customer or by Amazon FSx).
" } }, "com.amazonaws.fsx#DiskIopsConfigurationMode": { @@ -7773,6 +7776,12 @@ }, "WeeklyMaintenanceStartTime": { "target": "com.amazonaws.fsx#WeeklyTime" + }, + "FsxAdminPassword": { + "target": "com.amazonaws.fsx#AdminPassword", + "traits": { + "smithy.api#documentation": "You can use the fsxadmin
user account to access the NetApp ONTAP CLI and \n REST API. The password value is always redacted in the response.
The current percent of progress of an asynchronous task.
", + "smithy.api#documentation": "Displays the current percent of progress of an asynchronous task.
", "smithy.api#range": { "min": 0, "max": 100 @@ -8796,7 +8805,7 @@ } }, "traits": { - "smithy.api#documentation": "The configuration that Amazon FSx uses to join a FSx for Windows File Server file system or an ONTAP storage virtual machine (SVM) to\n a self-managed (including on-premises) Microsoft Active Directory (AD)\n directory. For more information, see \n \n Using Amazon FSx with your self-managed Microsoft Active Directory or \n Managing SVMs.
" + "smithy.api#documentation": "The configuration that Amazon FSx uses to join a FSx for Windows File Server file system or an FSx for ONTAP storage virtual machine (SVM) to\n a self-managed (including on-premises) Microsoft Active Directory (AD)\n directory. For more information, see \n \n Using Amazon FSx for Windows with your self-managed Microsoft Active Directory or \n Managing FSx for ONTAP SVMs.
" } }, "com.amazonaws.fsx#SelfManagedActiveDirectoryConfigurationUpdates": { @@ -8805,24 +8814,42 @@ "UserName": { "target": "com.amazonaws.fsx#DirectoryUserName", "traits": { - "smithy.api#documentation": "The user name for the service account on your self-managed AD domain that Amazon FSx will use to join to\n your AD domain. This account must have the permission to join\n computers to the domain in the organizational unit provided in\n OrganizationalUnitDistinguishedName
.
Specifies the updated user name for the service account on your self-managed AD domain.\n Amazon FSx uses this account to join to your self-managed AD domain.
\nThis account must have the permissions required to join\n computers to the domain in the organizational unit provided in\n OrganizationalUnitDistinguishedName
.
The password for the service account on your self-managed AD domain that Amazon FSx will use to join to\n your AD domain.
" + "smithy.api#documentation": "Specifies the updated password for the service account on your self-managed AD domain. \n Amazon FSx uses this account to join to your self-managed AD domain.
" } }, "DnsIps": { "target": "com.amazonaws.fsx#DnsIps", "traits": { - "smithy.api#documentation": "A list of up to three IP addresses of DNS servers or domain controllers in the\n self-managed AD directory.
" + "smithy.api#documentation": "A list of up to three DNS server or domain controller IP addresses in your\n self-managed AD domain.
" + } + }, + "DomainName": { + "target": "com.amazonaws.fsx#ActiveDirectoryFullyQualifiedName", + "traits": { + "smithy.api#documentation": "Specifies an updated fully qualified domain name of your self-managed AD configuration.
" + } + }, + "OrganizationalUnitDistinguishedName": { + "target": "com.amazonaws.fsx#OrganizationalUnitDistinguishedName", + "traits": { + "smithy.api#documentation": "Specifies an updated fully qualified distinguished name of the organization unit within your self-managed AD.
" + } + }, + "FileSystemAdministratorsGroup": { + "target": "com.amazonaws.fsx#FileSystemAdministratorsGroupName", + "traits": { + "smithy.api#documentation": "Specifies the updated name of the self-managed AD domain group whose members are granted administrative privileges\n for the Amazon FSx resource.
" } } }, "traits": { - "smithy.api#documentation": "The configuration that Amazon FSx uses to join the Windows File Server instance to a\n self-managed Microsoft Active Directory (AD) directory.
" + "smithy.api#documentation": "Specifies changes you are making to the self-managed Microsoft Active Directory (AD) configuration to which \n an FSx for Windows File Server file system or an FSx for ONTAP SVM is joined.
" } }, "com.amazonaws.fsx#ServiceLimit": { @@ -9193,7 +9220,7 @@ "com.amazonaws.fsx#StorageCapacity": { "type": "integer", "traits": { - "smithy.api#documentation": "The storage capacity for your Amazon FSx file system, in gibibytes.
", + "smithy.api#documentation": "Specifies the file system's storage capacity, in gibibytes (GiB).
", "smithy.api#range": { "min": 0, "max": 2147483647 @@ -9217,7 +9244,7 @@ } }, "traits": { - "smithy.api#documentation": "The storage type for your Amazon FSx file system.
" + "smithy.api#documentation": "Specifies the file system's storage type.
" } }, "com.amazonaws.fsx#StorageVirtualMachine": { @@ -9553,7 +9580,7 @@ "NetBiosName": { "target": "com.amazonaws.fsx#NetBiosAlias", "traits": { - "smithy.api#documentation": "The NetBIOS name of the Active Directory computer object that is joined to your SVM.
" + "smithy.api#documentation": "The NetBIOS name of the AD computer object to which the SVM is joined.
" } }, "SelfManagedActiveDirectoryConfiguration": { @@ -9561,7 +9588,7 @@ } }, "traits": { - "smithy.api#documentation": "Describes the configuration of the Microsoft Active Directory (AD) \n directory to which the Amazon FSx for ONTAP storage virtual machine (SVM) is joined.\n Pleae note, account credentials are not returned in the response payload.
" + "smithy.api#documentation": "Describes the Microsoft Active Directory (AD) directory configuration to which the FSx for ONTAP storage virtual machine (SVM) is joined.\n Note that account credentials are not returned in the response payload.
" } }, "com.amazonaws.fsx#SvmEndpoint": { @@ -9927,7 +9954,7 @@ } ], "traits": { - "smithy.api#documentation": "Updates the configuration of an existing data repository association\n on an Amazon FSx for Lustre file system. Data repository associations\n are supported for all file systems except for Scratch_1
\n deployment type.
Updates the configuration of an existing data repository association\n on an Amazon FSx for Lustre file system. Data repository associations\n are supported on all FSx for Lustre 2.12 and newer file systems,\n excluding scratch_1
deployment type.
The ONTAP administrative password for the fsxadmin
user.
Update the password for the fsxadmin
user by entering a new password. \n You use the fsxadmin
user to access the NetApp ONTAP CLI and REST API to manage your file system resources. \n For more information, see \n Managing resources using NetApp Applicaton.
The SSD IOPS (input/output operations per second) configuration for an Amazon FSx for NetApp ONTAP file system. The default is 3 IOPS per GB of storage capacity,\n but you can provision additional IOPS per GB of storage. The configuration consists\n of an IOPS mode (AUTOMATIC
or USER_PROVISIONED
), and in\n the case of USER_PROVISIONED
IOPS, the total number of SSD IOPS provisioned.
The SSD IOPS (input output operations per second) configuration for an Amazon FSx for NetApp ONTAP file system. The default is 3 IOPS per GB of storage capacity,\n but you can provision additional IOPS per GB of storage. The configuration consists\n of an IOPS mode (AUTOMATIC
or USER_PROVISIONED
), and in\n the case of USER_PROVISIONED
IOPS, the total number of SSD IOPS provisioned. \n For more information, see \n Updating SSD storage capacity and IOPS.
Specifies the throughput of an FSx for NetApp ONTAP file system, measured in megabytes per second\n (MBps). Valid values are 128, 256, 512, 1024, 2048, and 4096 MBps.
" + "smithy.api#documentation": "Enter a new value to change the amount of throughput capacity for the file system. Throughput capacity is measured in megabytes per second\n (MBps). Valid values are 128, 256, 512, 1024, 2048, and 4096 MBps. For more information, see \n Managing throughput capacity \n in the FSx for ONTAP User Guide.
" } }, "AddRouteTableIds": { @@ -10253,7 +10280,7 @@ "StorageCapacity": { "target": "com.amazonaws.fsx#StorageCapacity", "traits": { - "smithy.api#documentation": "Use this parameter to increase the storage capacity of an FSx for Windows File Server,\n FSx for Lustre, FSx for OpenZFS, or FSx for ONTAP file system.\n Specifies the storage capacity target value, in GiB, to increase the storage capacity for\n the file system that you're updating.
\nYou can't make a storage capacity increase request if there is an existing storage\n capacity increase request in progress.
\nFor Lustre file systems, the storage capacity target value can be the following:
\nFor SCRATCH_2
, PERSISTENT_1
, and PERSISTENT_2 SSD
deployment types, valid values\n are in multiples of 2400 GiB. The value must be greater than the current storage capacity.
For PERSISTENT HDD
file systems, valid values are multiples of 6000 GiB for\n 12-MBps throughput per TiB file systems and multiples of 1800 GiB for 40-MBps throughput\n per TiB file systems. The values must be greater than the current storage capacity.
For SCRATCH_1
file systems, you can't increase the storage capacity.
For more information, see Managing storage and throughput\n capacity in the FSx for Lustre User Guide.
\nFor FSx for OpenZFS file systems, the storage capacity target value must be at least 10 percent\n greater than the current storage capacity value. For more information, see\n Managing storage capacity in the FSx for OpenZFS User\n Guide.
\nFor Windows file systems, the storage capacity target value must be at least 10 percent\n greater than the current storage capacity value. To increase storage capacity, the file system\n must have at least 16 MBps of throughput capacity. For more information, see Managing storage\n capacity in the Amazon FSx for Windows File Server User\n Guide.
\nFor ONTAP file systems, the storage capacity target value must be at least 10 percent\n greater than the current storage capacity value. For more information, see\n Managing storage capacity and provisioned IOPS in the Amazon FSx for NetApp ONTAP User\n Guide.
" + "smithy.api#documentation": "Use this parameter to increase the storage capacity of an FSx for Windows File Server,\n FSx for Lustre, FSx for OpenZFS, or FSx for ONTAP file system.\n Specifies the storage capacity target value, in GiB, to increase the storage capacity for\n the file system that you're updating.
\nYou can't make a storage capacity increase request if there is an existing storage\n capacity increase request in progress.
\nFor Lustre file systems, the storage capacity target value can be the following:
\nFor SCRATCH_2
, PERSISTENT_1
, and PERSISTENT_2 SSD
deployment types, valid values\n are in multiples of 2400 GiB. The value must be greater than the current storage capacity.
For PERSISTENT HDD
file systems, valid values are multiples of 6000 GiB for\n 12-MBps throughput per TiB file systems and multiples of 1800 GiB for 40-MBps throughput\n per TiB file systems. The values must be greater than the current storage capacity.
For SCRATCH_1
file systems, you can't increase the storage capacity.
For more information, see Managing storage and throughput\n capacity in the FSx for Lustre User Guide.
\nFor FSx for OpenZFS file systems, the storage capacity target value must be at least 10 percent\n greater than the current storage capacity value. For more information, see\n Managing storage capacity in the FSx for OpenZFS User\n Guide.
\nFor Windows file systems, the storage capacity target value must be at least 10 percent\n greater than the current storage capacity value. To increase storage capacity, the file system\n must have at least 16 MBps of throughput capacity. For more information, see Managing storage\n capacity in the Amazon FSxfor Windows File Server User\n Guide.
\nFor ONTAP file systems, the storage capacity target value must be at least 10 percent\n greater than the current storage capacity value. For more information, see\n Managing storage capacity and provisioned IOPS in the Amazon FSx for NetApp ONTAP User\n Guide.
" } }, "WindowsConfiguration": { @@ -10271,7 +10298,7 @@ "OpenZFSConfiguration": { "target": "com.amazonaws.fsx#UpdateFileSystemOpenZFSConfiguration", "traits": { - "smithy.api#documentation": "The configuration updates for an Amazon FSx for OpenZFS file system.
" + "smithy.api#documentation": "The configuration updates for an FSx for OpenZFS file system.
" } } }, @@ -10531,7 +10558,7 @@ } ], "traits": { - "smithy.api#documentation": "Updates an Amazon FSx for ONTAP storage virtual machine (SVM).
" + "smithy.api#documentation": "Updates an FSx for ONTAP storage virtual machine (SVM).
" } }, "com.amazonaws.fsx#UpdateStorageVirtualMachineRequest": { @@ -10540,7 +10567,7 @@ "ActiveDirectoryConfiguration": { "target": "com.amazonaws.fsx#UpdateSvmActiveDirectoryConfiguration", "traits": { - "smithy.api#documentation": "Updates the Microsoft Active Directory (AD) configuration for an SVM that is joined to an AD.
" + "smithy.api#documentation": "Specifies updates to an SVM's Microsoft Active Directory (AD) configuration.
" } }, "ClientRequestToken": { @@ -10559,7 +10586,7 @@ "SvmAdminPassword": { "target": "com.amazonaws.fsx#AdminPassword", "traits": { - "smithy.api#documentation": "Enter a new SvmAdminPassword if you are updating it.
" + "smithy.api#documentation": "Specifies a new SvmAdminPassword.
" } } }, @@ -10583,10 +10610,16 @@ "members": { "SelfManagedActiveDirectoryConfiguration": { "target": "com.amazonaws.fsx#SelfManagedActiveDirectoryConfigurationUpdates" + }, + "NetBiosName": { + "target": "com.amazonaws.fsx#NetBiosAlias", + "traits": { + "smithy.api#documentation": "Specifies an updated NetBIOS name of the AD computer object NetBiosName
to which an SVM is joined.
Updates the Microsoft Active Directory (AD) configuration of an SVM joined to an AD. \n Please note, account credentials are not returned in the response payload.
" + "smithy.api#documentation": "Specifies updates to an FSx for ONTAP storage virtual machine's (SVM) Microsoft Active Directory (AD) configuration. \n Note that account credentials are not returned in the response payload.
" } }, "com.amazonaws.fsx#UpdateVolume": { diff --git a/codegen/sdk-codegen/aws-models/opensearch.json b/codegen/sdk-codegen/aws-models/opensearch.json index 4929aa573e0..bcce7710be5 100644 --- a/codegen/sdk-codegen/aws-models/opensearch.json +++ b/codegen/sdk-codegen/aws-models/opensearch.json @@ -2766,7 +2766,13 @@ "Endpoint": { "target": "com.amazonaws.opensearch#Endpoint", "traits": { - "smithy.api#documentation": "The endpoint of the remote domain.
" + "smithy.api#documentation": "The Endpoint attribute cannot be modified.
\nThe endpoint of the remote domain. Applicable for VPC_ENDPOINT connection mode.
" + } + }, + "CrossClusterSearch": { + "target": "com.amazonaws.opensearch#CrossClusterSearchConnectionProperties", + "traits": { + "smithy.api#documentation": "The connection properties for cross cluster search.
" } } }, @@ -3009,6 +3015,12 @@ "traits": { "smithy.api#documentation": "The connection mode.
" } + }, + "ConnectionProperties": { + "target": "com.amazonaws.opensearch#ConnectionProperties", + "traits": { + "smithy.api#documentation": "The ConnectionProperties
for the outbound connection.
Status of SkipUnavailable param for outbound connection.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Cross cluster search specific connection properties.
" + } + }, "com.amazonaws.opensearch#DeleteDomain": { "type": "operation", "input": { @@ -4597,11 +4623,6 @@ "type": "list", "member": { "target": "com.amazonaws.opensearch#DescribePackagesFilterValue" - }, - "traits": { - "smithy.api#length": { - "min": 1 - } } }, "com.amazonaws.opensearch#DescribePackagesRequest": { @@ -9929,6 +9950,26 @@ "smithy.api#documentation": "The domain endpoint to which index and search requests are submitted. For example,\n search-imdb-movies-oopcnjfn6ugo.eu-west-1.es.amazonaws.com
or\n doc-imdb-movies-oopcnjfn6u.eu-west-1.es.amazonaws.com
.
Status of SkipUnavailable param for outbound connection.
\n\n ENABLED - The SkipUnavailable param is enabled\n for the connection.
\n\n DISABLED - The SkipUnavailable param is disabled\n for the connection.
\nAssociates one or more faces with an existing UserID. Takes an array of\n FaceIds
. Each FaceId
that are present in the FaceIds
\n list is associated with the provided UserID. The maximum number of total FaceIds
\n per UserID is 100.
The UserMatchThreshold
parameter specifies the minimum user match confidence\n required for the face to be associated with a UserID that has at least one FaceID
\n already associated. This ensures that the FaceIds
are associated with the right\n UserID. The value ranges from 0-100 and default value is 75.
If successful, an array of AssociatedFace
objects containing the associated\n FaceIds
is returned. If a given face is already associated with the given\n UserID
, it will be ignored and will not be returned in the response. If a given\n face is already associated to a different UserID
, isn't found in the collection,\n doesn’t meet the UserMatchThreshold
, or there are already 100 faces associated\n with the UserID
, it will be returned as part of an array of\n UnsuccessfulFaceAssociations.
\n
The UserStatus
reflects the status of an operation which updates a UserID\n representation with a list of given faces. The UserStatus
can be:
ACTIVE - All associations or disassociations of FaceID(s) for a UserID are\n complete.
\nCREATED - A UserID has been created, but has no FaceID(s) associated with it.
\nUPDATING - A UserID is being updated and there are current associations or\n disassociations of FaceID(s) taking place.
\nThe ID of an existing collection containing the UserID.
", + "smithy.api#required": {} + } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "The ID for the existing UserID.
", + "smithy.api#required": {} + } + }, + "FaceIds": { + "target": "com.amazonaws.rekognition#UserFaceIdList", + "traits": { + "smithy.api#documentation": "An array of FaceIDs to associate with the UserID.
", + "smithy.api#required": {} + } + }, + "UserMatchThreshold": { + "target": "com.amazonaws.rekognition#Percent", + "traits": { + "smithy.api#documentation": "An optional value specifying the minimum confidence in the UserID match to return. The\n default value is 75.
" + } + }, + "ClientRequestToken": { + "target": "com.amazonaws.rekognition#ClientRequestToken", + "traits": { + "smithy.api#documentation": "Idempotent token used to identify the request to AssociateFaces
. If you use\n the same token with multiple AssociateFaces
requests, the same response is returned.\n Use ClientRequestToken to prevent the same request from being processed more than\n once.
An array of AssociatedFace objects containing FaceIDs that are successfully associated\n with the UserID is returned. Returned if the AssociateFaces action is successful.
" + } + }, + "UnsuccessfulFaceAssociations": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceAssociationList", + "traits": { + "smithy.api#documentation": "An array of UnsuccessfulAssociation objects containing FaceIDs that are not successfully\n associated along with the reasons. Returned if the AssociateFaces action is successful.
" + } + }, + "UserStatus": { + "target": "com.amazonaws.rekognition#UserStatus", + "traits": { + "smithy.api#documentation": "The status of an update made to a UserID. Reflects if the UserID has been updated for\n every requested change.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.rekognition#AssociatedFace": { + "type": "structure", + "members": { + "FaceId": { + "target": "com.amazonaws.rekognition#FaceId", + "traits": { + "smithy.api#documentation": "Unique identifier assigned to the face.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Provides face metadata for the faces that are associated to a specific UserID.
" + } + }, + "com.amazonaws.rekognition#AssociatedFacesList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#AssociatedFace" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 100 + } + } + }, "com.amazonaws.rekognition#Attribute": { "type": "enum", "members": { @@ -589,7 +765,53 @@ } ], "traits": { - "smithy.api#documentation": "Compares a face in the source input image with each of the 100\n largest faces detected in the target input image.
\nIf the source image contains multiple faces, the service detects the largest face and\n compares it with each face detected in the target image.
\nCompareFaces uses machine learning algorithms, which are probabilistic. A false negative\n is an incorrect prediction that a face in the target image has a low similarity confidence\n score when compared to the face in the source image. To reduce the probability of false\n negatives, we recommend that you compare the target image against multiple source images. If\n you plan to use CompareFaces
to make a decision that impacts an individual's\n rights, privacy, or access to services, we recommend that you pass the result to a human for\n review and further validation before taking action.
You pass the input and target images either as base64-encoded image bytes or as\n references to images in an Amazon S3 bucket. If you use the\n AWS\n CLI to call Amazon Rekognition operations, passing image bytes isn't\n supported. The image must be formatted as a PNG or JPEG file.
\nIn response, the operation returns an array of face matches ordered by similarity score\n in descending order. For each face match, the response provides a bounding box of the face,\n facial landmarks, pose details (pitch, roll, and yaw), quality (brightness and sharpness), and\n confidence value (indicating the level of confidence that the bounding box contains a face).\n The response also provides a similarity score, which indicates how closely the faces match.
\nBy default, only faces with a similarity score of greater than or equal to 80% are\n returned in the response. You can change this value by specifying the\n SimilarityThreshold
parameter.
\n CompareFaces
also returns an array of faces that don't match the source\n image. For each face, it returns a bounding box, confidence value, landmarks, pose details,\n and quality. The response also returns information about the face in the source image,\n including the bounding box of the face and confidence value.
The QualityFilter
input parameter allows you to filter out detected faces\n that don’t meet a required quality bar. The quality bar is based on a variety of common use\n cases. Use QualityFilter
to set the quality bar by specifying LOW
,\n MEDIUM
, or HIGH
. If you do not want to filter detected faces,\n specify NONE
. The default value is NONE
.
If the image doesn't contain Exif metadata, CompareFaces
returns\n orientation information for the source and target images. Use these values to display the\n images with the correct image orientation.
If no faces are detected in the source or target images, CompareFaces
\n returns an InvalidParameterException
error.
This is a stateless API operation. That is, data returned by this operation doesn't\n persist.
\nFor an example, see Comparing Faces in Images in the Amazon Rekognition Developer\n Guide.
\nThis operation requires permissions to perform the\n rekognition:CompareFaces
action.
Compares a face in the source input image with each of the 100\n largest faces detected in the target input image.
\nIf the source image contains multiple faces, the service detects the largest face and\n compares it with each face detected in the target image.
\nCompareFaces uses machine learning algorithms, which are probabilistic. A false negative\n is an incorrect prediction that a face in the target image has a low similarity confidence\n score when compared to the face in the source image. To reduce the probability of false\n negatives, we recommend that you compare the target image against multiple source images. If\n you plan to use CompareFaces
to make a decision that impacts an individual's\n rights, privacy, or access to services, we recommend that you pass the result to a human for\n review and further validation before taking action.
You pass the input and target images either as base64-encoded image bytes or as\n references to images in an Amazon S3 bucket. If you use the\n AWS\n CLI to call Amazon Rekognition operations, passing image bytes isn't\n supported. The image must be formatted as a PNG or JPEG file.
\nIn response, the operation returns an array of face matches ordered by similarity score\n in descending order. For each face match, the response provides a bounding box of the face,\n facial landmarks, pose details (pitch, roll, and yaw), quality (brightness and sharpness), and\n confidence value (indicating the level of confidence that the bounding box contains a face).\n The response also provides a similarity score, which indicates how closely the faces match.
\nBy default, only faces with a similarity score of greater than or equal to 80% are\n returned in the response. You can change this value by specifying the\n SimilarityThreshold
parameter.
\n CompareFaces
also returns an array of faces that don't match the source\n image. For each face, it returns a bounding box, confidence value, landmarks, pose details,\n and quality. The response also returns information about the face in the source image,\n including the bounding box of the face and confidence value.
The QualityFilter
input parameter allows you to filter out detected faces\n that don’t meet a required quality bar. The quality bar is based on a variety of common use\n cases. Use QualityFilter
to set the quality bar by specifying LOW
,\n MEDIUM
, or HIGH
. If you do not want to filter detected faces,\n specify NONE
. The default value is NONE
.
If the image doesn't contain Exif metadata, CompareFaces
returns\n orientation information for the source and target images. Use these values to display the\n images with the correct image orientation.
If no faces are detected in the source or target images, CompareFaces
\n returns an InvalidParameterException
error.
This is a stateless API operation. That is, data returned by this operation doesn't\n persist.
\nFor an example, see Comparing Faces in Images in the Amazon Rekognition Developer\n Guide.
\nThis operation requires permissions to perform the\n rekognition:CompareFaces
action.
Type that describes the face Amazon Rekognition chose to compare with the faces in the target.\n This contains a bounding box for the selected face and confidence level that the bounding box\n contains a face. Note that Amazon Rekognition selects the largest face in the source image for this\n comparison.
" } }, + "com.amazonaws.rekognition#ConflictException": { + "type": "structure", + "members": { + "Message": { + "target": "com.amazonaws.rekognition#String" + }, + "Code": { + "target": "com.amazonaws.rekognition#String" + }, + "Logref": { + "target": "com.amazonaws.rekognition#String", + "traits": { + "smithy.api#documentation": "A universally unique identifier (UUID) for the request.
" + } + } + }, + "traits": { + "smithy.api#documentation": "\n A User with the same Id already exists within the collection, or the update or deletion of the User caused an inconsistent state. ** \n
", + "smithy.api#error": "client" + } + }, "com.amazonaws.rekognition#ConnectedHomeLabel": { "type": "string" }, @@ -974,7 +1217,30 @@ } ], "traits": { - "smithy.api#documentation": "Copies a version of an Amazon Rekognition Custom Labels model from a source project to a destination project. The source and\n destination projects can be in different AWS accounts but must be in the same AWS Region.\n You can't copy a model to another AWS service.\n \n
\nTo copy a model version to a different AWS account, you need to create a resource-based policy known as a\n project policy. You attach the project policy to the\n source project by calling PutProjectPolicy. The project policy\n gives permission to copy the model version from a trusting AWS account to a trusted account.
\nFor more information creating and attaching a project policy, see Attaching a project policy (SDK)\n in the Amazon Rekognition Custom Labels Developer Guide.\n
\nIf you are copying a model version to a project in the same AWS account, you don't need to create a project policy.
\nTo copy a model, the destination project, source project, and source model version must already exist.
\nCopying a model version takes a while to complete. To get the current status, call DescribeProjectVersions and check the value of Status
in the\n ProjectVersionDescription object. The copy operation has finished when\n the value of Status
is COPYING_COMPLETED
.
This operation requires permissions to perform the rekognition:CopyProjectVersion
action.
Copies a version of an Amazon Rekognition Custom Labels model from a source project to a destination project. The source and\n destination projects can be in different AWS accounts but must be in the same AWS Region.\n You can't copy a model to another AWS service.\n \n
\nTo copy a model version to a different AWS account, you need to create a resource-based policy known as a\n project policy. You attach the project policy to the\n source project by calling PutProjectPolicy. The project policy\n gives permission to copy the model version from a trusting AWS account to a trusted account.
\nFor more information creating and attaching a project policy, see Attaching a project policy (SDK)\n in the Amazon Rekognition Custom Labels Developer Guide.\n
\nIf you are copying a model version to a project in the same AWS account, you don't need to create a project policy.
\nTo copy a model, the destination project, source project, and source model version must already exist.
\nCopying a model version takes a while to complete. To get the current status, call DescribeProjectVersions and check the value of Status
in the\n ProjectVersionDescription object. The copy operation has finished when\n the value of Status
is COPYING_COMPLETED
.
This operation requires permissions to perform the rekognition:CopyProjectVersion
action.
Creates a collection in an AWS Region. You can add faces to the collection using the\n IndexFaces operation.
\nFor example, you might create collections, one for each of your application users. A\n user can then index faces using the IndexFaces
operation and persist results in a\n specific collection. Then, a user can search the collection for faces in the user-specific\n container.
When you create a collection, it is associated with the latest version of the face model\n version.
\nCollection names are case-sensitive.
\nThis operation requires permissions to perform the\n rekognition:CreateCollection
action. If you want to tag your collection, you\n also require permission to perform the rekognition:TagResource
\n operation.
Creates a collection in an AWS Region. You can add faces to the collection using the\n IndexFaces operation.
\nFor example, you might create collections, one for each of your application users. A\n user can then index faces using the IndexFaces
operation and persist results in a\n specific collection. Then, a user can search the collection for faces in the user-specific\n container.
When you create a collection, it is associated with the latest version of the face model\n version.
\nCollection names are case-sensitive.
\nThis operation requires permissions to perform the\n rekognition:CreateCollection
action. If you want to tag your collection, you\n also require permission to perform the rekognition:TagResource
\n operation.
Creates a new User within a collection specified by CollectionId
. Takes\n UserId
as a parameter, which is a user provided ID which should be unique\n within the collection. The provided UserId
will alias the system generated\n UUID to make the UserId
more user friendly.
Uses a ClientToken
, an idempotency token that ensures a call to\n CreateUser
completes only once. If the value is not supplied, the AWS SDK\n generates an idempotency token for the requests. This prevents retries after a network\n error results from making multiple CreateUser
calls.
The ID of an existing collection to which the new UserID needs to be created.
", + "smithy.api#required": {} + } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "ID for the UserID to be created. This ID needs to be unique within the collection.
", + "smithy.api#required": {} + } + }, + "ClientRequestToken": { + "target": "com.amazonaws.rekognition#ClientRequestToken", + "traits": { + "smithy.api#documentation": "Idempotent token used to identify the request to CreateUser
. If you use the\n same token with multiple CreateUser
requests, the same response is returned. \n Use ClientRequestToken to prevent the same request from being processed more than\n once.
Deletes the specified collection. Note that this operation removes all faces in the\n collection. For an example, see Deleting a\n collection.
\nThis operation requires permissions to perform the\n rekognition:DeleteCollection
action.
Deletes the specified collection. Note that this operation removes all faces in the\n collection. For an example, see Deleting a\n collection.
\nThis operation requires permissions to perform the\n rekognition:DeleteCollection
action.
Deletes faces from a collection. You specify a collection ID and an array of face IDs\n to remove from the collection.
\nThis operation requires permissions to perform the rekognition:DeleteFaces
\n action.
Deletes faces from a collection. You specify a collection ID and an array of face IDs\n to remove from the collection.
\nThis operation requires permissions to perform the rekognition:DeleteFaces
\n action.
An array of strings (face IDs) of the faces that were deleted.
" } + }, + "UnsuccessfulFaceDeletions": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDeletionsList", + "traits": { + "smithy.api#documentation": "An array of any faces that weren't deleted.
" + } } }, "traits": { @@ -2256,7 +2657,19 @@ } ], "traits": { - "smithy.api#documentation": "Deletes an existing project policy.
\nTo get a list of project policies attached to a project, call ListProjectPolicies. To attach a project policy to a project, call PutProjectPolicy.
\nThis operation requires permissions to perform the rekognition:DeleteProjectPolicy
action.
Deletes an existing project policy.
\nTo get a list of project policies attached to a project, call ListProjectPolicies. To attach a project policy to a project, call PutProjectPolicy.
\nThis operation requires permissions to perform the rekognition:DeleteProjectPolicy
action.
Describes the specified collection. You can use DescribeCollection
to get\n information, such as the number of faces indexed into a collection and the version of the\n model used by the collection for face detection.
For more information, see Describing a Collection in the \n Amazon Rekognition Developer Guide.
" + "smithy.api#documentation": "Deletes the specified UserID within the collection. Faces that are associated with the\n UserID are disassociated from the UserID before deleting the specified UserID. If the\n specified Collection
or UserID
is already deleted or not found, a\n ResourceNotFoundException
will be thrown. If the action is successful with a\n 200 response, an empty HTTP body is returned.
The ID of the collection to describe.
", + "smithy.api#documentation": "The ID of an existing collection from which the UserID needs to be deleted.
", "smithy.api#required": {} } - } - }, - "traits": { - "smithy.api#input": {} - } - }, + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "ID for the UserID to be deleted.
", + "smithy.api#required": {} + } + }, + "ClientRequestToken": { + "target": "com.amazonaws.rekognition#ClientRequestToken", + "traits": { + "smithy.api#documentation": "Idempotent token used to identify the request to DeleteUser
. If you use the\n same token with multiple DeleteUser
requests, the same response is returned. \n Use ClientRequestToken to prevent the same request from being processed more than\n once.
Describes the specified collection. You can use DescribeCollection
to get\n information, such as the number of faces indexed into a collection and the version of the\n model used by the collection for face detection.
For more information, see Describing a Collection in the \n Amazon Rekognition Developer Guide.
" + } + }, + "com.amazonaws.rekognition#DescribeCollectionRequest": { + "type": "structure", + "members": { + "CollectionId": { + "target": "com.amazonaws.rekognition#CollectionId", + "traits": { + "smithy.api#documentation": "The ID of the collection to describe.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, "com.amazonaws.rekognition#DescribeCollectionResponse": { "type": "structure", "members": { @@ -2517,6 +3015,12 @@ "traits": { "smithy.api#documentation": "The number of milliseconds since the Unix epoch time until the creation of the collection.\n The Unix epoch time is 00:00:00 Coordinated Universal Time (UTC), Thursday, 1 January 1970.
" } + }, + "UserCount": { + "target": "com.amazonaws.rekognition#ULong", + "traits": { + "smithy.api#documentation": "The number of UserIDs assigned to the specified colleciton.
" + } } }, "traits": { @@ -3081,7 +3585,71 @@ } ], "traits": { - "smithy.api#documentation": "Detects faces within an image that is provided as input.
\n\n DetectFaces
detects the 100 largest faces in the image. For each face\n detected, the operation returns face details. These details include a bounding box of the\n face, a confidence value (that the bounding box contains a face), and a fixed set of\n attributes such as facial landmarks (for example, coordinates of eye and mouth), pose,\n presence of facial occlusion, and so on.
The face-detection algorithm is most effective on frontal faces. For non-frontal or\n obscured faces, the algorithm might not detect the faces or might detect faces with lower\n confidence.
\nYou pass the input image either as base64-encoded image bytes or as a reference to an\n image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations,\n passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.
\nThis is a stateless API operation. That is, the operation does not persist any\n data.
\nThis operation requires permissions to perform the rekognition:DetectFaces
\n action.
Detects faces within an image that is provided as input.
\n\n DetectFaces
detects the 100 largest faces in the image. For each face\n detected, the operation returns face details. These details include a bounding box of the\n face, a confidence value (that the bounding box contains a face), and a fixed set of\n attributes such as facial landmarks (for example, coordinates of eye and mouth), pose,\n presence of facial occlusion, and so on.
The face-detection algorithm is most effective on frontal faces. For non-frontal or\n obscured faces, the algorithm might not detect the faces or might detect faces with lower\n confidence.
\nYou pass the input image either as base64-encoded image bytes or as a reference to an\n image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations,\n passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.
\nThis is a stateless API operation. That is, the operation does not persist any\n data.
\nThis operation requires permissions to perform the rekognition:DetectFaces
\n action.
Detects instances of real-world entities within an image (JPEG or PNG) provided as\n input. This includes objects like flower, tree, and table; events like wedding, graduation,\n and birthday party; and concepts like landscape, evening, and nature.
\nFor an example, see Analyzing images stored in an Amazon S3 bucket in the\n Amazon Rekognition Developer Guide.
\nYou pass the input image as base64-encoded image bytes or as a reference to an image in\n an Amazon S3 bucket. If you use the\n AWS\n CLI to call Amazon Rekognition operations, passing image bytes is not\n supported. The image must be either a PNG or JPEG formatted file.
\n\n Optional Parameters\n
\nYou can specify one or both of the GENERAL_LABELS
and\n IMAGE_PROPERTIES
feature types when calling the DetectLabels API. Including\n GENERAL_LABELS
will ensure the response includes the labels detected in the\n input image, while including IMAGE_PROPERTIES
will ensure the response includes\n information about the image quality and color.
When using GENERAL_LABELS
and/or IMAGE_PROPERTIES
you can\n provide filtering criteria to the Settings parameter. You can filter with sets of individual\n labels or with label categories. You can specify inclusive filters, exclusive filters, or a\n combination of inclusive and exclusive filters. For more information on filtering see Detecting\n Labels in an Image.
You can specify MinConfidence
to control the confidence threshold for the\n labels returned. The default is 55%. You can also add the MaxLabels
parameter to\n limit the number of labels returned. The default and upper limit is 1000 labels.
\n Response Elements\n
\nFor each object, scene, and concept the API returns one or more labels. The API\n returns the following types of information about labels:
\nName - The name of the detected label.
\nConfidence - The level of confidence in the label assigned to a detected object.\n
\nParents - The ancestor labels for a detected label. DetectLabels returns a\n hierarchical taxonomy of detected labels. For example, a detected car might be assigned\n the label car. The label car has two parent labels: Vehicle (its parent) and\n Transportation (its grandparent). The response includes the all ancestors for a label,\n where every ancestor is a unique label. In the previous example, Car, Vehicle, and\n Transportation are returned as unique labels in the response.
\nAliases - Possible Aliases for the label.
\nCategories - The label categories that the detected label belongs to.
\nBoundingBox — Bounding boxes are described for all instances of detected common\n object labels, returned in an array of Instance objects. An Instance object contains a\n BoundingBox object, describing the location of the label on the input image. It also\n includes the confidence for the accuracy of the detected bounding box.
\nThe API returns the following information regarding the image, as part of the\n ImageProperties structure:
\nQuality - Information about the Sharpness, Brightness, and Contrast of the input\n image, scored between 0 to 100. Image quality is returned for the entire image, as well as\n the background and the foreground.
\nDominant Color - An array of the dominant colors in the image.
\nForeground - Information about the sharpness, brightness, and dominant colors of the input image’s foreground.
\nBackground - Information about the sharpness, brightness, and dominant colors of the input image’s background.
\nThe list of returned labels will include at least one label for every detected object,\n along with information about that label. In the following example, suppose the input image has\n a lighthouse, the sea, and a rock. The response includes all three labels, one for each\n object, as well as the confidence in the label:
\n\n {Name: lighthouse, Confidence: 98.4629}
\n
\n {Name: rock,Confidence: 79.2097}
\n
\n {Name: sea,Confidence: 75.061}
\n
The list of labels can include multiple labels for the same object. For example, if the\n input image shows a flower (for example, a tulip), the operation might return the following\n three labels.
\n\n {Name: flower,Confidence: 99.0562}
\n
\n {Name: plant,Confidence: 99.0562}
\n
\n {Name: tulip,Confidence: 99.0562}
\n
In this example, the detection algorithm more precisely identifies the flower as a\n tulip.
\nIf the object detected is a person, the operation doesn't provide the same facial\n details that the DetectFaces operation provides.
\nThis is a stateless API operation that doesn't return any data.
\nThis operation requires permissions to perform the\n rekognition:DetectLabels
action.
Detects instances of real-world entities within an image (JPEG or PNG) provided as\n input. This includes objects like flower, tree, and table; events like wedding, graduation,\n and birthday party; and concepts like landscape, evening, and nature.
\nFor an example, see Analyzing images stored in an Amazon S3 bucket in the\n Amazon Rekognition Developer Guide.
\nYou pass the input image as base64-encoded image bytes or as a reference to an image in\n an Amazon S3 bucket. If you use the\n AWS\n CLI to call Amazon Rekognition operations, passing image bytes is not\n supported. The image must be either a PNG or JPEG formatted file.
\n\n Optional Parameters\n
\nYou can specify one or both of the GENERAL_LABELS
and\n IMAGE_PROPERTIES
feature types when calling the DetectLabels API. Including\n GENERAL_LABELS
will ensure the response includes the labels detected in the\n input image, while including IMAGE_PROPERTIES
will ensure the response includes\n information about the image quality and color.
When using GENERAL_LABELS
and/or IMAGE_PROPERTIES
you can\n provide filtering criteria to the Settings parameter. You can filter with sets of individual\n labels or with label categories. You can specify inclusive filters, exclusive filters, or a\n combination of inclusive and exclusive filters. For more information on filtering see Detecting\n Labels in an Image.
You can specify MinConfidence
to control the confidence threshold for the\n labels returned. The default is 55%. You can also add the MaxLabels
parameter to\n limit the number of labels returned. The default and upper limit is 1000 labels.
\n Response Elements\n
\nFor each object, scene, and concept the API returns one or more labels. The API\n returns the following types of information about labels:
\nName - The name of the detected label.
\nConfidence - The level of confidence in the label assigned to a detected object.\n
\nParents - The ancestor labels for a detected label. DetectLabels returns a\n hierarchical taxonomy of detected labels. For example, a detected car might be assigned\n the label car. The label car has two parent labels: Vehicle (its parent) and\n Transportation (its grandparent). The response includes the all ancestors for a label,\n where every ancestor is a unique label. In the previous example, Car, Vehicle, and\n Transportation are returned as unique labels in the response.
\nAliases - Possible Aliases for the label.
\nCategories - The label categories that the detected label belongs to.
\nBoundingBox — Bounding boxes are described for all instances of detected common\n object labels, returned in an array of Instance objects. An Instance object contains a\n BoundingBox object, describing the location of the label on the input image. It also\n includes the confidence for the accuracy of the detected bounding box.
\nThe API returns the following information regarding the image, as part of the\n ImageProperties structure:
\nQuality - Information about the Sharpness, Brightness, and Contrast of the input\n image, scored between 0 to 100. Image quality is returned for the entire image, as well as\n the background and the foreground.
\nDominant Color - An array of the dominant colors in the image.
\nForeground - Information about the sharpness, brightness, and dominant colors of the input image’s foreground.
\nBackground - Information about the sharpness, brightness, and dominant colors of the input image’s background.
\nThe list of returned labels will include at least one label for every detected object,\n along with information about that label. In the following example, suppose the input image has\n a lighthouse, the sea, and a rock. The response includes all three labels, one for each\n object, as well as the confidence in the label:
\n\n {Name: lighthouse, Confidence: 98.4629}
\n
\n {Name: rock,Confidence: 79.2097}
\n
\n {Name: sea,Confidence: 75.061}
\n
The list of labels can include multiple labels for the same object. For example, if the\n input image shows a flower (for example, a tulip), the operation might return the following\n three labels.
\n\n {Name: flower,Confidence: 99.0562}
\n
\n {Name: plant,Confidence: 99.0562}
\n
\n {Name: tulip,Confidence: 99.0562}
\n
In this example, the detection algorithm more precisely identifies the flower as a\n tulip.
\nIf the object detected is a person, the operation doesn't provide the same facial\n details that the DetectFaces operation provides.
\nThis is a stateless API operation that doesn't return any data.
\nThis operation requires permissions to perform the\n rekognition:DetectLabels
action.
A set of parameters that allow you to filter out certain results from your returned results.
" } }, + "com.amazonaws.rekognition#DisassociateFaces": { + "type": "operation", + "input": { + "target": "com.amazonaws.rekognition#DisassociateFacesRequest" + }, + "output": { + "target": "com.amazonaws.rekognition#DisassociateFacesResponse" + }, + "errors": [ + { + "target": "com.amazonaws.rekognition#AccessDeniedException" + }, + { + "target": "com.amazonaws.rekognition#ConflictException" + }, + { + "target": "com.amazonaws.rekognition#IdempotentParameterMismatchException" + }, + { + "target": "com.amazonaws.rekognition#InternalServerError" + }, + { + "target": "com.amazonaws.rekognition#InvalidParameterException" + }, + { + "target": "com.amazonaws.rekognition#ProvisionedThroughputExceededException" + }, + { + "target": "com.amazonaws.rekognition#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.rekognition#ThrottlingException" + } + ], + "traits": { + "smithy.api#documentation": "Removes the association between a Face
supplied in an array of\n FaceIds
and the User. If the User is not present already, then a\n ResourceNotFound
exception is thrown. If successful, an array of faces that are\n disassociated from the User is returned. If a given face is already disassociated from the\n given UserID, it will be ignored and not be returned in the response. If a given face is\n already associated with a different User or not found in the collection it will be returned as\n part of UnsuccessfulDisassociations
. You can remove 1 - 100 face IDs from a user\n at one time.
The ID of an existing collection containing the UserID.
", + "smithy.api#required": {} + } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "ID for the existing UserID.
", + "smithy.api#required": {} + } + }, + "ClientRequestToken": { + "target": "com.amazonaws.rekognition#ClientRequestToken", + "traits": { + "smithy.api#documentation": "Idempotent token used to identify the request to DisassociateFaces
. If you\n use the same token with multiple DisassociateFaces
requests, the same response is\n returned. Use ClientRequestToken to prevent the same request from being processed more than\n once.
An array of face IDs to disassociate from the UserID.
", + "smithy.api#required": {} + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.rekognition#DisassociateFacesResponse": { + "type": "structure", + "members": { + "DisassociatedFaces": { + "target": "com.amazonaws.rekognition#DisassociatedFacesList", + "traits": { + "smithy.api#documentation": "An array of DissociatedFace objects containing FaceIds that are successfully disassociated\n with the UserID is returned. Returned if the DisassociatedFaces action is successful.
" + } + }, + "UnsuccessfulFaceDisassociations": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDisassociationList", + "traits": { + "smithy.api#documentation": "An array of UnsuccessfulDisassociation objects containing FaceIds that are not\n successfully associated, along with the reasons for the failure to associate. Returned if the\n DisassociateFaces action is successful.
" + } + }, + "UserStatus": { + "target": "com.amazonaws.rekognition#UserStatus", + "traits": { + "smithy.api#documentation": "The status of an update made to a User. Reflects if the User has been updated for every requested change.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.rekognition#DisassociatedFace": { + "type": "structure", + "members": { + "FaceId": { + "target": "com.amazonaws.rekognition#FaceId", + "traits": { + "smithy.api#documentation": "Unique identifier assigned to the face.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Provides face metadata for the faces that are disassociated from a specific UserID.
" + } + }, + "com.amazonaws.rekognition#DisassociatedFacesList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#DisassociatedFace" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 100 + } + } + }, "com.amazonaws.rekognition#DistributeDataset": { "type": "structure", "members": { @@ -4115,6 +4869,12 @@ "traits": { "smithy.api#documentation": "The version of the face detect and storage model that was used when indexing the face\n vector.
" } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "Unique identifier assigned to the user.
" + } } }, "traits": { @@ -6087,7 +6847,145 @@ } ], "traits": { - "smithy.api#documentation": "Detects faces in the input image and adds them to the specified collection.
\nAmazon Rekognition doesn't save the actual faces that are detected. Instead, the underlying\n detection algorithm first detects the faces in the input image. For each face, the algorithm\n extracts facial features into a feature vector, and stores it in the backend database.\n Amazon Rekognition uses feature vectors when it performs face match and search operations using the\n SearchFaces and SearchFacesByImage operations.
\nFor more information, see Adding faces to a collection in the Amazon Rekognition\n Developer Guide.
\nTo get the number of faces in a collection, call DescribeCollection.
\nIf you're using version 1.0 of the face detection model, IndexFaces
\n indexes the 15 largest faces in the input image. Later versions of the face detection model\n index the 100 largest faces in the input image.
If you're using version 4 or later of the face model, image orientation information is not\n returned in the OrientationCorrection
field.
To determine which version of the model you're using, call DescribeCollection and supply the collection ID. You can also get the model\n version from the value of FaceModelVersion
in the response from\n IndexFaces
\n
For more information, see Model Versioning in the Amazon Rekognition Developer\n Guide.
\nIf you provide the optional ExternalImageId
for the input image you\n provided, Amazon Rekognition associates this ID with all faces that it detects. When you call the ListFaces operation, the response returns the external ID. You can use this\n external image ID to create a client-side index to associate the faces with each image. You\n can then use the index to find all faces in an image.
You can specify the maximum number of faces to index with the MaxFaces
input\n parameter. This is useful when you want to index the largest faces in an image and don't want\n to index smaller faces, such as those belonging to people standing in the background.
The QualityFilter
input parameter allows you to filter out detected faces\n that don’t meet a required quality bar. The quality bar is based on a variety of common use\n cases. By default, IndexFaces
chooses the quality bar that's used to filter\n faces. You can also explicitly choose the quality bar. Use QualityFilter
, to set\n the quality bar by specifying LOW
, MEDIUM
, or HIGH
. If\n you do not want to filter detected faces, specify NONE
.
To use quality filtering, you need a collection associated with version 3 of the face\n model or higher. To get the version of the face model associated with a collection, call\n DescribeCollection.
\nInformation about faces detected in an image, but not indexed, is returned in an array of\n UnindexedFace objects, UnindexedFaces
. Faces aren't indexed\n for reasons such as:
The number of faces detected exceeds the value of the MaxFaces
request\n parameter.
The face is too small compared to the image dimensions.
\nThe face is too blurry.
\nThe image is too dark.
\nThe face has an extreme pose.
\nThe face doesn’t have enough detail to be suitable for face search.
\nIn response, the IndexFaces
operation returns an array of metadata for all\n detected faces, FaceRecords
. This includes:
The bounding box, BoundingBox
, of the detected face.
A confidence value, Confidence
, which indicates the confidence that the\n bounding box contains a face.
A face ID, FaceId
, assigned by the service for each face that's detected\n and stored.
An image ID, ImageId
, assigned by the service for the input image.
If you request ALL
or specific facial attributes (e.g.,\n FACE_OCCLUDED
) by using the detectionAttributes parameter, Amazon Rekognition\n returns detailed facial attributes, such as facial landmarks (for example, location of eye and\n mouth), facial occlusion, and other facial attributes.
If you provide the same image, specify the same collection, and use the same external ID\n in the IndexFaces
operation, Amazon Rekognition doesn't save duplicate face\n metadata.
The input image is passed either as base64-encoded image bytes, or as a reference to an\n image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations,\n passing image bytes isn't supported. The image must be formatted as a PNG or JPEG file.
\nThis operation requires permissions to perform the rekognition:IndexFaces
\n action.
Detects faces in the input image and adds them to the specified collection.
\nAmazon Rekognition doesn't save the actual faces that are detected. Instead, the underlying\n detection algorithm first detects the faces in the input image. For each face, the algorithm\n extracts facial features into a feature vector, and stores it in the backend database.\n Amazon Rekognition uses feature vectors when it performs face match and search operations using the\n SearchFaces and SearchFacesByImage operations.
\nFor more information, see Adding faces to a collection in the Amazon Rekognition\n Developer Guide.
\nTo get the number of faces in a collection, call DescribeCollection.
\nIf you're using version 1.0 of the face detection model, IndexFaces
\n indexes the 15 largest faces in the input image. Later versions of the face detection model\n index the 100 largest faces in the input image.
If you're using version 4 or later of the face model, image orientation information is not\n returned in the OrientationCorrection
field.
To determine which version of the model you're using, call DescribeCollection and supply the collection ID. You can also get the model\n version from the value of FaceModelVersion
in the response from\n IndexFaces
\n
For more information, see Model Versioning in the Amazon Rekognition Developer\n Guide.
\nIf you provide the optional ExternalImageId
for the input image you\n provided, Amazon Rekognition associates this ID with all faces that it detects. When you call the ListFaces operation, the response returns the external ID. You can use this\n external image ID to create a client-side index to associate the faces with each image. You\n can then use the index to find all faces in an image.
You can specify the maximum number of faces to index with the MaxFaces
input\n parameter. This is useful when you want to index the largest faces in an image and don't want\n to index smaller faces, such as those belonging to people standing in the background.
The QualityFilter
input parameter allows you to filter out detected faces\n that don’t meet a required quality bar. The quality bar is based on a variety of common use\n cases. By default, IndexFaces
chooses the quality bar that's used to filter\n faces. You can also explicitly choose the quality bar. Use QualityFilter
, to set\n the quality bar by specifying LOW
, MEDIUM
, or HIGH
. If\n you do not want to filter detected faces, specify NONE
.
To use quality filtering, you need a collection associated with version 3 of the face\n model or higher. To get the version of the face model associated with a collection, call\n DescribeCollection.
\nInformation about faces detected in an image, but not indexed, is returned in an array of\n UnindexedFace objects, UnindexedFaces
. Faces aren't indexed\n for reasons such as:
The number of faces detected exceeds the value of the MaxFaces
request\n parameter.
The face is too small compared to the image dimensions.
\nThe face is too blurry.
\nThe image is too dark.
\nThe face has an extreme pose.
\nThe face doesn’t have enough detail to be suitable for face search.
\nIn response, the IndexFaces
operation returns an array of metadata for all\n detected faces, FaceRecords
. This includes:
The bounding box, BoundingBox
, of the detected face.
A confidence value, Confidence
, which indicates the confidence that the\n bounding box contains a face.
A face ID, FaceId
, assigned by the service for each face that's detected\n and stored.
An image ID, ImageId
, assigned by the service for the input image.
If you request ALL
or specific facial attributes (e.g.,\n FACE_OCCLUDED
) by using the detectionAttributes parameter, Amazon Rekognition\n returns detailed facial attributes, such as facial landmarks (for example, location of eye and\n mouth), facial occlusion, and other facial attributes.
If you provide the same image, specify the same collection, and use the same external ID\n in the IndexFaces
operation, Amazon Rekognition doesn't save duplicate face\n metadata.
The input image is passed either as base64-encoded image bytes, or as a reference to an\n image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations,\n passing image bytes isn't supported. The image must be formatted as a PNG or JPEG file.
\nThis operation requires permissions to perform the rekognition:IndexFaces
\n action.
Returns list of collection IDs in your account. If the result is truncated, the\n response also provides a NextToken
that you can use in the subsequent request to\n fetch the next set of collection IDs.
For an example, see Listing collections in the Amazon Rekognition Developer\n Guide.
\nThis operation requires permissions to perform the\n rekognition:ListCollections
action.
Returns metadata for faces in the specified collection. This metadata\n includes information such as the bounding box coordinates, the confidence (that the bounding\n box contains a face), and face ID. For an example, see Listing Faces in a Collection in the\n Amazon Rekognition Developer Guide.
\nThis operation requires permissions to perform the rekognition:ListFaces
\n action.
Maximum number of faces to return.
" } - } - }, - "traits": { + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "An array of user IDs to match when listing faces in a collection.
" + } + }, + "FaceIds": { + "target": "com.amazonaws.rekognition#FaceIdList", + "traits": { + "smithy.api#documentation": "An array of face IDs to match when listing faces in a collection.
" + } + } + }, + "traits": { "smithy.api#input": {} } }, @@ -7576,6 +8549,116 @@ "smithy.api#output": {} } }, + "com.amazonaws.rekognition#ListUsers": { + "type": "operation", + "input": { + "target": "com.amazonaws.rekognition#ListUsersRequest" + }, + "output": { + "target": "com.amazonaws.rekognition#ListUsersResponse" + }, + "errors": [ + { + "target": "com.amazonaws.rekognition#AccessDeniedException" + }, + { + "target": "com.amazonaws.rekognition#InternalServerError" + }, + { + "target": "com.amazonaws.rekognition#InvalidPaginationTokenException" + }, + { + "target": "com.amazonaws.rekognition#InvalidParameterException" + }, + { + "target": "com.amazonaws.rekognition#ProvisionedThroughputExceededException" + }, + { + "target": "com.amazonaws.rekognition#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.rekognition#ThrottlingException" + } + ], + "traits": { + "smithy.api#documentation": "Returns metadata of the User such as UserID
in the specified collection.\n Anonymous User (to reserve faces without any identity) is not returned as part of this\n request. The results are sorted by system generated primary key ID. If the response is\n truncated, NextToken
is returned in the response that can be used in the\n subsequent request to retrieve the next set of identities.
The ID of an existing collection.
", + "smithy.api#required": {} + } + }, + "MaxResults": { + "target": "com.amazonaws.rekognition#MaxUserResults", + "traits": { + "smithy.api#documentation": "Maximum number of UsersID to return.
" + } + }, + "NextToken": { + "target": "com.amazonaws.rekognition#PaginationToken", + "traits": { + "smithy.api#documentation": "Pagingation token to receive the next set of UsersID.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.rekognition#ListUsersResponse": { + "type": "structure", + "members": { + "Users": { + "target": "com.amazonaws.rekognition#UserList", + "traits": { + "smithy.api#documentation": "List of UsersID associated with the specified collection.
" + } + }, + "NextToken": { + "target": "com.amazonaws.rekognition#PaginationToken", + "traits": { + "smithy.api#documentation": "A pagination token to be used with the subsequent request if the response is truncated.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, "com.amazonaws.rekognition#LivenessImageBlob": { "type": "blob", "traits": { @@ -7683,6 +8766,26 @@ "smithy.api#error": "client" } }, + "com.amazonaws.rekognition#MatchedUser": { + "type": "structure", + "members": { + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "A provided ID for the UserID. Unique within the collection.
" + } + }, + "UserStatus": { + "target": "com.amazonaws.rekognition#UserStatus", + "traits": { + "smithy.api#documentation": "The status of the user matched to a provided FaceID.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Contains metadata for a UserID matched with a given face.
" + } + }, "com.amazonaws.rekognition#MaxDurationInSecondsULong": { "type": "long", "traits": { @@ -7726,6 +8829,15 @@ } } }, + "com.amazonaws.rekognition#MaxUserResults": { + "type": "integer", + "traits": { + "smithy.api#range": { + "min": 1, + "max": 500 + } + } + }, "com.amazonaws.rekognition#MinCoveragePercentage": { "type": "float", "traits": { @@ -8683,7 +9795,22 @@ } ], "traits": { - "smithy.api#documentation": "Attaches a project policy to a Amazon Rekognition Custom Labels project in a trusting AWS account. A\n project policy specifies that a trusted AWS account can copy a model version from a\n trusting AWS account to a project in the trusted AWS account. To copy a model version you use\n the CopyProjectVersion operation.
\nFor more information about the format of a project policy document, see Attaching a project policy (SDK)\n in the Amazon Rekognition Custom Labels Developer Guide.\n
\nThe response from PutProjectPolicy
is a revision ID for the project policy.\n You can attach multiple project policies to a project. You can also update an existing\n project policy by specifying the policy revision ID of the existing policy.
To remove a project policy from a project, call DeleteProjectPolicy.\n To get a list of project policies attached to a project, call ListProjectPolicies.
\nYou copy a model version by calling CopyProjectVersion.
\nThis operation requires permissions to perform the rekognition:PutProjectPolicy
action.
Attaches a project policy to a Amazon Rekognition Custom Labels project in a trusting AWS account. A\n project policy specifies that a trusted AWS account can copy a model version from a\n trusting AWS account to a project in the trusted AWS account. To copy a model version you use\n the CopyProjectVersion operation.
\nFor more information about the format of a project policy document, see Attaching a project policy (SDK)\n in the Amazon Rekognition Custom Labels Developer Guide.\n
\nThe response from PutProjectPolicy
is a revision ID for the project policy.\n You can attach multiple project policies to a project. You can also update an existing\n project policy by specifying the policy revision ID of the existing policy.
To remove a project policy from a project, call DeleteProjectPolicy.\n To get a list of project policies attached to a project, call ListProjectPolicies.
\nYou copy a model version by calling CopyProjectVersion.
\nThis operation requires permissions to perform the rekognition:PutProjectPolicy
action.
This is the API Reference for Amazon Rekognition Image, Amazon Rekognition Custom Labels,\n Amazon Rekognition Stored\n Video, Amazon Rekognition Streaming Video. It provides descriptions of actions, data types, common\n parameters, and common errors.
\n\n Amazon Rekognition Image\n
\n\n CompareFaces\n
\n\n CreateCollection\n
\n\n DeleteCollection\n
\n\n DeleteFaces\n
\n\n DescribeCollection\n
\n\n DetectFaces\n
\n\n DetectLabels\n
\n\n DetectText\n
\n\n GetCelebrityInfo\n
\n\n IndexFaces\n
\n\n ListCollections\n
\n\n ListFaces\n
\n\n RecognizeCelebrities\n
\n\n SearchFaces\n
\n\n SearchFacesByImage\n
\n\n Amazon Rekognition Custom Labels\n
\n\n CopyProjectVersion\n
\n\n CreateDataset\n
\n\n CreateProject\n
\n\n CreateProjectVersion\n
\n\n DeleteDataset\n
\n\n DeleteProject\n
\n\n DeleteProjectPolicy\n
\n\n DeleteProjectVersion\n
\n\n DescribeDataset\n
\n\n DescribeProjects\n
\n\n DetectCustomLabels\n
\n\n ListDatasetEntries\n
\n\n ListDatasetLabels\n
\n\n ListProjectPolicies\n
\n\n PutProjectPolicy\n
\n\n StartProjectVersion\n
\n\n StopProjectVersion\n
\n\n UpdateDatasetEntries\n
\n\n Amazon Rekognition Video Stored Video\n
\n\n GetContentModeration\n
\n\n GetFaceDetection\n
\n\n GetFaceSearch\n
\n\n GetLabelDetection\n
\n\n GetPersonTracking\n
\n\n GetSegmentDetection\n
\n\n GetTextDetection\n
\n\n StartFaceDetection\n
\n\n StartFaceSearch\n
\n\n StartLabelDetection\n
\n\n StartPersonTracking\n
\n\n StartTextDetection\n
\n\n Amazon Rekognition Video Streaming Video\n
\n\n ListStreamProcessors\n
\n\n StartStreamProcessor\n
\n\n StopStreamProcessor\n
\nThis is the API Reference for Amazon Rekognition Image, Amazon Rekognition Custom Labels,\n Amazon Rekognition Stored\n Video, Amazon Rekognition Streaming Video. It provides descriptions of actions, data types, common\n parameters, and common errors.
\n\n Amazon Rekognition Image\n
\n\n AssociateFaces\n
\n\n CompareFaces\n
\n\n CreateCollection\n
\n\n CreateUser\n
\n\n DeleteCollection\n
\n\n DeleteFaces\n
\n\n DeleteUser\n
\n\n DescribeCollection\n
\n\n DetectFaces\n
\n\n DetectLabels\n
\n\n DetectText\n
\n\n DisassociateFaces\n
\n\n GetCelebrityInfo\n
\n\n IndexFaces\n
\n\n ListCollections\n
\n\n ListFaces\n
\n\n ListUsers\n
\n\n RecognizeCelebrities\n
\n\n SearchFaces\n
\n\n SearchFacesByImage\n
\n\n SearchUsers\n
\n\n SearchUsersByImage\n
\n\n Amazon Rekognition Custom Labels\n
\n\n CopyProjectVersion\n
\n\n CreateDataset\n
\n\n CreateProject\n
\n\n CreateProjectVersion\n
\n\n DeleteDataset\n
\n\n DeleteProject\n
\n\n DeleteProjectPolicy\n
\n\n DeleteProjectVersion\n
\n\n DescribeDataset\n
\n\n DescribeProjects\n
\n\n DetectCustomLabels\n
\n\n ListDatasetEntries\n
\n\n ListDatasetLabels\n
\n\n ListProjectPolicies\n
\n\n PutProjectPolicy\n
\n\n StartProjectVersion\n
\n\n StopProjectVersion\n
\n\n UpdateDatasetEntries\n
\n\n Amazon Rekognition Video Stored Video\n
\n\n GetContentModeration\n
\n\n GetFaceDetection\n
\n\n GetFaceSearch\n
\n\n GetLabelDetection\n
\n\n GetPersonTracking\n
\n\n GetSegmentDetection\n
\n\n GetTextDetection\n
\n\n StartFaceDetection\n
\n\n StartFaceSearch\n
\n\n StartLabelDetection\n
\n\n StartPersonTracking\n
\n\n StartTextDetection\n
\n\n Amazon Rekognition Video Streaming Video\n
\n\n ListStreamProcessors\n
\n\n StartStreamProcessor\n
\n\n StopStreamProcessor\n
\nFor a given input face ID, searches for matching faces in the collection the face\n belongs to. You get a face ID when you add a face to the collection using the IndexFaces operation. The operation compares the features of the input face with\n faces in the specified collection.
\nYou can also search faces without indexing faces by using the\n SearchFacesByImage
operation.
The operation response returns an array of faces that match, ordered by similarity\n score with the highest similarity first. More specifically, it is an array of metadata for\n each face match that is found. Along with the metadata, the response also includes a\n confidence
value for each face match, indicating the confidence that the\n specific face matches the input face.
For an example, see Searching for a face using its face ID in the Amazon Rekognition\n Developer Guide.
\nThis operation requires permissions to perform the rekognition:SearchFaces
\n action.
For a given input face ID, searches for matching faces in the collection the face\n belongs to. You get a face ID when you add a face to the collection using the IndexFaces operation. The operation compares the features of the input face with\n faces in the specified collection.
\nYou can also search faces without indexing faces by using the\n SearchFacesByImage
operation.
The operation response returns an array of faces that match, ordered by similarity\n score with the highest similarity first. More specifically, it is an array of metadata for\n each face match that is found. Along with the metadata, the response also includes a\n confidence
value for each face match, indicating the confidence that the\n specific face matches the input face.
For an example, see Searching for a face using its face ID in the Amazon Rekognition\n Developer Guide.
\nThis operation requires permissions to perform the rekognition:SearchFaces
\n action.
For a given input image, first detects the largest face in the image, and then searches\n the specified collection for matching faces. The operation compares the features of the input\n face with faces in the specified collection.
\nTo search for all faces in an input image, you might first call the IndexFaces operation, and then use the face IDs returned in subsequent calls\n to the SearchFaces operation.
\n You can also call the DetectFaces
operation and use the bounding boxes\n in the response to make face crops, which then you can pass in to the\n SearchFacesByImage
operation.
You pass the input image either as base64-encoded image bytes or as a reference to an\n image in an Amazon S3 bucket. If you use the\n AWS\n CLI to call Amazon Rekognition operations, passing image bytes is not\n supported. The image must be either a PNG or JPEG formatted file.
\n The response returns an array of faces that match, ordered by similarity score with\n the highest similarity first. More specifically, it is an array of metadata for each face\n match found. Along with the metadata, the response also includes a similarity
\n indicating how similar the face is to the input face. In the response, the operation also\n returns the bounding box (and a confidence level that the bounding box contains a face) of the\n face that Amazon Rekognition used for the input image.
If no faces are detected in the input image, SearchFacesByImage
returns an\n InvalidParameterException
error.
For an example, Searching for a Face Using an Image in the Amazon Rekognition\n Developer Guide.
\nThe QualityFilter
input parameter allows you to filter out detected faces\n that don’t meet a required quality bar. The quality bar is based on a variety of common use\n cases. Use QualityFilter
to set the quality bar for filtering by specifying\n LOW
, MEDIUM
, or HIGH
. If you do not want to filter\n detected faces, specify NONE
. The default value is NONE
.
To use quality filtering, you need a collection associated with version 3 of the face\n model or higher. To get the version of the face model associated with a collection, call\n DescribeCollection.
\nThis operation requires permissions to perform the\n rekognition:SearchFacesByImage
action.
For a given input image, first detects the largest face in the image, and then searches\n the specified collection for matching faces. The operation compares the features of the input\n face with faces in the specified collection.
\nTo search for all faces in an input image, you might first call the IndexFaces operation, and then use the face IDs returned in subsequent calls\n to the SearchFaces operation.
\n You can also call the DetectFaces
operation and use the bounding boxes\n in the response to make face crops, which then you can pass in to the\n SearchFacesByImage
operation.
You pass the input image either as base64-encoded image bytes or as a reference to an\n image in an Amazon S3 bucket. If you use the\n AWS\n CLI to call Amazon Rekognition operations, passing image bytes is not\n supported. The image must be either a PNG or JPEG formatted file.
\n The response returns an array of faces that match, ordered by similarity score with\n the highest similarity first. More specifically, it is an array of metadata for each face\n match found. Along with the metadata, the response also includes a similarity
\n indicating how similar the face is to the input face. In the response, the operation also\n returns the bounding box (and a confidence level that the bounding box contains a face) of the\n face that Amazon Rekognition used for the input image.
If no faces are detected in the input image, SearchFacesByImage
returns an\n InvalidParameterException
error.
For an example, Searching for a Face Using an Image in the Amazon Rekognition\n Developer Guide.
\nThe QualityFilter
input parameter allows you to filter out detected faces\n that don’t meet a required quality bar. The quality bar is based on a variety of common use\n cases. Use QualityFilter
to set the quality bar for filtering by specifying\n LOW
, MEDIUM
, or HIGH
. If you do not want to filter\n detected faces, specify NONE
. The default value is NONE
.
To use quality filtering, you need a collection associated with version 3 of the face\n model or higher. To get the version of the face model associated with a collection, call\n DescribeCollection.
\nThis operation requires permissions to perform the\n rekognition:SearchFacesByImage
action.
Searches for UserIDs within a collection based on a FaceId
or\n UserId
. This API can be used to find the closest UserID (with a highest\n similarity) to associate a face. The request must be provided with either FaceId
\n or UserId
. The operation returns an array of UserID that match the\n FaceId
or UserId
, ordered by similarity score with the highest\n similarity first.
The type of the segment. Valid values are TECHNICAL_CUE
and SHOT
.
The start time of the detected segment in milliseconds from the start of the video. This value\n is rounded down. For example, if the actual timestamp is 100.6667 milliseconds, Amazon Rekognition Video returns a value of\n 100 millis.
" - } + { + "target": "com.amazonaws.rekognition#ImageTooLargeException" }, - "EndTimestampMillis": { - "target": "com.amazonaws.rekognition#Timestamp", - "traits": { - "smithy.api#default": 0, - "smithy.api#documentation": "The end time of the detected segment, in milliseconds, from the start of the video.\n This value is rounded down.
" - } + { + "target": "com.amazonaws.rekognition#InternalServerError" }, - "DurationMillis": { - "target": "com.amazonaws.rekognition#ULong", - "traits": { - "smithy.api#documentation": "The duration of the detected segment in milliseconds.
" - } + { + "target": "com.amazonaws.rekognition#InvalidImageFormatException" }, - "StartTimecodeSMPTE": { - "target": "com.amazonaws.rekognition#Timecode", - "traits": { - "smithy.api#documentation": "The frame-accurate SMPTE timecode, from the start of a video, for the start of a detected segment.\n StartTimecode
is in HH:MM:SS:fr format\n (and ;fr for drop frame-rates).
The frame-accurate SMPTE timecode, from the start of a video, for the end of a detected segment. \n EndTimecode
is in HH:MM:SS:fr format\n (and ;fr for drop frame-rates).
The duration of the timecode for the detected segment in SMPTE format.
" - } + { + "target": "com.amazonaws.rekognition#ProvisionedThroughputExceededException" }, - "TechnicalCueSegment": { - "target": "com.amazonaws.rekognition#TechnicalCueSegment", - "traits": { - "smithy.api#documentation": "If the segment is a technical cue, contains information about the technical cue.
" - } + { + "target": "com.amazonaws.rekognition#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.rekognition#ThrottlingException" + } + ], + "traits": { + "smithy.api#documentation": "Searches for UserIDs using a supplied image. It first detects the largest face in the\n image, and then searches a specified collection for matching UserIDs.
\nThe operation returns an array of UserIDs that match the face in the supplied image,\n ordered by similarity score with the highest similarity first. It also returns a bounding box\n for the face found in the input image.
\nInformation about faces detected in the supplied image, but not used for the search, is\n returned in an array of UnsearchedFace
objects. If no valid face is detected\n in the image, the response will contain an empty UserMatches
list and no\n SearchedFace
object.
The ID of an existing collection containing the UserID.
", + "smithy.api#required": {} + } + }, + "Image": { + "target": "com.amazonaws.rekognition#Image", + "traits": { + "smithy.api#required": {} + } + }, + "UserMatchThreshold": { + "target": "com.amazonaws.rekognition#Percent", + "traits": { + "smithy.api#documentation": "Specifies the minimum confidence in the UserID match to return. Default value is\n 80.
" + } + }, + "MaxUsers": { + "target": "com.amazonaws.rekognition#MaxUserResults", + "traits": { + "smithy.api#documentation": "Maximum number of UserIDs to return.
" + } + }, + "QualityFilter": { + "target": "com.amazonaws.rekognition#QualityFilter", + "traits": { + "smithy.api#documentation": "A filter that specifies a quality bar for how much filtering is done to identify faces.\n Filtered faces aren't searched for in the collection. The default value is NONE.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.rekognition#SearchUsersByImageResponse": { + "type": "structure", + "members": { + "UserMatches": { + "target": "com.amazonaws.rekognition#UserMatchList", + "traits": { + "smithy.api#documentation": "An array of UserID objects that matched the input face, along with the confidence in the\n match. The returned structure will be empty if there are no matches. Returned if the\n SearchUsersByImageResponse action is successful.
" + } + }, + "FaceModelVersion": { + "target": "com.amazonaws.rekognition#String", + "traits": { + "smithy.api#documentation": "Version number of the face detection model associated with the input collection\n CollectionId.
" + } + }, + "SearchedFace": { + "target": "com.amazonaws.rekognition#SearchedFaceDetails", + "traits": { + "smithy.api#documentation": "A list of FaceDetail objects containing the BoundingBox for the largest face in image,\n as well as the confidence in the bounding box, that was searched for matches. If no valid\n face is detected in the image the response will contain no SearchedFace object.
" + } + }, + "UnsearchedFaces": { + "target": "com.amazonaws.rekognition#UnsearchedFacesList", + "traits": { + "smithy.api#documentation": "List of UnsearchedFace objects. Contains the face details infered from the specified\n image but not used for search. Contains reasons that describe why a face wasn't used for\n Search.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.rekognition#SearchUsersRequest": { + "type": "structure", + "members": { + "CollectionId": { + "target": "com.amazonaws.rekognition#CollectionId", + "traits": { + "smithy.api#documentation": "The ID of an existing collection containing the UserID, used with a UserId or FaceId. If a\n FaceId is provided, UserId isn’t required to be present in the Collection.
", + "smithy.api#required": {} + } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "ID for the existing User.
" + } + }, + "FaceId": { + "target": "com.amazonaws.rekognition#FaceId", + "traits": { + "smithy.api#documentation": "ID for the existing face.
" + } + }, + "UserMatchThreshold": { + "target": "com.amazonaws.rekognition#Percent", + "traits": { + "smithy.api#documentation": "Optional value that specifies the minimum confidence in the matched UserID to return.\n Default value of 80.
" + } + }, + "MaxUsers": { + "target": "com.amazonaws.rekognition#MaxUserResults", + "traits": { + "smithy.api#documentation": "Maximum number of identities to return.
" + } + } + }, + "traits": { + "smithy.api#input": {} + } + }, + "com.amazonaws.rekognition#SearchUsersResponse": { + "type": "structure", + "members": { + "UserMatches": { + "target": "com.amazonaws.rekognition#UserMatchList", + "traits": { + "smithy.api#documentation": "An array of UserMatch objects that matched the input face along with the confidence in\n the match. Array will be empty if there are no matches.
" + } + }, + "FaceModelVersion": { + "target": "com.amazonaws.rekognition#String", + "traits": { + "smithy.api#documentation": "Version number of the face detection model associated with the input\n CollectionId.
" + } + }, + "SearchedFace": { + "target": "com.amazonaws.rekognition#SearchedFace", + "traits": { + "smithy.api#documentation": "Contains the ID of a face that was used to search for matches in a collection.
" + } + }, + "SearchedUser": { + "target": "com.amazonaws.rekognition#SearchedUser", + "traits": { + "smithy.api#documentation": "Contains the ID of the UserID that was used to search for matches in a collection.
" + } + } + }, + "traits": { + "smithy.api#output": {} + } + }, + "com.amazonaws.rekognition#SearchedFace": { + "type": "structure", + "members": { + "FaceId": { + "target": "com.amazonaws.rekognition#FaceId", + "traits": { + "smithy.api#documentation": "Unique identifier assigned to the face.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Provides face metadata such as FaceId, BoundingBox, Confidence of the input face used for search.
" + } + }, + "com.amazonaws.rekognition#SearchedFaceDetails": { + "type": "structure", + "members": { + "FaceDetail": { + "target": "com.amazonaws.rekognition#FaceDetail" + } + }, + "traits": { + "smithy.api#documentation": "Contains data regarding the input face used for a search.
" + } + }, + "com.amazonaws.rekognition#SearchedUser": { + "type": "structure", + "members": { + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "A provided ID for the UserID. Unique within the collection.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Contains metadata about a User searched for within a collection.
" + } + }, + "com.amazonaws.rekognition#SegmentConfidence": { + "type": "float", + "traits": { + "smithy.api#range": { + "min": 50, + "max": 100 + } + } + }, + "com.amazonaws.rekognition#SegmentDetection": { + "type": "structure", + "members": { + "Type": { + "target": "com.amazonaws.rekognition#SegmentType", + "traits": { + "smithy.api#documentation": "The type of the segment. Valid values are TECHNICAL_CUE
and SHOT
.
The start time of the detected segment in milliseconds from the start of the video. This value\n is rounded down. For example, if the actual timestamp is 100.6667 milliseconds, Amazon Rekognition Video returns a value of\n 100 millis.
" + } + }, + "EndTimestampMillis": { + "target": "com.amazonaws.rekognition#Timestamp", + "traits": { + "smithy.api#default": 0, + "smithy.api#documentation": "The end time of the detected segment, in milliseconds, from the start of the video.\n This value is rounded down.
" + } + }, + "DurationMillis": { + "target": "com.amazonaws.rekognition#ULong", + "traits": { + "smithy.api#documentation": "The duration of the detected segment in milliseconds.
" + } + }, + "StartTimecodeSMPTE": { + "target": "com.amazonaws.rekognition#Timecode", + "traits": { + "smithy.api#documentation": "The frame-accurate SMPTE timecode, from the start of a video, for the start of a detected segment.\n StartTimecode
is in HH:MM:SS:fr format\n (and ;fr for drop frame-rates).
The frame-accurate SMPTE timecode, from the start of a video, for the end of a detected segment. \n EndTimecode
is in HH:MM:SS:fr format\n (and ;fr for drop frame-rates).
The duration of the timecode for the detected segment in SMPTE format.
" + } + }, + "TechnicalCueSegment": { + "target": "com.amazonaws.rekognition#TechnicalCueSegment", + "traits": { + "smithy.api#documentation": "If the segment is a technical cue, contains information about the technical cue.
" + } }, "ShotSegment": { "target": "com.amazonaws.rekognition#ShotSegment", @@ -12526,42 +14124,319 @@ "target": "com.amazonaws.rekognition#UnindexedFace" } }, - "com.amazonaws.rekognition#UntagResource": { - "type": "operation", - "input": { - "target": "com.amazonaws.rekognition#UntagResourceRequest" - }, - "output": { - "target": "com.amazonaws.rekognition#UntagResourceResponse" - }, - "errors": [ - { - "target": "com.amazonaws.rekognition#AccessDeniedException" - }, - { - "target": "com.amazonaws.rekognition#InternalServerError" - }, - { - "target": "com.amazonaws.rekognition#InvalidParameterException" - }, - { - "target": "com.amazonaws.rekognition#ProvisionedThroughputExceededException" - }, - { - "target": "com.amazonaws.rekognition#ResourceNotFoundException" + "com.amazonaws.rekognition#UnsearchedFace": { + "type": "structure", + "members": { + "FaceDetails": { + "target": "com.amazonaws.rekognition#FaceDetail" }, - { - "target": "com.amazonaws.rekognition#ThrottlingException" + "Reasons": { + "target": "com.amazonaws.rekognition#UnsearchedFaceReasons", + "traits": { + "smithy.api#documentation": "Reasons why a face wasn't used for Search.
" + } } - ], + }, "traits": { - "smithy.api#documentation": "Removes one or more tags from an Amazon Rekognition collection, stream processor, or Custom Labels\n model.
\nThis operation requires permissions to perform the\n rekognition:UntagResource
action.
Face details inferred from the image but not used for search. The response attribute\n contains reasons for why a face wasn't used for Search.
" } }, - "com.amazonaws.rekognition#UntagResourceRequest": { - "type": "structure", + "com.amazonaws.rekognition#UnsearchedFaceReason": { + "type": "enum", "members": { - "ResourceArn": { + "FACE_NOT_LARGEST": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "FACE_NOT_LARGEST" + } + }, + "EXCEEDS_MAX_FACES": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "EXCEEDS_MAX_FACES" + } + }, + "EXTREME_POSE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "EXTREME_POSE" + } + }, + "LOW_BRIGHTNESS": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "LOW_BRIGHTNESS" + } + }, + "LOW_SHARPNESS": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "LOW_SHARPNESS" + } + }, + "LOW_CONFIDENCE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "LOW_CONFIDENCE" + } + }, + "SMALL_BOUNDING_BOX": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "SMALL_BOUNDING_BOX" + } + }, + "LOW_FACE_QUALITY": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "LOW_FACE_QUALITY" + } + } + } + }, + "com.amazonaws.rekognition#UnsearchedFaceReasons": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsearchedFaceReason" + } + }, + "com.amazonaws.rekognition#UnsearchedFacesList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsearchedFace" + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceAssociation": { + "type": "structure", + "members": { + "FaceId": { + "target": "com.amazonaws.rekognition#FaceId", + "traits": { + "smithy.api#documentation": "A unique identifier assigned to the face.
" + } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "A provided ID for the UserID. Unique within the collection.
" + } + }, + "Confidence": { + "target": "com.amazonaws.rekognition#Percent", + "traits": { + "smithy.api#documentation": "Match confidence with the UserID, provides information regarding if a face association was\n unsuccessful because it didn't meet UserMatchThreshold.
" + } + }, + "Reasons": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceAssociationReasons", + "traits": { + "smithy.api#documentation": "The reason why the association was unsuccessful.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Contains metadata like FaceId, UserID, and Reasons, for a face that was unsuccessfully associated.
" + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceAssociationList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceAssociation" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 500 + } + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceAssociationReason": { + "type": "enum", + "members": { + "FACE_NOT_FOUND": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "FACE_NOT_FOUND" + } + }, + "ASSOCIATED_TO_A_DIFFERENT_USER": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ASSOCIATED_TO_A_DIFFERENT_USER" + } + }, + "LOW_MATCH_CONFIDENCE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "LOW_MATCH_CONFIDENCE" + } + } + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceAssociationReasons": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceAssociationReason" + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDeletion": { + "type": "structure", + "members": { + "FaceId": { + "target": "com.amazonaws.rekognition#FaceId", + "traits": { + "smithy.api#documentation": "A unique identifier assigned to the face.
" + } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "A provided ID for the UserID. Unique within the collection.
" + } + }, + "Reasons": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDeletionReasons", + "traits": { + "smithy.api#documentation": "The reason why the deletion was unsuccessful.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Contains metadata like FaceId, UserID, and Reasons, for a face that was unsuccessfully deleted.
" + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDeletionReason": { + "type": "enum", + "members": { + "ASSOCIATED_TO_AN_EXISTING_USER": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ASSOCIATED_TO_AN_EXISTING_USER" + } + }, + "FACE_NOT_FOUND": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "FACE_NOT_FOUND" + } + } + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDeletionReasons": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDeletionReason" + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDeletionsList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDeletion" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 4096 + } + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDisassociation": { + "type": "structure", + "members": { + "FaceId": { + "target": "com.amazonaws.rekognition#FaceId", + "traits": { + "smithy.api#documentation": "A unique identifier assigned to the face.
" + } + }, + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "A provided ID for the UserID. Unique within the collection.
" + } + }, + "Reasons": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDisassociationReasons", + "traits": { + "smithy.api#documentation": "The reason why the deletion was unsuccessful.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Contains metadata like FaceId, UserID, and Reasons, for a face that was unsuccessfully disassociated.
" + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDisassociationList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDisassociation" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 500 + } + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDisassociationReason": { + "type": "enum", + "members": { + "FACE_NOT_FOUND": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "FACE_NOT_FOUND" + } + }, + "ASSOCIATED_TO_A_DIFFERENT_USER": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ASSOCIATED_TO_A_DIFFERENT_USER" + } + } + } + }, + "com.amazonaws.rekognition#UnsuccessfulFaceDisassociationReasons": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UnsuccessfulFaceDisassociationReason" + } + }, + "com.amazonaws.rekognition#UntagResource": { + "type": "operation", + "input": { + "target": "com.amazonaws.rekognition#UntagResourceRequest" + }, + "output": { + "target": "com.amazonaws.rekognition#UntagResourceResponse" + }, + "errors": [ + { + "target": "com.amazonaws.rekognition#AccessDeniedException" + }, + { + "target": "com.amazonaws.rekognition#InternalServerError" + }, + { + "target": "com.amazonaws.rekognition#InvalidParameterException" + }, + { + "target": "com.amazonaws.rekognition#ProvisionedThroughputExceededException" + }, + { + "target": "com.amazonaws.rekognition#ResourceNotFoundException" + }, + { + "target": "com.amazonaws.rekognition#ThrottlingException" + } + ], + "traits": { + "smithy.api#documentation": "Removes one or more tags from an Amazon Rekognition collection, stream processor, or Custom Labels\n model.
\nThis operation requires permissions to perform the\n rekognition:UntagResource
action.
Amazon Resource Name (ARN) of the model, collection, or stream processor that you want to\n remove the tags from.
", @@ -12750,6 +14625,121 @@ } } }, + "com.amazonaws.rekognition#User": { + "type": "structure", + "members": { + "UserId": { + "target": "com.amazonaws.rekognition#UserId", + "traits": { + "smithy.api#documentation": "A provided ID for the User. Unique within the collection.
" + } + }, + "UserStatus": { + "target": "com.amazonaws.rekognition#UserStatus", + "traits": { + "smithy.api#documentation": "Communicates if the UserID has been updated with latest set of faces to be associated\n with the UserID.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Metadata of the user stored in a collection.
" + } + }, + "com.amazonaws.rekognition#UserFaceIdList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#FaceId" + }, + "traits": { + "smithy.api#length": { + "min": 1, + "max": 100 + } + } + }, + "com.amazonaws.rekognition#UserId": { + "type": "string", + "traits": { + "smithy.api#length": { + "min": 1, + "max": 128 + }, + "smithy.api#pattern": "^[a-zA-Z0-9_.\\-:]+$" + } + }, + "com.amazonaws.rekognition#UserList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#User" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 500 + } + } + }, + "com.amazonaws.rekognition#UserMatch": { + "type": "structure", + "members": { + "Similarity": { + "target": "com.amazonaws.rekognition#Percent", + "traits": { + "smithy.api#documentation": "Describes the UserID metadata.
" + } + }, + "User": { + "target": "com.amazonaws.rekognition#MatchedUser", + "traits": { + "smithy.api#documentation": "Confidence in the match of this UserID with the input face.
" + } + } + }, + "traits": { + "smithy.api#documentation": "Provides UserID metadata along with the confidence in the match of this UserID with the\n input face.
" + } + }, + "com.amazonaws.rekognition#UserMatchList": { + "type": "list", + "member": { + "target": "com.amazonaws.rekognition#UserMatch" + }, + "traits": { + "smithy.api#length": { + "min": 0, + "max": 500 + } + } + }, + "com.amazonaws.rekognition#UserStatus": { + "type": "enum", + "members": { + "ACTIVE": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ACTIVE" + } + }, + "UPDATING": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "UPDATING" + } + }, + "CREATING": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "CREATING" + } + }, + "CREATED": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "CREATED" + } + } + } + }, "com.amazonaws.rekognition#ValidationData": { "type": "structure", "members": { diff --git a/codegen/sdk-codegen/aws-models/sagemaker.json b/codegen/sdk-codegen/aws-models/sagemaker.json index b110ed74984..b7a89b71c70 100644 --- a/codegen/sdk-codegen/aws-models/sagemaker.json +++ b/codegen/sdk-codegen/aws-models/sagemaker.json @@ -256,7 +256,7 @@ "target": "com.amazonaws.sagemaker#AddTagsOutput" }, "traits": { - "smithy.api#documentation": "Adds or overwrites one or more tags for the specified SageMaker resource. You can add\n tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform\n jobs, models, labeling jobs, work teams, endpoint configurations, and\n endpoints.
\nEach tag consists of a key and an optional value. Tag keys must be unique per\n resource. For more information about tags, see For more information, see Amazon Web Services Tagging Strategies.
\nTags that you add to a hyperparameter tuning job by calling this API are also\n added to any training jobs that the hyperparameter tuning job launches after you\n call this API, but not to training jobs that the hyperparameter tuning job launched\n before you called this API. To make sure that the tags associated with a\n hyperparameter tuning job are also added to all training jobs that the\n hyperparameter tuning job launches, add the tags when you first create the tuning\n job by specifying them in the Tags
parameter of CreateHyperParameterTuningJob\n
Tags that you add to a SageMaker Studio Domain or User Profile by calling this API\n are also added to any Apps that the Domain or User Profile launches after you call\n this API, but not to Apps that the Domain or User Profile launched before you called\n this API. To make sure that the tags associated with a Domain or User Profile are\n also added to all Apps that the Domain or User Profile launches, add the tags when\n you first create the Domain or User Profile by specifying them in the\n Tags
parameter of CreateDomain or CreateUserProfile.
Adds or overwrites one or more tags for the specified SageMaker resource. You can add\n tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform\n jobs, models, labeling jobs, work teams, endpoint configurations, and\n endpoints.
\nEach tag consists of a key and an optional value. Tag keys must be unique per\n resource. For more information about tags, see For more information, see Amazon Web Services Tagging Strategies.
\nTags that you add to a hyperparameter tuning job by calling this API are also\n added to any training jobs that the hyperparameter tuning job launches after you\n call this API, but not to training jobs that the hyperparameter tuning job launched\n before you called this API. To make sure that the tags associated with a\n hyperparameter tuning job are also added to all training jobs that the\n hyperparameter tuning job launches, add the tags when you first create the tuning\n job by specifying them in the Tags
parameter of CreateHyperParameterTuningJob\n
Tags that you add to a SageMaker Studio Domain or User Profile by calling this API\n are also added to any Apps that the Domain or User Profile launches after you call\n this API, but not to Apps that the Domain or User Profile launched before you called\n this API. To make sure that the tags associated with a Domain or User Profile are\n also added to all Apps that the Domain or User Profile launches, add the tags when\n you first create the Domain or User Profile by specifying them in the\n Tags
parameter of CreateDomain\n or CreateUserProfile.
Specifies the training algorithm to use in a CreateTrainingJob\n request.
\nFor more information about algorithms provided by SageMaker, see Algorithms. For\n information about using your own algorithms, see Using Your Own Algorithms with\n Amazon SageMaker.
" + "smithy.api#documentation": "Specifies the training algorithm to use in a CreateTrainingJob request.
\nFor more information about algorithms provided by SageMaker, see Algorithms. For\n information about using your own algorithms, see Using Your Own Algorithms with\n Amazon SageMaker.
" } }, "com.amazonaws.sagemaker#AlgorithmStatus": { @@ -1820,7 +1820,7 @@ "IncludeInferenceResponseIn": { "target": "com.amazonaws.sagemaker#AsyncNotificationTopicTypeList", "traits": { - "smithy.api#documentation": "The Amazon SNS topics where you want the inference response to be included.
\nThe inference response is included only if the response size is less than or equal to 128 KB.
\nThe Amazon SNS topics where you want the inference response to be included.
\nThe inference response is included only if the response size is less than or equal\n to 128 KB.
\nThe content type of the data from the input source. The following are the allowed\n content types for different problems:
\nImageClassification: image/png
, image/jpeg
,\n image/*
\n
TextClassification: text/csv;header=present
\n
The content type of the data from the input source. The following are the allowed\n content types for different problems:
\nImageClassification: image/png
, image/jpeg
, or\n image/*
. The default value is image/*
.
TextClassification: text/csv;header=present
or x-application/vnd.amazon+parquet
. \n The default value is text/csv;header=present
.
The name and an example value of the hyperparameter that\n you want to use in Autotune. If Automatic model tuning (AMT) determines that your hyperparameter is eligible\n for Autotune, an optimal hyperparameter range is selected for you.
" + "smithy.api#documentation": "The name and an example value of the hyperparameter that you want to use in Autotune.\n If Automatic model tuning (AMT) determines that your hyperparameter is eligible for\n Autotune, an optimal hyperparameter range is selected for you.
" } }, "com.amazonaws.sagemaker#AutoParameters": { @@ -3397,7 +3397,7 @@ } }, "traits": { - "smithy.api#documentation": "A flag to indicate if you want to use Autotune to automatically find optimal values for\n the following fields:
\n\n ParameterRanges: The names and ranges of parameters that a hyperparameter\n tuning job can optimize.
\n\n ResourceLimits: The maximum resources that can be used for a training job.\n These resources include the maximum number of training jobs, the maximum runtime of a\n tuning job, and the maximum number of training jobs to run at the same time.
\n\n TrainingJobEarlyStoppingType: A flag that specifies whether or not to use\n early stopping for training jobs launched by a hyperparameter tuning job.
\n\n RetryStrategy: The number of times to retry a training job.
\n\n Strategy: Specifies how hyperparameter tuning chooses the combinations of\n hyperparameter values to use for the training jobs that it launches.
\n\n ConvergenceDetected: A flag to indicate that Automatic model tuning (AMT) has detected model\n convergence.
\nA flag to indicate if you want to use Autotune to automatically find optimal values\n for the following fields:
\n\n ParameterRanges: The names and ranges of parameters that a\n hyperparameter tuning job can optimize.
\n\n ResourceLimits: The maximum resources that can be used for a\n training job. These resources include the maximum number of training jobs, the\n maximum runtime of a tuning job, and the maximum number of training jobs to run\n at the same time.
\n\n TrainingJobEarlyStoppingType: A flag that specifies whether or not\n to use early stopping for training jobs launched by a hyperparameter tuning\n job.
\n\n RetryStrategy: The number of times to retry a training job.
\n\n Strategy: Specifies how hyperparameter tuning chooses the\n combinations of hyperparameter values to use for the training jobs that it\n launches.
\n\n ConvergenceDetected: A flag to indicate that Automatic model tuning\n (AMT) has detected model convergence.
\n(Optional) The input mode to use for the data channel in a training job. If you don't\n set a value for InputMode
, SageMaker uses the value set for\n TrainingInputMode
. Use this parameter to override the\n TrainingInputMode
setting in a AlgorithmSpecification \n request when you have a channel that needs a different input mode from the training\n job's general setting. To download the data from Amazon Simple Storage Service (Amazon S3) to the provisioned ML\n storage volume, and mount the directory to a Docker volume, use File
input\n mode. To stream data directly from Amazon S3 to the container, choose Pipe
input\n mode.
To use a model for incremental training, choose File
input model.
(Optional) The input mode to use for the data channel in a training job. If you don't\n set a value for InputMode
, SageMaker uses the value set for\n TrainingInputMode
. Use this parameter to override the\n TrainingInputMode
setting in a AlgorithmSpecification request when you have a channel that needs a\n different input mode from the training job's general setting. To download the data from\n Amazon Simple Storage Service (Amazon S3) to the provisioned ML storage volume, and mount the directory to a\n Docker volume, use File
input mode. To stream data directly from Amazon S3 to\n the container, choose Pipe
input mode.
To use a model for incremental training, choose File
input model.
The allowed input mode, either FILE or PIPE.
\nIn FILE mode, Amazon SageMaker copies the data from the input source onto the local\n Amazon Elastic Block Store (Amazon EBS) volumes before starting your training algorithm.\n This is the most commonly used input mode.
\nIn PIPE mode, Amazon SageMaker streams input data from the source directly to your\n algorithm without using the EBS volume.
", + "smithy.api#documentation": "The allowed input mode, either FILE or PIPE.
\nIn FILE mode, Amazon SageMaker copies the data from the input source onto the local Amazon\n Elastic Block Store (Amazon EBS) volumes before starting your training algorithm. This\n is the most commonly used input mode.
\nIn PIPE mode, Amazon SageMaker streams input data from the source directly to your algorithm\n without using the EBS volume.
", "smithy.api#required": {} } } @@ -5873,13 +5873,13 @@ "Image": { "target": "com.amazonaws.sagemaker#ContainerImage", "traits": { - "smithy.api#documentation": "The path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a\n Docker registry that is accessible from the same VPC that you configure for your\n endpoint. If you are using your own custom algorithm instead of an algorithm provided by\n SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both\n registry/repository[:tag]
and registry/repository[@digest]
\n image path formats. For more information, see Using Your Own Algorithms with\n Amazon SageMaker.\n
The model artifacts in an Amazon S3 bucket and the Docker image for inference container\n in Amazon EC2 Container Registry must be in the same region as the model or endpoint you are\n creating.
\nThe path where inference code is stored. This can be either in Amazon EC2 Container Registry or in a\n Docker registry that is accessible from the same VPC that you configure for your\n endpoint. If you are using your own custom algorithm instead of an algorithm provided by\n SageMaker, the inference code must meet SageMaker requirements. SageMaker supports both\n registry/repository[:tag]
and registry/repository[@digest]
\n image path formats. For more information, see Using Your Own Algorithms with\n Amazon SageMaker.
The model artifacts in an Amazon S3 bucket and the Docker image for inference container\n in Amazon EC2 Container Registry must be in the same region as the model or endpoint you are\n creating.
\nSpecifies whether the model container is in Amazon ECR or a private Docker registry\n accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a\n private Docker registry, see Use a\n Private Docker Registry for Real-Time Inference Containers.\n
\nThe model artifacts in an Amazon S3 bucket and the Docker image for inference container\n in Amazon EC2 Container Registry must be in the same region as the model or endpoint you are\n creating.
\nSpecifies whether the model container is in Amazon ECR or a private Docker registry\n accessible from your Amazon Virtual Private Cloud (VPC). For information about storing containers in a\n private Docker registry, see Use a\n Private Docker Registry for Real-Time Inference Containers.
\nThe model artifacts in an Amazon S3 bucket and the Docker image for inference container\n in Amazon EC2 Container Registry must be in the same region as the model or endpoint you are\n creating.
\nThe entity that creates and manages the required security groups for inter-app\n communication in VPCOnly
mode. Required when\n CreateDomain.AppNetworkAccessType
is VPCOnly
and\n DomainSettings.RStudioServerProDomainSettings.DomainExecutionRoleArn
is\n provided. If setting up the domain for use with RStudio, this value must be set to Service
.
The entity that creates and manages the required security groups for inter-app\n communication in VPCOnly
mode. Required when\n CreateDomain.AppNetworkAccessType
is VPCOnly
and\n DomainSettings.RStudioServerProDomainSettings.DomainExecutionRoleArn
is\n provided. If setting up the domain for use with RStudio, this value must be set to\n Service
.
Creates an endpoint using the endpoint configuration specified in the request. SageMaker\n uses the endpoint to provision resources and deploy models. You create the endpoint\n configuration with the CreateEndpointConfig API.
\nUse this API to deploy models using SageMaker hosting services.
\nFor an example that calls this method when deploying a model to SageMaker hosting services,\n see the Create Endpoint example notebook.\n
\n You must not delete an EndpointConfig
that is in use by an endpoint\n that is live or while the UpdateEndpoint
or CreateEndpoint
\n operations are being performed on the endpoint. To update an endpoint, you must\n create a new EndpointConfig
.
The endpoint name must be unique within an Amazon Web Services Region in your\n Amazon Web Services account.
\nWhen it receives the request, SageMaker creates the endpoint, launches the resources (ML\n compute instances), and deploys the model(s) on them.
\nWhen you call CreateEndpoint, a load call is made to DynamoDB to\n verify that your endpoint configuration exists. When you read data from a DynamoDB\n table supporting \n Eventually Consistent Reads
\n , the response might not\n reflect the results of a recently completed write operation. The response might\n include some stale data. If the dependent entities are not yet in DynamoDB, this\n causes a validation error. If you repeat your read request after a short time, the\n response should return the latest data. So retry logic is recommended to handle\n these possible issues. We also recommend that customers call DescribeEndpointConfig before calling CreateEndpoint to minimize the potential impact of a DynamoDB eventually consistent read.
When SageMaker receives the request, it sets the endpoint status to\n Creating
. After it creates the endpoint, it sets the status to\n InService
. SageMaker can then process incoming requests for inferences. To\n check the status of an endpoint, use the DescribeEndpoint\n API.
If any of the models hosted at this endpoint get model data from an Amazon S3 location,\n SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the\n S3 path you provided. Amazon Web Services STS is activated in your Amazon Web Services\n account by default. If you previously deactivated Amazon Web Services STS for a region,\n you need to reactivate Amazon Web Services STS for that region. For more information, see\n Activating and\n Deactivating Amazon Web Services STS in an Amazon Web Services Region in the\n Amazon Web Services Identity and Access Management User\n Guide.
\nTo add the IAM role policies for using this API operation, go to the IAM console, and choose\n Roles in the left navigation pane. Search the IAM role that you want to grant\n access to use the CreateEndpoint and CreateEndpointConfig API operations, add the following policies to\n the role.
\nOption 1: For a full SageMaker access, search and attach the\n AmazonSageMakerFullAccess
policy.
Option 2: For granting a limited access to an IAM role, paste the\n following Action elements manually into the JSON file of the IAM role:
\n\n \"Action\": [\"sagemaker:CreateEndpoint\",\n \"sagemaker:CreateEndpointConfig\"]
\n
\n \"Resource\": [
\n
\n \"arn:aws:sagemaker:region:account-id:endpoint/endpointName\"
\n
\n \"arn:aws:sagemaker:region:account-id:endpoint-config/endpointConfigName\"
\n
\n ]
\n
For more information, see SageMaker API\n Permissions: Actions, Permissions, and Resources\n Reference.
\nCreates an endpoint using the endpoint configuration specified in the request. SageMaker\n uses the endpoint to provision resources and deploy models. You create the endpoint\n configuration with the CreateEndpointConfig API.
\nUse this API to deploy models using SageMaker hosting services.
\nFor an example that calls this method when deploying a model to SageMaker hosting services,\n see the Create Endpoint example notebook.\n
\n You must not delete an EndpointConfig
that is in use by an endpoint\n that is live or while the UpdateEndpoint
or CreateEndpoint
\n operations are being performed on the endpoint. To update an endpoint, you must\n create a new EndpointConfig
.
The endpoint name must be unique within an Amazon Web Services Region in your\n Amazon Web Services account.
\nWhen it receives the request, SageMaker creates the endpoint, launches the resources (ML\n compute instances), and deploys the model(s) on them.
\nWhen you call CreateEndpoint, a load call is made to DynamoDB to verify that your\n endpoint configuration exists. When you read data from a DynamoDB table supporting\n \n Eventually Consistent Reads
\n , the response might not\n reflect the results of a recently completed write operation. The response might\n include some stale data. If the dependent entities are not yet in DynamoDB, this\n causes a validation error. If you repeat your read request after a short time, the\n response should return the latest data. So retry logic is recommended to handle\n these possible issues. We also recommend that customers call DescribeEndpointConfig before calling CreateEndpoint to minimize the potential impact of a DynamoDB\n eventually consistent read.
When SageMaker receives the request, it sets the endpoint status to\n Creating
. After it creates the endpoint, it sets the status to\n InService
. SageMaker can then process incoming requests for inferences. To\n check the status of an endpoint, use the DescribeEndpoint API.
If any of the models hosted at this endpoint get model data from an Amazon S3 location,\n SageMaker uses Amazon Web Services Security Token Service to download model artifacts from the\n S3 path you provided. Amazon Web Services STS is activated in your Amazon Web Services\n account by default. If you previously deactivated Amazon Web Services STS for a region,\n you need to reactivate Amazon Web Services STS for that region. For more information, see\n Activating and\n Deactivating Amazon Web Services STS in an Amazon Web Services Region in the\n Amazon Web Services Identity and Access Management User\n Guide.
\nTo add the IAM role policies for using this API operation, go to the IAM console, and choose\n Roles in the left navigation pane. Search the IAM role that you want to grant\n access to use the CreateEndpoint and CreateEndpointConfig API operations, add the following policies to the\n role.
\nOption 1: For a full SageMaker access, search and attach the\n AmazonSageMakerFullAccess
policy.
Option 2: For granting a limited access to an IAM role, paste the\n following Action elements manually into the JSON file of the IAM role:
\n\n \"Action\": [\"sagemaker:CreateEndpoint\",\n \"sagemaker:CreateEndpointConfig\"]
\n
\n \"Resource\": [
\n
\n \"arn:aws:sagemaker:region:account-id:endpoint/endpointName\"
\n
\n \"arn:aws:sagemaker:region:account-id:endpoint-config/endpointConfigName\"
\n
\n ]
\n
For more information, see SageMaker API\n Permissions: Actions, Permissions, and Resources\n Reference.
\nCreates an endpoint configuration that SageMaker hosting services uses to deploy models. In\n the configuration, you identify one or more models, created using the\n CreateModel
API, to deploy and the resources that you want SageMaker to\n provision. Then you call the CreateEndpoint API.
Use this API if you want to use SageMaker hosting services to deploy models into\n production.
\nIn the request, you define a ProductionVariant
, for each model that you\n want to deploy. Each ProductionVariant
parameter also describes the\n resources that you want SageMaker to provision. This includes the number and type of ML\n compute instances to deploy.
If you are hosting multiple models, you also assign a VariantWeight
to\n specify how much traffic you want to allocate to each model. For example, suppose that\n you want to host two models, A and B, and you assign traffic weight 2 for model A and 1\n for model B. SageMaker distributes two-thirds of the traffic to Model A, and one-third to\n model B.
When you call CreateEndpoint, a load call is made to DynamoDB to\n verify that your endpoint configuration exists. When you read data from a DynamoDB\n table supporting \n Eventually Consistent Reads
\n , the response might not\n reflect the results of a recently completed write operation. The response might\n include some stale data. If the dependent entities are not yet in DynamoDB, this\n causes a validation error. If you repeat your read request after a short time, the\n response should return the latest data. So retry logic is recommended to handle\n these possible issues. We also recommend that customers call DescribeEndpointConfig before calling CreateEndpoint to minimize the potential impact of a DynamoDB eventually consistent read.
Creates an endpoint configuration that SageMaker hosting services uses to deploy models. In\n the configuration, you identify one or more models, created using the\n CreateModel
API, to deploy and the resources that you want SageMaker to\n provision. Then you call the CreateEndpoint\n API.
Use this API if you want to use SageMaker hosting services to deploy models into\n production.
\nIn the request, you define a ProductionVariant
, for each model that you\n want to deploy. Each ProductionVariant
parameter also describes the\n resources that you want SageMaker to provision. This includes the number and type of ML\n compute instances to deploy.
If you are hosting multiple models, you also assign a VariantWeight
to\n specify how much traffic you want to allocate to each model. For example, suppose that\n you want to host two models, A and B, and you assign traffic weight 2 for model A and 1\n for model B. SageMaker distributes two-thirds of the traffic to Model A, and one-third to\n model B.
When you call CreateEndpoint, a load call is made to DynamoDB to verify that your\n endpoint configuration exists. When you read data from a DynamoDB table supporting\n \n Eventually Consistent Reads
\n , the response might not\n reflect the results of a recently completed write operation. The response might\n include some stale data. If the dependent entities are not yet in DynamoDB, this\n causes a validation error. If you repeat your read request after a short time, the\n response should return the latest data. So retry logic is recommended to handle\n these possible issues. We also recommend that customers call DescribeEndpointConfig before calling CreateEndpoint to minimize the potential impact of a DynamoDB\n eventually consistent read.
The name of the endpoint.The name must be unique within an Amazon Web Services\n Region in your Amazon Web Services account. The name is case-insensitive in\n CreateEndpoint
, but the case is preserved and must be matched in InvokeEndpoint.
The name of the endpoint.The name must be unique within an Amazon Web Services\n Region in your Amazon Web Services account. The name is case-insensitive in\n CreateEndpoint
, but the case is preserved and must be matched in InvokeEndpoint.
The HyperParameterTuningJobConfig object that describes the tuning\n job, including the search strategy, the objective metric used to evaluate training jobs,\n ranges of parameters to search, and resource limits for the tuning job. For more\n information, see How\n Hyperparameter Tuning Works.
", + "smithy.api#documentation": "The HyperParameterTuningJobConfig object that describes the tuning job,\n including the search strategy, the objective metric used to evaluate training jobs,\n ranges of parameters to search, and resource limits for the tuning job. For more\n information, see How\n Hyperparameter Tuning Works.
", "smithy.api#required": {} } }, "TrainingJobDefinition": { "target": "com.amazonaws.sagemaker#HyperParameterTrainingJobDefinition", "traits": { - "smithy.api#documentation": "The HyperParameterTrainingJobDefinition object that describes the\n training jobs that this tuning job launches, including static hyperparameters, input\n data configuration, output data configuration, resource configuration, and stopping\n condition.
" + "smithy.api#documentation": "The HyperParameterTrainingJobDefinition object that describes the training jobs\n that this tuning job launches, including static hyperparameters, input data\n configuration, output data configuration, resource configuration, and stopping\n condition.
" } }, "TrainingJobDefinitions": { "target": "com.amazonaws.sagemaker#HyperParameterTrainingJobDefinitions", "traits": { - "smithy.api#documentation": "A list of the HyperParameterTrainingJobDefinition objects launched\n for this tuning job.
" + "smithy.api#documentation": "A list of the HyperParameterTrainingJobDefinition objects launched for this tuning\n job.
" } }, "WarmStartConfig": { @@ -8158,7 +8158,7 @@ "Autotune": { "target": "com.amazonaws.sagemaker#Autotune", "traits": { - "smithy.api#documentation": "Configures SageMaker Automatic model tuning (AMT) to automatically find optimal parameters\n for the following fields:
\n\n ParameterRanges: The names and ranges of parameters that a hyperparameter\n tuning job can optimize.
\n\n ResourceLimits: The maximum resources that can be used for a training job.\n These resources include the maximum number of training jobs, the maximum runtime of a\n tuning job, and the maximum number of training jobs to run at the same time.
\n\n TrainingJobEarlyStoppingType: A flag that specifies whether or not to use\n early stopping for training jobs launched by a hyperparameter tuning job.
\n\n RetryStrategy: The number of times to retry a training job.
\n\n Strategy: Specifies how hyperparameter tuning chooses the combinations of\n hyperparameter values to use for the training jobs that it launches.
\n\n ConvergenceDetected: A flag to indicate that Automatic model tuning (AMT) has detected model\n convergence.
\nConfigures SageMaker Automatic model tuning (AMT) to automatically find optimal\n parameters for the following fields:
\n\n ParameterRanges: The names and ranges of parameters that a\n hyperparameter tuning job can optimize.
\n\n ResourceLimits: The maximum resources that can be used for a\n training job. These resources include the maximum number of training jobs, the\n maximum runtime of a tuning job, and the maximum number of training jobs to run\n at the same time.
\n\n TrainingJobEarlyStoppingType: A flag that specifies whether or not\n to use early stopping for training jobs launched by a hyperparameter tuning\n job.
\n\n RetryStrategy: The number of times to retry a training job.
\n\n Strategy: Specifies how hyperparameter tuning chooses the\n combinations of hyperparameter values to use for the training jobs that it\n launches.
\n\n ConvergenceDetected: A flag to indicate that Automatic model tuning\n (AMT) has detected model convergence.
\nA VpcConfig object that specifies the VPC that you want your model\n to connect to. Control access to and from your model container by configuring the VPC.\n VpcConfig
is used in hosting services and in batch transform. For more\n information, see Protect Endpoints by Using an Amazon Virtual Private Cloud and Protect Data in Batch\n Transform Jobs by Using an Amazon Virtual Private Cloud.
A VpcConfig object that specifies the VPC that you want your model to connect\n to. Control access to and from your model container by configuring the VPC.\n VpcConfig
is used in hosting services and in batch transform. For more\n information, see Protect Endpoints by Using an Amazon Virtual Private Cloud and Protect Data in Batch\n Transform Jobs by Using an Amazon Virtual Private Cloud.
Returns a URL that you can use to connect to the Jupyter server from a notebook\n instance. In the SageMaker console, when you choose Open
next to a notebook\n instance, SageMaker opens a new tab showing the Jupyter server home page from the notebook\n instance. The console uses this API to get the URL and show the page.
The IAM role or user used to call this API defines the permissions to access the\n notebook instance. Once the presigned URL is created, no additional permission is\n required to access this URL. IAM authorization policies for this API are also enforced\n for every HTTP request and WebSocket frame that attempts to connect to the notebook\n instance.
\nYou can restrict access to this API and to the URL that it returns to a list of IP\n addresses that you specify. Use the NotIpAddress
condition operator and the\n aws:SourceIP
condition context key to specify the list of IP addresses\n that you want to have access to the notebook instance. For more information, see Limit Access to a Notebook Instance by IP Address.
The URL that you get from a call to CreatePresignedNotebookInstanceUrl is valid only for 5 minutes. If\n you try to use the URL after the 5-minute limit expires, you are directed to the\n Amazon Web Services console sign-in page.
\nReturns a URL that you can use to connect to the Jupyter server from a notebook\n instance. In the SageMaker console, when you choose Open
next to a notebook\n instance, SageMaker opens a new tab showing the Jupyter server home page from the notebook\n instance. The console uses this API to get the URL and show the page.
The IAM role or user used to call this API defines the permissions to access the\n notebook instance. Once the presigned URL is created, no additional permission is\n required to access this URL. IAM authorization policies for this API are also enforced\n for every HTTP request and WebSocket frame that attempts to connect to the notebook\n instance.
\nYou can restrict access to this API and to the URL that it returns to a list of IP\n addresses that you specify. Use the NotIpAddress
condition operator and the\n aws:SourceIP
condition context key to specify the list of IP addresses\n that you want to have access to the notebook instance. For more information, see Limit Access to a Notebook Instance by IP Address.
The URL that you get from a call to CreatePresignedNotebookInstanceUrl is valid only for 5 minutes. If you\n try to use the URL after the 5-minute limit expires, you are directed to the Amazon Web Services console sign-in page.
\nThe registry path of the Docker image that contains the training algorithm and\n algorithm-specific metadata, including the input mode. For more information about\n algorithms provided by SageMaker, see Algorithms. For information about\n providing your own algorithms, see Using Your Own Algorithms with Amazon SageMaker.\n
", + "smithy.api#documentation": "The registry path of the Docker image that contains the training algorithm and\n algorithm-specific metadata, including the input mode. For more information about\n algorithms provided by SageMaker, see Algorithms. For information about\n providing your own algorithms, see Using Your Own Algorithms with\n Amazon SageMaker.
", "smithy.api#required": {} } }, @@ -10305,7 +10305,7 @@ "InputDataConfig": { "target": "com.amazonaws.sagemaker#InputDataConfig", "traits": { - "smithy.api#documentation": "An array of Channel
objects. Each channel is a named input source.\n InputDataConfig
describes the input data and its location.
Algorithms can accept input data from one or more channels. For example, an\n algorithm might have two channels of input data, training_data
and\n validation_data
. The configuration for each channel provides the S3,\n EFS, or FSx location where the input data is stored. It also provides information about\n the stored data: the MIME type, compression method, and whether the data is wrapped in\n RecordIO format.
Depending on the input mode that the algorithm supports, SageMaker either copies input\n data files from an S3 bucket to a local directory in the Docker container, or makes it\n available as input streams. For example, if you specify an EFS location, input data\n files are available as input streams. They do not need to be downloaded.
\nYour input must be in the same Amazon Web Services region as your training job.
" + "smithy.api#documentation": "An array of Channel
objects. Each channel is a named input source.\n InputDataConfig
describes the input data and its location.
Algorithms can accept input data from one or more channels. For example, an\n algorithm might have two channels of input data, training_data
and\n validation_data
. The configuration for each channel provides the S3,\n EFS, or FSx location where the input data is stored. It also provides information about\n the stored data: the MIME type, compression method, and whether the data is wrapped in\n RecordIO format.
Depending on the input mode that the algorithm supports, SageMaker either copies input\n data files from an S3 bucket to a local directory in the Docker container, or makes it\n available as input streams. For example, if you specify an EFS location, input data\n files are available as input streams. They do not need to be downloaded.
\nYour input must be in the same Amazon Web Services region as your training\n job.
" } }, "OutputDataConfig": { @@ -10325,7 +10325,7 @@ "VpcConfig": { "target": "com.amazonaws.sagemaker#VpcConfig", "traits": { - "smithy.api#documentation": "A VpcConfig object that specifies the VPC that you want your\n training job to connect to. Control access to and from your training container by\n configuring the VPC. For more information, see Protect Training Jobs by Using an Amazon\n Virtual Private Cloud.
" + "smithy.api#documentation": "A VpcConfig object that specifies the VPC that you want your training job to\n connect to. Control access to and from your training container by configuring the VPC.\n For more information, see Protect Training Jobs by Using an Amazon\n Virtual Private Cloud.
" } }, "StoppingCondition": { @@ -15586,7 +15586,7 @@ "KmsKeyId": { "target": "com.amazonaws.sagemaker#KmsKeyId", "traits": { - "smithy.api#documentation": "Amazon Web Services KMS key ID Amazon SageMaker uses to encrypt data when storing it on\n the ML storage volume attached to the instance.
" + "smithy.api#documentation": "Amazon Web Services KMS key ID Amazon SageMaker uses to encrypt data when storing it on the ML\n storage volume attached to the instance.
" } }, "CreationTime": { @@ -15661,7 +15661,7 @@ "ProductionVariants": { "target": "com.amazonaws.sagemaker#ProductionVariantSummaryList", "traits": { - "smithy.api#documentation": "An array of ProductionVariantSummary objects, one for each model\n hosted behind this endpoint.
" + "smithy.api#documentation": "An array of ProductionVariantSummary objects, one for each model hosted behind this\n endpoint.
" } }, "DataCaptureConfig": { @@ -15670,7 +15670,7 @@ "EndpointStatus": { "target": "com.amazonaws.sagemaker#EndpointStatus", "traits": { - "smithy.api#documentation": "The status of the endpoint.
\n\n OutOfService
: Endpoint is not available to take incoming\n requests.
\n Creating
: CreateEndpoint is executing.
\n Updating
: UpdateEndpoint or UpdateEndpointWeightsAndCapacities is executing.
\n SystemUpdating
: Endpoint is undergoing maintenance and cannot be\n updated or deleted or re-scaled until it has completed. This maintenance\n operation does not change any customer-specified values such as VPC config, KMS\n encryption, model, instance type, or instance count.
\n RollingBack
: Endpoint fails to scale up or down or change its\n variant weight and is in the process of rolling back to its previous\n configuration. Once the rollback completes, endpoint returns to an\n InService
status. This transitional status only applies to an\n endpoint that has autoscaling enabled and is undergoing variant weight or\n capacity changes as part of an UpdateEndpointWeightsAndCapacities call or when the UpdateEndpointWeightsAndCapacities operation is called\n explicitly.
\n InService
: Endpoint is available to process incoming\n requests.
\n Deleting
: DeleteEndpoint is executing.
\n Failed
: Endpoint could not be created, updated, or re-scaled. Use the \n FailureReason
value returned by DescribeEndpoint for information about\n the failure. DeleteEndpoint is the only operation that can be\n performed on a failed endpoint.
The status of the endpoint.
\n\n OutOfService
: Endpoint is not available to take incoming\n requests.
\n Creating
: CreateEndpoint is executing.
\n Updating
: UpdateEndpoint or UpdateEndpointWeightsAndCapacities is executing.
\n SystemUpdating
: Endpoint is undergoing maintenance and cannot be\n updated or deleted or re-scaled until it has completed. This maintenance\n operation does not change any customer-specified values such as VPC config, KMS\n encryption, model, instance type, or instance count.
\n RollingBack
: Endpoint fails to scale up or down or change its\n variant weight and is in the process of rolling back to its previous\n configuration. Once the rollback completes, endpoint returns to an\n InService
status. This transitional status only applies to an\n endpoint that has autoscaling enabled and is undergoing variant weight or\n capacity changes as part of an UpdateEndpointWeightsAndCapacities call or when the UpdateEndpointWeightsAndCapacities operation is called\n explicitly.
\n InService
: Endpoint is available to process incoming\n requests.
\n Deleting
: DeleteEndpoint is executing.
\n Failed
: Endpoint could not be created, updated, or re-scaled. Use\n the FailureReason
value returned by DescribeEndpoint for information about the failure. DeleteEndpoint is the only operation that can be performed on a\n failed endpoint.
An array of ProductionVariantSummary objects, one for each model\n that you want to host at this endpoint in shadow mode with production traffic replicated\n from the model specified on ProductionVariants
.
An array of ProductionVariantSummary objects, one for each model that you want to host\n at this endpoint in shadow mode with production traffic replicated from the model\n specified on ProductionVariants
.
The HyperParameterTuningJobConfig object that specifies the\n configuration of the tuning job.
", + "smithy.api#documentation": "The HyperParameterTuningJobConfig object that specifies the configuration of\n the tuning job.
", "smithy.api#required": {} } }, "TrainingJobDefinition": { "target": "com.amazonaws.sagemaker#HyperParameterTrainingJobDefinition", "traits": { - "smithy.api#documentation": "The HyperParameterTrainingJobDefinition object that specifies the\n definition of the training jobs that this tuning job launches.
" + "smithy.api#documentation": "The HyperParameterTrainingJobDefinition object that specifies the definition of\n the training jobs that this tuning job launches.
" } }, "TrainingJobDefinitions": { "target": "com.amazonaws.sagemaker#HyperParameterTrainingJobDefinitions", "traits": { - "smithy.api#documentation": "A list of the HyperParameterTrainingJobDefinition objects launched\n for this tuning job.
" + "smithy.api#documentation": "A list of the HyperParameterTrainingJobDefinition objects launched for this tuning\n job.
" } }, "HyperParameterTuningJobStatus": { @@ -16627,27 +16627,27 @@ "TrainingJobStatusCounters": { "target": "com.amazonaws.sagemaker#TrainingJobStatusCounters", "traits": { - "smithy.api#documentation": "The TrainingJobStatusCounters object that specifies the number of\n training jobs, categorized by status, that this tuning job launched.
", + "smithy.api#documentation": "The TrainingJobStatusCounters object that specifies the number of training\n jobs, categorized by status, that this tuning job launched.
", "smithy.api#required": {} } }, "ObjectiveStatusCounters": { "target": "com.amazonaws.sagemaker#ObjectiveStatusCounters", "traits": { - "smithy.api#documentation": "The ObjectiveStatusCounters object that specifies the number of\n training jobs, categorized by the status of their final objective metric, that this\n tuning job launched.
", + "smithy.api#documentation": "The ObjectiveStatusCounters object that specifies the number of training jobs,\n categorized by the status of their final objective metric, that this tuning job\n launched.
", "smithy.api#required": {} } }, "BestTrainingJob": { "target": "com.amazonaws.sagemaker#HyperParameterTrainingJobSummary", "traits": { - "smithy.api#documentation": "A TrainingJobSummary object that describes the training job that\n completed with the best current HyperParameterTuningJobObjective.
" + "smithy.api#documentation": "A TrainingJobSummary object that describes the training job that completed\n with the best current HyperParameterTuningJobObjective.
" } }, "OverallBestTrainingJob": { "target": "com.amazonaws.sagemaker#HyperParameterTrainingJobSummary", "traits": { - "smithy.api#documentation": "If the hyperparameter tuning job is an warm start tuning job with a\n WarmStartType
of IDENTICAL_DATA_AND_ALGORITHM
, this is the\n TrainingJobSummary for the training job with the best objective\n metric value of all training jobs launched by this tuning job and all parent jobs\n specified for the warm start tuning job.
If the hyperparameter tuning job is an warm start tuning job with a\n WarmStartType
of IDENTICAL_DATA_AND_ALGORITHM
, this is the\n TrainingJobSummary for the training job with the best objective metric\n value of all training jobs launched by this tuning job and all parent jobs specified for\n the warm start tuning job.
A VpcConfig object that specifies the VPC that this model has access\n to. For more information, see Protect Endpoints by Using an Amazon Virtual\n Private Cloud\n
" + "smithy.api#documentation": "A VpcConfig object that specifies the VPC that this model has access to. For\n more information, see Protect Endpoints by Using an Amazon Virtual\n Private Cloud\n
" } }, "CreationTime": { @@ -18889,7 +18889,7 @@ "AcceleratorTypes": { "target": "com.amazonaws.sagemaker#NotebookInstanceAcceleratorTypes", "traits": { - "smithy.api#documentation": "A list of the Elastic Inference (EI) instance types associated with this notebook\n instance. Currently only one EI instance type can be associated with a notebook\n instance. For more information, see Using Elastic Inference in Amazon SageMaker.
" + "smithy.api#documentation": "A list of the Elastic Inference (EI) instance types associated with this notebook\n instance. Currently only one EI instance type can be associated with a notebook\n instance. For more information, see Using Elastic Inference in\n Amazon SageMaker.
" } }, "DefaultCodeRepository": { @@ -19111,7 +19111,7 @@ "PipelineName": { "target": "com.amazonaws.sagemaker#PipelineNameOrArn", "traits": { - "smithy.api#documentation": "The name of the pipeline to describe.
", + "smithy.api#documentation": "The name or Amazon Resource Name (ARN) of the pipeline to describe.
", "smithy.api#required": {} } } @@ -19919,7 +19919,7 @@ "VpcConfig": { "target": "com.amazonaws.sagemaker#VpcConfig", "traits": { - "smithy.api#documentation": "A VpcConfig object that specifies the VPC that this training job has\n access to. For more information, see Protect Training Jobs by Using an Amazon\n Virtual Private Cloud.
" + "smithy.api#documentation": "A VpcConfig object that specifies the VPC that this training job has access\n to. For more information, see Protect Training Jobs by Using an Amazon\n Virtual Private Cloud.
" } }, "StoppingCondition": { @@ -21556,7 +21556,7 @@ "RStudioServerProDomainSettingsForUpdate": { "target": "com.amazonaws.sagemaker#RStudioServerProDomainSettingsForUpdate", "traits": { - "smithy.api#documentation": "A collection of RStudioServerPro
Domain-level app settings to\n update. A single RStudioServerPro
application is created for a domain.
A collection of RStudioServerPro
Domain-level app settings to update. A\n single RStudioServerPro
application is created for a domain.
The status of the endpoint.
\n\n OutOfService
: Endpoint is not available to take incoming\n requests.
\n Creating
: CreateEndpoint is executing.
\n Updating
: UpdateEndpoint or UpdateEndpointWeightsAndCapacities is executing.
\n SystemUpdating
: Endpoint is undergoing maintenance and cannot be\n updated or deleted or re-scaled until it has completed. This maintenance\n operation does not change any customer-specified values such as VPC config, KMS\n encryption, model, instance type, or instance count.
\n RollingBack
: Endpoint fails to scale up or down or change its\n variant weight and is in the process of rolling back to its previous\n configuration. Once the rollback completes, endpoint returns to an\n InService
status. This transitional status only applies to an\n endpoint that has autoscaling enabled and is undergoing variant weight or\n capacity changes as part of an UpdateEndpointWeightsAndCapacities call or when the UpdateEndpointWeightsAndCapacities operation is called\n explicitly.
\n InService
: Endpoint is available to process incoming\n requests.
\n Deleting
: DeleteEndpoint is executing.
\n Failed
: Endpoint could not be created, updated, or re-scaled. Use\n DescribeEndpointOutput$FailureReason
for information about\n the failure. DeleteEndpoint is the only operation that can be\n performed on a failed endpoint.
To get a list of endpoints with a specified status, use the StatusEquals
\n filter with a call to ListEndpoints.
The status of the endpoint.
\n\n OutOfService
: Endpoint is not available to take incoming\n requests.
\n Creating
: CreateEndpoint is executing.
\n Updating
: UpdateEndpoint or UpdateEndpointWeightsAndCapacities is executing.
\n SystemUpdating
: Endpoint is undergoing maintenance and cannot be\n updated or deleted or re-scaled until it has completed. This maintenance\n operation does not change any customer-specified values such as VPC config, KMS\n encryption, model, instance type, or instance count.
\n RollingBack
: Endpoint fails to scale up or down or change its\n variant weight and is in the process of rolling back to its previous\n configuration. Once the rollback completes, endpoint returns to an\n InService
status. This transitional status only applies to an\n endpoint that has autoscaling enabled and is undergoing variant weight or\n capacity changes as part of an UpdateEndpointWeightsAndCapacities call or when the UpdateEndpointWeightsAndCapacities operation is called\n explicitly.
\n InService
: Endpoint is available to process incoming\n requests.
\n Deleting
: DeleteEndpoint is executing.
\n Failed
: Endpoint could not be created, updated, or re-scaled. Use\n DescribeEndpointOutput$FailureReason
for information about the\n failure. DeleteEndpoint is the only operation that can be performed on a\n failed endpoint.
To get a list of endpoints with a specified status, use the StatusEquals
\n filter with a call to ListEndpoints.
Shows the latest objective metric emitted by a training job that was launched by a\n hyperparameter tuning job. You define the objective metric in the\n HyperParameterTuningJobObjective
parameter of HyperParameterTuningJobConfig.
Shows the latest objective metric emitted by a training job that was launched by a\n hyperparameter tuning job. You define the objective metric in the\n HyperParameterTuningJobObjective
parameter of HyperParameterTuningJobConfig.
The registry path of the Docker image that contains the training algorithm. For\n information about Docker registry paths for built-in algorithms, see Algorithms\n Provided by Amazon SageMaker: Common Parameters. SageMaker supports both\n registry/repository[:tag]
and registry/repository[@digest]
\n image path formats. For more information, see Using Your Own Algorithms with\n Amazon SageMaker.
The registry path of the Docker image that contains the training algorithm. For\n information about Docker registry paths for built-in algorithms, see Algorithms\n Provided by Amazon SageMaker: Common Parameters. SageMaker supports both\n registry/repository[:tag]
and registry/repository[@digest]
\n image path formats. For more information, see Using Your Own Algorithms with\n Amazon SageMaker.
An array of Channel objects that specify\n the\n input for the training jobs that the tuning job launches.
" + "smithy.api#documentation": "An array of Channel objects that\n specify\n the\n input for the training jobs that the tuning job launches.
" } }, "VpcConfig": { "target": "com.amazonaws.sagemaker#VpcConfig", "traits": { - "smithy.api#documentation": "The VpcConfig object that specifies the VPC that you want the\n training jobs that this hyperparameter tuning job launches to connect to. Control access\n to and from your training container by configuring the VPC. For more information, see\n Protect\n Training Jobs by Using an Amazon Virtual Private Cloud.
" + "smithy.api#documentation": "The VpcConfig object that specifies the VPC that you want the training jobs\n that this hyperparameter tuning job launches to connect to. Control access to and from\n your training container by configuring the VPC. For more information, see Protect Training Jobs\n by Using an Amazon Virtual Private Cloud.
" } }, "OutputDataConfig": { @@ -25799,7 +25799,7 @@ "Environment": { "target": "com.amazonaws.sagemaker#HyperParameterTrainingJobEnvironmentMap", "traits": { - "smithy.api#documentation": "An environment variable that you can pass into the SageMaker CreateTrainingJob API. You can use an existing environment variable from the training container or use your own. See\n Define metrics\n and variables for more information.
\nThe maximum number of items specified for Map Entries
refers to the\n maximum number of environment variables for each TrainingJobDefinition
\n and also the maximum for the hyperparameter tuning job itself. That is, the sum of\n the number of environment variables for all the training job definitions can't\n exceed the maximum number specified.
An environment variable that you can pass into the SageMaker CreateTrainingJob API. You can use an existing environment variable from the training container or use your own. See\n Define metrics and variables for more information.
\nThe maximum number of items specified for Map Entries
refers to the\n maximum number of environment variables for each TrainingJobDefinition
\n and also the maximum for the hyperparameter tuning job itself. That is, the sum of\n the number of environment variables for all the training job definitions can't\n exceed the maximum number specified.
The FinalHyperParameterTuningJobObjectiveMetric object that\n specifies the\n value\n of the\n objective\n metric of the tuning job that launched this training job.
" + "smithy.api#documentation": "The FinalHyperParameterTuningJobObjectiveMetric object that specifies the\n value\n of the\n objective\n metric of the tuning job that launched this training job.
" } }, "ObjectiveStatus": { @@ -25980,7 +25980,7 @@ "target": "com.amazonaws.sagemaker#TrainingInstanceCount", "traits": { "smithy.api#default": 0, - "smithy.api#documentation": "The number of instances of the type specified by InstanceType
. Choose an\n instance count larger than 1 for distributed training algorithms. See SageMaker distributed training\n jobs for more information.
The number of instances of the type specified by InstanceType
. Choose an\n instance count larger than 1 for distributed training algorithms. See Step 2:\n Launch a SageMaker Distributed Training Job Using the SageMaker Python SDK for more information.
The HyperParameterTuningJobObjective specifies the objective metric\n used to evaluate the performance of training jobs launched by this tuning job.
" + "smithy.api#documentation": "The HyperParameterTuningJobObjective specifies the objective metric used to\n evaluate the performance of training jobs launched by this tuning job.
" } }, "ResourceLimits": { "target": "com.amazonaws.sagemaker#ResourceLimits", "traits": { - "smithy.api#documentation": "The ResourceLimits object that specifies the maximum number of\n training and parallel training jobs that can be used for this hyperparameter tuning\n job.
", + "smithy.api#documentation": "The ResourceLimits\n object that specifies the maximum number of training and parallel training jobs that can\n be used for this hyperparameter tuning job.
", "smithy.api#required": {} } }, "ParameterRanges": { "target": "com.amazonaws.sagemaker#ParameterRanges", "traits": { - "smithy.api#documentation": "The ParameterRanges object that specifies the ranges of\n hyperparameters that this tuning job searches over to find the optimal configuration for\n the highest model performance against your chosen objective metric.
" + "smithy.api#documentation": "The ParameterRanges\n object that specifies the ranges of hyperparameters that this tuning job searches over\n to find the optimal configuration for the highest model performance against your chosen\n objective metric.
" } }, "TrainingJobEarlyStoppingType": { @@ -26426,21 +26426,21 @@ "TrainingJobStatusCounters": { "target": "com.amazonaws.sagemaker#TrainingJobStatusCounters", "traits": { - "smithy.api#documentation": "The TrainingJobStatusCounters object that specifies the numbers of\n training jobs, categorized by status, that this tuning job launched.
", + "smithy.api#documentation": "The TrainingJobStatusCounters object that specifies the numbers of training\n jobs, categorized by status, that this tuning job launched.
", "smithy.api#required": {} } }, "ObjectiveStatusCounters": { "target": "com.amazonaws.sagemaker#ObjectiveStatusCounters", "traits": { - "smithy.api#documentation": "The ObjectiveStatusCounters object that specifies the numbers of\n training jobs, categorized by objective metric status, that this tuning job\n launched.
", + "smithy.api#documentation": "The ObjectiveStatusCounters object that specifies the numbers of training jobs,\n categorized by objective metric status, that this tuning job launched.
", "smithy.api#required": {} } }, "ResourceLimits": { "target": "com.amazonaws.sagemaker#ResourceLimits", "traits": { - "smithy.api#documentation": "The ResourceLimits object that specifies the maximum number of\n training jobs and parallel training jobs allowed for this tuning job.
" + "smithy.api#documentation": "The ResourceLimits\n object that specifies the maximum number of training jobs and parallel training jobs\n allowed for this tuning job.
" } } }, @@ -26502,7 +26502,7 @@ "InstanceType": { "target": "com.amazonaws.sagemaker#TrainingInstanceType", "traits": { - "smithy.api#documentation": "The instance type used to run hyperparameter optimization tuning jobs. See descriptions of\n instance types for more information.
" + "smithy.api#documentation": "The instance type used to run hyperparameter optimization tuning jobs. See descriptions of\n instance types for more information.
" } }, "InstanceCount": { @@ -29755,7 +29755,7 @@ "MaxResults": { "target": "com.amazonaws.sagemaker#MaxResults", "traits": { - "smithy.api#documentation": "The maximum number of AppImageConfigs to return in the response. The default value is\n 10.
" + "smithy.api#documentation": "The total number of items to return in the response. If the total\n number of items available is more than the value specified, a NextToken
\n is provided in the response. To resume pagination, provide the NextToken
\n value in the as part of a subsequent call. The default value is 10.
Returns a list up to a specified limit.
" + "smithy.api#documentation": "The total number of items to return in the response. If the total\n number of items available is more than the value specified, a NextToken
\n is provided in the response. To resume pagination, provide the NextToken
\n value in the as part of a subsequent call. The default value is 10.
Returns a list up to a specified limit.
" + "smithy.api#documentation": "The total number of items to return in the response. If the total\n number of items available is more than the value specified, a NextToken
\n is provided in the response. To resume pagination, provide the NextToken
\n value in the as part of a subsequent call. The default value is 10.
The name of the pipeline.
", + "smithy.api#documentation": "The name or Amazon Resource Name (ARN) of the pipeline.
", "smithy.api#required": {} } }, @@ -35274,7 +35274,7 @@ "MaxResults": { "target": "com.amazonaws.sagemaker#MaxResults", "traits": { - "smithy.api#documentation": "Returns a list up to a specified limit.
" + "smithy.api#documentation": "The total number of items to return in the response. If the total\n number of items available is more than the value specified, a NextToken
\n is provided in the response. To resume pagination, provide the NextToken
\n value in the as part of a subsequent call. The default value is 10.
The maximum number of Studio Lifecycle Configurations to return in the response. The default value is 10.
" + "smithy.api#documentation": "The total number of items to return in the response. If the total\n number of items available is more than the value specified, a NextToken
\n is provided in the response. To resume pagination, provide the NextToken
\n value in the as part of a subsequent call. The default value is 10.
A token for getting the next set of actions, if there are any.
" + "smithy.api#documentation": "If the previous response was truncated, you will receive this token.\n Use it in your next request to receive the next set of results.
" } }, "StudioLifecycleConfigs": { @@ -35688,7 +35688,7 @@ } ], "traits": { - "smithy.api#documentation": "Gets a list of TrainingJobSummary objects that describe the training\n jobs that a hyperparameter tuning job launched.
", + "smithy.api#documentation": "Gets a list of TrainingJobSummary objects that describe the training jobs that a\n hyperparameter tuning job launched.
", "smithy.api#paginated": { "inputToken": "NextToken", "outputToken": "NextToken", @@ -36205,7 +36205,7 @@ "MaxResults": { "target": "com.amazonaws.sagemaker#MaxResults", "traits": { - "smithy.api#documentation": "Returns a list up to a specified limit.
" + "smithy.api#documentation": "The total number of items to return in the response. If the total\n number of items available is more than the value specified, a NextToken
\n is provided in the response. To resume pagination, provide the NextToken
\n value in the as part of a subsequent call. The default value is 10.
A regular expression that searches the output of a training job and gets the value of\n the metric. For more information about using regular expressions to define metrics, see\n Defining\n metrics and environment variables.
", + "smithy.api#documentation": "A regular expression that searches the output of a training job and gets the value of\n the metric. For more information about using regular expressions to define metrics, see\n Defining metrics and environment variables.
", "smithy.api#required": {} } } @@ -41742,25 +41742,25 @@ "IntegerParameterRanges": { "target": "com.amazonaws.sagemaker#IntegerParameterRanges", "traits": { - "smithy.api#documentation": "The array of IntegerParameterRange objects that specify ranges of\n integer hyperparameters that a hyperparameter tuning job searches.
" + "smithy.api#documentation": "The array of IntegerParameterRange objects that specify ranges of integer\n hyperparameters that a hyperparameter tuning job searches.
" } }, "ContinuousParameterRanges": { "target": "com.amazonaws.sagemaker#ContinuousParameterRanges", "traits": { - "smithy.api#documentation": "The array of ContinuousParameterRange objects that specify ranges of\n continuous hyperparameters that a hyperparameter tuning job searches.
" + "smithy.api#documentation": "The array of ContinuousParameterRange objects that specify ranges of continuous\n hyperparameters that a hyperparameter tuning job searches.
" } }, "CategoricalParameterRanges": { "target": "com.amazonaws.sagemaker#CategoricalParameterRanges", "traits": { - "smithy.api#documentation": "The array of CategoricalParameterRange objects that specify ranges\n of categorical hyperparameters that a hyperparameter tuning job searches.
" + "smithy.api#documentation": "The array of CategoricalParameterRange objects that specify ranges of categorical\n hyperparameters that a hyperparameter tuning job searches.
" } }, "AutoParameters": { "target": "com.amazonaws.sagemaker#AutoParameters", "traits": { - "smithy.api#documentation": "A list containing hyperparameter names and example values to be used by Autotune to determine optimal ranges for your tuning job.
" + "smithy.api#documentation": "A list containing hyperparameter names and example values to be used by Autotune to\n determine optimal ranges for your tuning job.
" } } }, @@ -41884,7 +41884,7 @@ "ProductionVariants": { "target": "com.amazonaws.sagemaker#PendingProductionVariantSummaryList", "traits": { - "smithy.api#documentation": "An array of PendingProductionVariantSummary objects, one for each\n model hosted behind this endpoint for the in-progress deployment.
" + "smithy.api#documentation": "An array of PendingProductionVariantSummary objects, one for each model hosted behind\n this endpoint for the in-progress deployment.
" } }, "StartTime": { @@ -41896,7 +41896,7 @@ "ShadowProductionVariants": { "target": "com.amazonaws.sagemaker#PendingProductionVariantSummaryList", "traits": { - "smithy.api#documentation": "An array of PendingProductionVariantSummary objects, one for each\n model hosted behind this endpoint in shadow mode with production traffic replicated from\n the model specified on ProductionVariants
for the in-progress\n deployment.
An array of PendingProductionVariantSummary objects, one for each model hosted behind\n this endpoint in shadow mode with production traffic replicated from the model specified\n on ProductionVariants
for the in-progress deployment.
The production variant summary for a deployment when an endpoint is creating or\n updating with the CreateEndpoint or UpdateEndpoint operations. Describes the VariantStatus\n
, weight and capacity for a production variant associated with an endpoint.\n
The production variant summary for a deployment when an endpoint is creating or\n updating with the CreateEndpoint\n or UpdateEndpoint\n operations. Describes the VariantStatus
, weight and capacity for a\n production variant associated with an endpoint.
The size of the Elastic Inference (EI) instance to use for the production variant. EI\n instances provide on-demand GPU computing for inference. For more information, see\n Using Elastic\n Inference in Amazon SageMaker.
" + "smithy.api#documentation": "The size of the Elastic Inference (EI) instance to use for the production variant. EI\n instances provide on-demand GPU computing for inference. For more information, see\n Using Elastic\n Inference in Amazon SageMaker.
" } }, "CoreDumpConfig": { @@ -43805,12 +43805,12 @@ "EnableSSMAccess": { "target": "com.amazonaws.sagemaker#ProductionVariantSSMAccess", "traits": { - "smithy.api#documentation": "\n You can use this parameter to turn on native Amazon Web Services Systems Manager (SSM)\n access for a production variant behind an endpoint. By default, SSM access is disabled\n for all production variants behind an endpoint. You can turn on or turn off SSM access\n for a production variant behind an existing endpoint by creating a new endpoint\n configuration and calling UpdateEndpoint
.\n
You can use this parameter to turn on native Amazon Web Services Systems Manager (SSM)\n access for a production variant behind an endpoint. By default, SSM access is disabled\n for all production variants behind an endpoint. You can turn on or turn off SSM access\n for a production variant behind an existing endpoint by creating a new endpoint\n configuration and calling UpdateEndpoint
.
\n Identifies a model that you want to host and the resources chosen to deploy for\n hosting it. If you are deploying multiple models, tell SageMaker how to distribute traffic\n among the models by specifying variant weights. For more information on production\n variants, check \n Production variants.\n
" + "smithy.api#documentation": "Identifies a model that you want to host and the resources chosen to deploy for\n hosting it. If you are deploying multiple models, tell SageMaker how to distribute traffic\n among the models by specifying variant weights. For more information on production\n variants, check Production variants.\n
" } }, "com.amazonaws.sagemaker#ProductionVariantAcceleratorType": { @@ -47628,7 +47628,7 @@ "S3Uri": { "target": "com.amazonaws.sagemaker#S3Uri", "traits": { - "smithy.api#documentation": "Depending on the value specified for the S3DataType
, identifies either\n a key name prefix or a manifest. For example:
A key name prefix might look like this:\n s3://bucketname/exampleprefix
\n
A manifest might look like this:\n s3://bucketname/example.manifest
\n
A manifest is an S3 object which is a JSON file consisting of an array of\n elements. The first element is a prefix which is followed by one or more\n suffixes. SageMaker appends the suffix elements to the prefix to get a full set\n of S3Uri
. Note that the prefix must be a valid non-empty\n S3Uri
that precludes users from specifying a manifest whose\n individual S3Uri
is sourced from different S3 buckets.
The following code example shows a valid manifest format:
\n\n [ {\"prefix\": \"s3://customer_bucket/some/prefix/\"},
\n
\n \"relative/path/to/custdata-1\",
\n
\n \"relative/path/custdata-2\",
\n
\n ...
\n
\n \"relative/path/custdata-N\"
\n
\n ]
\n
This JSON is equivalent to the following S3Uri
\n list:
\n s3://customer_bucket/some/prefix/relative/path/to/custdata-1
\n
\n s3://customer_bucket/some/prefix/relative/path/custdata-2
\n
\n ...
\n
\n s3://customer_bucket/some/prefix/relative/path/custdata-N
\n
The complete set of S3Uri
in this manifest is the input data\n for the channel for this data source. The object that each S3Uri
\n points to must be readable by the IAM role that SageMaker uses to perform tasks on\n your behalf.
Your input bucket must be located in same Amazon Web Services region as your training job.
", + "smithy.api#documentation": "Depending on the value specified for the S3DataType
, identifies either\n a key name prefix or a manifest. For example:
A key name prefix might look like this:\n s3://bucketname/exampleprefix
\n
A manifest might look like this:\n s3://bucketname/example.manifest
\n
A manifest is an S3 object which is a JSON file consisting of an array of\n elements. The first element is a prefix which is followed by one or more\n suffixes. SageMaker appends the suffix elements to the prefix to get a full set of\n S3Uri
. Note that the prefix must be a valid non-empty\n S3Uri
that precludes users from specifying a manifest whose\n individual S3Uri
is sourced from different S3 buckets.
The following code example shows a valid manifest format:
\n\n [ {\"prefix\": \"s3://customer_bucket/some/prefix/\"},
\n
\n \"relative/path/to/custdata-1\",
\n
\n \"relative/path/custdata-2\",
\n
\n ...
\n
\n \"relative/path/custdata-N\"
\n
\n ]
\n
This JSON is equivalent to the following S3Uri
\n list:
\n s3://customer_bucket/some/prefix/relative/path/to/custdata-1
\n
\n s3://customer_bucket/some/prefix/relative/path/custdata-2
\n
\n ...
\n
\n s3://customer_bucket/some/prefix/relative/path/custdata-N
\n
The complete set of S3Uri
in this manifest is the input data\n for the channel for this data source. The object that each S3Uri
\n points to must be readable by the IAM role that SageMaker uses to perform tasks on\n your behalf.
Your input bucket must be located in same Amazon Web Services region as your\n training job.
", "smithy.api#required": {} } }, @@ -47652,7 +47652,7 @@ } }, "traits": { - "smithy.api#documentation": "Describes the S3 data source.
\nYour input bucket must be in the same Amazon Web Services region as your training job.
" + "smithy.api#documentation": "Describes the S3 data source.
\nYour input bucket must be in the same Amazon Web Services region as your training\n job.
" } }, "com.amazonaws.sagemaker#S3DataType": { @@ -50166,12 +50166,12 @@ "StatusMessage": { "target": "com.amazonaws.sagemaker#StatusMessage", "traits": { - "smithy.api#documentation": "A detailed description of the progress within a secondary status.\n
\nSageMaker provides secondary statuses and status messages that apply to each of\n them:
\nStarting the training job.
\nLaunching requested ML\n instances.
\nInsufficient\n capacity error from EC2 while launching instances,\n retrying!
\nLaunched\n instance was unhealthy, replacing it!
\nPreparing the instances for training.
\nDownloading the training image.
\nTraining\n image download completed. Training in\n progress.
\nStatus messages are subject to change. Therefore, we recommend not including them\n in code that programmatically initiates actions. For examples, don't use status\n messages in if statements.
\nTo have an overview of your training job's progress, view\n TrainingJobStatus
and SecondaryStatus
in DescribeTrainingJob, and StatusMessage
together. For\n example, at the start of a training job, you might see the following:
\n TrainingJobStatus
- InProgress
\n SecondaryStatus
- Training
\n StatusMessage
- Downloading the training image
A detailed description of the progress within a secondary status.\n
\nSageMaker provides secondary statuses and status messages that apply to each of\n them:
\nStarting the training job.
\nLaunching requested ML\n instances.
\nInsufficient\n capacity error from EC2 while launching instances,\n retrying!
\nLaunched\n instance was unhealthy, replacing it!
\nPreparing the instances for training.
\nDownloading the training image.
\nTraining\n image download completed. Training in\n progress.
\nStatus messages are subject to change. Therefore, we recommend not including them\n in code that programmatically initiates actions. For examples, don't use status\n messages in if statements.
\nTo have an overview of your training job's progress, view\n TrainingJobStatus
and SecondaryStatus
in DescribeTrainingJob, and StatusMessage
together. For example,\n at the start of a training job, you might see the following:
\n TrainingJobStatus
- InProgress
\n SecondaryStatus
- Training
\n StatusMessage
- Downloading the training image
An array element of SecondaryStatusTransitions
for DescribeTrainingJob. It provides\n additional details about a status that the training job has transitioned through. A\n training job can be in one of several states, for example, starting, downloading,\n training, or uploading. Within each state, there are a number of intermediate states.\n For example, within the starting state, SageMaker could be starting the training job or\n launching the ML instances. These transitional states are referred to as the job's\n secondary\n status.\n
An array element of SecondaryStatusTransitions
for DescribeTrainingJob. It provides additional details about a status that the\n training job has transitioned through. A training job can be in one of several states,\n for example, starting, downloading, training, or uploading. Within each state, there are\n a number of intermediate states. For example, within the starting state, SageMaker could be\n starting the training job or launching the ML instances. These transitional states are\n referred to as the job's secondary\n status.\n
The Amazon S3 path where the model artifacts, which result from model training, are stored.\n This path must point to a single gzip
compressed tar archive\n (.tar.gz
suffix).
The model artifacts must be in an S3 bucket that is in the same Amazon Web Services region as the\n algorithm.
\nThe Amazon S3 path where the model artifacts, which result from model training, are stored.\n This path must point to a single gzip
compressed tar archive\n (.tar.gz
suffix).
The model artifacts must be in an S3 bucket that is in the same Amazon Web Services\n region as the algorithm.
\nThe name of the pipeline.
", + "smithy.api#documentation": "The name or Amazon Resource Name (ARN) of the pipeline.
", "smithy.api#required": {} } }, @@ -52507,6 +52507,18 @@ "smithy.api#enumValue": "ml_inf1" } }, + "ML_INF2": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ml_inf2" + } + }, + "ML_TRN1": { + "target": "smithy.api#Unit", + "traits": { + "smithy.api#enumValue": "ml_trn1" + } + }, "ML_EIA2": { "target": "smithy.api#Unit", "traits": { @@ -53569,7 +53581,7 @@ "SecondaryStatus": { "target": "com.amazonaws.sagemaker#SecondaryStatus", "traits": { - "smithy.api#documentation": " Provides detailed information about the state of the training job. For detailed\n information about the secondary status of the training job, see\n StatusMessage
under SecondaryStatusTransition.
SageMaker provides primary statuses and secondary statuses that apply to each of\n them:
\n\n Starting
\n - Starting the training job.
\n Downloading
- An optional stage for algorithms that\n support File
training input mode. It indicates that\n data is being downloaded to the ML storage volumes.
\n Training
- Training is in progress.
\n Uploading
- Training is complete and the model\n artifacts are being uploaded to the S3 location.
\n Completed
- The training job has completed.
\n Failed
- The training job has failed. The reason for\n the failure is returned in the FailureReason
field of\n DescribeTrainingJobResponse
.
\n MaxRuntimeExceeded
- The job stopped because it\n exceeded the maximum allowed runtime.
\n Stopped
- The training job has stopped.
\n Stopping
- Stopping the training job.
Valid values for SecondaryStatus
are subject to change.
We no longer support the following secondary statuses:
\n\n LaunchingMLInstances
\n
\n PreparingTrainingStack
\n
\n DownloadingTrainingImage
\n
Provides detailed information about the state of the training job. For detailed\n information about the secondary status of the training job, see\n StatusMessage
under SecondaryStatusTransition.
SageMaker provides primary statuses and secondary statuses that apply to each of\n them:
\n\n Starting
\n - Starting the training job.
\n Downloading
- An optional stage for algorithms that\n support File
training input mode. It indicates that\n data is being downloaded to the ML storage volumes.
\n Training
- Training is in progress.
\n Uploading
- Training is complete and the model\n artifacts are being uploaded to the S3 location.
\n Completed
- The training job has completed.
\n Failed
- The training job has failed. The reason for\n the failure is returned in the FailureReason
field of\n DescribeTrainingJobResponse
.
\n MaxRuntimeExceeded
- The job stopped because it\n exceeded the maximum allowed runtime.
\n Stopped
- The training job has stopped.
\n Stopping
- Stopping the training job.
Valid values for SecondaryStatus
are subject to change.
We no longer support the following secondary statuses:
\n\n LaunchingMLInstances
\n
\n PreparingTrainingStack
\n
\n DownloadingTrainingImage
\n
A VpcConfig object that specifies the VPC that this training job has\n access to. For more information, see Protect Training Jobs by Using an Amazon\n Virtual Private Cloud.
" + "smithy.api#documentation": "A VpcConfig object that specifies the VPC that this training job has access\n to. For more information, see Protect Training Jobs by Using an Amazon\n Virtual Private Cloud.
" } }, "StoppingCondition": { @@ -54589,7 +54601,7 @@ "TransformOutput": { "target": "com.amazonaws.sagemaker#TransformOutput", "traits": { - "smithy.api#documentation": "Identifies the Amazon S3 location where you want Amazon SageMaker to save the results\n from the transform job.
", + "smithy.api#documentation": "Identifies the Amazon S3 location where you want Amazon SageMaker to save the results from the\n transform job.
", "smithy.api#required": {} } }, @@ -56131,7 +56143,7 @@ } ], "traits": { - "smithy.api#documentation": "Deploys the new EndpointConfig
specified in the request, switches to\n using newly created endpoint, and then deletes resources provisioned for the endpoint\n using the previous EndpointConfig
(there is no availability loss).
When SageMaker receives the request, it sets the endpoint status to\n Updating
. After updating the endpoint, it sets the status to\n InService
. To check the status of an endpoint, use the DescribeEndpoint API.\n \n
You must not delete an EndpointConfig
in use by an endpoint that is\n live or while the UpdateEndpoint
or CreateEndpoint
\n operations are being performed on the endpoint. To update an endpoint, you must\n create a new EndpointConfig
.
If you delete the EndpointConfig
of an endpoint that is active or\n being created or updated you may lose visibility into the instance type the endpoint\n is using. The endpoint must be deleted in order to stop incurring charges.
Deploys the new EndpointConfig
specified in the request, switches to\n using newly created endpoint, and then deletes resources provisioned for the endpoint\n using the previous EndpointConfig
(there is no availability loss).
When SageMaker receives the request, it sets the endpoint status to\n Updating
. After updating the endpoint, it sets the status to\n InService
. To check the status of an endpoint, use the DescribeEndpoint API.\n \n
You must not delete an EndpointConfig
in use by an endpoint that is\n live or while the UpdateEndpoint
or CreateEndpoint
\n operations are being performed on the endpoint. To update an endpoint, you must\n create a new EndpointConfig
.
If you delete the EndpointConfig
of an endpoint that is active or\n being created or updated you may lose visibility into the instance type the endpoint\n is using. The endpoint must be deleted in order to stop incurring charges.
When you are updating endpoint resources with RetainAllVariantProperties
, whose value is set to\n true
, ExcludeRetainedVariantProperties
specifies the list\n of type VariantProperty to override with the values provided by\n EndpointConfig
. If you don't specify a value for\n ExcludeRetainedVariantProperties
, no variant properties are overridden.\n
When you are updating endpoint resources with RetainAllVariantProperties
,\n whose value is set to true
, ExcludeRetainedVariantProperties
\n specifies the list of type VariantProperty\n to override with the values provided by EndpointConfig
. If you don't\n specify a value for ExcludeRetainedVariantProperties
, no variant properties\n are overridden.
A list of the Elastic Inference (EI) instance types to associate with this notebook\n instance. Currently only one EI instance type can be associated with a notebook\n instance. For more information, see Using Elastic Inference in Amazon SageMaker.
" + "smithy.api#documentation": "A list of the Elastic Inference (EI) instance types to associate with this notebook\n instance. Currently only one EI instance type can be associated with a notebook\n instance. For more information, see Using Elastic Inference in\n Amazon SageMaker.
" } }, "DisassociateAcceleratorTypes": { @@ -58159,13 +58171,13 @@ "VariantPropertyType": { "target": "com.amazonaws.sagemaker#VariantPropertyType", "traits": { - "smithy.api#documentation": "The type of variant property. The supported values are:
\n\n DesiredInstanceCount
: Overrides the existing variant instance\n counts using the InitialInstanceCount
values\n in the ProductionVariants
of CreateEndpointConfig.
\n DesiredWeight
: Overrides the existing variant weights using the\n InitialVariantWeight
values in the ProductionVariants
of CreateEndpointConfig.
\n DataCaptureConfig
: (Not currently supported.)
The type of variant property. The supported values are:
\n\n DesiredInstanceCount
: Overrides the existing variant instance\n counts using the InitialInstanceCount
values in the\n ProductionVariants
of CreateEndpointConfig.
\n DesiredWeight
: Overrides the existing variant weights using the\n InitialVariantWeight
values in the\n ProductionVariants
of CreateEndpointConfig.
\n DataCaptureConfig
: (Not currently supported.)
Specifies a production variant property type for an Endpoint.
\nIf you are updating an endpoint with the RetainAllVariantProperties
\n option of UpdateEndpointInput set to\n true
, the VariantProperty
objects listed in the ExcludeRetainedVariantProperties
parameter of UpdateEndpointInput\n override the\n existing variant properties of the endpoint.
Specifies a production variant property type for an Endpoint.
\nIf you are updating an endpoint with the RetainAllVariantProperties
\n option of UpdateEndpointInput set to true
, the\n VariantProperty
objects listed in the\n ExcludeRetainedVariantProperties
parameter of UpdateEndpointInput override the existing variant properties of the\n endpoint.