diff --git a/CHANGELOG.md b/CHANGELOG.md index bd58d1167c2..d15c441a63d 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -1,3 +1,11 @@ +Release v1.42.39 (2022-01-21) +=== + +### Service Client Updates +* `service/mediaconvert`: Updates service API and documentation + * AWS Elemental MediaConvert SDK has added support for 4K AV1 output resolutions & 10-bit AV1 color, the ability to ingest sidecar Dolby Vision XML metadata files, and the ability to flag WebVTT and IMSC tracks for accessibility in HLS. +* `service/transcribe`: Updates service API and documentation + Release v1.42.38 (2022-01-20) === diff --git a/aws/version.go b/aws/version.go index b578df04658..36cfb1b4e37 100644 --- a/aws/version.go +++ b/aws/version.go @@ -5,4 +5,4 @@ package aws const SDKName = "aws-sdk-go" // SDKVersion is the version of this SDK -const SDKVersion = "1.42.38" +const SDKVersion = "1.42.39" diff --git a/models/apis/mediaconvert/2017-08-29/api-2.json b/models/apis/mediaconvert/2017-08-29/api-2.json index 7564b228c82..d3959cc3ce8 100644 --- a/models/apis/mediaconvert/2017-08-29/api-2.json +++ b/models/apis/mediaconvert/2017-08-29/api-2.json @@ -1615,6 +1615,13 @@ "MAX" ] }, + "Av1BitDepth": { + "type": "string", + "enum": [ + "BIT_8", + "BIT_10" + ] + }, "Av1FramerateControl": { "type": "string", "enum": [ @@ -1656,6 +1663,10 @@ "shape": "Av1AdaptiveQuantization", "locationName": "adaptiveQuantization" }, + "BitDepth": { + "shape": "Av1BitDepth", + "locationName": "bitDepth" + }, "FramerateControl": { "shape": "Av1FramerateControl", "locationName": "framerateControl" @@ -2626,9 +2637,20 @@ "Scte35Source": { "shape": "CmfcScte35Source", "locationName": "scte35Source" + }, + "TimedMetadata": { + "shape": "CmfcTimedMetadata", + "locationName": "timedMetadata" } } }, + "CmfcTimedMetadata": { + "type": "string", + "enum": [ + "PASSTHROUGH", + "NONE" + ] + }, "ColorCorrector": { "type": "structure", "members": { @@ -4765,7 +4787,8 @@ "MEDIUM", "HIGH", "HIGHER", - "MAX" + "MAX", + "AUTO" ] }, "H265AlternateTransferFunctionSei": { @@ -5703,9 +5726,20 @@ } } }, + "ImscAccessibilitySubs": { + "type": "string", + "enum": [ + "DISABLED", + "ENABLED" + ] + }, "ImscDestinationSettings": { "type": "structure", "members": { + "Accessibility": { + "shape": "ImscAccessibilitySubs", + "locationName": "accessibility" + }, "StylePassthrough": { "shape": "ImscStylePassthrough", "locationName": "stylePassthrough" @@ -5750,6 +5784,10 @@ "shape": "InputDenoiseFilter", "locationName": "denoiseFilter" }, + "DolbyVisionMetadataXml": { + "shape": "__stringMin14PatternS3XmlXMLHttpsXmlXML", + "locationName": "dolbyVisionMetadataXml" + }, "FileInput": { "shape": "__stringPatternS3MM2PPMM2VVMMPPEEGGMMPP3AAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMKKAAMMOOVVMMTTSSMM2TTWWMMVVaAAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8WWEEBBMMLLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MMXXMMLLOOGGGGaAAATTMMOOSSHttpsMM2VVMMPPEEGGMMPP3AAVVIIMMPP4FFLLVVMMPPTTMMPPGGMM4VVTTRRPPFF4VVMM2TTSSTTSS264HH264MMKKVVMMKKAAMMOOVVMMTTSSMM2TTWWMMVVaAAASSFFVVOOBB3GGPP3GGPPPPMMXXFFDDIIVVXXXXVVIIDDRRAAWWDDVVGGXXFFMM1VV3GG2VVMMFFMM3UU8WWEEBBMMLLCCHHGGXXFFMMPPEEGG2MMXXFFMMPPEEGG2MMXXFFHHDDWWAAVVYY4MMXXMMLLOOGGGGaAAATTMMOOSS", "locationName": "fileInput" @@ -5926,6 +5964,10 @@ "shape": "InputDenoiseFilter", "locationName": "denoiseFilter" }, + "DolbyVisionMetadataXml": { + "shape": "__stringMin14PatternS3XmlXMLHttpsXmlXML", + "locationName": "dolbyVisionMetadataXml" + }, "FilterEnable": { "shape": "InputFilterEnable", "locationName": "filterEnable" @@ -6607,7 +6649,8 @@ "ZUL", "ORJ", "QPC", - "TNG" + "TNG", + "SRP" ] }, "ListJobTemplatesRequest": { @@ -7468,9 +7511,20 @@ "Scte35Source": { "shape": "MpdScte35Source", "locationName": "scte35Source" + }, + "TimedMetadata": { + "shape": "MpdTimedMetadata", + "locationName": "timedMetadata" } } }, + "MpdTimedMetadata": { + "type": "string", + "enum": [ + "PASSTHROUGH", + "NONE" + ] + }, "Mpeg2AdaptiveQuantization": { "type": "string", "enum": [ @@ -8773,7 +8827,7 @@ "locationName": "kmsEncryptionContext" }, "KmsKeyArn": { - "shape": "__stringPatternArnAwsUsGovCnKmsAZ26EastWestCentralNorthSouthEastWest1912D12KeyAFAF098AFAF094AFAF094AFAF094AFAF0912", + "shape": "__stringPatternArnAwsUsGovCnKmsAZ26EastWestCentralNorthSouthEastWest1912D12KeyAFAF098AFAF094AFAF094AFAF094AFAF0912MrkAFAF0932", "locationName": "kmsKeyArn" } } @@ -9839,9 +9893,20 @@ } } }, + "WebvttAccessibilitySubs": { + "type": "string", + "enum": [ + "DISABLED", + "ENABLED" + ] + }, "WebvttDestinationSettings": { "type": "structure", "members": { + "Accessibility": { + "shape": "WebvttAccessibilitySubs", + "locationName": "accessibility" + }, "StylePassthrough": { "shape": "WebvttStylePassthrough", "locationName": "stylePassthrough" @@ -10988,6 +11053,11 @@ "min": 14, "pattern": "^((s3://(.*?)\\.(scc|SCC|ttml|TTML|dfxp|DFXP|stl|STL|srt|SRT|xml|XML|smi|SMI|vtt|VTT|webvtt|WEBVTT))|(https?://(.*?)\\.(scc|SCC|ttml|TTML|dfxp|DFXP|stl|STL|srt|SRT|xml|XML|smi|SMI|vtt|VTT|webvtt|WEBVTT)(\\?([^&=]+=[^&]+&)*[^&=]+=[^&]+)?))$" }, + "__stringMin14PatternS3XmlXMLHttpsXmlXML": { + "type": "string", + "min": 14, + "pattern": "^((s3://(.*?)\\.(xml|XML))|(https?://(.*?)\\.(xml|XML)(\\?([^&=]+=[^&]+&)*[^&=]+=[^&]+)?))$" + }, "__stringMin16Max24PatternAZaZ0922AZaZ0916": { "type": "string", "min": 16, @@ -11108,9 +11178,9 @@ "type": "string", "pattern": "^arn:aws(-us-gov)?:acm:" }, - "__stringPatternArnAwsUsGovCnKmsAZ26EastWestCentralNorthSouthEastWest1912D12KeyAFAF098AFAF094AFAF094AFAF094AFAF0912": { + "__stringPatternArnAwsUsGovCnKmsAZ26EastWestCentralNorthSouthEastWest1912D12KeyAFAF098AFAF094AFAF094AFAF094AFAF0912MrkAFAF0932": { "type": "string", - "pattern": "^arn:aws(-us-gov|-cn)?:kms:[a-z-]{2,6}-(east|west|central|((north|south)(east|west)?))-[1-9]{1,2}:\\d{12}:key/[a-fA-F0-9]{8}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{12}$" + "pattern": "^arn:aws(-us-gov|-cn)?:kms:[a-z-]{2,6}-(east|west|central|((north|south)(east|west)?))-[1-9]{1,2}:\\d{12}:key/([a-fA-F0-9]{8}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{4}-[a-fA-F0-9]{12}|mrk-[a-fA-F0-9]{32})$" }, "__stringPatternDD": { "type": "string", diff --git a/models/apis/mediaconvert/2017-08-29/docs-2.json b/models/apis/mediaconvert/2017-08-29/docs-2.json index 090fe1c51ea..ffe1f709985 100644 --- a/models/apis/mediaconvert/2017-08-29/docs-2.json +++ b/models/apis/mediaconvert/2017-08-29/docs-2.json @@ -316,6 +316,12 @@ "Av1Settings$AdaptiveQuantization": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to Spatial adaptive quantization (spatialAdaptiveQuantization)." } }, + "Av1BitDepth": { + "base": "Specify the Bit depth (Av1BitDepth). You can choose 8-bit (BIT_8) or 10-bit (BIT_10).", + "refs": { + "Av1Settings$BitDepth": "Specify the Bit depth (Av1BitDepth). You can choose 8-bit (BIT_8) or 10-bit (BIT_10)." + } + }, "Av1FramerateControl": { "base": "If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction. If you are creating your transcoding job specification as a JSON file without the console, use FramerateControl to specify which value the service uses for the frame rate for this output. Choose INITIALIZE_FROM_SOURCE if you want the service to use the frame rate from the input. Choose SPECIFIED if you want the service to use the frame rate you specify in the settings FramerateNumerator and FramerateDenominator.", "refs": { @@ -731,6 +737,12 @@ "ContainerSettings$CmfcSettings": "These settings relate to the fragmented MP4 container for the segments in your CMAF outputs." } }, + "CmfcTimedMetadata": { + "base": "Applies to CMAF outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output.", + "refs": { + "CmfcSettings$TimedMetadata": "Applies to CMAF outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output." + } + }, "ColorCorrector": { "base": "Settings for color correction.", "refs": { @@ -1626,9 +1638,9 @@ } }, "H265AdaptiveQuantization": { - "base": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization).", + "base": "When you set Adaptive Quantization (H265AdaptiveQuantization) to Auto (AUTO), or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low (LOW), Medium (MEDIUM), High (HIGH), Higher (HIGHER), or Max (MAX) to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization (H265SpatialAdaptiveQuantization), Temporal Adaptive Quantization (H265TemporalAdaptiveQuantization), and Flicker Adaptive Quantization (H265FlickerAdaptiveQuantization), to further control the quantization filter. Set Adaptive Quantization to Off (OFF) to apply no quantization to your output.", "refs": { - "H265Settings$AdaptiveQuantization": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization)." + "H265Settings$AdaptiveQuantization": "When you set Adaptive Quantization (H265AdaptiveQuantization) to Auto (AUTO), or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low (LOW), Medium (MEDIUM), High (HIGH), Higher (HIGHER), or Max (MAX) to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization (H265SpatialAdaptiveQuantization), Temporal Adaptive Quantization (H265TemporalAdaptiveQuantization), and Flicker Adaptive Quantization (H265FlickerAdaptiveQuantization), to further control the quantization filter. Set Adaptive Quantization to Off (OFF) to apply no quantization to your output." } }, "H265AlternateTransferFunctionSei": { @@ -2012,6 +2024,12 @@ "VideoPreprocessor$ImageInserter": "Enable the Image inserter (ImageInserter) feature to include a graphic overlay on your video. Enable or disable this feature for each output individually. This setting is disabled by default." } }, + "ImscAccessibilitySubs": { + "base": "Specify whether to flag this caption track as accessibility in your HLS/CMAF parent manifest. When you choose ENABLED, MediaConvert includes the parameters CHARACTERISTICS=\"public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound\" and AUTOSELECT=\"YES\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, DISABLED, MediaConvert leaves this parameter out.", + "refs": { + "ImscDestinationSettings$Accessibility": "Specify whether to flag this caption track as accessibility in your HLS/CMAF parent manifest. When you choose ENABLED, MediaConvert includes the parameters CHARACTERISTICS=\"public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound\" and AUTOSELECT=\"YES\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, DISABLED, MediaConvert leaves this parameter out." + } + }, "ImscDestinationSettings": { "base": "Settings related to IMSC captions. IMSC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to IMSC.", "refs": { @@ -2525,6 +2543,12 @@ "ContainerSettings$MpdSettings": "These settings relate to the fragmented MP4 container for the segments in your DASH outputs." } }, + "MpdTimedMetadata": { + "base": "Applies to DASH outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output.", + "refs": { + "MpdSettings$TimedMetadata": "Applies to DASH outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output." + } + }, "Mpeg2AdaptiveQuantization": { "base": "Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization).", "refs": { @@ -3222,9 +3246,9 @@ } }, "TimedMetadata": { - "base": "Applies only to HLS outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output.", + "base": "Applies to HLS outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output.", "refs": { - "M3u8Settings$TimedMetadata": "Applies only to HLS outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output." + "M3u8Settings$TimedMetadata": "Applies to HLS outputs. Use this setting to specify whether the service inserts the ID3 timed metadata from the input in this output." } }, "TimedMetadataInsertion": { @@ -3505,6 +3529,12 @@ "AudioCodecSettings$WavSettings": "Required when you set (Codec) under (AudioDescriptions)>(CodecSettings) to the value WAV." } }, + "WebvttAccessibilitySubs": { + "base": "Specify whether to flag this caption track as accessibility in your HLS/CMAF parent manifest. When you choose ENABLED, MediaConvert includes the parameters CHARACTERISTICS=\"public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound\" and AUTOSELECT=\"YES\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, DISABLED, MediaConvert leaves this parameter out.", + "refs": { + "WebvttDestinationSettings$Accessibility": "Specify whether to flag this caption track as accessibility in your HLS/CMAF parent manifest. When you choose ENABLED, MediaConvert includes the parameters CHARACTERISTICS=\"public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound\" and AUTOSELECT=\"YES\" in the EXT-X-MEDIA entry for this track. When you keep the default choice, DISABLED, MediaConvert leaves this parameter out." + } + }, "WebvttDestinationSettings": { "base": "Settings related to WebVTT captions. WebVTT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html. When you work directly in your JSON job specification, include this object and any required children when you set destinationType to WebVTT.", "refs": { @@ -4938,6 +4968,13 @@ "FileSourceSettings$SourceFile": "External caption file used for loading captions. Accepted file extensions are 'scc', 'ttml', 'dfxp', 'stl', 'srt', 'xml', 'smi', 'webvtt', and 'vtt'." } }, + "__stringMin14PatternS3XmlXMLHttpsXmlXML": { + "base": null, + "refs": { + "Input$DolbyVisionMetadataXml": "Use this setting only when your video source has Dolby Vision studio mastering metadata that is carried in a separate XML file. Specify the Amazon S3 location for the metadata XML file. MediaConvert uses this file to provide global and frame-level metadata for Dolby Vision preprocessing. When you specify a file here and your input also has interleaved global and frame level metadata, MediaConvert ignores the interleaved metadata and uses only the the metadata from this external XML file. Note that your IAM service role must grant MediaConvert read permissions to this file. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html.", + "InputTemplate$DolbyVisionMetadataXml": "Use this setting only when your video source has Dolby Vision studio mastering metadata that is carried in a separate XML file. Specify the Amazon S3 location for the metadata XML file. MediaConvert uses this file to provide global and frame-level metadata for Dolby Vision preprocessing. When you specify a file here and your input also has interleaved global and frame level metadata, MediaConvert ignores the interleaved metadata and uses only the the metadata from this external XML file. Note that your IAM service role must grant MediaConvert read permissions to this file. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html." + } + }, "__stringMin16Max24PatternAZaZ0922AZaZ0916": { "base": null, "refs": { @@ -5106,7 +5143,7 @@ "SpekeKeyProviderCmaf$CertificateArn": "If you want your key provider to encrypt the content keys that it provides to MediaConvert, set up a certificate with a master key using AWS Certificate Manager. Specify the certificate's Amazon Resource Name (ARN) here." } }, - "__stringPatternArnAwsUsGovCnKmsAZ26EastWestCentralNorthSouthEastWest1912D12KeyAFAF098AFAF094AFAF094AFAF094AFAF0912": { + "__stringPatternArnAwsUsGovCnKmsAZ26EastWestCentralNorthSouthEastWest1912D12KeyAFAF098AFAF094AFAF094AFAF094AFAF0912MrkAFAF0932": { "base": null, "refs": { "S3EncryptionSettings$KmsKeyArn": "Optionally, specify the customer master key (CMK) that you want to use to encrypt the data key that AWS uses to encrypt your output content. Enter the Amazon Resource Name (ARN) of the CMK. To use this setting, you must also set Server-side encryption (S3ServerSideEncryptionType) to AWS KMS (SERVER_SIDE_ENCRYPTION_KMS). If you set Server-side encryption to AWS KMS but don't specify a CMK here, AWS uses the AWS managed CMK associated with Amazon S3." diff --git a/models/apis/transcribe/2017-10-26/api-2.json b/models/apis/transcribe/2017-10-26/api-2.json index 26698538d0f..9ce621d54b1 100644 --- a/models/apis/transcribe/2017-10-26/api-2.json +++ b/models/apis/transcribe/2017-10-26/api-2.json @@ -736,7 +736,8 @@ ], "members":{ "RedactionType":{"shape":"RedactionType"}, - "RedactionOutput":{"shape":"RedactionOutput"} + "RedactionOutput":{"shape":"RedactionOutput"}, + "PiiEntityTypes":{"shape":"PiiEntityTypes"} } }, "CreateCallAnalyticsCategoryRequest":{ @@ -1516,6 +1517,29 @@ "type":"list", "member":{"shape":"Phrase"} }, + "PiiEntityType":{ + "type":"string", + "enum":[ + "BANK_ACCOUNT_NUMBER", + "BANK_ROUTING", + "CREDIT_DEBIT_NUMBER", + "CREDIT_DEBIT_CVV", + "CREDIT_DEBIT_EXPIRY", + "PIN", + "EMAIL", + "ADDRESS", + "NAME", + "PHONE", + "SSN", + "ALL" + ] + }, + "PiiEntityTypes":{ + "type":"list", + "member":{"shape":"PiiEntityType"}, + "max":11, + "min":0 + }, "RedactionOutput":{ "type":"string", "enum":[ diff --git a/models/apis/transcribe/2017-10-26/docs-2.json b/models/apis/transcribe/2017-10-26/docs-2.json index 31a8fb2d3dc..05d19837388 100644 --- a/models/apis/transcribe/2017-10-26/docs-2.json +++ b/models/apis/transcribe/2017-10-26/docs-2.json @@ -871,6 +871,18 @@ "UpdateVocabularyRequest$Phrases": "

An array of strings containing the vocabulary entries.

" } }, + "PiiEntityType": { + "base": null, + "refs": { + "PiiEntityTypes$member": null + } + }, + "PiiEntityTypes": { + "base": null, + "refs": { + "ContentRedaction$PiiEntityTypes": "

The types of personally identifiable information (PII) you want to redact in your transcript.

" + } + }, "RedactionOutput": { "base": null, "refs": { diff --git a/service/mediaconvert/api.go b/service/mediaconvert/api.go index 5214f75adf2..14f12ad4f13 100644 --- a/service/mediaconvert/api.go +++ b/service/mediaconvert/api.go @@ -4339,6 +4339,10 @@ type Av1Settings struct { // The value that you choose here applies to Spatial adaptive quantization (spatialAdaptiveQuantization). AdaptiveQuantization *string `locationName:"adaptiveQuantization" type:"string" enum:"Av1AdaptiveQuantization"` + // Specify the Bit depth (Av1BitDepth). You can choose 8-bit (BIT_8) or 10-bit + // (BIT_10). + BitDepth *string `locationName:"bitDepth" type:"string" enum:"Av1BitDepth"` + // If you are using the console, use the Framerate setting to specify the frame // rate for this output. If you want to keep the same frame rate as the input // video, choose Follow source. If you want to do frame rate conversion, choose @@ -4478,6 +4482,12 @@ func (s *Av1Settings) SetAdaptiveQuantization(v string) *Av1Settings { return s } +// SetBitDepth sets the BitDepth field's value. +func (s *Av1Settings) SetBitDepth(v string) *Av1Settings { + s.BitDepth = &v + return s +} + // SetFramerateControl sets the FramerateControl field's value. func (s *Av1Settings) SetFramerateControl(v string) *Av1Settings { s.FramerateControl = &v @@ -6744,6 +6754,10 @@ type CmfcSettings struct { // in your input to also appear in this output. Choose None (NONE) if you don't // want those SCTE-35 markers in this output. Scte35Source *string `locationName:"scte35Source" type:"string" enum:"CmfcScte35Source"` + + // Applies to CMAF outputs. Use this setting to specify whether the service + // inserts the ID3 timed metadata from the input in this output. + TimedMetadata *string `locationName:"timedMetadata" type:"string" enum:"CmfcTimedMetadata"` } // String returns the string representation. @@ -6812,6 +6826,12 @@ func (s *CmfcSettings) SetScte35Source(v string) *CmfcSettings { return s } +// SetTimedMetadata sets the TimedMetadata field's value. +func (s *CmfcSettings) SetTimedMetadata(v string) *CmfcSettings { + s.TimedMetadata = &v + return s +} + // Settings for color correction. type ColorCorrector struct { _ struct{} `type:"structure"` @@ -12204,10 +12224,16 @@ func (s *H265QvbrSettings) SetQvbrQualityLevelFineTune(v float64) *H265QvbrSetti type H265Settings struct { _ struct{} `type:"structure"` - // Specify the strength of any adaptive quantization filters that you enable. - // The value that you choose here applies to the following settings: Flicker - // adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization - // (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization). + // When you set Adaptive Quantization (H265AdaptiveQuantization) to Auto (AUTO), + // or leave blank, MediaConvert automatically applies quantization to improve + // the video quality of your output. Set Adaptive Quantization to Low (LOW), + // Medium (MEDIUM), High (HIGH), Higher (HIGHER), or Max (MAX) to manually control + // the strength of the quantization filter. When you do, you can specify a value + // for Spatial Adaptive Quantization (H265SpatialAdaptiveQuantization), Temporal + // Adaptive Quantization (H265TemporalAdaptiveQuantization), and Flicker Adaptive + // Quantization (H265FlickerAdaptiveQuantization), to further control the quantization + // filter. Set Adaptive Quantization to Off (OFF) to apply no quantization to + // your output. AdaptiveQuantization *string `locationName:"adaptiveQuantization" type:"string" enum:"H265AdaptiveQuantization"` // Enables Alternate Transfer Function SEI message for outputs using Hybrid @@ -14154,6 +14180,13 @@ func (s *ImageInserter) SetInsertableImages(v []*InsertableImage) *ImageInserter type ImscDestinationSettings struct { _ struct{} `type:"structure"` + // Specify whether to flag this caption track as accessibility in your HLS/CMAF + // parent manifest. When you choose ENABLED, MediaConvert includes the parameters + // CHARACTERISTICS="public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound" + // and AUTOSELECT="YES" in the EXT-X-MEDIA entry for this track. When you keep + // the default choice, DISABLED, MediaConvert leaves this parameter out. + Accessibility *string `locationName:"accessibility" type:"string" enum:"ImscAccessibilitySubs"` + // Keep this setting enabled to have MediaConvert use the font style and position // information from the captions source in the output. This option is available // only when your input captions are IMSC, SMPTE-TT, or TTML. Disable this setting @@ -14179,6 +14212,12 @@ func (s ImscDestinationSettings) GoString() string { return s.String() } +// SetAccessibility sets the Accessibility field's value. +func (s *ImscDestinationSettings) SetAccessibility(v string) *ImscDestinationSettings { + s.Accessibility = &v + return s +} + // SetStylePassthrough sets the StylePassthrough field's value. func (s *ImscDestinationSettings) SetStylePassthrough(v string) *ImscDestinationSettings { s.StylePassthrough = &v @@ -14230,6 +14269,16 @@ type Input struct { // inputs. DenoiseFilter *string `locationName:"denoiseFilter" type:"string" enum:"InputDenoiseFilter"` + // Use this setting only when your video source has Dolby Vision studio mastering + // metadata that is carried in a separate XML file. Specify the Amazon S3 location + // for the metadata XML file. MediaConvert uses this file to provide global + // and frame-level metadata for Dolby Vision preprocessing. When you specify + // a file here and your input also has interleaved global and frame level metadata, + // MediaConvert ignores the interleaved metadata and uses only the the metadata + // from this external XML file. Note that your IAM service role must grant MediaConvert + // read permissions to this file. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html. + DolbyVisionMetadataXml *string `locationName:"dolbyVisionMetadataXml" min:"14" type:"string"` + // Specify the source file for your transcoding job. You can use multiple inputs // in a single job. The service concatenates these inputs, in the order that // you specify them in the job, to create the outputs. If your input format @@ -14347,6 +14396,9 @@ func (s Input) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *Input) Validate() error { invalidParams := request.ErrInvalidParams{Context: "Input"} + if s.DolbyVisionMetadataXml != nil && len(*s.DolbyVisionMetadataXml) < 14 { + invalidParams.Add(request.NewErrParamMinLen("DolbyVisionMetadataXml", 14)) + } if s.FilterStrength != nil && *s.FilterStrength < -5 { invalidParams.Add(request.NewErrParamMinValue("FilterStrength", -5)) } @@ -14450,6 +14502,12 @@ func (s *Input) SetDenoiseFilter(v string) *Input { return s } +// SetDolbyVisionMetadataXml sets the DolbyVisionMetadataXml field's value. +func (s *Input) SetDolbyVisionMetadataXml(v string) *Input { + s.DolbyVisionMetadataXml = &v + return s +} + // SetFileInput sets the FileInput field's value. func (s *Input) SetFileInput(v string) *Input { s.FileInput = &v @@ -14714,6 +14772,16 @@ type InputTemplate struct { // inputs. DenoiseFilter *string `locationName:"denoiseFilter" type:"string" enum:"InputDenoiseFilter"` + // Use this setting only when your video source has Dolby Vision studio mastering + // metadata that is carried in a separate XML file. Specify the Amazon S3 location + // for the metadata XML file. MediaConvert uses this file to provide global + // and frame-level metadata for Dolby Vision preprocessing. When you specify + // a file here and your input also has interleaved global and frame level metadata, + // MediaConvert ignores the interleaved metadata and uses only the the metadata + // from this external XML file. Note that your IAM service role must grant MediaConvert + // read permissions to this file. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html. + DolbyVisionMetadataXml *string `locationName:"dolbyVisionMetadataXml" min:"14" type:"string"` + // Specify how the transcoding service applies the denoise and deblock filters. // You must also enable the filters separately, with Denoise (InputDenoiseFilter) // and Deblock (InputDeblockFilter). * Auto - The transcoding service determines @@ -14814,6 +14882,9 @@ func (s InputTemplate) GoString() string { // Validate inspects the fields of the type to determine if they are valid. func (s *InputTemplate) Validate() error { invalidParams := request.ErrInvalidParams{Context: "InputTemplate"} + if s.DolbyVisionMetadataXml != nil && len(*s.DolbyVisionMetadataXml) < 14 { + invalidParams.Add(request.NewErrParamMinLen("DolbyVisionMetadataXml", 14)) + } if s.FilterStrength != nil && *s.FilterStrength < -5 { invalidParams.Add(request.NewErrParamMinValue("FilterStrength", -5)) } @@ -14906,6 +14977,12 @@ func (s *InputTemplate) SetDenoiseFilter(v string) *InputTemplate { return s } +// SetDolbyVisionMetadataXml sets the DolbyVisionMetadataXml field's value. +func (s *InputTemplate) SetDolbyVisionMetadataXml(v string) *InputTemplate { + s.DolbyVisionMetadataXml = &v + return s +} + // SetFilterEnable sets the FilterEnable field's value. func (s *InputTemplate) SetFilterEnable(v string) *InputTemplate { s.FilterEnable = &v @@ -17574,8 +17651,8 @@ type M3u8Settings struct { // XML (sccXml). Scte35Source *string `locationName:"scte35Source" type:"string" enum:"M3u8Scte35Source"` - // Applies only to HLS outputs. Use this setting to specify whether the service - // inserts the ID3 timed metadata from the input in this output. + // Applies to HLS outputs. Use this setting to specify whether the service inserts + // the ID3 timed metadata from the input in this output. TimedMetadata *string `locationName:"timedMetadata" type:"string" enum:"TimedMetadata"` // Packet Identifier (PID) of the timed metadata stream in the transport stream. @@ -18374,6 +18451,10 @@ type MpdSettings struct { // in your input to also appear in this output. Choose None (NONE) if you don't // want those SCTE-35 markers in this output. Scte35Source *string `locationName:"scte35Source" type:"string" enum:"MpdScte35Source"` + + // Applies to DASH outputs. Use this setting to specify whether the service + // inserts the ID3 timed metadata from the input in this output. + TimedMetadata *string `locationName:"timedMetadata" type:"string" enum:"MpdTimedMetadata"` } // String returns the string representation. @@ -18424,6 +18505,12 @@ func (s *MpdSettings) SetScte35Source(v string) *MpdSettings { return s } +// SetTimedMetadata sets the TimedMetadata field's value. +func (s *MpdSettings) SetTimedMetadata(v string) *MpdSettings { + s.TimedMetadata = &v + return s +} + // Required when you set (Codec) under (VideoDescription)>(CodecSettings) to // the value MPEG2. type Mpeg2Settings struct { @@ -24934,6 +25021,13 @@ func (s *WavSettings) SetSampleRate(v int64) *WavSettings { type WebvttDestinationSettings struct { _ struct{} `type:"structure"` + // Specify whether to flag this caption track as accessibility in your HLS/CMAF + // parent manifest. When you choose ENABLED, MediaConvert includes the parameters + // CHARACTERISTICS="public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound" + // and AUTOSELECT="YES" in the EXT-X-MEDIA entry for this track. When you keep + // the default choice, DISABLED, MediaConvert leaves this parameter out. + Accessibility *string `locationName:"accessibility" type:"string" enum:"WebvttAccessibilitySubs"` + // Set Style passthrough (StylePassthrough) to ENABLED to use the available // style, color, and position information from your input captions. MediaConvert // uses default settings for any missing style and position information in your @@ -24961,6 +25055,12 @@ func (s WebvttDestinationSettings) GoString() string { return s.String() } +// SetAccessibility sets the Accessibility field's value. +func (s *WebvttDestinationSettings) SetAccessibility(v string) *WebvttDestinationSettings { + s.Accessibility = &v + return s +} + // SetStylePassthrough sets the StylePassthrough field's value. func (s *WebvttDestinationSettings) SetStylePassthrough(v string) *WebvttDestinationSettings { s.StylePassthrough = &v @@ -26590,6 +26690,24 @@ func Av1AdaptiveQuantization_Values() []string { } } +// Specify the Bit depth (Av1BitDepth). You can choose 8-bit (BIT_8) or 10-bit +// (BIT_10). +const ( + // Av1BitDepthBit8 is a Av1BitDepth enum value + Av1BitDepthBit8 = "BIT_8" + + // Av1BitDepthBit10 is a Av1BitDepth enum value + Av1BitDepthBit10 = "BIT_10" +) + +// Av1BitDepth_Values returns all elements of the Av1BitDepth enum +func Av1BitDepth_Values() []string { + return []string{ + Av1BitDepthBit8, + Av1BitDepthBit10, + } +} + // If you are using the console, use the Framerate setting to specify the frame // rate for this output. If you want to keep the same frame rate as the input // video, choose Follow source. If you want to do frame rate conversion, choose @@ -27861,6 +27979,24 @@ func CmfcScte35Source_Values() []string { } } +// Applies to CMAF outputs. Use this setting to specify whether the service +// inserts the ID3 timed metadata from the input in this output. +const ( + // CmfcTimedMetadataPassthrough is a CmfcTimedMetadata enum value + CmfcTimedMetadataPassthrough = "PASSTHROUGH" + + // CmfcTimedMetadataNone is a CmfcTimedMetadata enum value + CmfcTimedMetadataNone = "NONE" +) + +// CmfcTimedMetadata_Values returns all elements of the CmfcTimedMetadata enum +func CmfcTimedMetadata_Values() []string { + return []string{ + CmfcTimedMetadataPassthrough, + CmfcTimedMetadataNone, + } +} + // Choose Insert (INSERT) for this setting to include color metadata in this // output. Choose Ignore (IGNORE) to exclude color metadata from this output. // If you don't specify a value, the service sets this to Insert by default. @@ -30206,10 +30342,16 @@ func H264UnregisteredSeiTimecode_Values() []string { } } -// Specify the strength of any adaptive quantization filters that you enable. -// The value that you choose here applies to the following settings: Flicker -// adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization -// (spatialAdaptiveQuantization), and Temporal adaptive quantization (temporalAdaptiveQuantization). +// When you set Adaptive Quantization (H265AdaptiveQuantization) to Auto (AUTO), +// or leave blank, MediaConvert automatically applies quantization to improve +// the video quality of your output. Set Adaptive Quantization to Low (LOW), +// Medium (MEDIUM), High (HIGH), Higher (HIGHER), or Max (MAX) to manually control +// the strength of the quantization filter. When you do, you can specify a value +// for Spatial Adaptive Quantization (H265SpatialAdaptiveQuantization), Temporal +// Adaptive Quantization (H265TemporalAdaptiveQuantization), and Flicker Adaptive +// Quantization (H265FlickerAdaptiveQuantization), to further control the quantization +// filter. Set Adaptive Quantization to Off (OFF) to apply no quantization to +// your output. const ( // H265AdaptiveQuantizationOff is a H265AdaptiveQuantization enum value H265AdaptiveQuantizationOff = "OFF" @@ -30228,6 +30370,9 @@ const ( // H265AdaptiveQuantizationMax is a H265AdaptiveQuantization enum value H265AdaptiveQuantizationMax = "MAX" + + // H265AdaptiveQuantizationAuto is a H265AdaptiveQuantization enum value + H265AdaptiveQuantizationAuto = "AUTO" ) // H265AdaptiveQuantization_Values returns all elements of the H265AdaptiveQuantization enum @@ -30239,6 +30384,7 @@ func H265AdaptiveQuantization_Values() []string { H265AdaptiveQuantizationHigh, H265AdaptiveQuantizationHigher, H265AdaptiveQuantizationMax, + H265AdaptiveQuantizationAuto, } } @@ -31421,6 +31567,27 @@ func HlsTimedMetadataId3Frame_Values() []string { } } +// Specify whether to flag this caption track as accessibility in your HLS/CMAF +// parent manifest. When you choose ENABLED, MediaConvert includes the parameters +// CHARACTERISTICS="public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound" +// and AUTOSELECT="YES" in the EXT-X-MEDIA entry for this track. When you keep +// the default choice, DISABLED, MediaConvert leaves this parameter out. +const ( + // ImscAccessibilitySubsDisabled is a ImscAccessibilitySubs enum value + ImscAccessibilitySubsDisabled = "DISABLED" + + // ImscAccessibilitySubsEnabled is a ImscAccessibilitySubs enum value + ImscAccessibilitySubsEnabled = "ENABLED" +) + +// ImscAccessibilitySubs_Values returns all elements of the ImscAccessibilitySubs enum +func ImscAccessibilitySubs_Values() []string { + return []string{ + ImscAccessibilitySubsDisabled, + ImscAccessibilitySubsEnabled, + } +} + // Keep this setting enabled to have MediaConvert use the font style and position // information from the captions source in the output. This option is available // only when your input captions are IMSC, SMPTE-TT, or TTML. Disable this setting @@ -32310,6 +32477,9 @@ const ( // LanguageCodeTng is a LanguageCode enum value LanguageCodeTng = "TNG" + + // LanguageCodeSrp is a LanguageCode enum value + LanguageCodeSrp = "SRP" ) // LanguageCode_Values returns all elements of the LanguageCode enum @@ -32506,6 +32676,7 @@ func LanguageCode_Values() []string { LanguageCodeOrj, LanguageCodeQpc, LanguageCodeTng, + LanguageCodeSrp, } } @@ -33244,6 +33415,24 @@ func MpdScte35Source_Values() []string { } } +// Applies to DASH outputs. Use this setting to specify whether the service +// inserts the ID3 timed metadata from the input in this output. +const ( + // MpdTimedMetadataPassthrough is a MpdTimedMetadata enum value + MpdTimedMetadataPassthrough = "PASSTHROUGH" + + // MpdTimedMetadataNone is a MpdTimedMetadata enum value + MpdTimedMetadataNone = "NONE" +) + +// MpdTimedMetadata_Values returns all elements of the MpdTimedMetadata enum +func MpdTimedMetadata_Values() []string { + return []string{ + MpdTimedMetadataPassthrough, + MpdTimedMetadataNone, + } +} + // Specify the strength of any adaptive quantization filters that you enable. // The value that you choose here applies to the following settings: Spatial // adaptive quantization (spatialAdaptiveQuantization), and Temporal adaptive @@ -34857,8 +35046,8 @@ func TimecodeSource_Values() []string { } } -// Applies only to HLS outputs. Use this setting to specify whether the service -// inserts the ID3 timed metadata from the input in this output. +// Applies to HLS outputs. Use this setting to specify whether the service inserts +// the ID3 timed metadata from the input in this output. const ( // TimedMetadataPassthrough is a TimedMetadata enum value TimedMetadataPassthrough = "PASSTHROUGH" @@ -35462,6 +35651,27 @@ func WavFormat_Values() []string { } } +// Specify whether to flag this caption track as accessibility in your HLS/CMAF +// parent manifest. When you choose ENABLED, MediaConvert includes the parameters +// CHARACTERISTICS="public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound" +// and AUTOSELECT="YES" in the EXT-X-MEDIA entry for this track. When you keep +// the default choice, DISABLED, MediaConvert leaves this parameter out. +const ( + // WebvttAccessibilitySubsDisabled is a WebvttAccessibilitySubs enum value + WebvttAccessibilitySubsDisabled = "DISABLED" + + // WebvttAccessibilitySubsEnabled is a WebvttAccessibilitySubs enum value + WebvttAccessibilitySubsEnabled = "ENABLED" +) + +// WebvttAccessibilitySubs_Values returns all elements of the WebvttAccessibilitySubs enum +func WebvttAccessibilitySubs_Values() []string { + return []string{ + WebvttAccessibilitySubsDisabled, + WebvttAccessibilitySubsEnabled, + } +} + // Set Style passthrough (StylePassthrough) to ENABLED to use the available // style, color, and position information from your input captions. MediaConvert // uses default settings for any missing style and position information in your diff --git a/service/transcribeservice/api.go b/service/transcribeservice/api.go index a73fef3cb36..673f5fa0f18 100644 --- a/service/transcribeservice/api.go +++ b/service/transcribeservice/api.go @@ -4917,6 +4917,10 @@ func (s *ConflictException) RequestID() string { type ContentRedaction struct { _ struct{} `type:"structure"` + // The types of personally identifiable information (PII) you want to redact + // in your transcript. + PiiEntityTypes []*string `type:"list"` + // The output transcript file stored in either the default S3 bucket or in a // bucket you specify. // @@ -4969,6 +4973,12 @@ func (s *ContentRedaction) Validate() error { return nil } +// SetPiiEntityTypes sets the PiiEntityTypes field's value. +func (s *ContentRedaction) SetPiiEntityTypes(v []*string) *ContentRedaction { + s.PiiEntityTypes = v + return s +} + // SetRedactionOutput sets the RedactionOutput field's value. func (s *ContentRedaction) SetRedactionOutput(v string) *ContentRedaction { s.RedactionOutput = &v @@ -12847,6 +12857,62 @@ func ParticipantRole_Values() []string { } } +const ( + // PiiEntityTypeBankAccountNumber is a PiiEntityType enum value + PiiEntityTypeBankAccountNumber = "BANK_ACCOUNT_NUMBER" + + // PiiEntityTypeBankRouting is a PiiEntityType enum value + PiiEntityTypeBankRouting = "BANK_ROUTING" + + // PiiEntityTypeCreditDebitNumber is a PiiEntityType enum value + PiiEntityTypeCreditDebitNumber = "CREDIT_DEBIT_NUMBER" + + // PiiEntityTypeCreditDebitCvv is a PiiEntityType enum value + PiiEntityTypeCreditDebitCvv = "CREDIT_DEBIT_CVV" + + // PiiEntityTypeCreditDebitExpiry is a PiiEntityType enum value + PiiEntityTypeCreditDebitExpiry = "CREDIT_DEBIT_EXPIRY" + + // PiiEntityTypePin is a PiiEntityType enum value + PiiEntityTypePin = "PIN" + + // PiiEntityTypeEmail is a PiiEntityType enum value + PiiEntityTypeEmail = "EMAIL" + + // PiiEntityTypeAddress is a PiiEntityType enum value + PiiEntityTypeAddress = "ADDRESS" + + // PiiEntityTypeName is a PiiEntityType enum value + PiiEntityTypeName = "NAME" + + // PiiEntityTypePhone is a PiiEntityType enum value + PiiEntityTypePhone = "PHONE" + + // PiiEntityTypeSsn is a PiiEntityType enum value + PiiEntityTypeSsn = "SSN" + + // PiiEntityTypeAll is a PiiEntityType enum value + PiiEntityTypeAll = "ALL" +) + +// PiiEntityType_Values returns all elements of the PiiEntityType enum +func PiiEntityType_Values() []string { + return []string{ + PiiEntityTypeBankAccountNumber, + PiiEntityTypeBankRouting, + PiiEntityTypeCreditDebitNumber, + PiiEntityTypeCreditDebitCvv, + PiiEntityTypeCreditDebitExpiry, + PiiEntityTypePin, + PiiEntityTypeEmail, + PiiEntityTypeAddress, + PiiEntityTypeName, + PiiEntityTypePhone, + PiiEntityTypeSsn, + PiiEntityTypeAll, + } +} + const ( // RedactionOutputRedacted is a RedactionOutput enum value RedactionOutputRedacted = "redacted"