Class: Aws::ChimeSDKMediaPipelines::Client

Inherits:
Seahorse::Client::Base show all
Includes:
Aws::ClientStubs
Defined in:
gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb

Overview

An API client for ChimeSDKMediaPipelines. To construct a client, you need to configure a :region and :credentials.

client = Aws::ChimeSDKMediaPipelines::Client.new(
  region: region_name,
  credentials: credentials,
  # ...
)

For details on configuring region and credentials see the developer guide.

See #initialize for a full list of supported configuration options.

Instance Attribute Summary

Attributes inherited from Seahorse::Client::Base

#config, #handlers

API Operations collapse

Instance Method Summary collapse

Methods included from Aws::ClientStubs

#api_requests, #stub_data, #stub_responses

Methods inherited from Seahorse::Client::Base

add_plugin, api, clear_plugins, define, new, #operation_names, plugins, remove_plugin, set_api, set_plugins

Methods included from Seahorse::Client::HandlerBuilder

#handle, #handle_request, #handle_response

Constructor Details

#initialize(options) ⇒ Client

Returns a new instance of Client.

Parameters:

  • options (Hash)

Options Hash (options):

  • :credentials (required, Aws::CredentialProvider)

    Your AWS credentials. This can be an instance of any one of the following classes:

    • Aws::Credentials - Used for configuring static, non-refreshing credentials.

    • Aws::SharedCredentials - Used for loading static credentials from a shared file, such as ~/.aws/config.

    • Aws::AssumeRoleCredentials - Used when you need to assume a role.

    • Aws::AssumeRoleWebIdentityCredentials - Used when you need to assume a role after providing credentials via the web.

    • Aws::SSOCredentials - Used for loading credentials from AWS SSO using an access token generated from aws login.

    • Aws::ProcessCredentials - Used for loading credentials from a process that outputs to stdout.

    • Aws::InstanceProfileCredentials - Used for loading credentials from an EC2 IMDS on an EC2 instance.

    • Aws::ECSCredentials - Used for loading credentials from instances running in ECS.

    • Aws::CognitoIdentityCredentials - Used for loading credentials from the Cognito Identity service.

    When :credentials are not configured directly, the following locations will be searched for credentials:

    • Aws.config[:credentials]
    • The :access_key_id, :secret_access_key, and :session_token options.
    • ENV['AWS_ACCESS_KEY_ID'], ENV['AWS_SECRET_ACCESS_KEY']
    • ~/.aws/credentials
    • ~/.aws/config
    • EC2/ECS IMDS instance profile - When used by default, the timeouts are very aggressive. Construct and pass an instance of Aws::InstanceProfileCredentails or Aws::ECSCredentials to enable retries and extended timeouts. Instance profile credential fetching can be disabled by setting ENV['AWS_EC2_METADATA_DISABLED'] to true.
  • :region (required, String)

    The AWS region to connect to. The configured :region is used to determine the service :endpoint. When not passed, a default :region is searched for in the following locations:

    • Aws.config[:region]
    • ENV['AWS_REGION']
    • ENV['AMAZON_REGION']
    • ENV['AWS_DEFAULT_REGION']
    • ~/.aws/credentials
    • ~/.aws/config
  • :access_key_id (String)
  • :active_endpoint_cache (Boolean) — default: false

    When set to true, a thread polling for endpoints will be running in the background every 60 secs (default). Defaults to false.

  • :adaptive_retry_wait_to_fill (Boolean) — default: true

    Used only in adaptive retry mode. When true, the request will sleep until there is sufficent client side capacity to retry the request. When false, the request will raise a RetryCapacityNotAvailableError and will not retry instead of sleeping.

  • :client_side_monitoring (Boolean) — default: false

    When true, client-side metrics will be collected for all API requests from this client.

  • :client_side_monitoring_client_id (String) — default: ""

    Allows you to provide an identifier for this client which will be attached to all generated client side metrics. Defaults to an empty string.

  • :client_side_monitoring_host (String) — default: "127.0.0.1"

    Allows you to specify the DNS hostname or IPv4 or IPv6 address that the client side monitoring agent is running on, where client metrics will be published via UDP.

  • :client_side_monitoring_port (Integer) — default: 31000

    Required for publishing client metrics. The port that the client side monitoring agent is running on, where client metrics will be published via UDP.

  • :client_side_monitoring_publisher (Aws::ClientSideMonitoring::Publisher) — default: Aws::ClientSideMonitoring::Publisher

    Allows you to provide a custom client-side monitoring publisher class. By default, will use the Client Side Monitoring Agent Publisher.

  • :convert_params (Boolean) — default: true

    When true, an attempt is made to coerce request parameters into the required types.

  • :correct_clock_skew (Boolean) — default: true

    Used only in standard and adaptive retry modes. Specifies whether to apply a clock skew correction and retry requests with skewed client clocks.

  • :defaults_mode (String) — default: "legacy"

    See DefaultsModeConfiguration for a list of the accepted modes and the configuration defaults that are included.

  • :disable_host_prefix_injection (Boolean) — default: false

    Set to true to disable SDK automatically adding host prefix to default service endpoint when available.

  • :disable_request_compression (Boolean) — default: false

    When set to 'true' the request body will not be compressed for supported operations.

  • :endpoint (String, URI::HTTPS, URI::HTTP)

    Normally you should not configure the :endpoint option directly. This is normally constructed from the :region option. Configuring :endpoint is normally reserved for connecting to test or custom endpoints. The endpoint should be a URI formatted like:

    'http://example.com'
    'https://example.com'
    'http://example.com:123'
    
  • :endpoint_cache_max_entries (Integer) — default: 1000

    Used for the maximum size limit of the LRU cache storing endpoints data for endpoint discovery enabled operations. Defaults to 1000.

  • :endpoint_cache_max_threads (Integer) — default: 10

    Used for the maximum threads in use for polling endpoints to be cached, defaults to 10.

  • :endpoint_cache_poll_interval (Integer) — default: 60

    When :endpoint_discovery and :active_endpoint_cache is enabled, Use this option to config the time interval in seconds for making requests fetching endpoints information. Defaults to 60 sec.

  • :endpoint_discovery (Boolean) — default: false

    When set to true, endpoint discovery will be enabled for operations when available.

  • :ignore_configured_endpoint_urls (Boolean)

    Setting to true disables use of endpoint URLs provided via environment variables and the shared configuration file.

  • :log_formatter (Aws::Log::Formatter) — default: Aws::Log::Formatter.default

    The log formatter.

  • :log_level (Symbol) — default: :info

    The log level to send messages to the :logger at.

  • :logger (Logger)

    The Logger instance to send log messages to. If this option is not set, logging will be disabled.

  • :max_attempts (Integer) — default: 3

    An integer representing the maximum number attempts that will be made for a single request, including the initial attempt. For example, setting this value to 5 will result in a request being retried up to 4 times. Used in standard and adaptive retry modes.

  • :profile (String) — default: "default"

    Used when loading credentials from the shared credentials file at HOME/.aws/credentials. When not specified, 'default' is used.

  • :request_min_compression_size_bytes (Integer) — default: 10240

    The minimum size in bytes that triggers compression for request bodies. The value must be non-negative integer value between 0 and 10485780 bytes inclusive.

  • :retry_backoff (Proc)

    A proc or lambda used for backoff. Defaults to 2**retries * retry_base_delay. This option is only used in the legacy retry mode.

  • :retry_base_delay (Float) — default: 0.3

    The base delay in seconds used by the default backoff function. This option is only used in the legacy retry mode.

  • :retry_jitter (Symbol) — default: :none

    A delay randomiser function used by the default backoff function. Some predefined functions can be referenced by name - :none, :equal, :full, otherwise a Proc that takes and returns a number. This option is only used in the legacy retry mode.

    @see https://www.awsarchitectureblog.com/2015/03/backoff.html

  • :retry_limit (Integer) — default: 3

    The maximum number of times to retry failed requests. Only ~ 500 level server errors and certain ~ 400 level client errors are retried. Generally, these are throttling errors, data checksum errors, networking errors, timeout errors, auth errors, endpoint discovery, and errors from expired credentials. This option is only used in the legacy retry mode.

  • :retry_max_delay (Integer) — default: 0

    The maximum number of seconds to delay between retries (0 for no limit) used by the default backoff function. This option is only used in the legacy retry mode.

  • :retry_mode (String) — default: "legacy"

    Specifies which retry algorithm to use. Values are:

    • legacy - The pre-existing retry behavior. This is default value if no retry mode is provided.

    • standard - A standardized set of retry rules across the AWS SDKs. This includes support for retry quotas, which limit the number of unsuccessful retries a client can make.

    • adaptive - An experimental retry mode that includes all the functionality of standard mode along with automatic client side throttling. This is a provisional mode that may change behavior in the future.

  • :sdk_ua_app_id (String)

    A unique and opaque application ID that is appended to the User-Agent header as app/. It should have a maximum length of 50.

  • :secret_access_key (String)
  • :session_token (String)
  • :stub_responses (Boolean) — default: false

    Causes the client to return stubbed responses. By default fake responses are generated and returned. You can specify the response data to return or errors to raise by calling Aws::ClientStubs#stub_responses. See Aws::ClientStubs for more information.

    Please note When response stubbing is enabled, no HTTP requests are made, and retries are disabled.

  • :token_provider (Aws::TokenProvider)

    A Bearer Token Provider. This can be an instance of any one of the following classes:

    • Aws::StaticTokenProvider - Used for configuring static, non-refreshing tokens.

    • Aws::SSOTokenProvider - Used for loading tokens from AWS SSO using an access token generated from aws login.

    When :token_provider is not configured directly, the Aws::TokenProviderChain will be used to search for tokens configured for your profile in shared configuration files.

  • :use_dualstack_endpoint (Boolean)

    When set to true, dualstack enabled endpoints (with .aws TLD) will be used if available.

  • :use_fips_endpoint (Boolean)

    When set to true, fips compatible endpoints will be used if available. When a fips region is used, the region is normalized and this config is set to true.

  • :validate_params (Boolean) — default: true

    When true, request parameters are validated before sending the request.

  • :endpoint_provider (Aws::ChimeSDKMediaPipelines::EndpointProvider)

    The endpoint provider used to resolve endpoints. Any object that responds to #resolve_endpoint(parameters) where parameters is a Struct similar to Aws::ChimeSDKMediaPipelines::EndpointParameters

  • :http_continue_timeout (Float) — default: 1

    The number of seconds to wait for a 100-continue response before sending the request body. This option has no effect unless the request has "Expect" header set to "100-continue". Defaults to nil which disables this behaviour. This value can safely be set per request on the session.

  • :http_idle_timeout (Float) — default: 5

    The number of seconds a connection is allowed to sit idle before it is considered stale. Stale connections are closed and removed from the pool before making a request.

  • :http_open_timeout (Float) — default: 15

    The default number of seconds to wait for response data. This value can safely be set per-request on the session.

  • :http_proxy (URI::HTTP, String)

    A proxy to send requests through. Formatted like 'http://proxy.com:123'.

  • :http_read_timeout (Float) — default: 60

    The default number of seconds to wait for response data. This value can safely be set per-request on the session.

  • :http_wire_trace (Boolean) — default: false

    When true, HTTP debug output will be sent to the :logger.

  • :on_chunk_received (Proc)

    When a Proc object is provided, it will be used as callback when each chunk of the response body is received. It provides three arguments: the chunk, the number of bytes received, and the total number of bytes in the response (or nil if the server did not send a content-length).

  • :on_chunk_sent (Proc)

    When a Proc object is provided, it will be used as callback when each chunk of the request body is sent. It provides three arguments: the chunk, the number of bytes read from the body, and the total number of bytes in the body.

  • :raise_response_errors (Boolean) — default: true

    When true, response errors are raised.

  • :ssl_ca_bundle (String)

    Full path to the SSL certificate authority bundle file that should be used when verifying peer certificates. If you do not pass :ssl_ca_bundle or :ssl_ca_directory the the system default will be used if available.

  • :ssl_ca_directory (String)

    Full path of the directory that contains the unbundled SSL certificate authority files for verifying peer certificates. If you do not pass :ssl_ca_bundle or :ssl_ca_directory the the system default will be used if available.

  • :ssl_ca_store (String)

    Sets the X509::Store to verify peer certificate.

  • :ssl_timeout (Float)

    Sets the SSL timeout in seconds

  • :ssl_verify_peer (Boolean) — default: true

    When true, SSL peer certificates are verified when establishing a connection.



409
410
411
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 409

def initialize(*args)
  super
end

Instance Method Details

#create_media_capture_pipeline(params = {}) ⇒ Types::CreateMediaCapturePipelineResponse

Creates a media pipeline.

Examples:

Request syntax with placeholder values


resp = client.create_media_capture_pipeline({
  source_type: "ChimeSdkMeeting", # required, accepts ChimeSdkMeeting
  source_arn: "Arn", # required
  sink_type: "S3Bucket", # required, accepts S3Bucket
  sink_arn: "Arn", # required
  client_request_token: "ClientRequestToken",
  chime_sdk_meeting_configuration: {
    source_configuration: {
      selected_video_streams: {
        attendee_ids: ["GuidString"],
        external_user_ids: ["ExternalUserIdType"],
      },
    },
    artifacts_configuration: {
      audio: { # required
        mux_type: "AudioOnly", # required, accepts AudioOnly, AudioWithActiveSpeakerVideo, AudioWithCompositedVideo
      },
      video: { # required
        state: "Enabled", # required, accepts Enabled, Disabled
        mux_type: "VideoOnly", # accepts VideoOnly
      },
      content: { # required
        state: "Enabled", # required, accepts Enabled, Disabled
        mux_type: "ContentOnly", # accepts ContentOnly
      },
      composited_video: {
        layout: "GridView", # accepts GridView
        resolution: "HD", # accepts HD, FHD
        grid_view_configuration: { # required
          content_share_layout: "PresenterOnly", # required, accepts PresenterOnly, Horizontal, Vertical, ActiveSpeakerOnly
          presenter_only_configuration: {
            presenter_position: "TopLeft", # accepts TopLeft, TopRight, BottomLeft, BottomRight
          },
          active_speaker_only_configuration: {
            active_speaker_position: "TopLeft", # accepts TopLeft, TopRight, BottomLeft, BottomRight
          },
          horizontal_layout_configuration: {
            tile_order: "JoinSequence", # accepts JoinSequence, SpeakerSequence
            tile_position: "Top", # accepts Top, Bottom
            tile_count: 1,
            tile_aspect_ratio: "TileAspectRatio",
          },
          vertical_layout_configuration: {
            tile_order: "JoinSequence", # accepts JoinSequence, SpeakerSequence
            tile_position: "Left", # accepts Left, Right
            tile_count: 1,
            tile_aspect_ratio: "TileAspectRatio",
          },
          video_attribute: {
            corner_radius: 1,
            border_color: "Black", # accepts Black, Blue, Red, Green, White, Yellow
            highlight_color: "Black", # accepts Black, Blue, Red, Green, White, Yellow
            border_thickness: 1,
          },
          canvas_orientation: "Landscape", # accepts Landscape, Portrait
        },
      },
    },
  },
  tags: [
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
})

Response structure


resp.media_capture_pipeline.media_pipeline_id #=> String
resp.media_capture_pipeline.media_pipeline_arn #=> String
resp.media_capture_pipeline.source_type #=> String, one of "ChimeSdkMeeting"
resp.media_capture_pipeline.source_arn #=> String
resp.media_capture_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_capture_pipeline.sink_type #=> String, one of "S3Bucket"
resp.media_capture_pipeline.sink_arn #=> String
resp.media_capture_pipeline.created_timestamp #=> Time
resp.media_capture_pipeline.updated_timestamp #=> Time
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.attendee_ids #=> Array
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.attendee_ids[0] #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.external_user_ids #=> Array
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.external_user_ids[0] #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.audio.mux_type #=> String, one of "AudioOnly", "AudioWithActiveSpeakerVideo", "AudioWithCompositedVideo"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.video.state #=> String, one of "Enabled", "Disabled"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.video.mux_type #=> String, one of "VideoOnly"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.content.state #=> String, one of "Enabled", "Disabled"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.content.mux_type #=> String, one of "ContentOnly"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.layout #=> String, one of "GridView"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.resolution #=> String, one of "HD", "FHD"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.content_share_layout #=> String, one of "PresenterOnly", "Horizontal", "Vertical", "ActiveSpeakerOnly"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.presenter_only_configuration.presenter_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.active_speaker_only_configuration.active_speaker_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_position #=> String, one of "Top", "Bottom"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_count #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_aspect_ratio #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_position #=> String, one of "Left", "Right"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_count #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_aspect_ratio #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.corner_radius #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.border_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.highlight_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.border_thickness #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.canvas_orientation #=> String, one of "Landscape", "Portrait"

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :source_type (required, String)

    Source type from which the media artifacts are captured. A Chime SDK Meeting is the only supported source.

  • :source_arn (required, String)

    ARN of the source from which the media artifacts are captured.

  • :sink_type (required, String)

    Destination type to which the media artifacts are saved. You must use an S3 bucket.

  • :sink_arn (required, String)

    The ARN of the sink type.

  • :client_request_token (String)

    The unique identifier for the client request. The token makes the API request idempotent. Use a unique token for each media pipeline request.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

  • :chime_sdk_meeting_configuration (Types::ChimeSdkMeetingConfiguration)

    The configuration for a specified media pipeline. SourceType must be ChimeSdkMeeting.

  • :tags (Array<Types::Tag>)

    The tag key-value pairs.

Returns:

See Also:



562
563
564
565
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 562

def create_media_capture_pipeline(params = {}, options = {})
  req = build_request(:create_media_capture_pipeline, params)
  req.send_request(options)
end

#create_media_concatenation_pipeline(params = {}) ⇒ Types::CreateMediaConcatenationPipelineResponse

Creates a media concatenation pipeline.

Examples:

Request syntax with placeholder values


resp = client.create_media_concatenation_pipeline({
  sources: [ # required
    {
      type: "MediaCapturePipeline", # required, accepts MediaCapturePipeline
      media_capture_pipeline_source_configuration: { # required
        media_pipeline_arn: "Arn", # required
        chime_sdk_meeting_configuration: { # required
          artifacts_configuration: { # required
            audio: { # required
              state: "Enabled", # required, accepts Enabled
            },
            video: { # required
              state: "Enabled", # required, accepts Enabled, Disabled
            },
            content: { # required
              state: "Enabled", # required, accepts Enabled, Disabled
            },
            data_channel: { # required
              state: "Enabled", # required, accepts Enabled, Disabled
            },
            transcription_messages: { # required
              state: "Enabled", # required, accepts Enabled, Disabled
            },
            meeting_events: { # required
              state: "Enabled", # required, accepts Enabled, Disabled
            },
            composited_video: { # required
              state: "Enabled", # required, accepts Enabled, Disabled
            },
          },
        },
      },
    },
  ],
  sinks: [ # required
    {
      type: "S3Bucket", # required, accepts S3Bucket
      s3_bucket_sink_configuration: { # required
        destination: "Arn", # required
      },
    },
  ],
  client_request_token: "ClientRequestToken",
  tags: [
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
})

Response structure


resp.media_concatenation_pipeline.media_pipeline_id #=> String
resp.media_concatenation_pipeline.media_pipeline_arn #=> String
resp.media_concatenation_pipeline.sources #=> Array
resp.media_concatenation_pipeline.sources[0].type #=> String, one of "MediaCapturePipeline"
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.media_pipeline_arn #=> String
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.audio.state #=> String, one of "Enabled"
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.video.state #=> String, one of "Enabled", "Disabled"
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.content.state #=> String, one of "Enabled", "Disabled"
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.data_channel.state #=> String, one of "Enabled", "Disabled"
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.transcription_messages.state #=> String, one of "Enabled", "Disabled"
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.meeting_events.state #=> String, one of "Enabled", "Disabled"
resp.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.state #=> String, one of "Enabled", "Disabled"
resp.media_concatenation_pipeline.sinks #=> Array
resp.media_concatenation_pipeline.sinks[0].type #=> String, one of "S3Bucket"
resp.media_concatenation_pipeline.sinks[0].s3_bucket_sink_configuration.destination #=> String
resp.media_concatenation_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_concatenation_pipeline.created_timestamp #=> Time
resp.media_concatenation_pipeline.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :sources (required, Array<Types::ConcatenationSource>)

    An object that specifies the sources for the media concatenation pipeline.

  • :sinks (required, Array<Types::ConcatenationSink>)

    An object that specifies the data sinks for the media concatenation pipeline.

  • :client_request_token (String)

    The unique identifier for the client request. The token makes the API request idempotent. Use a unique token for each media concatenation pipeline request.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

  • :tags (Array<Types::Tag>)

    The tags associated with the media concatenation pipeline.

Returns:

See Also:



670
671
672
673
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 670

def create_media_concatenation_pipeline(params = {}, options = {})
  req = build_request(:create_media_concatenation_pipeline, params)
  req.send_request(options)
end

#create_media_insights_pipeline(params = {}) ⇒ Types::CreateMediaInsightsPipelineResponse

Creates a media insights pipeline.

Examples:

Request syntax with placeholder values


resp = client.create_media_insights_pipeline({
  media_insights_pipeline_configuration_arn: "Arn", # required
  kinesis_video_stream_source_runtime_configuration: {
    streams: [ # required
      {
        stream_arn: "KinesisVideoStreamArn", # required
        fragment_number: "FragmentNumberString",
        stream_channel_definition: { # required
          number_of_channels: 1, # required
          channel_definitions: [
            {
              channel_id: 1, # required
              participant_role: "AGENT", # accepts AGENT, CUSTOMER
            },
          ],
        },
      },
    ],
    media_encoding: "pcm", # required, accepts pcm
    media_sample_rate: 1, # required
  },
  media_insights_runtime_metadata: {
    "NonEmptyString" => "String",
  },
  kinesis_video_stream_recording_source_runtime_configuration: {
    streams: [ # required
      {
        stream_arn: "KinesisVideoStreamArn",
      },
    ],
    fragment_selector: { # required
      fragment_selector_type: "ProducerTimestamp", # required, accepts ProducerTimestamp, ServerTimestamp
      timestamp_range: { # required
        start_timestamp: Time.now, # required
        end_timestamp: Time.now, # required
      },
    },
  },
  s3_recording_sink_runtime_configuration: {
    destination: "Arn", # required
    recording_file_format: "Wav", # required, accepts Wav, Opus
  },
  tags: [
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
  client_request_token: "ClientRequestToken",
})

Response structure


resp.media_insights_pipeline.media_pipeline_id #=> String
resp.media_insights_pipeline.media_pipeline_arn #=> String
resp.media_insights_pipeline.media_insights_pipeline_configuration_arn #=> String
resp.media_insights_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams #=> Array
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_arn #=> String
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].fragment_number #=> String
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.number_of_channels #=> Integer
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.channel_definitions #=> Array
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.channel_definitions[0].channel_id #=> Integer
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.channel_definitions[0].participant_role #=> String, one of "AGENT", "CUSTOMER"
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.media_encoding #=> String, one of "pcm"
resp.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.media_sample_rate #=> Integer
resp.media_insights_pipeline. #=> Hash
resp.media_insights_pipeline.["NonEmptyString"] #=> String
resp.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.streams #=> Array
resp.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.streams[0].stream_arn #=> String
resp.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.fragment_selector.fragment_selector_type #=> String, one of "ProducerTimestamp", "ServerTimestamp"
resp.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.fragment_selector.timestamp_range.start_timestamp #=> Time
resp.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.fragment_selector.timestamp_range.end_timestamp #=> Time
resp.media_insights_pipeline.s3_recording_sink_runtime_configuration.destination #=> String
resp.media_insights_pipeline.s3_recording_sink_runtime_configuration.recording_file_format #=> String, one of "Wav", "Opus"
resp.media_insights_pipeline.created_timestamp #=> Time
resp.media_insights_pipeline.element_statuses #=> Array
resp.media_insights_pipeline.element_statuses[0].type #=> String, one of "AmazonTranscribeCallAnalyticsProcessor", "VoiceAnalyticsProcessor", "AmazonTranscribeProcessor", "KinesisDataStreamSink", "LambdaFunctionSink", "SqsQueueSink", "SnsTopicSink", "S3RecordingSink", "VoiceEnhancementSink"
resp.media_insights_pipeline.element_statuses[0].status #=> String, one of "NotStarted", "NotSupported", "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused"

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :media_insights_pipeline_configuration_arn (required, String)

    The ARN of the pipeline's configuration.

  • :kinesis_video_stream_source_runtime_configuration (Types::KinesisVideoStreamSourceRuntimeConfiguration)

    The runtime configuration for the Kinesis video stream source of the media insights pipeline.

  • :media_insights_runtime_metadata (Hash<String,String>)

    The runtime metadata for the media insights pipeline. Consists of a key-value map of strings.

  • :kinesis_video_stream_recording_source_runtime_configuration (Types::KinesisVideoStreamRecordingSourceRuntimeConfiguration)

    The runtime configuration for the Kinesis video recording stream source.

  • :s3_recording_sink_runtime_configuration (Types::S3RecordingSinkRuntimeConfiguration)

    The runtime configuration for the S3 recording sink. If specified, the settings in this structure override any settings in S3RecordingSinkConfiguration.

  • :tags (Array<Types::Tag>)

    The tags assigned to the media insights pipeline.

  • :client_request_token (String)

    The unique identifier for the media insights pipeline request.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

Returns:

See Also:



796
797
798
799
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 796

def create_media_insights_pipeline(params = {}, options = {})
  req = build_request(:create_media_insights_pipeline, params)
  req.send_request(options)
end

#create_media_insights_pipeline_configuration(params = {}) ⇒ Types::CreateMediaInsightsPipelineConfigurationResponse

A structure that contains the static configurations for a media insights pipeline.

Examples:

Request syntax with placeholder values


resp = client.create_media_insights_pipeline_configuration({
  media_insights_pipeline_configuration_name: "MediaInsightsPipelineConfigurationNameString", # required
  resource_access_role_arn: "Arn", # required
  real_time_alert_configuration: {
    disabled: false,
    rules: [
      {
        type: "KeywordMatch", # required, accepts KeywordMatch, Sentiment, IssueDetection
        keyword_match_configuration: {
          rule_name: "RuleName", # required
          keywords: ["Keyword"], # required
          negate: false,
        },
        sentiment_configuration: {
          rule_name: "RuleName", # required
          sentiment_type: "NEGATIVE", # required, accepts NEGATIVE
          time_period: 1, # required
        },
        issue_detection_configuration: {
          rule_name: "RuleName", # required
        },
      },
    ],
  },
  elements: [ # required
    {
      type: "AmazonTranscribeCallAnalyticsProcessor", # required, accepts AmazonTranscribeCallAnalyticsProcessor, VoiceAnalyticsProcessor, AmazonTranscribeProcessor, KinesisDataStreamSink, LambdaFunctionSink, SqsQueueSink, SnsTopicSink, S3RecordingSink, VoiceEnhancementSink
      amazon_transcribe_call_analytics_processor_configuration: {
        language_code: "en-US", # required, accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR
        vocabulary_name: "VocabularyName",
        vocabulary_filter_name: "VocabularyFilterName",
        vocabulary_filter_method: "remove", # accepts remove, mask, tag
        language_model_name: "ModelName",
        enable_partial_results_stabilization: false,
        partial_results_stability: "high", # accepts high, medium, low
        content_identification_type: "PII", # accepts PII
        content_redaction_type: "PII", # accepts PII
        pii_entity_types: "PiiEntityTypes",
        filter_partial_results: false,
        post_call_analytics_settings: {
          output_location: "String", # required
          data_access_role_arn: "String", # required
          content_redaction_output: "redacted", # accepts redacted, redacted_and_unredacted
          output_encryption_kms_key_id: "String",
        },
        call_analytics_stream_categories: ["CategoryName"],
      },
      amazon_transcribe_processor_configuration: {
        language_code: "en-US", # accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR
        vocabulary_name: "VocabularyName",
        vocabulary_filter_name: "VocabularyFilterName",
        vocabulary_filter_method: "remove", # accepts remove, mask, tag
        show_speaker_label: false,
        enable_partial_results_stabilization: false,
        partial_results_stability: "high", # accepts high, medium, low
        content_identification_type: "PII", # accepts PII
        content_redaction_type: "PII", # accepts PII
        pii_entity_types: "PiiEntityTypes",
        language_model_name: "ModelName",
        filter_partial_results: false,
        identify_language: false,
        language_options: "LanguageOptions",
        preferred_language: "en-US", # accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR
        vocabulary_names: "VocabularyNames",
        vocabulary_filter_names: "VocabularyFilterNames",
      },
      kinesis_data_stream_sink_configuration: {
        insights_target: "Arn",
      },
      s3_recording_sink_configuration: {
        destination: "Arn",
        recording_file_format: "Wav", # accepts Wav, Opus
      },
      voice_analytics_processor_configuration: {
        speaker_search_status: "Enabled", # accepts Enabled, Disabled
        voice_tone_analysis_status: "Enabled", # accepts Enabled, Disabled
      },
      lambda_function_sink_configuration: {
        insights_target: "Arn",
      },
      sqs_queue_sink_configuration: {
        insights_target: "Arn",
      },
      sns_topic_sink_configuration: {
        insights_target: "Arn",
      },
      voice_enhancement_sink_configuration: {
        disabled: false,
      },
    },
  ],
  tags: [
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
  client_request_token: "ClientRequestToken",
})

Response structure


resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_name #=> String
resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_arn #=> String
resp.media_insights_pipeline_configuration.resource_access_role_arn #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.disabled #=> Boolean
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules #=> Array
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].type #=> String, one of "KeywordMatch", "Sentiment", "IssueDetection"
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.keywords #=> Array
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.keywords[0] #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.negate #=> Boolean
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.sentiment_type #=> String, one of "NEGATIVE"
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.time_period #=> Integer
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].issue_detection_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.elements #=> Array
resp.media_insights_pipeline_configuration.elements[0].type #=> String, one of "AmazonTranscribeCallAnalyticsProcessor", "VoiceAnalyticsProcessor", "AmazonTranscribeProcessor", "KinesisDataStreamSink", "LambdaFunctionSink", "SqsQueueSink", "SnsTopicSink", "S3RecordingSink", "VoiceEnhancementSink"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.language_code #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_filter_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_filter_method #=> String, one of "remove", "mask", "tag"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.language_model_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.enable_partial_results_stabilization #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.partial_results_stability #=> String, one of "high", "medium", "low"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.content_identification_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.content_redaction_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.pii_entity_types #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.filter_partial_results #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.output_location #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.data_access_role_arn #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.content_redaction_output #=> String, one of "redacted", "redacted_and_unredacted"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.output_encryption_kms_key_id #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.call_analytics_stream_categories #=> Array
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.call_analytics_stream_categories[0] #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_code #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_method #=> String, one of "remove", "mask", "tag"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.show_speaker_label #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.enable_partial_results_stabilization #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.partial_results_stability #=> String, one of "high", "medium", "low"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.content_identification_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.content_redaction_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.pii_entity_types #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_model_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.filter_partial_results #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.identify_language #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_options #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.preferred_language #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_names #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_names #=> String
resp.media_insights_pipeline_configuration.elements[0].kinesis_data_stream_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].s3_recording_sink_configuration.destination #=> String
resp.media_insights_pipeline_configuration.elements[0].s3_recording_sink_configuration.recording_file_format #=> String, one of "Wav", "Opus"
resp.media_insights_pipeline_configuration.elements[0].voice_analytics_processor_configuration.speaker_search_status #=> String, one of "Enabled", "Disabled"
resp.media_insights_pipeline_configuration.elements[0].voice_analytics_processor_configuration.voice_tone_analysis_status #=> String, one of "Enabled", "Disabled"
resp.media_insights_pipeline_configuration.elements[0].lambda_function_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].sqs_queue_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].sns_topic_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].voice_enhancement_sink_configuration.disabled #=> Boolean
resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_id #=> String
resp.media_insights_pipeline_configuration.created_timestamp #=> Time
resp.media_insights_pipeline_configuration.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :media_insights_pipeline_configuration_name (required, String)

    The name of the media insights pipeline configuration.

  • :resource_access_role_arn (required, String)

    The ARN of the role used by the service to access Amazon Web Services resources, including Transcribe and Transcribe Call Analytics, on the caller’s behalf.

  • :real_time_alert_configuration (Types::RealTimeAlertConfiguration)

    The configuration settings for the real-time alerts in a media insights pipeline configuration.

  • :elements (required, Array<Types::MediaInsightsPipelineConfigurationElement>)

    The elements in the request, such as a processor for Amazon Transcribe or a sink for a Kinesis Data Stream.

  • :tags (Array<Types::Tag>)

    The tags assigned to the media insights pipeline configuration.

  • :client_request_token (String)

    The unique identifier for the media insights pipeline configuration request.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

Returns:

See Also:



1005
1006
1007
1008
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1005

def create_media_insights_pipeline_configuration(params = {}, options = {})
  req = build_request(:create_media_insights_pipeline_configuration, params)
  req.send_request(options)
end

#create_media_live_connector_pipeline(params = {}) ⇒ Types::CreateMediaLiveConnectorPipelineResponse

Creates a media live connector pipeline in an Amazon Chime SDK meeting.

Examples:

Request syntax with placeholder values


resp = client.create_media_live_connector_pipeline({
  sources: [ # required
    {
      source_type: "ChimeSdkMeeting", # required, accepts ChimeSdkMeeting
      chime_sdk_meeting_live_connector_configuration: { # required
        arn: "Arn", # required
        mux_type: "AudioWithCompositedVideo", # required, accepts AudioWithCompositedVideo, AudioWithActiveSpeakerVideo
        composited_video: {
          layout: "GridView", # accepts GridView
          resolution: "HD", # accepts HD, FHD
          grid_view_configuration: { # required
            content_share_layout: "PresenterOnly", # required, accepts PresenterOnly, Horizontal, Vertical, ActiveSpeakerOnly
            presenter_only_configuration: {
              presenter_position: "TopLeft", # accepts TopLeft, TopRight, BottomLeft, BottomRight
            },
            active_speaker_only_configuration: {
              active_speaker_position: "TopLeft", # accepts TopLeft, TopRight, BottomLeft, BottomRight
            },
            horizontal_layout_configuration: {
              tile_order: "JoinSequence", # accepts JoinSequence, SpeakerSequence
              tile_position: "Top", # accepts Top, Bottom
              tile_count: 1,
              tile_aspect_ratio: "TileAspectRatio",
            },
            vertical_layout_configuration: {
              tile_order: "JoinSequence", # accepts JoinSequence, SpeakerSequence
              tile_position: "Left", # accepts Left, Right
              tile_count: 1,
              tile_aspect_ratio: "TileAspectRatio",
            },
            video_attribute: {
              corner_radius: 1,
              border_color: "Black", # accepts Black, Blue, Red, Green, White, Yellow
              highlight_color: "Black", # accepts Black, Blue, Red, Green, White, Yellow
              border_thickness: 1,
            },
            canvas_orientation: "Landscape", # accepts Landscape, Portrait
          },
        },
        source_configuration: {
          selected_video_streams: {
            attendee_ids: ["GuidString"],
            external_user_ids: ["ExternalUserIdType"],
          },
        },
      },
    },
  ],
  sinks: [ # required
    {
      sink_type: "RTMP", # required, accepts RTMP
      rtmp_configuration: { # required
        url: "SensitiveString", # required
        audio_channels: "Stereo", # accepts Stereo, Mono
        audio_sample_rate: "AudioSampleRateOption",
      },
    },
  ],
  client_request_token: "ClientRequestToken",
  tags: [
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
})

Response structure


resp.media_live_connector_pipeline.sources #=> Array
resp.media_live_connector_pipeline.sources[0].source_type #=> String, one of "ChimeSdkMeeting"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.arn #=> String
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.mux_type #=> String, one of "AudioWithCompositedVideo", "AudioWithActiveSpeakerVideo"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.layout #=> String, one of "GridView"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.resolution #=> String, one of "HD", "FHD"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.content_share_layout #=> String, one of "PresenterOnly", "Horizontal", "Vertical", "ActiveSpeakerOnly"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.presenter_only_configuration.presenter_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.active_speaker_only_configuration.active_speaker_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_position #=> String, one of "Top", "Bottom"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_count #=> Integer
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_aspect_ratio #=> String
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_position #=> String, one of "Left", "Right"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_count #=> Integer
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_aspect_ratio #=> String
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.corner_radius #=> Integer
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.border_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.highlight_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.border_thickness #=> Integer
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.canvas_orientation #=> String, one of "Landscape", "Portrait"
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.attendee_ids #=> Array
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.attendee_ids[0] #=> String
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.external_user_ids #=> Array
resp.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.external_user_ids[0] #=> String
resp.media_live_connector_pipeline.sinks #=> Array
resp.media_live_connector_pipeline.sinks[0].sink_type #=> String, one of "RTMP"
resp.media_live_connector_pipeline.sinks[0].rtmp_configuration.url #=> String
resp.media_live_connector_pipeline.sinks[0].rtmp_configuration.audio_channels #=> String, one of "Stereo", "Mono"
resp.media_live_connector_pipeline.sinks[0].rtmp_configuration.audio_sample_rate #=> String
resp.media_live_connector_pipeline.media_pipeline_id #=> String
resp.media_live_connector_pipeline.media_pipeline_arn #=> String
resp.media_live_connector_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_live_connector_pipeline.created_timestamp #=> Time
resp.media_live_connector_pipeline.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :sources (required, Array<Types::LiveConnectorSourceConfiguration>)

    The media live connector pipeline's data sources.

  • :sinks (required, Array<Types::LiveConnectorSinkConfiguration>)

    The media live connector pipeline's data sinks.

  • :client_request_token (String)

    The token assigned to the client making the request.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

  • :tags (Array<Types::Tag>)

    The tags associated with the media live connector pipeline.

Returns:

See Also:



1144
1145
1146
1147
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1144

def create_media_live_connector_pipeline(params = {}, options = {})
  req = build_request(:create_media_live_connector_pipeline, params)
  req.send_request(options)
end

#create_media_pipeline_kinesis_video_stream_pool(params = {}) ⇒ Types::CreateMediaPipelineKinesisVideoStreamPoolResponse

Creates an Kinesis video stream pool for the media pipeline.

Examples:

Request syntax with placeholder values


resp = client.create_media_pipeline_kinesis_video_stream_pool({
  stream_configuration: { # required
    region: "AwsRegion", # required
    data_retention_in_hours: 1,
  },
  pool_name: "KinesisVideoStreamPoolName", # required
  client_request_token: "ClientRequestToken",
  tags: [
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
})

Response structure


resp.kinesis_video_stream_pool_configuration.pool_arn #=> String
resp.kinesis_video_stream_pool_configuration.pool_name #=> String
resp.kinesis_video_stream_pool_configuration.pool_id #=> String
resp.kinesis_video_stream_pool_configuration.pool_status #=> String, one of "CREATING", "ACTIVE", "UPDATING", "DELETING", "FAILED"
resp.kinesis_video_stream_pool_configuration.pool_size #=> Integer
resp.kinesis_video_stream_pool_configuration.stream_configuration.region #=> String
resp.kinesis_video_stream_pool_configuration.stream_configuration.data_retention_in_hours #=> Integer
resp.kinesis_video_stream_pool_configuration.created_timestamp #=> Time
resp.kinesis_video_stream_pool_configuration.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :stream_configuration (required, Types::KinesisVideoStreamConfiguration)

    The configuration settings for the video stream.

  • :pool_name (required, String)

    The name of the video stream pool.

  • :client_request_token (String)

    The token assigned to the client making the request.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

  • :tags (Array<Types::Tag>)

    The tags assigned to the video stream pool.

Returns:

See Also:



1203
1204
1205
1206
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1203

def create_media_pipeline_kinesis_video_stream_pool(params = {}, options = {})
  req = build_request(:create_media_pipeline_kinesis_video_stream_pool, params)
  req.send_request(options)
end

#create_media_stream_pipeline(params = {}) ⇒ Types::CreateMediaStreamPipelineResponse

Creates a streaming media pipeline.

Examples:

Request syntax with placeholder values


resp = client.create_media_stream_pipeline({
  sources: [ # required
    {
      source_type: "ChimeSdkMeeting", # required, accepts ChimeSdkMeeting
      source_arn: "Arn", # required
    },
  ],
  sinks: [ # required
    {
      sink_arn: "Arn", # required
      sink_type: "KinesisVideoStreamPool", # required, accepts KinesisVideoStreamPool
      reserved_stream_capacity: 1, # required
      media_stream_type: "MixedAudio", # required, accepts MixedAudio, IndividualAudio
    },
  ],
  client_request_token: "ClientRequestToken",
  tags: [
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
})

Response structure


resp.media_stream_pipeline.media_pipeline_id #=> String
resp.media_stream_pipeline.media_pipeline_arn #=> String
resp.media_stream_pipeline.created_timestamp #=> Time
resp.media_stream_pipeline.updated_timestamp #=> Time
resp.media_stream_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_stream_pipeline.sources #=> Array
resp.media_stream_pipeline.sources[0].source_type #=> String, one of "ChimeSdkMeeting"
resp.media_stream_pipeline.sources[0].source_arn #=> String
resp.media_stream_pipeline.sinks #=> Array
resp.media_stream_pipeline.sinks[0].sink_arn #=> String
resp.media_stream_pipeline.sinks[0].sink_type #=> String, one of "KinesisVideoStreamPool"
resp.media_stream_pipeline.sinks[0].reserved_stream_capacity #=> Integer
resp.media_stream_pipeline.sinks[0].media_stream_type #=> String, one of "MixedAudio", "IndividualAudio"

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :sources (required, Array<Types::MediaStreamSource>)

    The data sources for the media pipeline.

  • :sinks (required, Array<Types::MediaStreamSink>)

    The data sink for the media pipeline.

  • :client_request_token (String)

    The token assigned to the client making the request.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

  • :tags (Array<Types::Tag>)

    The tags assigned to the media pipeline.

Returns:

See Also:



1275
1276
1277
1278
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1275

def create_media_stream_pipeline(params = {}, options = {})
  req = build_request(:create_media_stream_pipeline, params)
  req.send_request(options)
end

#delete_media_capture_pipeline(params = {}) ⇒ Struct

Deletes the media pipeline.

Examples:

Request syntax with placeholder values


resp = client.delete_media_capture_pipeline({
  media_pipeline_id: "GuidString", # required
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :media_pipeline_id (required, String)

    The ID of the media pipeline being deleted.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



1297
1298
1299
1300
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1297

def delete_media_capture_pipeline(params = {}, options = {})
  req = build_request(:delete_media_capture_pipeline, params)
  req.send_request(options)
end

#delete_media_insights_pipeline_configuration(params = {}) ⇒ Struct

Deletes the specified configuration settings.

Examples:

Request syntax with placeholder values


resp = client.delete_media_insights_pipeline_configuration({
  identifier: "NonEmptyString", # required
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be deleted. Valid values include the name and ARN of the media insights pipeline configuration.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



1320
1321
1322
1323
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1320

def delete_media_insights_pipeline_configuration(params = {}, options = {})
  req = build_request(:delete_media_insights_pipeline_configuration, params)
  req.send_request(options)
end

#delete_media_pipeline(params = {}) ⇒ Struct

Deletes the media pipeline.

Examples:

Request syntax with placeholder values


resp = client.delete_media_pipeline({
  media_pipeline_id: "GuidString", # required
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :media_pipeline_id (required, String)

    The ID of the media pipeline to delete.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



1342
1343
1344
1345
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1342

def delete_media_pipeline(params = {}, options = {})
  req = build_request(:delete_media_pipeline, params)
  req.send_request(options)
end

#delete_media_pipeline_kinesis_video_stream_pool(params = {}) ⇒ Struct

Deletes an Kinesis video stream pool.

Examples:

Request syntax with placeholder values


resp = client.delete_media_pipeline_kinesis_video_stream_pool({
  identifier: "NonEmptyString", # required
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The ID of the pool being deleted.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



1364
1365
1366
1367
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1364

def delete_media_pipeline_kinesis_video_stream_pool(params = {}, options = {})
  req = build_request(:delete_media_pipeline_kinesis_video_stream_pool, params)
  req.send_request(options)
end

#get_media_capture_pipeline(params = {}) ⇒ Types::GetMediaCapturePipelineResponse

Gets an existing media pipeline.

Examples:

Request syntax with placeholder values


resp = client.get_media_capture_pipeline({
  media_pipeline_id: "GuidString", # required
})

Response structure


resp.media_capture_pipeline.media_pipeline_id #=> String
resp.media_capture_pipeline.media_pipeline_arn #=> String
resp.media_capture_pipeline.source_type #=> String, one of "ChimeSdkMeeting"
resp.media_capture_pipeline.source_arn #=> String
resp.media_capture_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_capture_pipeline.sink_type #=> String, one of "S3Bucket"
resp.media_capture_pipeline.sink_arn #=> String
resp.media_capture_pipeline.created_timestamp #=> Time
resp.media_capture_pipeline.updated_timestamp #=> Time
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.attendee_ids #=> Array
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.attendee_ids[0] #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.external_user_ids #=> Array
resp.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.external_user_ids[0] #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.audio.mux_type #=> String, one of "AudioOnly", "AudioWithActiveSpeakerVideo", "AudioWithCompositedVideo"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.video.state #=> String, one of "Enabled", "Disabled"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.video.mux_type #=> String, one of "VideoOnly"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.content.state #=> String, one of "Enabled", "Disabled"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.content.mux_type #=> String, one of "ContentOnly"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.layout #=> String, one of "GridView"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.resolution #=> String, one of "HD", "FHD"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.content_share_layout #=> String, one of "PresenterOnly", "Horizontal", "Vertical", "ActiveSpeakerOnly"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.presenter_only_configuration.presenter_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.active_speaker_only_configuration.active_speaker_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_position #=> String, one of "Top", "Bottom"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_count #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_aspect_ratio #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_position #=> String, one of "Left", "Right"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_count #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_aspect_ratio #=> String
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.corner_radius #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.border_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.highlight_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.border_thickness #=> Integer
resp.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.canvas_orientation #=> String, one of "Landscape", "Portrait"

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :media_pipeline_id (required, String)

    The ID of the pipeline that you want to get.

Returns:

See Also:



1427
1428
1429
1430
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1427

def get_media_capture_pipeline(params = {}, options = {})
  req = build_request(:get_media_capture_pipeline, params)
  req.send_request(options)
end

#get_media_insights_pipeline_configuration(params = {}) ⇒ Types::GetMediaInsightsPipelineConfigurationResponse

Gets the configuration settings for a media insights pipeline.

Examples:

Request syntax with placeholder values


resp = client.get_media_insights_pipeline_configuration({
  identifier: "NonEmptyString", # required
})

Response structure


resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_name #=> String
resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_arn #=> String
resp.media_insights_pipeline_configuration.resource_access_role_arn #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.disabled #=> Boolean
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules #=> Array
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].type #=> String, one of "KeywordMatch", "Sentiment", "IssueDetection"
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.keywords #=> Array
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.keywords[0] #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.negate #=> Boolean
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.sentiment_type #=> String, one of "NEGATIVE"
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.time_period #=> Integer
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].issue_detection_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.elements #=> Array
resp.media_insights_pipeline_configuration.elements[0].type #=> String, one of "AmazonTranscribeCallAnalyticsProcessor", "VoiceAnalyticsProcessor", "AmazonTranscribeProcessor", "KinesisDataStreamSink", "LambdaFunctionSink", "SqsQueueSink", "SnsTopicSink", "S3RecordingSink", "VoiceEnhancementSink"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.language_code #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_filter_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_filter_method #=> String, one of "remove", "mask", "tag"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.language_model_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.enable_partial_results_stabilization #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.partial_results_stability #=> String, one of "high", "medium", "low"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.content_identification_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.content_redaction_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.pii_entity_types #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.filter_partial_results #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.output_location #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.data_access_role_arn #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.content_redaction_output #=> String, one of "redacted", "redacted_and_unredacted"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.output_encryption_kms_key_id #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.call_analytics_stream_categories #=> Array
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.call_analytics_stream_categories[0] #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_code #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_method #=> String, one of "remove", "mask", "tag"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.show_speaker_label #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.enable_partial_results_stabilization #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.partial_results_stability #=> String, one of "high", "medium", "low"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.content_identification_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.content_redaction_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.pii_entity_types #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_model_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.filter_partial_results #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.identify_language #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_options #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.preferred_language #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_names #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_names #=> String
resp.media_insights_pipeline_configuration.elements[0].kinesis_data_stream_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].s3_recording_sink_configuration.destination #=> String
resp.media_insights_pipeline_configuration.elements[0].s3_recording_sink_configuration.recording_file_format #=> String, one of "Wav", "Opus"
resp.media_insights_pipeline_configuration.elements[0].voice_analytics_processor_configuration.speaker_search_status #=> String, one of "Enabled", "Disabled"
resp.media_insights_pipeline_configuration.elements[0].voice_analytics_processor_configuration.voice_tone_analysis_status #=> String, one of "Enabled", "Disabled"
resp.media_insights_pipeline_configuration.elements[0].lambda_function_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].sqs_queue_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].sns_topic_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].voice_enhancement_sink_configuration.disabled #=> Boolean
resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_id #=> String
resp.media_insights_pipeline_configuration.created_timestamp #=> Time
resp.media_insights_pipeline_configuration.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the requested resource. Valid values include the name and ARN of the media insights pipeline configuration.

Returns:

See Also:



1517
1518
1519
1520
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1517

def get_media_insights_pipeline_configuration(params = {}, options = {})
  req = build_request(:get_media_insights_pipeline_configuration, params)
  req.send_request(options)
end

#get_media_pipeline(params = {}) ⇒ Types::GetMediaPipelineResponse

Gets an existing media pipeline.

Examples:

Request syntax with placeholder values


resp = client.get_media_pipeline({
  media_pipeline_id: "GuidString", # required
})

Response structure


resp.media_pipeline.media_capture_pipeline.media_pipeline_id #=> String
resp.media_pipeline.media_capture_pipeline.media_pipeline_arn #=> String
resp.media_pipeline.media_capture_pipeline.source_type #=> String, one of "ChimeSdkMeeting"
resp.media_pipeline.media_capture_pipeline.source_arn #=> String
resp.media_pipeline.media_capture_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_pipeline.media_capture_pipeline.sink_type #=> String, one of "S3Bucket"
resp.media_pipeline.media_capture_pipeline.sink_arn #=> String
resp.media_pipeline.media_capture_pipeline.created_timestamp #=> Time
resp.media_pipeline.media_capture_pipeline.updated_timestamp #=> Time
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.attendee_ids #=> Array
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.attendee_ids[0] #=> String
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.external_user_ids #=> Array
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.source_configuration.selected_video_streams.external_user_ids[0] #=> String
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.audio.mux_type #=> String, one of "AudioOnly", "AudioWithActiveSpeakerVideo", "AudioWithCompositedVideo"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.video.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.video.mux_type #=> String, one of "VideoOnly"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.content.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.content.mux_type #=> String, one of "ContentOnly"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.layout #=> String, one of "GridView"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.resolution #=> String, one of "HD", "FHD"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.content_share_layout #=> String, one of "PresenterOnly", "Horizontal", "Vertical", "ActiveSpeakerOnly"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.presenter_only_configuration.presenter_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.active_speaker_only_configuration.active_speaker_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_position #=> String, one of "Top", "Bottom"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_count #=> Integer
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_aspect_ratio #=> String
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_position #=> String, one of "Left", "Right"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_count #=> Integer
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_aspect_ratio #=> String
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.corner_radius #=> Integer
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.border_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.highlight_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.video_attribute.border_thickness #=> Integer
resp.media_pipeline.media_capture_pipeline.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.grid_view_configuration.canvas_orientation #=> String, one of "Landscape", "Portrait"
resp.media_pipeline.media_live_connector_pipeline.sources #=> Array
resp.media_pipeline.media_live_connector_pipeline.sources[0].source_type #=> String, one of "ChimeSdkMeeting"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.arn #=> String
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.mux_type #=> String, one of "AudioWithCompositedVideo", "AudioWithActiveSpeakerVideo"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.layout #=> String, one of "GridView"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.resolution #=> String, one of "HD", "FHD"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.content_share_layout #=> String, one of "PresenterOnly", "Horizontal", "Vertical", "ActiveSpeakerOnly"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.presenter_only_configuration.presenter_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.active_speaker_only_configuration.active_speaker_position #=> String, one of "TopLeft", "TopRight", "BottomLeft", "BottomRight"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_position #=> String, one of "Top", "Bottom"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_count #=> Integer
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.horizontal_layout_configuration.tile_aspect_ratio #=> String
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_order #=> String, one of "JoinSequence", "SpeakerSequence"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_position #=> String, one of "Left", "Right"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_count #=> Integer
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.vertical_layout_configuration.tile_aspect_ratio #=> String
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.corner_radius #=> Integer
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.border_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.highlight_color #=> String, one of "Black", "Blue", "Red", "Green", "White", "Yellow"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.video_attribute.border_thickness #=> Integer
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.composited_video.grid_view_configuration.canvas_orientation #=> String, one of "Landscape", "Portrait"
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.attendee_ids #=> Array
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.attendee_ids[0] #=> String
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.external_user_ids #=> Array
resp.media_pipeline.media_live_connector_pipeline.sources[0].chime_sdk_meeting_live_connector_configuration.source_configuration.selected_video_streams.external_user_ids[0] #=> String
resp.media_pipeline.media_live_connector_pipeline.sinks #=> Array
resp.media_pipeline.media_live_connector_pipeline.sinks[0].sink_type #=> String, one of "RTMP"
resp.media_pipeline.media_live_connector_pipeline.sinks[0].rtmp_configuration.url #=> String
resp.media_pipeline.media_live_connector_pipeline.sinks[0].rtmp_configuration.audio_channels #=> String, one of "Stereo", "Mono"
resp.media_pipeline.media_live_connector_pipeline.sinks[0].rtmp_configuration.audio_sample_rate #=> String
resp.media_pipeline.media_live_connector_pipeline.media_pipeline_id #=> String
resp.media_pipeline.media_live_connector_pipeline.media_pipeline_arn #=> String
resp.media_pipeline.media_live_connector_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_pipeline.media_live_connector_pipeline.created_timestamp #=> Time
resp.media_pipeline.media_live_connector_pipeline.updated_timestamp #=> Time
resp.media_pipeline.media_concatenation_pipeline.media_pipeline_id #=> String
resp.media_pipeline.media_concatenation_pipeline.media_pipeline_arn #=> String
resp.media_pipeline.media_concatenation_pipeline.sources #=> Array
resp.media_pipeline.media_concatenation_pipeline.sources[0].type #=> String, one of "MediaCapturePipeline"
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.media_pipeline_arn #=> String
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.audio.state #=> String, one of "Enabled"
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.video.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.content.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.data_channel.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.transcription_messages.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.meeting_events.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_concatenation_pipeline.sources[0].media_capture_pipeline_source_configuration.chime_sdk_meeting_configuration.artifacts_configuration.composited_video.state #=> String, one of "Enabled", "Disabled"
resp.media_pipeline.media_concatenation_pipeline.sinks #=> Array
resp.media_pipeline.media_concatenation_pipeline.sinks[0].type #=> String, one of "S3Bucket"
resp.media_pipeline.media_concatenation_pipeline.sinks[0].s3_bucket_sink_configuration.destination #=> String
resp.media_pipeline.media_concatenation_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_pipeline.media_concatenation_pipeline.created_timestamp #=> Time
resp.media_pipeline.media_concatenation_pipeline.updated_timestamp #=> Time
resp.media_pipeline.media_insights_pipeline.media_pipeline_id #=> String
resp.media_pipeline.media_insights_pipeline.media_pipeline_arn #=> String
resp.media_pipeline.media_insights_pipeline.media_insights_pipeline_configuration_arn #=> String
resp.media_pipeline.media_insights_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams #=> Array
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_arn #=> String
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].fragment_number #=> String
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.number_of_channels #=> Integer
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.channel_definitions #=> Array
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.channel_definitions[0].channel_id #=> Integer
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.streams[0].stream_channel_definition.channel_definitions[0].participant_role #=> String, one of "AGENT", "CUSTOMER"
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.media_encoding #=> String, one of "pcm"
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_source_runtime_configuration.media_sample_rate #=> Integer
resp.media_pipeline.media_insights_pipeline. #=> Hash
resp.media_pipeline.media_insights_pipeline.["NonEmptyString"] #=> String
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.streams #=> Array
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.streams[0].stream_arn #=> String
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.fragment_selector.fragment_selector_type #=> String, one of "ProducerTimestamp", "ServerTimestamp"
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.fragment_selector.timestamp_range.start_timestamp #=> Time
resp.media_pipeline.media_insights_pipeline.kinesis_video_stream_recording_source_runtime_configuration.fragment_selector.timestamp_range.end_timestamp #=> Time
resp.media_pipeline.media_insights_pipeline.s3_recording_sink_runtime_configuration.destination #=> String
resp.media_pipeline.media_insights_pipeline.s3_recording_sink_runtime_configuration.recording_file_format #=> String, one of "Wav", "Opus"
resp.media_pipeline.media_insights_pipeline.created_timestamp #=> Time
resp.media_pipeline.media_insights_pipeline.element_statuses #=> Array
resp.media_pipeline.media_insights_pipeline.element_statuses[0].type #=> String, one of "AmazonTranscribeCallAnalyticsProcessor", "VoiceAnalyticsProcessor", "AmazonTranscribeProcessor", "KinesisDataStreamSink", "LambdaFunctionSink", "SqsQueueSink", "SnsTopicSink", "S3RecordingSink", "VoiceEnhancementSink"
resp.media_pipeline.media_insights_pipeline.element_statuses[0].status #=> String, one of "NotStarted", "NotSupported", "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused"
resp.media_pipeline.media_stream_pipeline.media_pipeline_id #=> String
resp.media_pipeline.media_stream_pipeline.media_pipeline_arn #=> String
resp.media_pipeline.media_stream_pipeline.created_timestamp #=> Time
resp.media_pipeline.media_stream_pipeline.updated_timestamp #=> Time
resp.media_pipeline.media_stream_pipeline.status #=> String, one of "Initializing", "InProgress", "Failed", "Stopping", "Stopped", "Paused", "NotStarted"
resp.media_pipeline.media_stream_pipeline.sources #=> Array
resp.media_pipeline.media_stream_pipeline.sources[0].source_type #=> String, one of "ChimeSdkMeeting"
resp.media_pipeline.media_stream_pipeline.sources[0].source_arn #=> String
resp.media_pipeline.media_stream_pipeline.sinks #=> Array
resp.media_pipeline.media_stream_pipeline.sinks[0].sink_arn #=> String
resp.media_pipeline.media_stream_pipeline.sinks[0].sink_type #=> String, one of "KinesisVideoStreamPool"
resp.media_pipeline.media_stream_pipeline.sinks[0].reserved_stream_capacity #=> Integer
resp.media_pipeline.media_stream_pipeline.sinks[0].media_stream_type #=> String, one of "MixedAudio", "IndividualAudio"

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :media_pipeline_id (required, String)

    The ID of the pipeline that you want to get.

Returns:

See Also:



1673
1674
1675
1676
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1673

def get_media_pipeline(params = {}, options = {})
  req = build_request(:get_media_pipeline, params)
  req.send_request(options)
end

#get_media_pipeline_kinesis_video_stream_pool(params = {}) ⇒ Types::GetMediaPipelineKinesisVideoStreamPoolResponse

Gets an Kinesis video stream pool.

Examples:

Request syntax with placeholder values


resp = client.get_media_pipeline_kinesis_video_stream_pool({
  identifier: "NonEmptyString", # required
})

Response structure


resp.kinesis_video_stream_pool_configuration.pool_arn #=> String
resp.kinesis_video_stream_pool_configuration.pool_name #=> String
resp.kinesis_video_stream_pool_configuration.pool_id #=> String
resp.kinesis_video_stream_pool_configuration.pool_status #=> String, one of "CREATING", "ACTIVE", "UPDATING", "DELETING", "FAILED"
resp.kinesis_video_stream_pool_configuration.pool_size #=> Integer
resp.kinesis_video_stream_pool_configuration.stream_configuration.region #=> String
resp.kinesis_video_stream_pool_configuration.stream_configuration.data_retention_in_hours #=> Integer
resp.kinesis_video_stream_pool_configuration.created_timestamp #=> Time
resp.kinesis_video_stream_pool_configuration.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The ID of the video stream pool.

Returns:

See Also:



1709
1710
1711
1712
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1709

def get_media_pipeline_kinesis_video_stream_pool(params = {}, options = {})
  req = build_request(:get_media_pipeline_kinesis_video_stream_pool, params)
  req.send_request(options)
end

#get_speaker_search_task(params = {}) ⇒ Types::GetSpeakerSearchTaskResponse

Retrieves the details of the specified speaker search task.

Examples:

Request syntax with placeholder values


resp = client.get_speaker_search_task({
  identifier: "NonEmptyString", # required
  speaker_search_task_id: "GuidString", # required
})

Response structure


resp.speaker_search_task.speaker_search_task_id #=> String
resp.speaker_search_task.speaker_search_task_status #=> String, one of "NotStarted", "Initializing", "InProgress", "Failed", "Stopping", "Stopped"
resp.speaker_search_task.created_timestamp #=> Time
resp.speaker_search_task.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be updated. Valid values include the ID and ARN of the media insights pipeline.

  • :speaker_search_task_id (required, String)

    The ID of the speaker search task.

Returns:

See Also:



1745
1746
1747
1748
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1745

def get_speaker_search_task(params = {}, options = {})
  req = build_request(:get_speaker_search_task, params)
  req.send_request(options)
end

#get_voice_tone_analysis_task(params = {}) ⇒ Types::GetVoiceToneAnalysisTaskResponse

Retrieves the details of a voice tone analysis task.

Examples:

Request syntax with placeholder values


resp = client.get_voice_tone_analysis_task({
  identifier: "NonEmptyString", # required
  voice_tone_analysis_task_id: "GuidString", # required
})

Response structure


resp.voice_tone_analysis_task.voice_tone_analysis_task_id #=> String
resp.voice_tone_analysis_task.voice_tone_analysis_task_status #=> String, one of "NotStarted", "Initializing", "InProgress", "Failed", "Stopping", "Stopped"
resp.voice_tone_analysis_task.created_timestamp #=> Time
resp.voice_tone_analysis_task.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be updated. Valid values include the ID and ARN of the media insights pipeline.

  • :voice_tone_analysis_task_id (required, String)

    The ID of the voice tone analysis task.

Returns:

See Also:



1781
1782
1783
1784
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1781

def get_voice_tone_analysis_task(params = {}, options = {})
  req = build_request(:get_voice_tone_analysis_task, params)
  req.send_request(options)
end

#list_media_capture_pipelines(params = {}) ⇒ Types::ListMediaCapturePipelinesResponse

Returns a list of media pipelines.

The returned response is a pageable response and is Enumerable. For details on usage see PageableResponse.

Examples:

Request syntax with placeholder values


resp = client.list_media_capture_pipelines({
  next_token: "String",
  max_results: 1,
})

Response structure


resp.media_capture_pipelines #=> Array
resp.media_capture_pipelines[0].media_pipeline_id #=> String
resp.media_capture_pipelines[0].media_pipeline_arn #=> String
resp.next_token #=> String

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :next_token (String)

    The token used to retrieve the next page of results.

  • :max_results (Integer)

    The maximum number of results to return in a single call. Valid Range: 1 - 99.

Returns:

See Also:



1820
1821
1822
1823
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1820

def list_media_capture_pipelines(params = {}, options = {})
  req = build_request(:list_media_capture_pipelines, params)
  req.send_request(options)
end

#list_media_insights_pipeline_configurations(params = {}) ⇒ Types::ListMediaInsightsPipelineConfigurationsResponse

Lists the available media insights pipeline configurations.

The returned response is a pageable response and is Enumerable. For details on usage see PageableResponse.

Examples:

Request syntax with placeholder values


resp = client.list_media_insights_pipeline_configurations({
  next_token: "String",
  max_results: 1,
})

Response structure


resp.media_insights_pipeline_configurations #=> Array
resp.media_insights_pipeline_configurations[0].media_insights_pipeline_configuration_name #=> String
resp.media_insights_pipeline_configurations[0].media_insights_pipeline_configuration_id #=> String
resp.media_insights_pipeline_configurations[0].media_insights_pipeline_configuration_arn #=> String
resp.next_token #=> String

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :next_token (String)

    The token used to return the next page of results.

  • :max_results (Integer)

    The maximum number of results to return in a single call.

Returns:

See Also:



1859
1860
1861
1862
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1859

def list_media_insights_pipeline_configurations(params = {}, options = {})
  req = build_request(:list_media_insights_pipeline_configurations, params)
  req.send_request(options)
end

#list_media_pipeline_kinesis_video_stream_pools(params = {}) ⇒ Types::ListMediaPipelineKinesisVideoStreamPoolsResponse

Lists the video stream pools in the media pipeline.

The returned response is a pageable response and is Enumerable. For details on usage see PageableResponse.

Examples:

Request syntax with placeholder values


resp = client.list_media_pipeline_kinesis_video_stream_pools({
  next_token: "String",
  max_results: 1,
})

Response structure


resp.kinesis_video_stream_pools #=> Array
resp.kinesis_video_stream_pools[0].pool_name #=> String
resp.kinesis_video_stream_pools[0].pool_id #=> String
resp.kinesis_video_stream_pools[0].pool_arn #=> String
resp.next_token #=> String

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :next_token (String)

    The token used to return the next page of results.

  • :max_results (Integer)

    The maximum number of results to return in a single call.

Returns:

See Also:



1898
1899
1900
1901
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1898

def list_media_pipeline_kinesis_video_stream_pools(params = {}, options = {})
  req = build_request(:list_media_pipeline_kinesis_video_stream_pools, params)
  req.send_request(options)
end

#list_media_pipelines(params = {}) ⇒ Types::ListMediaPipelinesResponse

Returns a list of media pipelines.

The returned response is a pageable response and is Enumerable. For details on usage see PageableResponse.

Examples:

Request syntax with placeholder values


resp = client.list_media_pipelines({
  next_token: "String",
  max_results: 1,
})

Response structure


resp.media_pipelines #=> Array
resp.media_pipelines[0].media_pipeline_id #=> String
resp.media_pipelines[0].media_pipeline_arn #=> String
resp.next_token #=> String

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :next_token (String)

    The token used to retrieve the next page of results.

  • :max_results (Integer)

    The maximum number of results to return in a single call. Valid Range: 1 - 99.

Returns:

See Also:



1937
1938
1939
1940
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1937

def list_media_pipelines(params = {}, options = {})
  req = build_request(:list_media_pipelines, params)
  req.send_request(options)
end

#list_tags_for_resource(params = {}) ⇒ Types::ListTagsForResourceResponse

Lists the tags available for a media pipeline.

Examples:

Request syntax with placeholder values


resp = client.list_tags_for_resource({
  resource_arn: "AmazonResourceName", # required
})

Response structure


resp.tags #=> Array
resp.tags[0].key #=> String
resp.tags[0].value #=> String

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :resource_arn (required, String)

    The ARN of the media pipeline associated with any tags. The ARN consists of the pipeline's region, resource ID, and pipeline ID.

Returns:

See Also:



1968
1969
1970
1971
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 1968

def list_tags_for_resource(params = {}, options = {})
  req = build_request(:list_tags_for_resource, params)
  req.send_request(options)
end

#start_speaker_search_task(params = {}) ⇒ Types::StartSpeakerSearchTaskResponse

Starts a speaker search task.

Before starting any speaker search tasks, you must provide all notices and obtain all consents from the speaker as required under applicable privacy and biometrics laws, and as required under the AWS service terms for the Amazon Chime SDK.

Examples:

Request syntax with placeholder values


resp = client.start_speaker_search_task({
  identifier: "NonEmptyString", # required
  voice_profile_domain_arn: "Arn", # required
  kinesis_video_stream_source_task_configuration: {
    stream_arn: "KinesisVideoStreamArn", # required
    channel_id: 1, # required
    fragment_number: "FragmentNumberString",
  },
  client_request_token: "ClientRequestToken",
})

Response structure


resp.speaker_search_task.speaker_search_task_id #=> String
resp.speaker_search_task.speaker_search_task_status #=> String, one of "NotStarted", "Initializing", "InProgress", "Failed", "Stopping", "Stopped"
resp.speaker_search_task.created_timestamp #=> Time
resp.speaker_search_task.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be updated. Valid values include the ID and ARN of the media insights pipeline.

  • :voice_profile_domain_arn (required, String)

    The ARN of the voice profile domain that will store the voice profile.

  • :kinesis_video_stream_source_task_configuration (Types::KinesisVideoStreamSourceTaskConfiguration)

    The task configuration for the Kinesis video stream source of the media insights pipeline.

  • :client_request_token (String)

    The unique identifier for the client request. Use a different token for different speaker search tasks.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

Returns:

See Also:



2030
2031
2032
2033
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2030

def start_speaker_search_task(params = {}, options = {})
  req = build_request(:start_speaker_search_task, params)
  req.send_request(options)
end

#start_voice_tone_analysis_task(params = {}) ⇒ Types::StartVoiceToneAnalysisTaskResponse

Starts a voice tone analysis task. For more information about voice tone analysis, see Using Amazon Chime SDK voice analytics in the Amazon Chime SDK Developer Guide.

Before starting any voice tone analysis tasks, you must provide all notices and obtain all consents from the speaker as required under applicable privacy and biometrics laws, and as required under the AWS service terms for the Amazon Chime SDK.

Examples:

Request syntax with placeholder values


resp = client.start_voice_tone_analysis_task({
  identifier: "NonEmptyString", # required
  language_code: "en-US", # required, accepts en-US
  kinesis_video_stream_source_task_configuration: {
    stream_arn: "KinesisVideoStreamArn", # required
    channel_id: 1, # required
    fragment_number: "FragmentNumberString",
  },
  client_request_token: "ClientRequestToken",
})

Response structure


resp.voice_tone_analysis_task.voice_tone_analysis_task_id #=> String
resp.voice_tone_analysis_task.voice_tone_analysis_task_status #=> String, one of "NotStarted", "Initializing", "InProgress", "Failed", "Stopping", "Stopped"
resp.voice_tone_analysis_task.created_timestamp #=> Time
resp.voice_tone_analysis_task.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be updated. Valid values include the ID and ARN of the media insights pipeline.

  • :language_code (required, String)

    The language code.

  • :kinesis_video_stream_source_task_configuration (Types::KinesisVideoStreamSourceTaskConfiguration)

    The task configuration for the Kinesis video stream source of the media insights pipeline.

  • :client_request_token (String)

    The unique identifier for the client request. Use a different token for different voice tone analysis tasks.

    A suitable default value is auto-generated. You should normally not need to pass this option.**

Returns:

See Also:



2095
2096
2097
2098
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2095

def start_voice_tone_analysis_task(params = {}, options = {})
  req = build_request(:start_voice_tone_analysis_task, params)
  req.send_request(options)
end

#stop_speaker_search_task(params = {}) ⇒ Struct

Stops a speaker search task.

Examples:

Request syntax with placeholder values


resp = client.stop_speaker_search_task({
  identifier: "NonEmptyString", # required
  speaker_search_task_id: "GuidString", # required
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be updated. Valid values include the ID and ARN of the media insights pipeline.

  • :speaker_search_task_id (required, String)

    The speaker search task ID.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



2122
2123
2124
2125
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2122

def stop_speaker_search_task(params = {}, options = {})
  req = build_request(:stop_speaker_search_task, params)
  req.send_request(options)
end

#stop_voice_tone_analysis_task(params = {}) ⇒ Struct

Stops a voice tone analysis task.

Examples:

Request syntax with placeholder values


resp = client.stop_voice_tone_analysis_task({
  identifier: "NonEmptyString", # required
  voice_tone_analysis_task_id: "GuidString", # required
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be updated. Valid values include the ID and ARN of the media insights pipeline.

  • :voice_tone_analysis_task_id (required, String)

    The ID of the voice tone analysis task.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



2149
2150
2151
2152
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2149

def stop_voice_tone_analysis_task(params = {}, options = {})
  req = build_request(:stop_voice_tone_analysis_task, params)
  req.send_request(options)
end

#tag_resource(params = {}) ⇒ Struct

The ARN of the media pipeline that you want to tag. Consists of the pipeline's endpoint region, resource ID, and pipeline ID.

Examples:

Request syntax with placeholder values


resp = client.tag_resource({
  resource_arn: "AmazonResourceName", # required
  tags: [ # required
    {
      key: "TagKey", # required
      value: "TagValue", # required
    },
  ],
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :resource_arn (required, String)

    The ARN of the media pipeline associated with any tags. The ARN consists of the pipeline's endpoint region, resource ID, and pipeline ID.

  • :tags (required, Array<Types::Tag>)

    The tags associated with the specified media pipeline.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



2183
2184
2185
2186
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2183

def tag_resource(params = {}, options = {})
  req = build_request(:tag_resource, params)
  req.send_request(options)
end

#untag_resource(params = {}) ⇒ Struct

Removes any tags from a media pipeline.

Examples:

Request syntax with placeholder values


resp = client.untag_resource({
  resource_arn: "AmazonResourceName", # required
  tag_keys: ["TagKey"], # required
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :resource_arn (required, String)

    The ARN of the pipeline that you want to untag.

  • :tag_keys (required, Array<String>)

    The key/value pairs in the tag that you want to remove.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



2209
2210
2211
2212
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2209

def untag_resource(params = {}, options = {})
  req = build_request(:untag_resource, params)
  req.send_request(options)
end

#update_media_insights_pipeline_configuration(params = {}) ⇒ Types::UpdateMediaInsightsPipelineConfigurationResponse

Updates the media insights pipeline's configuration settings.

Examples:

Request syntax with placeholder values


resp = client.update_media_insights_pipeline_configuration({
  identifier: "NonEmptyString", # required
  resource_access_role_arn: "Arn", # required
  real_time_alert_configuration: {
    disabled: false,
    rules: [
      {
        type: "KeywordMatch", # required, accepts KeywordMatch, Sentiment, IssueDetection
        keyword_match_configuration: {
          rule_name: "RuleName", # required
          keywords: ["Keyword"], # required
          negate: false,
        },
        sentiment_configuration: {
          rule_name: "RuleName", # required
          sentiment_type: "NEGATIVE", # required, accepts NEGATIVE
          time_period: 1, # required
        },
        issue_detection_configuration: {
          rule_name: "RuleName", # required
        },
      },
    ],
  },
  elements: [ # required
    {
      type: "AmazonTranscribeCallAnalyticsProcessor", # required, accepts AmazonTranscribeCallAnalyticsProcessor, VoiceAnalyticsProcessor, AmazonTranscribeProcessor, KinesisDataStreamSink, LambdaFunctionSink, SqsQueueSink, SnsTopicSink, S3RecordingSink, VoiceEnhancementSink
      amazon_transcribe_call_analytics_processor_configuration: {
        language_code: "en-US", # required, accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR
        vocabulary_name: "VocabularyName",
        vocabulary_filter_name: "VocabularyFilterName",
        vocabulary_filter_method: "remove", # accepts remove, mask, tag
        language_model_name: "ModelName",
        enable_partial_results_stabilization: false,
        partial_results_stability: "high", # accepts high, medium, low
        content_identification_type: "PII", # accepts PII
        content_redaction_type: "PII", # accepts PII
        pii_entity_types: "PiiEntityTypes",
        filter_partial_results: false,
        post_call_analytics_settings: {
          output_location: "String", # required
          data_access_role_arn: "String", # required
          content_redaction_output: "redacted", # accepts redacted, redacted_and_unredacted
          output_encryption_kms_key_id: "String",
        },
        call_analytics_stream_categories: ["CategoryName"],
      },
      amazon_transcribe_processor_configuration: {
        language_code: "en-US", # accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR
        vocabulary_name: "VocabularyName",
        vocabulary_filter_name: "VocabularyFilterName",
        vocabulary_filter_method: "remove", # accepts remove, mask, tag
        show_speaker_label: false,
        enable_partial_results_stabilization: false,
        partial_results_stability: "high", # accepts high, medium, low
        content_identification_type: "PII", # accepts PII
        content_redaction_type: "PII", # accepts PII
        pii_entity_types: "PiiEntityTypes",
        language_model_name: "ModelName",
        filter_partial_results: false,
        identify_language: false,
        language_options: "LanguageOptions",
        preferred_language: "en-US", # accepts en-US, en-GB, es-US, fr-CA, fr-FR, en-AU, it-IT, de-DE, pt-BR
        vocabulary_names: "VocabularyNames",
        vocabulary_filter_names: "VocabularyFilterNames",
      },
      kinesis_data_stream_sink_configuration: {
        insights_target: "Arn",
      },
      s3_recording_sink_configuration: {
        destination: "Arn",
        recording_file_format: "Wav", # accepts Wav, Opus
      },
      voice_analytics_processor_configuration: {
        speaker_search_status: "Enabled", # accepts Enabled, Disabled
        voice_tone_analysis_status: "Enabled", # accepts Enabled, Disabled
      },
      lambda_function_sink_configuration: {
        insights_target: "Arn",
      },
      sqs_queue_sink_configuration: {
        insights_target: "Arn",
      },
      sns_topic_sink_configuration: {
        insights_target: "Arn",
      },
      voice_enhancement_sink_configuration: {
        disabled: false,
      },
    },
  ],
})

Response structure


resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_name #=> String
resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_arn #=> String
resp.media_insights_pipeline_configuration.resource_access_role_arn #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.disabled #=> Boolean
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules #=> Array
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].type #=> String, one of "KeywordMatch", "Sentiment", "IssueDetection"
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.keywords #=> Array
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.keywords[0] #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].keyword_match_configuration.negate #=> Boolean
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.sentiment_type #=> String, one of "NEGATIVE"
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].sentiment_configuration.time_period #=> Integer
resp.media_insights_pipeline_configuration.real_time_alert_configuration.rules[0].issue_detection_configuration.rule_name #=> String
resp.media_insights_pipeline_configuration.elements #=> Array
resp.media_insights_pipeline_configuration.elements[0].type #=> String, one of "AmazonTranscribeCallAnalyticsProcessor", "VoiceAnalyticsProcessor", "AmazonTranscribeProcessor", "KinesisDataStreamSink", "LambdaFunctionSink", "SqsQueueSink", "SnsTopicSink", "S3RecordingSink", "VoiceEnhancementSink"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.language_code #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_filter_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.vocabulary_filter_method #=> String, one of "remove", "mask", "tag"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.language_model_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.enable_partial_results_stabilization #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.partial_results_stability #=> String, one of "high", "medium", "low"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.content_identification_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.content_redaction_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.pii_entity_types #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.filter_partial_results #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.output_location #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.data_access_role_arn #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.content_redaction_output #=> String, one of "redacted", "redacted_and_unredacted"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.post_call_analytics_settings.output_encryption_kms_key_id #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.call_analytics_stream_categories #=> Array
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_call_analytics_processor_configuration.call_analytics_stream_categories[0] #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_code #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_method #=> String, one of "remove", "mask", "tag"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.show_speaker_label #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.enable_partial_results_stabilization #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.partial_results_stability #=> String, one of "high", "medium", "low"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.content_identification_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.content_redaction_type #=> String, one of "PII"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.pii_entity_types #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_model_name #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.filter_partial_results #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.identify_language #=> Boolean
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.language_options #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.preferred_language #=> String, one of "en-US", "en-GB", "es-US", "fr-CA", "fr-FR", "en-AU", "it-IT", "de-DE", "pt-BR"
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_names #=> String
resp.media_insights_pipeline_configuration.elements[0].amazon_transcribe_processor_configuration.vocabulary_filter_names #=> String
resp.media_insights_pipeline_configuration.elements[0].kinesis_data_stream_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].s3_recording_sink_configuration.destination #=> String
resp.media_insights_pipeline_configuration.elements[0].s3_recording_sink_configuration.recording_file_format #=> String, one of "Wav", "Opus"
resp.media_insights_pipeline_configuration.elements[0].voice_analytics_processor_configuration.speaker_search_status #=> String, one of "Enabled", "Disabled"
resp.media_insights_pipeline_configuration.elements[0].voice_analytics_processor_configuration.voice_tone_analysis_status #=> String, one of "Enabled", "Disabled"
resp.media_insights_pipeline_configuration.elements[0].lambda_function_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].sqs_queue_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].sns_topic_sink_configuration.insights_target #=> String
resp.media_insights_pipeline_configuration.elements[0].voice_enhancement_sink_configuration.disabled #=> Boolean
resp.media_insights_pipeline_configuration.media_insights_pipeline_configuration_id #=> String
resp.media_insights_pipeline_configuration.created_timestamp #=> Time
resp.media_insights_pipeline_configuration.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier for the resource to be updated. Valid values include the name and ARN of the media insights pipeline configuration.

  • :resource_access_role_arn (required, String)

    The ARN of the role used by the service to access Amazon Web Services resources.

  • :real_time_alert_configuration (Types::RealTimeAlertConfiguration)

    The configuration settings for real-time alerts for the media insights pipeline.

  • :elements (required, Array<Types::MediaInsightsPipelineConfigurationElement>)

    The elements in the request, such as a processor for Amazon Transcribe or a sink for a Kinesis Data Stream..

Returns:

See Also:



2400
2401
2402
2403
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2400

def update_media_insights_pipeline_configuration(params = {}, options = {})
  req = build_request(:update_media_insights_pipeline_configuration, params)
  req.send_request(options)
end

#update_media_insights_pipeline_status(params = {}) ⇒ Struct

Updates the status of a media insights pipeline.

Examples:

Request syntax with placeholder values


resp = client.update_media_insights_pipeline_status({
  identifier: "NonEmptyString", # required
  update_status: "Pause", # required, accepts Pause, Resume
})

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

  • :identifier (required, String)

    The unique identifier of the resource to be updated. Valid values include the ID and ARN of the media insights pipeline.

  • :update_status (required, String)

    The requested status of the media insights pipeline.

Returns:

  • (Struct)

    Returns an empty response.

See Also:



2427
2428
2429
2430
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2427

def update_media_insights_pipeline_status(params = {}, options = {})
  req = build_request(:update_media_insights_pipeline_status, params)
  req.send_request(options)
end

#update_media_pipeline_kinesis_video_stream_pool(params = {}) ⇒ Types::UpdateMediaPipelineKinesisVideoStreamPoolResponse

Updates an Kinesis video stream pool in a media pipeline.

Examples:

Request syntax with placeholder values


resp = client.update_media_pipeline_kinesis_video_stream_pool({
  identifier: "NonEmptyString", # required
  stream_configuration: {
    data_retention_in_hours: 1,
  },
})

Response structure


resp.kinesis_video_stream_pool_configuration.pool_arn #=> String
resp.kinesis_video_stream_pool_configuration.pool_name #=> String
resp.kinesis_video_stream_pool_configuration.pool_id #=> String
resp.kinesis_video_stream_pool_configuration.pool_status #=> String, one of "CREATING", "ACTIVE", "UPDATING", "DELETING", "FAILED"
resp.kinesis_video_stream_pool_configuration.pool_size #=> Integer
resp.kinesis_video_stream_pool_configuration.stream_configuration.region #=> String
resp.kinesis_video_stream_pool_configuration.stream_configuration.data_retention_in_hours #=> Integer
resp.kinesis_video_stream_pool_configuration.created_timestamp #=> Time
resp.kinesis_video_stream_pool_configuration.updated_timestamp #=> Time

Parameters:

  • params (Hash) (defaults to: {})

    ({})

Options Hash (params):

Returns:

See Also:



2469
2470
2471
2472
# File 'gems/aws-sdk-chimesdkmediapipelines/lib/aws-sdk-chimesdkmediapipelines/client.rb', line 2469

def update_media_pipeline_kinesis_video_stream_pool(params = {}, options = {})
  req = build_request(:update_media_pipeline_kinesis_video_stream_pool, params)
  req.send_request(options)
end