Module Smaws_Client_Kinesis

Kinesis client library built on EIO.

Types

module Types : sig ... end

Builders

val make_stream_mode_details : stream_mode:Types.stream_mode -> unit -> Types.stream_mode_details
val make_update_stream_mode_input : stream_mode_details:Types.stream_mode_details -> stream_ar_n:Types.stream_ar_n -> unit -> Types.update_stream_mode_input
val make_update_shard_count_output : ?stream_ar_n:Types.stream_ar_n -> ?target_shard_count:Types.positive_integer_object -> ?current_shard_count:Types.positive_integer_object -> ?stream_name:Types.stream_name -> unit -> Types.update_shard_count_output
val make_update_shard_count_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> scaling_type:Types.scaling_type -> target_shard_count:Types.positive_integer_object -> unit -> Types.update_shard_count_input
val make_untag_resource_input : resource_ar_n:Types.resource_ar_n -> tag_keys:Types.tag_key_list -> unit -> Types.untag_resource_input
val make_tag_resource_input : resource_ar_n:Types.resource_ar_n -> tags:Types.tag_map -> unit -> Types.tag_resource_input
val make_tag : ?value:Types.tag_value -> key:Types.tag_key -> unit -> Types.tag
val make_record : ?encryption_type:Types.encryption_type -> ?approximate_arrival_timestamp:Types.timestamp -> partition_key:Types.partition_key -> data:Types.data -> sequence_number:Types.sequence_number -> unit -> Types.record
val make_hash_key_range : ending_hash_key:Types.hash_key -> starting_hash_key:Types.hash_key -> unit -> Types.hash_key_range
val make_child_shard : hash_key_range:Types.hash_key_range -> parent_shards:Types.shard_id_list -> shard_id:Types.shard_id -> unit -> Types.child_shard
val make_subscribe_to_shard_event : ?child_shards:Types.child_shard_list -> millis_behind_latest:Types.millis_behind_latest -> continuation_sequence_number:Types.sequence_number -> records:Types.record_list -> unit -> Types.subscribe_to_shard_event
val make_subscribe_to_shard_output : event_stream:Types.subscribe_to_shard_event_stream -> unit -> Types.subscribe_to_shard_output
val make_starting_position : ?timestamp:Types.timestamp -> ?sequence_number:Types.sequence_number -> type_:Types.shard_iterator_type -> unit -> Types.starting_position
val make_subscribe_to_shard_input : starting_position:Types.starting_position -> shard_id:Types.shard_id -> consumer_ar_n:Types.consumer_ar_n -> unit -> Types.subscribe_to_shard_input
val make_stream_summary : ?stream_creation_timestamp:Types.timestamp -> ?stream_mode_details:Types.stream_mode_details -> stream_status:Types.stream_status -> stream_ar_n:Types.stream_ar_n -> stream_name:Types.stream_name -> unit -> Types.stream_summary
val make_enhanced_metrics : ?shard_level_metrics:Types.metrics_name_list -> unit -> Types.enhanced_metrics
val make_stream_description_summary : ?consumer_count:Types.consumer_count_object -> ?key_id:Types.key_id -> ?encryption_type:Types.encryption_type -> ?stream_mode_details:Types.stream_mode_details -> open_shard_count:Types.shard_count_object -> enhanced_monitoring:Types.enhanced_monitoring_list -> stream_creation_timestamp:Types.timestamp -> retention_period_hours:Types.retention_period_hours -> stream_status:Types.stream_status -> stream_ar_n:Types.stream_ar_n -> stream_name:Types.stream_name -> unit -> Types.stream_description_summary
val make_sequence_number_range : ?ending_sequence_number:Types.sequence_number -> starting_sequence_number:Types.sequence_number -> unit -> Types.sequence_number_range
val make_shard : ?adjacent_parent_shard_id:Types.shard_id -> ?parent_shard_id:Types.shard_id -> sequence_number_range:Types.sequence_number_range -> hash_key_range:Types.hash_key_range -> shard_id:Types.shard_id -> unit -> Types.shard
val make_stream_description : ?key_id:Types.key_id -> ?encryption_type:Types.encryption_type -> ?stream_mode_details:Types.stream_mode_details -> enhanced_monitoring:Types.enhanced_monitoring_list -> stream_creation_timestamp:Types.timestamp -> retention_period_hours:Types.retention_period_hours -> has_more_shards:Types.boolean_object -> shards:Types.shard_list -> stream_status:Types.stream_status -> stream_ar_n:Types.stream_ar_n -> stream_name:Types.stream_name -> unit -> Types.stream_description
val make_stop_stream_encryption_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> key_id:Types.key_id -> encryption_type:Types.encryption_type -> unit -> Types.stop_stream_encryption_input
val make_start_stream_encryption_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> key_id:Types.key_id -> encryption_type:Types.encryption_type -> unit -> Types.start_stream_encryption_input
val make_split_shard_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> new_starting_hash_key:Types.hash_key -> shard_to_split:Types.shard_id -> unit -> Types.split_shard_input
val make_shard_filter : ?timestamp:Types.timestamp -> ?shard_id:Types.shard_id -> type_:Types.shard_filter_type -> unit -> Types.shard_filter
val make_remove_tags_from_stream_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> tag_keys:Types.tag_key_list -> unit -> Types.remove_tags_from_stream_input
val make_consumer : consumer_creation_timestamp:Types.timestamp -> consumer_status:Types.consumer_status -> consumer_ar_n:Types.consumer_ar_n -> consumer_name:Types.consumer_name -> unit -> Types.consumer
val make_register_stream_consumer_output : consumer:Types.consumer -> unit -> Types.register_stream_consumer_output
val make_register_stream_consumer_input : ?tags:Types.tag_map -> consumer_name:Types.consumer_name -> stream_ar_n:Types.stream_ar_n -> unit -> Types.register_stream_consumer_input
val make_put_resource_policy_input : policy:Types.policy -> resource_ar_n:Types.resource_ar_n -> unit -> Types.put_resource_policy_input
val make_put_records_result_entry : ?error_message:Types.error_message -> ?error_code:Types.error_code -> ?shard_id:Types.shard_id -> ?sequence_number:Types.sequence_number -> unit -> Types.put_records_result_entry
val make_put_records_request_entry : ?explicit_hash_key:Types.hash_key -> partition_key:Types.partition_key -> data:Types.data -> unit -> Types.put_records_request_entry
val make_put_records_output : ?encryption_type:Types.encryption_type -> ?failed_record_count:Types.positive_integer_object -> records:Types.put_records_result_entry_list -> unit -> Types.put_records_output
val make_put_records_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> records:Types.put_records_request_entry_list -> unit -> Types.put_records_input
val make_put_record_output : ?encryption_type:Types.encryption_type -> sequence_number:Types.sequence_number -> shard_id:Types.shard_id -> unit -> Types.put_record_output
val make_put_record_input : ?stream_ar_n:Types.stream_ar_n -> ?sequence_number_for_ordering:Types.sequence_number -> ?explicit_hash_key:Types.hash_key -> ?stream_name:Types.stream_name -> partition_key:Types.partition_key -> data:Types.data -> unit -> Types.put_record_input
val make_merge_shards_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> adjacent_shard_to_merge:Types.shard_id -> shard_to_merge:Types.shard_id -> unit -> Types.merge_shards_input
val make_list_tags_for_stream_output : has_more_tags:Types.boolean_object -> tags:Types.tag_list -> unit -> Types.list_tags_for_stream_output
val make_list_tags_for_stream_input : ?stream_ar_n:Types.stream_ar_n -> ?limit:Types.list_tags_for_stream_input_limit -> ?exclusive_start_tag_key:Types.tag_key -> ?stream_name:Types.stream_name -> unit -> Types.list_tags_for_stream_input
val make_list_tags_for_resource_output : ?tags:Types.tag_list -> unit -> Types.list_tags_for_resource_output
val make_list_tags_for_resource_input : resource_ar_n:Types.resource_ar_n -> unit -> Types.list_tags_for_resource_input
val make_list_streams_output : ?stream_summaries:Types.stream_summary_list -> ?next_token:Types.next_token -> has_more_streams:Types.boolean_object -> stream_names:Types.stream_name_list -> unit -> Types.list_streams_output
val make_list_streams_input : ?next_token:Types.next_token -> ?exclusive_start_stream_name:Types.stream_name -> ?limit:Types.list_streams_input_limit -> unit -> Types.list_streams_input
val make_list_stream_consumers_output : ?next_token:Types.next_token -> ?consumers:Types.consumer_list -> unit -> Types.list_stream_consumers_output
val make_list_stream_consumers_input : ?stream_creation_timestamp:Types.timestamp -> ?max_results:Types.list_stream_consumers_input_limit -> ?next_token:Types.next_token -> stream_ar_n:Types.stream_ar_n -> unit -> Types.list_stream_consumers_input
val make_list_shards_output : ?next_token:Types.next_token -> ?shards:Types.shard_list -> unit -> Types.list_shards_output
val make_list_shards_input : ?stream_ar_n:Types.stream_ar_n -> ?shard_filter:Types.shard_filter -> ?stream_creation_timestamp:Types.timestamp -> ?max_results:Types.list_shards_input_limit -> ?exclusive_start_shard_id:Types.shard_id -> ?next_token:Types.next_token -> ?stream_name:Types.stream_name -> unit -> Types.list_shards_input
val make_increase_stream_retention_period_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> retention_period_hours:Types.retention_period_hours -> unit -> Types.increase_stream_retention_period_input
val make_get_shard_iterator_output : ?shard_iterator:Types.shard_iterator -> unit -> Types.get_shard_iterator_output
val make_get_shard_iterator_input : ?stream_ar_n:Types.stream_ar_n -> ?timestamp:Types.timestamp -> ?starting_sequence_number:Types.sequence_number -> ?stream_name:Types.stream_name -> shard_iterator_type:Types.shard_iterator_type -> shard_id:Types.shard_id -> unit -> Types.get_shard_iterator_input
val make_get_resource_policy_output : policy:Types.policy -> unit -> Types.get_resource_policy_output
val make_get_resource_policy_input : resource_ar_n:Types.resource_ar_n -> unit -> Types.get_resource_policy_input
val make_get_records_output : ?child_shards:Types.child_shard_list -> ?millis_behind_latest:Types.millis_behind_latest -> ?next_shard_iterator:Types.shard_iterator -> records:Types.record_list -> unit -> Types.get_records_output
val make_get_records_input : ?stream_ar_n:Types.stream_ar_n -> ?limit:Types.get_records_input_limit -> shard_iterator:Types.shard_iterator -> unit -> Types.get_records_input
val make_enhanced_monitoring_output : ?stream_ar_n:Types.stream_ar_n -> ?desired_shard_level_metrics:Types.metrics_name_list -> ?current_shard_level_metrics:Types.metrics_name_list -> ?stream_name:Types.stream_name -> unit -> Types.enhanced_monitoring_output
val make_enable_enhanced_monitoring_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> shard_level_metrics:Types.metrics_name_list -> unit -> Types.enable_enhanced_monitoring_input
val make_disable_enhanced_monitoring_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> shard_level_metrics:Types.metrics_name_list -> unit -> Types.disable_enhanced_monitoring_input
val make_describe_stream_summary_output : stream_description_summary:Types.stream_description_summary -> unit -> Types.describe_stream_summary_output
val make_describe_stream_summary_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> unit -> Types.describe_stream_summary_input
val make_consumer_description : stream_ar_n:Types.stream_ar_n -> consumer_creation_timestamp:Types.timestamp -> consumer_status:Types.consumer_status -> consumer_ar_n:Types.consumer_ar_n -> consumer_name:Types.consumer_name -> unit -> Types.consumer_description
val make_describe_stream_consumer_output : consumer_description:Types.consumer_description -> unit -> Types.describe_stream_consumer_output
val make_describe_stream_consumer_input : ?consumer_ar_n:Types.consumer_ar_n -> ?consumer_name:Types.consumer_name -> ?stream_ar_n:Types.stream_ar_n -> unit -> Types.describe_stream_consumer_input
val make_describe_stream_output : stream_description:Types.stream_description -> unit -> Types.describe_stream_output
val make_describe_stream_input : ?stream_ar_n:Types.stream_ar_n -> ?exclusive_start_shard_id:Types.shard_id -> ?limit:Types.describe_stream_input_limit -> ?stream_name:Types.stream_name -> unit -> Types.describe_stream_input
val make_describe_limits_output : on_demand_stream_count_limit:Types.on_demand_stream_count_limit_object -> on_demand_stream_count:Types.on_demand_stream_count_object -> open_shard_count:Types.shard_count_object -> shard_limit:Types.shard_count_object -> unit -> Types.describe_limits_output
val make_describe_limits_input : unit -> unit
val make_deregister_stream_consumer_input : ?consumer_ar_n:Types.consumer_ar_n -> ?consumer_name:Types.consumer_name -> ?stream_ar_n:Types.stream_ar_n -> unit -> Types.deregister_stream_consumer_input
val make_delete_stream_input : ?stream_ar_n:Types.stream_ar_n -> ?enforce_consumer_deletion:Types.boolean_object -> ?stream_name:Types.stream_name -> unit -> Types.delete_stream_input
val make_delete_resource_policy_input : resource_ar_n:Types.resource_ar_n -> unit -> Types.delete_resource_policy_input
val make_decrease_stream_retention_period_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> retention_period_hours:Types.retention_period_hours -> unit -> Types.decrease_stream_retention_period_input
val make_create_stream_input : ?tags:Types.tag_map -> ?stream_mode_details:Types.stream_mode_details -> ?shard_count:Types.positive_integer_object -> stream_name:Types.stream_name -> unit -> Types.create_stream_input
val make_add_tags_to_stream_input : ?stream_ar_n:Types.stream_ar_n -> ?stream_name:Types.stream_name -> tags:Types.tag_map -> unit -> Types.add_tags_to_stream_input

Operations

module AddTagsToStream : sig ... end

Adds or updates tags for the specified Kinesis data stream. You can assign up to 50 tags to a data stream.

module CreateStream : sig ... end

Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from different data sources or producers. Scale-out within a stream is explicitly supported by means of shards, which are uniquely identified groups of data records in a stream.

module DecreaseStreamRetentionPeriod : sig ... end

Decreases the Kinesis data stream's retention period, which is the length of time data records are accessible after they are added to the stream. The minimum value of a stream's retention period is 24 hours.

module DeleteResourcePolicy : sig ... end

Delete a policy for the specified data stream or consumer. Request patterns can be one of the following:

module DeleteStream : sig ... end

Deletes a Kinesis data stream and all its shards and data. You must shut down any applications that are operating on the stream before you delete the stream. If an application attempts to operate on a deleted stream, it receives the exception ResourceNotFoundException.

module DeregisterStreamConsumer : sig ... end

To deregister a consumer, provide its ARN. Alternatively, you can provide the ARN of the data stream and the name you gave the consumer when you registered it. You may also provide all three parameters, as long as they don't conflict with each other. If you don't know the name or ARN of the consumer that you want to deregister, you can use the ListStreamConsumers operation to get a list of the descriptions of all the consumers that are currently registered with a given data stream. The description of a consumer contains its name and ARN.

module DescribeLimits : sig ... end

Describes the shard limits and usage for the account.

module DescribeStream : sig ... end

Describes the specified Kinesis data stream.

module DescribeStreamConsumer : sig ... end

To get the description of a registered consumer, provide the ARN of the consumer. Alternatively, you can provide the ARN of the data stream and the name you gave the consumer when you registered it. You may also provide all three parameters, as long as they don't conflict with each other. If you don't know the name or ARN of the consumer that you want to describe, you can use the ListStreamConsumers operation to get a list of the descriptions of all the consumers that are currently registered with a given data stream.

module DescribeStreamSummary : sig ... end

Provides a summarized description of the specified Kinesis data stream without the shard list.

module DisableEnhancedMonitoring : sig ... end

Disables enhanced monitoring.

module EnableEnhancedMonitoring : sig ... end

Enables enhanced Kinesis data stream monitoring for shard-level metrics.

module GetRecords : sig ... end

Gets data records from a Kinesis data stream's shard.

module GetResourcePolicy : sig ... end

Returns a policy attached to the specified data stream or consumer. Request patterns can be one of the following:

module GetShardIterator : sig ... end

Gets an Amazon Kinesis shard iterator. A shard iterator expires 5 minutes after it is returned to the requester.

module IncreaseStreamRetentionPeriod : sig ... end

Increases the Kinesis data stream's retention period, which is the length of time data records are accessible after they are added to the stream. The maximum value of a stream's retention period is 8760 hours (365 days).

module ListShards : sig ... end

Lists the shards in a stream and provides information about each shard. This operation has a limit of 1000 transactions per second per data stream.

module ListStreamConsumers : sig ... end

Lists the consumers registered to receive data from a stream using enhanced fan-out, and provides information about each consumer.

module ListStreams : sig ... end

Lists your Kinesis data streams.

module ListTagsForResource : sig ... end

List all tags added to the specified Kinesis resource. Each tag is a label consisting of a user-defined key and value. Tags can help you manage, identify, organize, search for, and filter resources.

module ListTagsForStream : sig ... end

Lists the tags for the specified Kinesis data stream. This operation has a limit of five transactions per second per account.

module MergeShards : sig ... end

Merges two adjacent shards in a Kinesis data stream and combines them into a single shard to reduce the stream's capacity to ingest and transport data. This API is only supported for the data streams with the provisioned capacity mode. Two shards are considered adjacent if the union of the hash key ranges for the two shards form a contiguous set with no gaps. For example, if you have two shards, one with a hash key range of 276...381 and the other with a hash key range of 382...454, then you could merge these two shards into a single shard that would have a hash key range of 276...454. After the merge, the single child shard receives data for all hash key values covered by the two parent shards.

module PutRecord : sig ... end

Writes a single data record into an Amazon Kinesis data stream. Call PutRecord to send data into the stream for real-time ingestion and subsequent processing, one record at a time. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second.

module PutRecords : sig ... end

Writes multiple data records into a Kinesis data stream in a single call (also referred to as a PutRecords request). Use this operation to send data into the stream for data ingestion and processing.

module PutResourcePolicy : sig ... end

Attaches a resource-based policy to a data stream or registered consumer. If you are using an identity other than the root user of the Amazon Web Services account that owns the resource, the calling identity must have the PutResourcePolicy permissions on the specified Kinesis Data Streams resource and belong to the owner's account in order to use this operation. If you don't have PutResourcePolicy permissions, Amazon Kinesis Data Streams returns a 403 Access Denied error. If you receive a ResourceNotFoundException, check to see if you passed a valid stream or consumer resource.

module RegisterStreamConsumer : sig ... end

Registers a consumer with a Kinesis data stream. When you use this operation, the consumer you register can then call SubscribeToShard to receive data from the stream using enhanced fan-out, at a rate of up to 2 MiB per second for every shard you subscribe to. This rate is unaffected by the total number of consumers that read from the same stream.

module RemoveTagsFromStream : sig ... end

Removes tags from the specified Kinesis data stream. Removed tags are deleted and cannot be recovered after this operation successfully completes.

module SplitShard : sig ... end

Splits a shard into two new shards in the Kinesis data stream, to increase the stream's capacity to ingest and transport data. SplitShard is called when there is a need to increase the overall capacity of a stream because of an expected increase in the volume of data records being ingested. This API is only supported for the data streams with the provisioned capacity mode.

module StartStreamEncryption : sig ... end

Enables or updates server-side encryption using an Amazon Web Services KMS key for a specified stream.

module StopStreamEncryption : sig ... end

Disables server-side encryption for a specified stream.

module SubscribeToShard : sig ... end

This operation establishes an HTTP/2 connection between the consumer you specify in the ConsumerARN parameter and the shard you specify in the ShardId parameter. After the connection is successfully established, Kinesis Data Streams pushes records from the shard to the consumer over this connection. Before you call this operation, call RegisterStreamConsumer to register the consumer with Kinesis Data Streams.

module TagResource : sig ... end

Adds or updates tags for the specified Kinesis resource. Each tag is a label consisting of a user-defined key and value. Tags can help you manage, identify, organize, search for, and filter resources. You can assign up to 50 tags to a Kinesis resource.

module UntagResource : sig ... end

Removes tags from the specified Kinesis resource. Removed tags are deleted and can't be recovered after this operation completes successfully.

module UpdateShardCount : sig ... end

Updates the shard count of the specified stream to the specified number of shards. This API is only supported for the data streams with the provisioned capacity mode.

module UpdateStreamMode : sig ... end

Updates the capacity mode of the data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data stream.

module Json_serializers : sig ... end
module Json_deserializers : sig ... end