Struct aws_sdk_kinesis::client::fluent_builders::CreateStream
source · pub struct CreateStream { /* private fields */ }
Expand description
Fluent builder constructing a request to CreateStream
.
Creates a Kinesis data stream. A stream captures and transports data records that are continuously emitted from different data sources or producers. Scale-out within a stream is explicitly supported by means of shards, which are uniquely identified groups of data records in a stream.
You can create your data stream using either on-demand or provisioned capacity mode. Data streams with an on-demand mode require no capacity planning and automatically scale to handle gigabytes of write and read throughput per minute. With the on-demand mode, Kinesis Data Streams automatically manages the shards in order to provide the necessary throughput. For the data streams with a provisioned mode, you must specify the number of shards for the data stream. Each shard can support reads up to five transactions per second, up to a maximum data read total of 2 MiB per second. Each shard can support writes up to 1,000 records per second, up to a maximum data write total of 1 MiB per second. If the amount of data input increases or decreases, you can add or remove shards.
The stream name identifies the stream. The name is scoped to the Amazon Web Services account used by the application. It is also scoped by Amazon Web Services Region. That is, two streams in two different accounts can have the same name, and two streams in the same account, but in two different Regions, can have the same name.
CreateStream
is an asynchronous operation. Upon receiving a CreateStream
request, Kinesis Data Streams immediately returns and sets the stream status to CREATING
. After the stream is created, Kinesis Data Streams sets the stream status to ACTIVE
. You should perform read and write operations only on an ACTIVE
stream.
You receive a LimitExceededException
when making a CreateStream
request when you try to do one of the following:
-
Have more than five streams in the
CREATING
state at any point in time. -
Create more shards than are authorized for your account.
For the default shard limit for an Amazon Web Services account, see Amazon Kinesis Data Streams Limits in the Amazon Kinesis Data Streams Developer Guide. To increase this limit, contact Amazon Web Services Support.
You can use DescribeStreamSummary
to check the stream status, which is returned in StreamStatus
.
CreateStream
has a limit of five transactions per second per account.
Implementations§
source§impl CreateStream
impl CreateStream
sourcepub async fn customize(
self
) -> Result<CustomizableOperation<CreateStream, AwsResponseRetryClassifier>, SdkError<CreateStreamError>>
pub async fn customize(
self
) -> Result<CustomizableOperation<CreateStream, AwsResponseRetryClassifier>, SdkError<CreateStreamError>>
Consume this builder, creating a customizable operation that can be modified before being sent. The operation’s inner http::Request can be modified as well.
sourcepub async fn send(
self
) -> Result<CreateStreamOutput, SdkError<CreateStreamError>>
pub async fn send(
self
) -> Result<CreateStreamOutput, SdkError<CreateStreamError>>
Sends the request and returns the response.
If an error occurs, an SdkError
will be returned with additional details that
can be matched against.
By default, any retryable failures will be retried twice. Retry behavior is configurable with the RetryConfig, which can be set when configuring the client.
sourcepub fn stream_name(self, input: impl Into<String>) -> Self
pub fn stream_name(self, input: impl Into<String>) -> Self
A name to identify the stream. The stream name is scoped to the Amazon Web Services account used by the application that creates the stream. It is also scoped by Amazon Web Services Region. That is, two streams in two different Amazon Web Services accounts can have the same name. Two streams in the same Amazon Web Services account but in two different Regions can also have the same name.
sourcepub fn set_stream_name(self, input: Option<String>) -> Self
pub fn set_stream_name(self, input: Option<String>) -> Self
A name to identify the stream. The stream name is scoped to the Amazon Web Services account used by the application that creates the stream. It is also scoped by Amazon Web Services Region. That is, two streams in two different Amazon Web Services accounts can have the same name. Two streams in the same Amazon Web Services account but in two different Regions can also have the same name.
sourcepub fn shard_count(self, input: i32) -> Self
pub fn shard_count(self, input: i32) -> Self
The number of shards that the stream will use. The throughput of the stream is a function of the number of shards; more shards are required for greater provisioned throughput.
sourcepub fn set_shard_count(self, input: Option<i32>) -> Self
pub fn set_shard_count(self, input: Option<i32>) -> Self
The number of shards that the stream will use. The throughput of the stream is a function of the number of shards; more shards are required for greater provisioned throughput.
sourcepub fn stream_mode_details(self, input: StreamModeDetails) -> Self
pub fn stream_mode_details(self, input: StreamModeDetails) -> Self
Indicates the capacity mode of the data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data streams.
sourcepub fn set_stream_mode_details(self, input: Option<StreamModeDetails>) -> Self
pub fn set_stream_mode_details(self, input: Option<StreamModeDetails>) -> Self
Indicates the capacity mode of the data stream. Currently, in Kinesis Data Streams, you can choose between an on-demand capacity mode and a provisioned capacity mode for your data streams.
Trait Implementations§
source§impl Clone for CreateStream
impl Clone for CreateStream
source§fn clone(&self) -> CreateStream
fn clone(&self) -> CreateStream
1.0.0 · source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source
. Read more