Struct aws_sdk_kinesis::input::put_record_input::Builder
source · pub struct Builder { /* private fields */ }
Expand description
A builder for PutRecordInput
.
Implementations§
source§impl Builder
impl Builder
sourcepub fn stream_name(self, input: impl Into<String>) -> Self
pub fn stream_name(self, input: impl Into<String>) -> Self
The name of the stream to put the data record into.
sourcepub fn set_stream_name(self, input: Option<String>) -> Self
pub fn set_stream_name(self, input: Option<String>) -> Self
The name of the stream to put the data record into.
sourcepub fn data(self, input: Blob) -> Self
pub fn data(self, input: Blob) -> Self
The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the maximum record size (1 MiB).
sourcepub fn set_data(self, input: Option<Blob>) -> Self
pub fn set_data(self, input: Option<Blob>) -> Self
The data blob to put into the record, which is base64-encoded when the blob is serialized. When the data blob (the payload before base64-encoding) is added to the partition key size, the total size must not exceed the maximum record size (1 MiB).
sourcepub fn partition_key(self, input: impl Into<String>) -> Self
pub fn partition_key(self, input: impl Into<String>) -> Self
Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As a result of this hashing mechanism, all data records with the same partition key map to the same shard within the stream.
sourcepub fn set_partition_key(self, input: Option<String>) -> Self
pub fn set_partition_key(self, input: Option<String>) -> Self
Determines which shard in the stream the data record is assigned to. Partition keys are Unicode strings with a maximum length limit of 256 characters for each key. Amazon Kinesis Data Streams uses the partition key as input to a hash function that maps the partition key and associated data to a specific shard. Specifically, an MD5 hash function is used to map partition keys to 128-bit integer values and to map associated data records to shards. As a result of this hashing mechanism, all data records with the same partition key map to the same shard within the stream.
sourcepub fn explicit_hash_key(self, input: impl Into<String>) -> Self
pub fn explicit_hash_key(self, input: impl Into<String>) -> Self
The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition key hash.
sourcepub fn set_explicit_hash_key(self, input: Option<String>) -> Self
pub fn set_explicit_hash_key(self, input: Option<String>) -> Self
The hash value used to explicitly determine the shard the data record is assigned to by overriding the partition key hash.
sourcepub fn sequence_number_for_ordering(self, input: impl Into<String>) -> Self
pub fn sequence_number_for_ordering(self, input: impl Into<String>) -> Self
Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. Usage: set the SequenceNumberForOrdering
of record n to the sequence number of record n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are coarsely ordered based on arrival time.
sourcepub fn set_sequence_number_for_ordering(self, input: Option<String>) -> Self
pub fn set_sequence_number_for_ordering(self, input: Option<String>) -> Self
Guarantees strictly increasing sequence numbers, for puts from the same client and to the same partition key. Usage: set the SequenceNumberForOrdering
of record n to the sequence number of record n-1 (as returned in the result when putting record n-1). If this parameter is not set, records are coarsely ordered based on arrival time.
sourcepub fn stream_arn(self, input: impl Into<String>) -> Self
pub fn stream_arn(self, input: impl Into<String>) -> Self
The ARN of the stream.
sourcepub fn set_stream_arn(self, input: Option<String>) -> Self
pub fn set_stream_arn(self, input: Option<String>) -> Self
The ARN of the stream.
sourcepub fn build(self) -> Result<PutRecordInput, BuildError>
pub fn build(self) -> Result<PutRecordInput, BuildError>
Consumes the builder and constructs a PutRecordInput
.