Amazon Data Firehose
Amazon Data Firehose was previously known as Amazon Kinesis Data Firehose.
Amazon Data Firehose is a fully managed service that delivers real-time streaming data to destinations such as Amazon Simple Storage Service (Amazon S3), Amazon OpenSearch Service, Amazon Redshift, Splunk, and various other supported destinations.
firehose(config = list(), credentials = list(), endpoint = NULL, region = NULL)
A client for the service. You can call the service's operations using
syntax like svc$operation(...)
, where svc
is the name you've assigned
to the client. The available operations are listed in the
Operations section.
Optional configuration of credentials, endpoint, and/or region.
credentials:
creds:
access_key_id: AWS access key ID
secret_access_key: AWS secret access key
session_token: AWS temporary session token
profile: The name of a profile to use. If not given, then the default profile is used.
anonymous: Set anonymous credentials.
endpoint: The complete URL to use for the constructed client.
region: The AWS Region used in instantiating the client.
close_connection: Immediately close all HTTP connections.
timeout: The time in seconds till a timeout exception is thrown when attempting to make a connection. The default is 60 seconds.
s3_force_path_style: Set this to true
to force the request to use path-style addressing, i.e. http://s3.amazonaws.com/BUCKET/KEY
.
sts_regional_endpoint: Set sts regional endpoint resolver to regional or legacy https://docs.aws.amazon.com/sdkref/latest/guide/feature-sts-regionalized-endpoints.html
Optional credentials shorthand for the config parameter
creds:
access_key_id: AWS access key ID
secret_access_key: AWS secret access key
session_token: AWS temporary session token
profile: The name of a profile to use. If not given, then the default profile is used.
anonymous: Set anonymous credentials.
Optional shorthand for complete URL to use for the constructed client.
Optional shorthand for AWS Region used in instantiating the client.
svc <- firehose(
config = list(
credentials = list(
creds = list(
access_key_id = "string",
secret_access_key = "string",
session_token = "string"
),
profile = "string",
anonymous = "logical"
),
endpoint = "string",
region = "string",
close_connection = "logical",
timeout = "numeric",
s3_force_path_style = "logical",
sts_regional_endpoint = "string"
),
credentials = list(
creds = list(
access_key_id = "string",
secret_access_key = "string",
session_token = "string"
),
profile = "string",
anonymous = "logical"
),
endpoint = "string",
region = "string"
)
create_delivery_stream | Creates a Firehose stream |
delete_delivery_stream | Deletes a Firehose stream and its data |
describe_delivery_stream | Describes the specified Firehose stream and its status |
list_delivery_streams | Lists your Firehose streams in alphabetical order of their names |
list_tags_for_delivery_stream | Lists the tags for the specified Firehose stream |
put_record | Writes a single data record into an Firehose stream |
put_record_batch | Writes multiple data records into a Firehose stream in a single call, which can achieve higher throughput per producer than when writing single records |
start_delivery_stream_encryption | Enables server-side encryption (SSE) for the Firehose stream |
stop_delivery_stream_encryption | Disables server-side encryption (SSE) for the Firehose stream |
tag_delivery_stream | Adds or updates tags for the specified Firehose stream |
untag_delivery_stream | Removes tags from the specified Firehose stream |
update_destination | Updates the specified destination of the specified Firehose stream |
if (FALSE) {
svc <- firehose()
svc$create_delivery_stream(
Foo = 123
)
}
Run the code above in your browser using DataLab