aws s3 streaming upload sink AWS S3 Streaming upload Sink

Provided by: "Apache Software Foundation"

Support Level for this Kamelet is: "Preview"

Upload data to AWS S3 in streaming upload mode.

Configuration Options

The following table summarizes the configuration options available for the aws-s3-streaming-upload-sink Kamelet:

Property Name Description Type Default Example

accessKey *

Access Key

The access key obtained from AWS.

string

bucketNameOrArn *

Bucket Name

The S3 Bucket name or ARN.

string

keyName *

Key Name

Setting the key name for an element in the bucket through endpoint parameter. In Streaming Upload, with the default configuration, this will be the base for the progressive creation of files.

string

region *

AWS Region

The AWS region to connect to.

string

"eu-west-1"

secretKey *

Secret Key

The secret key obtained from AWS.

string

autoCreateBucket

Autocreate Bucket

Setting the autocreation of the S3 bucket bucketName.

boolean

false

batchMessageNumber

Batch Message Number

The number of messages composing a batch in streaming upload mode

int

10

batchSize

Batch Size

The batch size (in bytes) in streaming upload mode

int

"1000000"

namingStrategy

Naming Strategy

The naming strategy to use in streaming upload mode. There are 2 enums and the value can be one of progressive, random

string

"progressive"

restartingPolicy

Restarting Policy

The restarting policy to use in streaming upload mode. There are 2 enums and the value can be one of override, lastPart

string

"lastPart"

streamingUploadMode

Streaming Upload Mode

Setting the Streaming Upload Mode

boolean

true

streamingUploadTimeout

Streaming Upload Timeout

While streaming upload mode is true, this option set the timeout to complete upload

long

Fields marked with (*) are mandatory.

Usage

This section summarizes how the aws-s3-streaming-upload-sink can be used in various contexts.

Knative Sink

The aws-s3-streaming-upload-sink Kamelet can be used as Knative sink by binding it to a Knative object.

aws-s3-streaming-upload-sink-binding.yaml
apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
  name: aws-s3-streaming-upload-sink-binding
spec:
  source:
    ref:
      kind: Channel
      apiVersion: messaging.knative.dev/v1
      name: mychannel
  sink:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1alpha1
      name: aws-s3-streaming-upload-sink
    properties:
      accessKey: "The Access Key"
      bucketNameOrArn: "The Bucket Name"
      keyName: "The Key Name"
      region: "eu-west-1"
      secretKey: "The Secret Key"

Make sure you have Camel K installed into the Kubernetes cluster you’re connected to.

Save the aws-s3-streaming-upload-sink-binding.yaml file into your hard drive, then configure it according to your needs.

You can run the sink using the following command:

kubectl apply -f aws-s3-streaming-upload-sink-binding.yaml

Dependencies

The Kamelet needs the following dependencies:

  • camel:aws2-s3

  • camel:kamelet

Binding to Knative using the Kamel CLI:

The procedure described above can be simplified into a single execution of the kamel bind command:

kamel bind channel:mychannel aws-s3-streaming-upload-sink -p "sink.accessKey=The Access Key" -p "sink.bucketNameOrArn=The Bucket Name" -p "sink.keyName=The Key Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"

This will create the KameletBinding under the hood and apply it to the current namespace in the cluster.

Kafka Sink

The aws-s3-streaming-upload-sink Kamelet can be used as Kafka sink by binding it to a Kafka topic.

aws-s3-streaming-upload-sink-binding.yaml
apiVersion: camel.apache.org/v1alpha1
kind: KameletBinding
metadata:
  name: aws-s3-streaming-upload-sink-binding
spec:
  source:
    ref:
      kind: KafkaTopic
      apiVersion: kafka.strimzi.io/v1beta1
      name: my-topic
  sink:
    ref:
      kind: Kamelet
      apiVersion: camel.apache.org/v1alpha1
      name: aws-s3-streaming-upload-sink
    properties:
      accessKey: "The Access Key"
      bucketNameOrArn: "The Bucket Name"
      keyName: "The Key Name"
      region: "eu-west-1"
      secretKey: "The Secret Key"

Ensure that you’ve installed Strimzi and created a topic named my-topic in the current namespace. Make also sure you have Camel K installed into the Kubernetes cluster you’re connected to.

Save the aws-s3-streaming-upload-sink-binding.yaml file into your hard drive, then configure it according to your needs.

You can run the sink using the following command:

kubectl apply -f aws-s3-streaming-upload-sink-binding.yaml

Binding to Kafka using the Kamel CLI:

The procedure described above can be simplified into a single execution of the kamel bind command:

kamel bind kafka.strimzi.io/v1beta1:KafkaTopic:my-topic aws-s3-streaming-upload-sink -p "sink.accessKey=The Access Key" -p "sink.bucketNameOrArn=The Bucket Name" -p "sink.keyName=The Key Name" -p "sink.region=eu-west-1" -p "sink.secretKey=The Secret Key"

This will create the KameletBinding under the hood and apply it to the current namespace in the cluster.