Send data to Amazon Kinesis

Most recent version: v0.0.1

See the changelog of this Data Sink type here.

Overview

Onum supports integration with Amazon Kinesis Data Stream.

Amazon Kinesis Data Streams is a fully managed, serverless streaming data service that allows you to ingest, store, and process real-time data streams. It's designed for high-throughput, low-latency data ingestion from various sources, enabling real-time analytics and applications.

Prerequisites

You will need an IAM User, role or group with the correct permissions to access and manage Kinesis.

Amazon Kinesis Data Stream Setup

1

Go to IAM (Identity and Access Management) to manage users, groups, roles and permissions.

Under Permissions Policies, make sure you have assigned the policy AmazonKinesisFullAccess to give full access to Kinesis resources. Alternatively, if you have custom permissions, go to Policies - Create Policy and in the JSON tab, paste your custom JSON e.g.

  {
    "Version": "2012-10-17",
      "Statement": [
       {
         "Effect": "Allow"
         "Action": [
         "kinesis:CreateStream",
         "kinesis:DescribeStream",
         "kinesis:PutRecord"
         ],
         "Resource": "*"
 		}        
	]
  }      
2

Test the Configuration

Run aws kinesis list-streams

If you IAM permission are correct, you'll see a list of streams.

Onum Setup

1

Log in to your Onum tenant and click Data Sinks> New Data sink.

2

Double-click the Amazon Kinesis Data Stream Sink.

3

Enter a Name for the new Data Sink. Optionally, add a Description and some Tags to identify the Sink.

4

Decide whether or not to include this Data sink info in the metrics and graphs of the Home area.

5

In the AWS authentication section, enter the region of your AWS data center. Your region is displayed in the top right-hand corner of your AWS console.

6

Select the Access Key ID from your Secrets or click New secret to generate a new one.

The Access Key ID is found in the IAM Dashboard of the AWS Management Console.

  1. In the left panel, click on Users.

  2. Select your IAM user.

  3. Under the Security Credentials tab, scroll to Access Keys, and you will find existing Access Key IDs (but not the secret access key).

7

Select the Secret Access Key from your Secrets or click New secret to generate a new one.

Under Access keys, you can see your Access Key IDs, but AWS will not show the Secret Access Key. You must have it saved somewhere. If you don't have the secret key saved, you need to create a new one.

8

Configure your Data Stream.

  • Stream Name*

    1. Select Data Streams under Amazon Kinesis in the sidebar.

    2. The Stream Name will be in the first column e.g. my-kinesis-stream-prod

  • A Partition key* is a string that you assign to each record you put into a stream. It plays a critical role in determining which shard the data record is routed to within the stream e.g. user_456 means all login events for this user will be sent to the same shard, preserving the order of events for that user.

9

In the Advanced Configuration section, toggle on the Bulk configuration switch to enable bulk configuration. Configure the following settings:

  • Event time limit* - Time in milliseconds to wait before sending the batch.

  • Number of events* - Maximum number of events to send in a single batch. The minimum value is 1, and the maximum value is 500.

  • Batch size* - Maximum number of events to send in a single batch. The minimum value is 1024, and the maximum value is 5242880.

10

If you have non-default URL that directs API requests to a specific Kinesis service endpoint, enter it here in the Custom endpoint.

Click Create data sink when complete.

Your new Data sink will appear in the Data sinks area list.

Pipeline configuration

When it comes to using this Data sink in a Pipeline, you must configure the following output parameters. To do it, simply click the Data sink on the canvas and select Configuration.

Output configuration

Parameter
Description

Event Field*

Select the event field that contains the output message. The data type must be string.

Click Save to save your configuration.

Last updated

Was this helpful?