Getting Started: Flink 1.8.2 - Amazon Kinesis Data Analytics
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China (PDF).

Getting Started: Flink 1.8.2

This topic contains a version of the Getting Started (DataStream API) Tutorial that uses Apache Flink 1.8.2.

Components of Kinesis Data Analytics for Flink Application

To process data, your Kinesis Data Analytics application uses a Java/Apache Maven or Scala application that processes input and produces output using the Apache Flink runtime.

An Kinesis Data Analytics application has the following components:

  • Runtime properties: You can use runtime properties to configure your application without recompiling your application code.

  • Source: The application consumes data by using a source. A source connector reads data from a Kinesis data stream, an Amazon S3 bucket, etc. For more information, see Sources.

  • Operators: The application processes data by using one or more operators. An operator can transform, enrich, or aggregate data. For more information, see DataStream API Operators.

  • Sink: The application produces data to external sources by using sinks. A sink connector writes data to a Kinesis data stream, a Kinesis Data Firehose delivery stream, an Amazon S3 bucket, etc. For more information, see Sinks.

After you create, compile, and package your application code, you upload the code package to an Amazon Simple Storage Service (Amazon S3) bucket. You then create a Kinesis Data Analytics application. You pass in the code package location, a Kinesis data stream as the streaming data source, and typically a streaming or file location that receives the application's processed data.

Prerequisites for Completing the Exercises

To complete the steps in this guide, you must have the following:

To get started, go to Step 1: Set Up an Amazon Account and Create an Administrator User.