Stream Processing with Apache Flink
As data processing becomes more real-time, stream processing is becoming more important. Apache Flink makes it easier to build and manage scalable stream processing applications.
In this workshop you will learn the basics about stream processing with Apache Flink. You will learn how to implement stream processing applications that ingest events from Apache Kafka and submit them to a (Docker) local Flink cluster for execution.
We will show the basic commands to manage a continuously running application and how to access its metrics. Later, we will introduce you to Flink's streaming SQL interface. You will submit SQL queries that are evaluated over unbounded data streams, producing results that are continuously updated as more and more data is ingested.
The Workshop is designed in English, but Fabian Hueske will present in German, if attending participants prefer so.
Vorkenntnisse
• Basics of Java and SQL
• Basic knowledge of distributed data processing (MapReduce/Spark/etc.) will be helpful but is not required
• You will need a notebook with at least 8 GB RAM and the following software installed: Docker, Java 8, Java IDE (preferably IDEA IntelliJ), Apache Maven
Lernziele
Learn how to run SQL on streaming data