Stream Processing with Apache Flink

As data processing becomes more real-time, stream processing is becoming more important. Apache Flink makes it easier to build and manage scalable stream processing applications.

In this workshop you will learn the basics about stream processing with Apache Flink. You will learn how to implement stream processing applications that ingest events from Apache Kafka and submit them to a (Docker) local Flink cluster for execution.

We will show the basic commands to manage a continuously running application and how to access its metrics. Later, we will introduce you to Flink's streaming SQL interface. You will submit SQL queries that are evaluated over unbounded data streams, producing results that are continuously updated as more and more data is ingested.

The Workshop is presented in English, but Fabian Hueske speaks German, too.

Vorkenntnisse


  • Basics of Java and SQL.

  • Basic knowledge of distributed data processing (MapReduce/Spark/etc.) will be helpful but is not required.

  • You will need a notebook with at least 8 GB RAM and the following software installed:

    • Docker

    • Java 8

    • A Java IDE (preferably IDEA IntelliJ)

    • Apache Maven



Lernziele

Learn how to run SQL on streaming data

 

Speaker

 


Fabian Hueske is a committer and PMC member of the Apache Flink project and has been contributing to Flink since its earliest days. Fabian is a co-founder of Ververica, a Berlin-based startup devoted to fostering Flink, where he works as a software engineer and contributes to Apache Flink. He holds a PhD in computer science from TU Berlin and is currently writing a book about Stream Processing.

Gold-Sponsoren

HMS
Structr

Silber-Sponsoren

codecentric
Phytec

Bronze-Sponsor

incontext.technology GmbH

data2day-Newsletter

Sie möchten über die data2day
auf dem Laufenden gehalten werden?

 

Anmelden