StreamThoughts
REMOTE

An introduction to Apache Kafka

Price : 2000€ H.T. (for a group)

For more information about this training course, please free to contact :
training@streamthoughts.io

Description

During this instructor-led one-day workshop you will have the opportunity to have an overview to the Apache Kafka platform. We will introduce and discusss the architectures concepts and usages of Kafka Core, Kafka Connect and Kafka Streams.

Course Objectives

This course enables participants to acquire the following skills:

  • Understanding the ecosystem and uses of the Apache Kafka solution.
  • Understanding the fundamental concepts of Apache Kafka's architecture.
  • Understanding the uses of Kafka Connect solutions.
  • Understanding the uses of Kafka Streams solutions.

Pedagogy

80% theory, 20% practise

Who Should Attend ?

This course is designed for applications Developers, Architects, Data engineers and Technical Project managers.

Course Duration

1 Day

Course Prerequisites

There are no prerequisites for this course

Course Content

Module 1: Introduction

  • Event Streaming, the motivations ?
  • What is Apache Kafka
  • The Apache Kafka project
  • The key benefits of Kafka
  • What is it used for ?
  • The alternative solutions
  • The Confluent Streaming Platform

Module 2: Kafka Fundamentals

  • Broker, Message, Topic & Partitions
  • Producers Basics
  • Consumers & Consumer Groups
  • Replication & Fault-tolerance
  • Data retention and compression
  • Understanding Zookeeper’s roles
  • Understanding Kafka’s performance

Module 3: Data Integration with Kafka Connect

  • Introduction to Kafka Connect
  • Kafka Connect Architecture
  • Converters & Single Message Transforms
  • Deployment Modes
  • Connectors Configuration
  • Use Cases

Module 4: Kafka Streams

  • Introducton to Kafka Streams
  • Kafka Streams Concepts: Stream & Table
  • Stateless & Stateful applications
  • Windowed Aggregations, Join Operations
  • Interactive Queries
  • Introduction to ksqlDB

Module 5: Governance and data models

  • Data Serialization
  • Introduction Avro
  • Managing Schema Evolution
  • Why Using a Schema Registry ?
The Author's Avatar
Instructor

Florian travaille depuis plus de 8 ans dans le conseil, il est co-fondateur et CEO de StreamThoughts. Au cours de sa carrière, il a travaillé sur divers projets impliquant la mise en oeuvre de plateformes d’intégration et de traitement de la data à travers les mondes technologiques de Hadoop et de Spark. Passionné par les systèmes distribués, il se spécialise dans les technologies d’event-streaming comme Apache Kafka, Apache Pulsar. Aujourd’hui, il accompagne les entreprises dans leur transition vers les architectures orientées streaming d’événements. Florian est certifié Confluent Administrator & Developer pour Apache Kafka. Il est nommé deux années consécutive (2019 et 2020) “Confluent Community Catalyst” pour ses contributions sur le projet Apache Kafka Streams et son implication dans la communauté open-source. Il fait partie des organisateurs du Paris Apache Kafka Meetup.