Learn Apache Kafka with Python

In this online workshop you'll learn the basic components of Apache Kafka and how to get started with data streaming using Python.

  • 2 hour extended version with a brief introduction to transforming your data using Apache Flink (recording coming soon)
  • 1 hour workshop recording

See all workshops

Youtube thumbnail

What's in the Workshop Recipe?

Apache Kafka is the industry de-facto standard for data streaming. An open-source, scalable, highly available and reliable solution to move data across companies' departments, technologies or micro-services.

In this workshop you'll learn the basic components of Apache Kafka and how to get started with data streaming using Python. We'll dive deep, with the help of some prebuilt Jupyter notebooks, into how to produce and consume data, including how to have concurrent applications reading from the same source, empowering multiple use-cases with the same streaming data.

Finally we'll demonstrate how to use Apache Flink to reshape streaming data in Kafka in real time, enabling new data stakeholders and, for instance, allowing transformations to satisfy regulatory needs.

Prepare for a Brain Upgrade

You will learn:

  • Apache Kafka main components
  • How to Produce data to Apache Kafka
  • How to Consume data from Apache Kafka
  • How to setup concurrent consumers
  • How multiple applications can read from the same dataset
  • How to reshape the streaming data in real time

Prerequisities

Workshop host

Jay Miller portrait
Jay Miller

Staff Developer Advocate, Aiven

Jay Miller is a Staff Developer Advocate and long term advocate for building developer communities. Jay has been a contributor to the Python Community ecosystem and has participated in many user groups and conferences.

Boost your tech skills with our developer workshops

Live and interactive sessions to upgrade your skills with expert guidance covering a range of open source technologies.

Explore all workshops

Related resources