A Practical Introduction to Stream Processing Training Course
Stream Processing refers to the real-time processing of "data in motion", that is, performing computations on data as it is being received. Such data is read as continuous streams from data sources such as sensor events, website user activity, financial trades, credit card swipes, click streams, etc. Stream Processing frameworks are able to read large volumes of incoming data and provide valuable insights almost instantaneously.
In this instructor-led, live training (onsite or remote), participants will learn how to set up and integrate different Stream Processing frameworks with existing big data storage systems and related software applications and microservices.
By the end of this training, participants will be able to:
- Install and configure different Stream Processing frameworks, such as Spark Streaming and Kafka Streaming.
- Understand and select the most appropriate framework for the job.
- Process of data continuously, concurrently, and in a record-by-record fashion.
- Integrate Stream Processing solutions with existing databases, data warehouses, data lakes, etc.
- Integrate the most appropriate stream processing library with enterprise applications and microservices.
Audience
- Developers
- Software architects
Format of the Course
- Part lecture, part discussion, exercises and heavy hands-on practice
Course Outline
Introduction
- Stream processing vs batch processing
- Analytics-focused stream processing
Overview Frameworks and Programming Languages
- Spark Streaming (Scala)
- Kafka Streaming (Java)
- Flink
- Storm
- Comparison of Features and Strengths of Each Framework
Overview of Data Sources
- Live data as a series of events over time
- Historical data sources
Deployment Options
- In the cloud (AWS, etc.)
- On premise (private cloud, etc.)
Getting Started
- Setting up the Development Environment
- Installing and Configuring
- Assessing Your Data Analysis Needs
Operating a Streaming Framework
- Integrating the Streaming Framework with Big Data Tools
- Event Stream Processing (ESP) vs Complex Event Processing (CEP)
- Transforming the Input Data
- Inspecting the Output Data
- Integrating the Stream Processing Framework with Existing Applications and Microservices
Troubleshooting
Summary and Conclusion
Requirements
- Programming experience in any language
- An understanding of Big Data concepts (Hadoop, etc.)
Need help picking the right course?
A Practical Introduction to Stream Processing Training Course - Enquiry
Testimonials (1)
Sufficient hands on, trainer is knowledgable
Chris Tan
Course - A Practical Introduction to Stream Processing
Related Courses
Administration of Confluent Apache Kafka
21 HoursConfluent Apache Kafka is a distributed event streaming platform designed for high-throughput, fault-tolerant data pipelines and real-time analytics.
This instructor-led, live training (online or onsite) is aimed at intermediate-level system administrators and DevOps professionals who wish to install, configure, monitor, and troubleshoot Confluent Apache Kafka clusters.
By the end of this training, participants will be able to:
- Understand the components and architecture of Confluent Kafka.
- Deploy and manage Kafka brokers, Zookeeper quorums, and key services.
- Configure advanced features including security, replication, and performance tuning.
- Use management tools to monitor and maintain Kafka clusters.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
Apache Kafka Connect
7 HoursThis instructor-led, live training in Kenya (online or onsite) is aimed at developers who wish to integrate Apache Kafka with existing databases and applications for processing, analysis, etc.
By the end of this training, participants will be able to:
- Use Kafka Connect to ingest large amounts of data from a database into Kafka topics.
- Ingest log data generated by an application servers into Kafka topics.
- Make any collected data available for stream processing.
- Export data from Kafka topics into secondary systems for storage and analysis.
Big Data Streaming for Developers
14 HoursLearn to implement end-to-end big data streaming use cases. Real-time data preparation and maintenance with Informatica, Edge, Kafka and Spark. This training covers software versions 10.2.1 and up.
Confluent Apache Kafka: Cluster Operations and Configuration
16 HoursConfluent Apache Kafka is an enterprise-grade distributed event streaming platform built on Apache Kafka. It supports high-throughput, fault-tolerant data pipelines and real-time streaming applications.
This instructor-led, live training (online or onsite) is aimed at intermediate-level engineers and administrators who wish to deploy, configure, and optimize Confluent Kafka clusters in production environments.
By the end of this training, participants will be able to:
- Install, configure, and operate Confluent Kafka clusters with multiple brokers.
- Design high-availability setups using Zookeeper and replication techniques.
- Tune performance, monitor metrics, and apply recovery strategies.
- Secure, scale, and integrate Kafka with enterprise environments.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.
Building Kafka Solutions with Confluent
14 HoursThis instructor-led, live training (online or onsite) is aimed at engineers who wish to use Confluent (a distribution of Kafka) to build and manage a real-time data processing platform for their applications.
By the end of this training, participants will be able to:
- Install and configure Confluent Platform.
- Use Confluent's management tools and services to run Kafka more easily.
- Store and process incoming stream data.
- Optimize and manage Kafka clusters.
- Secure data streams.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- This course is based on the open source version of Confluent: Confluent Open Source.
- To request a customized training for this course, please contact us to arrange.
Building Data Pipelines with Apache Kafka
7 HoursApache Kafka is a distributed streaming platform. It is de facto a standard for building data pipelines and it solves a lot of different use-cases around data processing: it can be used as a message queue, distributed log, stream processor, etc.
We'll start with some theory behind data pipelines in general, then continue with fundamental concepts behind Kafka. We'll also discover important components like Kafka Streams and Kafka Connect.
Distributed Messaging with Apache Kafka
14 HoursThis course is for enterprise architects, developers, system administrators and anyone who wants to understand and use a high-throughput distributed messaging system. If you have more specific requirements (e.g. only system administration side), this course can be tailored to better suit your needs.
Kafka for Administrators
21 HoursThis instructor-led, live training in Kenya (online or onsite) is aimed at beginner-level / intermediate-level / advanced-level system administrators and operations engineers who wish to use Apache Kafka to deploy, secure, monitor, and troubleshoot Kafka clusters.
By the end of this training, participants will be able to: explain Kafka architecture and KRaft mode, operate and secure Kafka clusters, monitor performance and reliability, and resolve common production issues.
Apache Kafka for Developers
21 HoursThis instructor-led, live training in Kenya (online or onsite) is aimed at intermediate-level developers who wish to develop big data applications with Apache Kafka.
By the end of this training, participants will be able to:
- Develop Kafka producers and consumers to send and read data from Kafka.
- Integrate Kafka with external systems using Kafka Connect.
- Write streaming applications with Kafka Streams & ksqlDB.
- Integrate a Kafka client application with Confluent Cloud for cloud-based Kafka deployments.
- Gain practical experience through hands-on exercises and real-world use cases.
Apache Kafka for Python Programmers
7 HoursThis instructor-led, live training in Kenya (online or onsite) is aimed at data engineers, data scientists, and programmers who wish to use Apache Kafka features in data streaming with Python.
By the end of this training, participants will be able to use Apache Kafka to monitor and manage conditions in continuous data streams using Python programming.
Kafka Fundamentals for Java Developers
14 HoursThis instructor-led, live training in Kenya (online or onsite) is aimed at intermediate-level Java developers who wish to integrate Apache Kafka into their applications for reliable, scalable, and high-throughput messaging.
By the end of this training, participants will be able to:
- Understand the architecture and core components of Kafka.
- Set up and configure a Kafka cluster.
- Produce and consume messages using Java.
- Implement Kafka Streams for real-time data processing.
- Ensure fault tolerance and scalability in Kafka applications.
Administration of Kafka Message Queue
14 HoursThis instructor-led, live training in Kenya (online or onsite) is aimed at intermediate-level system administrators who wish to harness Kafka's message queuing features effectively.
By the end of this training, participants will be able to:
- Understand Kafka's message queuing capabilities and architecture.
- Configure Kafka topics for message queuing scenarios.
- Produce and consume messages using Kafka.
- Monitor and manage Kafka as a message queue.
PySpark and Machine Learning
21 HoursThis training provides a practical introduction to building scalable data processing and Machine Learning workflows using PySpark. Participants learn how Apache Spark operates within modern Big Data ecosystems and how to efficiently process large datasets using distributed computing principles.
Python and Spark for Big Data (PySpark)
21 HoursIn this instructor-led, live training in Kenya, participants will learn how to use Python and Spark together to analyze big data as they work on hands-on exercises.
By the end of this training, participants will be able to:
- Learn how to use Spark with Python to analyze Big Data.
- Work on exercises that mimic real world cases.
- Use different tools and techniques for big data analysis using PySpark.
Stratio: Rocket and Intelligence Modules with PySpark
14 HoursStratio is a data-centric platform that integrates big data, AI, and governance into a single solution. Its Rocket and Intelligence modules enable rapid data exploration, transformation, and advanced analytics in enterprise environments.
This instructor-led, live training (online or onsite) is aimed at intermediate-level data professionals who wish to use the Rocket and Intelligence modules in Stratio effectively with PySpark, focusing on looping structures, user-defined functions, and advanced data logic.
By the end of this training, participants will be able to:
- Navigate and work within the Stratio platform using Rocket and Intelligence modules.
- Apply PySpark in the context of data ingestion, transformation, and analysis.
- Use loops and conditional logic to control data workflows and feature engineering tasks.
- Create and manage user-defined functions (UDFs) for reusable data operations in PySpark.
Format of the Course
- Interactive lecture and discussion.
- Lots of exercises and practice.
- Hands-on implementation in a live-lab environment.
Course Customization Options
- To request a customized training for this course, please contact us to arrange.