Summer-Ready Savings: Find the Training Course You Need at a Price You'll Love

closeClose

Confluent Fundamentals for Apache Kafka

Course Details
Code: AK-FUN
Tuition (USD): $199.00 • Classroom (1 day)

This is a one-day course available On-demand (self-paced) or Instructor-led by request.

Skills Gained

  • Gain understanding of the Apache Kafka and Confluent Platform
  • Explore use cases
  • Receive an overview of Kafka’s core concepts that enable it to power a highly scalable, highly available and resilient real-time event streaming platform
  • Be introduced to the Confluent Platform, offering an enterprise-ready, real-time event streaming platform powered by Apache Kafka
  • Begin preparation to attend Confluent Operations for Apache Kafka and/or the Confluent Skills for Building Apache Kafka training courses.

Who Can Benefit

This course is designed for all professionals who work with a real-time event streaming platform powered by Apache Kafka. This course is a recommended prerequisite for students attending Confluent Operations for Apache Kafka and/or the Confluent Skills for Building Apache Kafka.

Prerequisites

Attendees are not expected to have any prior experience with Kafka. Some basic understanding of the Linux OS and experience in using a shell like Bash is beneficial.

Course Details

Hands-on Training

To help reinforce the concepts being taught, hands-on exercises are included throughout the training. Exercises include:

  • Launching and exploring a minimal Kafka cluster
  • Using Kafka command line tools to explore cluster meta data in ZooKeeper, create topics on the broker, and publish & consume messages
  • Running a Java based consumer and observe consumer lag when scaling the consumer
  • Configuring Kafka Connect with a MQTT Connector source to create a data pipeline
  • Using Confluent Control Center to monitor your cluster and execute KSQL queries.

Course Outline

Motivation and Customer Use Cases

  • Motivation for a paradigm change to “Event-driven”
  • How Kafka is the backbone of real-time event streaming
  • How other major players in the market use Kafka
  • Microservices, IoT and Edge Computing
  • Core Banking, payments engine and fraud detection
  • Cyber Data Collection and Dissemination
  • ESB Replacement
  • Data Pipelining
  • eCommerce and Customer 360
  • Mainframe offloading

Apache Kafka Fundamentals

  • Architecture
  • ZooKeeper’s role
  • Topics, Partitions and Segments
  • The commit log and streams
  • Brokers and Broker replication
  • Producers Basics
  • Consumers, Consumer groups and Offsets

How Kafka Works

  • High-level code overview for a basic producer and a basic consumer
  • High Availability through Replication
  • Data Retention Policies
  • Producer Design and Producer Guarantees
  • Delivery Guarantees, including Exactly Once Semantic
  • Partition strategies
  • Consumer group rebalances
  • Compacted Topics
  • Troubleshooting strategies
  • Security overview

Integrating Kafka into your Environment

  • Get streams of data into and out of Kafka with Kafka Connect and REST Proxy
  • Maintain data formats and ensure compatibility with Schema Registry and Avro
  • Build real-time streaming applications with Confluent KSQL & Kafka Streams

The Confluent Platform

  • The Streaming Platform as the Central Nervous System
  • Deployment Models — on premise versus SaaS
  • The Confluent Control Center
  • Role Based Access Control (RBAC)
  • The Confluent CLI
  • Confluent Operator
  • The Confluent Hub for Certified Connectors
Contact Us 1-800-803-3948
Contact Us Live Chat
FAQ Get immediate answers to our most frequently asked qestions. View FAQs arrow_forward