Almost 4 hours of video instruction to help you learn the key concepts and features behind Kafka and how it can be used to solve many problems related to the transfer of data.
Apache Kafka Fundamentals LiveLessons provides a complete overview of Kafka and Kafka-related topics. The course begins with a general overview of Kafka and then dives into use cases and the design of Kafka. Learners are introduced to the concept of clusters as well as how to create producers and consumers. Additional topics, such as security, connectors, and streams, are covered. The course finishes with a discussion of how Kafka can be used with Big Data software.
With detailed, hands-on code examples, provided with the intent to translate to a coding language of your choosing, learners are walked through multiple scenarios involving projects from Twitter, Netflix, Yahoo!, and other real-world scenarios.
* Kafka concepts
* Use cases
* Kafka design
* API overview
* Installation and configuration
* Writing producers
* Writing consumers
* Kafka operations
What You Will Learn
* A complete overview of Kafka
* How to install and configure Kafka
* How Kafka fits into the big picture of Big Data
Who Should Take This Course?
* IT professional who are responsible for implementing and maintaining their organizations’ Big Data solution
* IT managers who have an interest in understanding how Kafka can solve part of their Big Data issues
Requires a basic knowledge of Linux and a programming language such as Java, Perl, Python, C, or Ruby.
Lesson 1, “Kafka Concepts”: You learn the essentials of what Kafka is, it’s history, and some of the key concepts of the Kafka solution. We compare Kafka to other potential solutions and point out the major advantages of using it.
Lesson 2, “Use Cases”: In this lesson you see how Kafka is used in real-world scenarios using projects from Twitter, Netflix, and Yahoo!
Lesson 3, “Kafka Design”: In this lesson you learn the design principals of Kafka. You start by learning key features and concepts, including the Kafka APIs, topics, logs, producers, and consumers. You also learn different ways in which Kafka can be used, including as a message system, a storage system, and for stream processing.
Lesson 4, “API Overview”: This lesson explores some basics of the producer and consumer APIs. The goal of this lesson is to start exploring the Java libraries that are used while creating Kafka applications.
Lesson 5, “Installation and Configuration”: As its names suggests, this lesson covers how to install and configure Kafka. It starts by discussing hardware and operating system considerations. It then covers the installation of Zookeeper, a separate software tool that is needed for Kafka to perform correctly. Finally, it covers how to install Kafka and perform some basic configuration operations.
Lesson 6, “Clusters”: Clusters are a key component of Kafka. The purpose and replication of clusters are discussed in this lesson. You also see a demonstration of the configuration of clusters.
Lesson 7, “Writing Producers”: This lesson explains the concepts and code for creating producers. The lesson focuses on how producers communicate with Kafka and how to handle serialization.
Lesson 8, “Writing Consumers”: This lesson discusses the concepts and code for creating consumers. The lesson starts with key concepts of consumers, including consumer groups, and then discusses how to create a consumer and subscribe to topics. The lesson also covers the concepts of polling, commits, rebalancing listeners, and deserializers.
Lesson 9, “Kafka Operations”: This lesson focuses on several features of Kafka that can be used to modify the manner in which Kafka performs, and shows you how you can modify them. It starts with how to administer topics and then jumps into a discussion on balancing and mirroring. Bandwidth and quotas are also covered.
Lesson 10, “Connectors”: Kafka Connect is a newer feature of Kafka that provides an easy way to pull in data from non-Kafka sources. The goal of this lesson is to provide a good overview of what connectors are and discuss some basic configuration of Kafka Connect, transformations, and the Connect REST API interface.
Lesson 11, “Streams”: Streams are a newer feature of Kafka that provide an easy way to pull in data from non-Kafka sources. This is a rather large topic, so the goal of this lesson is to provide a good overview of what streams are. In addition to covering the basics concepts of streams, we discuss how streams might be implemented in a system that also uses Kafka Connect.
About Pearson Video Training
Pearson publishes expert-led video tutorials covering a wide selection of technology topics designed to teach you the skills you need to succeed. These professional and personal technology videos feature world-leading author instructors published by your trusted technology brands: Addison-Wesley, Cisco Press, Pearson IT Certification, Prentice Hall, Sams, and Que Topics include IT Certification, Network Security, Cisco Technology, Programming, Web Development, Mobile Development, and more. Learn more about Pearson Video training at http://www.informit.com/video.
Lesson 1: Kafka Concepts
Lesson 2: Use Cases
Lesson 3: Kafka Design
Lesson 4: API Overview
Lesson 5: Installation and Configuration
Lesson 6: Clusters
Lesson 7: Writing Producers
Lesson 8: Writing Consumers
Lesson 9: Kafka Operations
Lesson 10: Connectors
Lesson 11: Streams