


Is available in Apache Kafka to perform such data processing as described above.Īpart from Kafka Streams, alternative open source stream processing tools include Apache Storm andĮvent Sourcing Event sourcing is a style of application design where state changes are logged as a Starting in 0.10.0.0, a light-weight but powerful stream processing library called Kafka Streams Such processing pipelines create graphs of real-time data flows based on the individual topics. Many users of Kafka process data in processing pipelines consisting of multiple stages, where raw input data is consumed from Kafka topics and thenĪggregated, enriched, or otherwise transformed into new topics for further consumption or follow-up processing.įor example, a processing pipeline for recommending news articles might crawl article content from RSS feeds and publish it to an "articles" topic įurther processing might normalize or deduplicate this content and publish the cleansed article content to a new topic Ī final processing stage might attempt to recommend this content to users.

In comparison to log-centric systems like Scribe or Flume, Kafka offers equally good performance, stronger durability guarantees due to replication, This allows for lower-latency processing and easier support for multiple data sources and distributed data consumption. Kafka abstracts away the details of files and gives a cleaner abstraction of log or event data as a stream of messages. Log aggregation typically collects physical log files off servers and puts them in a central place (a file server or HDFS perhaps) for processing. Many people use Kafka as a replacement for a log aggregation solution. This involves aggregating statistics from distributed applications to produce centralized feeds of operational data.

Kafka is often used for operational monitoring data.
#INSYNC LOGIN OFFLINE#
Offline data warehousing systems for offline processing and reporting.Īctivity tracking is often very high volume as many activity messages are generated for each user page view. These feeds are available for subscription for a range of use cases including real-time processing, real-time monitoring, and loading into Hadoop or This means site activity (page views, searches, or other actions users may take) is published to central topics with one topic per activity type. The original use case for Kafka was to be able to rebuild a user activity tracking pipeline as a set of real-time publish-subscribe feeds. In this domain Kafka is comparable to traditional messaging systems such as ActiveMQ or In our experience messaging uses are often comparatively low-throughput, but may require low end-to-end latency and often depend on the strong Solution for large scale message processing applications. In comparison to most messaging systems Kafka has better throughput, built-in partitioning, replication, and fault-tolerance which makes it a good Message brokers are used for a variety of reasons (to decouple processing from data producers, to buffer unprocessed messages, etc). Kafka works well as a replacement for a more traditional message broker. Here is a description of a few of the popular use cases for Apache Kafka®.įor an overview of a number of these areas in action, see this blog post.
