Processing Real-time events with Apache Storm
In this course, we will explore Apache Storm and use it with Apache Kafka to develop a multi-stage event processing pipeline. In an event processing pipeline, each stage is a purpose-built step that performs some real-time processing against upstream event streams for downstream analysis. This produces increasingly richer event streams, as data flows through .
Real time Data Ingestion in HBase & Hive using Storm Bolt
In this tutorial, we will build a solution to ingest real time streaming data into HBase and HDFS.
In previous tutorial we have explored generating and processing streaming data with Apache Kafka and Apache Storm. In this tutorial we will create HDFS Bolt & HBase Bolt to read the streaming data from the Kafka Spout and persist in Hive & HBase tables.
Processing streaming data in Hadoop with Apache Storm
How to use Apache Storm to process real-time streaming data in Hadoop with Hortonworks Data Platform.