Apache Flume Tutorial


Apache Flume is tool/Service for efficiently collecting, combining and moving large amounts of log data. Flume is a highly distributed, reliable and available tool/service. Flume is an intermediate tool/service for data collection from external data sources and sends to centralized data storage like Hadoop HDFS or HBase.


The tutorial is intended for Hadoop developers who are having minimum Hadoop and java knowledge.

Audiences & Prerequisites:

The tutorial can be understanble by anyone as all the topics are covered with in-depth information.

Users having the Java and hadoop technologies experience in addition to computation background will be an advantage in understanding the concepts easily. Users without having computation background may need to go through the topic more than once to understand the concept clearly.

Table of contents:

  1. Flume Tutorial