Stream data to Hadoop using Apache Flume Overview * Integrate Flume with your data sources* Transcode your data en-route in Flume* Route and separate your data using regular expression matching* Configure failover paths and load-balancing to remove single points of failure* Utilize Gzip Compression for files written to HDFS In Detail Apache Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data. Its main goal is to deliver data from applications to Apache Hadoop's HDFS. It has a simple and flexible architecture...
Stream data to Hadoop using Apache Flume Overview * Integrate Flume with your data sources* Transcode your data en-route in Flume* Route and separate ...