Op werkdagen voor 23:00 besteld, morgen in huis Gratis verzending vanaf €20

Using Flume

Flexible, Scalable, and Reliable Data Streaming

Specificaties
Paperback, 221 blz. | Engels
O'Reilly | 1e druk, 2014
ISBN13: 9781449368302
Rubricering
Hoofdrubriek : Computer en informatica
O'Reilly 1e druk, 2014 9781449368302
Verwachte levertijd ongeveer 16 werkdagen

Samenvatting

How can you get your data from frontend servers to Hadoop in near real time? With this complete reference guide, you’ll learn Flume’s rich set of features for collecting, aggregating, and writing large amounts of streaming data to the Hadoop Distributed File System (HDFS), Apache HBase, SolrCloud, Elastic Search, and other systems.

'Using Flume' shows operations engineers how to configure, deploy, and monitor a Flume cluster, and teaches developers how to write Flume plugins and custom components for their specific use-cases. You’ll learn about Flume’s design and implementation, as well as various features that make it highly scalable, flexible, and reliable.

- Learn how Flume provides a steady rate of flow by acting as a buffer between data producers and consumers
- Dive into key Flume components, including sources that accept data and sinks that write and deliver it
- Write custom plugins to customize the way Flume receives, modifies, formats, and writes data
- Explore APIs for sending data to Flume agents from your own applications
- Plan and deploy Flume in a scalable and flexible way—and monitor your cluster once it’s running

Specificaties

ISBN13:9781449368302
Taal:Engels
Bindwijze:paperback
Aantal pagina's:221
Uitgever:O'Reilly
Druk:1
Verschijningsdatum:20-9-2014
Hoofdrubriek:IT-management / ICT

Over Hari Shreedharan

Hari Shreedharan is a PMC Member and Committer on the Apache Flume Project. As a PMC member, he is involved in making decisions on the direction of the project. Hari is also a Software Engineer at Cloudera where he works on Apache Flume and Apache Sqoop. He also ensures that customers can successfully deploy and manage Flume and Sqoop on their clusters, by helping them resolve any issues they are facing. Hari completed his Bachelors from Malaviya National Institute of Technology, Jaipur, India and his Masters in Computer Science from Cornell University in 2010.

Andere boeken door Hari Shreedharan

Inhoudsopgave

Foreword
Preface

1. Apache Hadoop and Apache HBase: An Introduction
-HDFS
-Apache HBase
-Summary
-References

2. Streaming Data Using Apache Flume
-The Need for Flume
-Is Flume a Good Fit?
-Inside a Flume Agent
-Configuring Flume Agents
-Getting Flume Agents to Talk to Each Other
-Complex Flows
-Replicating Data to Various Destinations
-Dynamic Routing
-Flume’s No Data Loss Guarantee, Channels, and Transactions
-Agent Failure and Data Loss
-The Importance of Batching
-What About Duplicates?
-Running a Flume Agent
-Summary
-References

3. Sources
-Lifecycle of a Source
-Sink-to-Source Communication
-HTTP Source
-Spooling Directory Source
-Syslog Sources
-Exec Source
-JMS Source
-Writing Your Own Sources*
-Summary
-References

4. Channels
-Transaction Workflow
-Channels Bundled with Flume
-Summary
-References

5. Sinks
-Lifecycle of a Sink
-Optimizing the Performance of Sinks
-Writing to HDFS: The HDFS Sink
-HBase Sinks
-RPC Sinks
-Morphline Solr Sink
-Elastic Search Sink
-Other Sinks: Null Sink, Rolling File Sink, Logger Sink
-Writing Your Own Sink*
-Summary
-References

6. Interceptors, Channel Selectors, Sink Groups, and Sink Processors
-Interceptors
-Channel Selectors
-Sink Groups and Sink Processors
-Summary
-References

7. Getting Data into Flume*
-Building Flume Events
-Flume Client SDK
-Embedded Agent
-log4j Appenders
-Summary
-References

8. Planning, Deploying, and Monitoring Flume
-Planning a Flume Deployment
-Deploying Flume
-Monitoring Flume
-Summary

Index

Rubrieken

Populaire producten

    Personen

      Trefwoorden

        Using Flume