• No products in the cart.

Import data to HDFS, HBase and Hive from a variety of sources, including Twitter and MySQL

Course Description

Flume and Sqoop play a special role in the Hadoop ecosystem. They transport data from sources like local file systems, HTTP, MySQL and Twitter which hold/produce data to data stores like HDFS, HBase and Hive. Both tools come with built-in functionality and abstract away users from the complexity of transporting data between these systems.
Flume: Flume Agents can transport data produced by a streaming application to data stores like HDFS and HBase.
Sqoop: Use Sqoop to bulk import data from traditional RDBMS to Hadoop storage architectures like HDFS or Hive.

Learning Outcomes

  • Use Flume to ingest data to HDFS and HBase
  • Use Sqoop to import data from MySQL to HDFS and Hive
  • Ingest data from a variety of sources including HTTP, Twitter and MySQL


  • Knowledge of HDFS is a prerequisite for the course
  • HBase and Hive examples assume basic understanding of HBase and Hive shells
  • HDFS is required to run most of the examples, so you’ll need to have a working installation of HDFS

Who is this course intended for?

Engineers building an application with HDFS/HBase/Hive as the data store

Engineers who want to port data from legacy data stores to HDFS


Your Instructor


Loonycorn is us, Janani Ravi and Vitthal Srinivasan. Between us, we have studied at Stanford, been admitted to IIM Ahmedabad and have spent years working in tech, in the Bay Area, New York, Singapore and Bangalore.

Janani: 7 years at Google (New York, Singapore); Studied at Stanford; also worked at Flipkart and Microsoft

Vitthal: Also Google (Singapore) and studied at Stanford; Flipkart, Credit Suisse and INSEAD too

We think we might have hit upon a neat way of teaching complicated tech courses in a funny, practical, engaging way, which is why we are so excited to be here on Learnsector!

We hope you will try our offerings and think you’ll like them 🙂

Course Curriculum

Introduction 00:00:00
Why do we need Flume and Sqoop?
Why do we need Flume and Sqoop? 00:00:00
Installing Flume 00:00:00
Flume Agent – the basic unit of Flume 00:00:00
Example 1 : Spool to Logger 00:00:00
Flume Events are how data is transported 00:00:00
Example 2 : Spool to HDFS 00:00:00
Example 3: HTTP to HDFS 00:00:00
Example 4: HTTP to HDFS with Event Bucketing 00:00:00
Example 5: Spool to HBase 00:00:00
Example 6: Using multiple sinks and Channel selectors 00:00:00
Example 7: Twitter Source with Interceptors 00:00:00
[For Linux/Mac OS Shell Newbies] Path and other Environment Variables 00:00:00
Installing Sqoop 00:00:00
Example 8: Sqoop Import from MySQL to HDFS 00:00:00
Example 9: Sqoop Import from MySQL to Hive 00:00:00
Example 10: Incremental Imports using Sqoop Jobs 00:00:00

Course Reviews


  • 5 stars0
  • 4 stars0
  • 3 stars0
  • 2 stars0
  • 1 stars0

No Reviews found for this course.

  • $99.00 $15.00
  • Course Certificate

    Related Courses

    © Learnsector