This is the code repository for Hands-On Beginner’s Guide on Big Data and Hadoop 3 [Video], published by Packt. It contains all the supporting project files necessary to work through the video course from start to finish.
Do you struggle to store and handle big data sets? This course will teach to smoothly handle big data sets using Hadoop 3.
The course starts by covering basic commands used by big data developers on a daily basis. Then, you'll focus on HDFS architecture and command lines that a developer uses frequently. Next, you'll use Flume to import data from other ecosystems into the Hadoop ecosystem, which plays a crucial role in the data available for storage and analysis using MapReduce. Also, you'll learn to import and export data from RDBMS to HDFS and vice-versa using SQOOP. Then, you'll learn about Apache Pig, which is used to deal with data using Flume and SQOOP. Here you'll also learn to load, transform, and store data in Pig relation. Finally, you'll dive into Hive functionality and learn to load, update, delete content in Hive.
By the end of the course, you'll have gained enough knowledge to work with big data using Hadoop. So, grab the course and handle big data sets with ease
- Focus on the Hadoop ecosystem to understand big data and how to manage it
- Learn the basic commands used by big data developers and the structure of the Unix OS.
- Understand the HDFS architecture and command line to deal with HDFS files and directories
- Import data using Flume and analyze it using MapReduce
- Export and import data from RDBMS to HDFS and vice-versa with SQOOP
- Use command-line language Pig Latin for data transformation operations
- Deal with stored data and learn to load, update, and delete data using Hive
To fully benefit from the coverage included in this course, you will need:
This course has the following software requirements: