Using Hive & Pig with Hadoop
Apache Hadoop 2.0 | Beginner
- 7 videos | 32m 57s
- Includes Assessment
- Earns a Badge
There are components other than MapReduce that let you write code to process large data sets stored in Hadoop. Let's see how to work with two such components - Hive and Pig.
WHAT YOU WILL LEARN
Understand the basics of apache hive and hiveql. describe how hiveql is similar to ansi sql and how it can be used to select data. understand how hiveql is implicitly transformed into mapreduce jobs.Understand usage of the four file formats supported in hive, which are textfile, sequencefile, orc and rcfile. demonstrate each, and be able to describe each of the four.Understand how to use custom hive data types such arrays and maps to write custom hive jobs. learn ddl hive commands.Understand pig and how it is used. demonstrate how to use pig latin like sql to obtain data. understand how to use pig as a component to build complex and large mapreduce applications.
Learn how to write pig scripts. also understand the pig modes, local, mapreduce, and batch.Learn pig command such as load, limit, dump, and store for data read/write operators in pig latin. understand grunt commands used for ddl.Compare and contrast the internals and performance of mapreduce, hive, and pig. understand the strengths and weaknesses of the three.
IN THIS COURSE
EARN A DIGITAL BADGE WHEN YOU COMPLETE THIS COURSE
Skillsoft is providing you the opportunity to earn a digital badge upon successful completion on some of our courses, which can be shared on any social network or business platform.Digital badges are yours to keep, forever.
YOU MIGHT ALSO LIKE
Rating 4.7 of 54 users (54)
Course Hadoop HDFS Getting Started
Rating 4.7 of 45 users (45)