Data Flow for the Hadoop Ecosystem
Apache Hadoop 2.0
| Intermediate
- 12 Videos | 59m 5s
- Earns a Badge
Data must move into and through Hadoop for it to function. Here we look at Hadoop and the data life cycle management, and use Sqoop and Hive to flow data.
WHAT YOU WILL LEARN
-
describe the data life cycle managementrecall the parameters that must be set in the Sqoop import statementcreate a table and load data into MySQL.use Sqoop to import data into Hiverecall the parameters that must be set in the Sqoop export statementuse Sqoop to export data from Hive
-
recall the three most common date datatypes and which systems support eachuse casting to import datetime stamps into Hiveexport datetime stamps from Hive into MySQLdescribe dirty data and how it should be preprocesseduse Hive to create tables outside the warehouseuse pig to sample data
IN THIS COURSE
-
1.The World of Data8m 41sUP NEXT
-
2.Sqoop and Hive2m 40s
-
3.Loading SQL Data Tables1m 26s
-
4.Importing Data into Hive6m 35s
-
5.Sqoop and Hive Exports3m 14s
-
6.Exporting Data from Hive7m 23s
-
7.Working with Date data types1m 58s
-
8.Importing Datetime Stamps5m 50s
-
9.Exporting Datetime Stamps4m 41s
-
10.Preprocessing Data4m 24s
-
11.Cleaning with Functions7m 45s
-
12.Cleaning with Regular Expressions4m 27s
EARN A DIGITAL BADGE WHEN YOU COMPLETE THIS COURSE
Skillsoft is providing you the opportunity to earn a digital badge upon successful completion of this course, which can be shared on any social network or business platform
Digital badges are yours to keep, forever.