Data Flow for the Hadoop Ecosystem

Apache Hadoop 2.0    |    Intermediate
  • 12 videos | 59m 5s
  • Earns a Badge
Likes 10 Likes 10
Data must move into and through Hadoop for it to function. Here we look at Hadoop and the data life cycle management, and use Sqoop and Hive to flow data.

WHAT YOU WILL LEARN

  • describe the data life cycle management
    recall the parameters that must be set in the Sqoop import statement
    create a table and load data into MySQL.
    use Sqoop to import data into Hive
    recall the parameters that must be set in the Sqoop export statement
    use Sqoop to export data from Hive
  • recall the three most common date datatypes and which systems support each
    use casting to import datetime stamps into Hive
    export datetime stamps from Hive into MySQL
    describe dirty data and how it should be preprocessed
    use Hive to create tables outside the warehouse
    use pig to sample data

IN THIS COURSE

EARN A DIGITAL BADGE WHEN YOU COMPLETE THIS COURSE

Skillsoft is providing you the opportunity to earn a digital badge upon successful completion on some of our courses, which can be shared on any social network or business platform.

Digital badges are yours to keep, forever.

YOU MIGHT ALSO LIKE

PEOPLE WHO VIEWED THIS ALSO VIEWED THESE

Likes 39 Likes 39  
Likes 26 Likes 26