Modeling Data Warehouses using Apache Hive

12 ratings
Offered By
Coursera Project Network
In this Guided Project, you will:

Design a data warehouse

Build a data warehouse using Apache Hive over apache Hadoop

Query a data warehouse for specific records

Clock1 hour 40 minutes
CloudNo download needed
VideoSplit-screen video
Comment DotsEnglish
LaptopDesktop only

By the end of this project, you will create your own data warehouse by using a star schema modeling technique that holds baseball raw data gathered from multiple baseball leagues in the States. During the course of this project, you will be using Apache Hadoop and Apache Hive as the main tools to build your very own Data Warehouse using HQL as the querying language, which is very similar to SQL. The learner will be introduced to the star schema modeling technique, database creation and handling, data insertion, and data storage using two different methods in Apache Hadoop. Data Warehouses are viewed as the core of any application, as any further step that requires dealing with the data stored such as Data Analytics, building a recommendation system, or applying any sort of Machine Learning will inevitably point out to the start point, which is the data warehouse modeling. This guided project is for students who are aiming to start their careers in the broad field of Data Engineering and the specific field of Data Warehouses Modeling. It will also be of interest to people who are keen on exploring Big Data platforms such as Apache Hadoop and Apache Hive. Note: This project works best for learners who are based in the North America region. We’re currently working on providing the same experience in other regions.

Skills you will develop

Data Warehouse DevelopmentDatabase (DB) DesignData WarehousingDatabase Management Systems

Learn step-by-step

In a video that plays in a split-screen with your work area, your instructor will walk you through these steps:

  1. By the end of Task 1,you will be able to identify the star schema format in our project description.

  2. By the end of Task 2, you will be able to identify the attributes and their data types.

  3. By the end of Task 3, you will be able to create a database and tables in the database using Apache Hadoop and Apache Hive.

  4. By the end of Task 4, you will be able to insert data into tables manually.

  5. By the end of Task 5, you will be able to amend tables by adding primary and foreign keys constraints.

  6. By the end of Task 6,   you will be able to apply useful queries to your data warehouse to fetch specific data records.

How Guided Projects work

Your workspace is a cloud desktop right in your browser, no download required

In a split-screen video, your instructor guides you step-by-step



View all reviews

Frequently asked questions

Frequently Asked Questions

More questions? Visit the Learner Help Center.