Coder Social home page Coder Social logo

big-data-movie-ratings's Introduction

big-data-movie-ratings

This Hadoop-based web app allows users to view movie ratings for any given movie and generate movie rankings.

There are two webpages in the app:

"/movie_ratings" to view movie ratings breakdown for any movie:

Ratings

"/movie_rankings" to display best or worst movies by genre:

Rankings

The app is modeled after the lambda architecture and consists of a batch layer (HDFS/Hive), serving layer (HBase), and speed layer (Spark and Kafka).

Background

This is a capstone project from the University of Chicago's "Big Data Application Architecture" MPCS course, taught by Professor Michael Spertus. The focus of the couse is how to architect large scale Big Data applications within the Hadoop ecosystem. Although the size of my dataset is on the order of gigabytes, the design is easily extensible to much larger datasets e.g. on the order of petabytes.

The Hadoop cluster and Web servers were all hosted on Google Cloud and managed internally by the UChicago teaching staff.

Data

The data is sourced primarily from the MovieLens database. One portion of the data comes from their website, https://grouplens.org/datasets/movielens/, and the other from a kaggle competition https://www.kaggle.com/rounakbanik/the-movies-dataset. Together the data is 1.7 GB in size. The csvs I used from these datasets were:

  • movies.csv
  • movies_metadata.csv
  • ratings.csv
  • links.csv

I did some pre-parsing on the data using Pandas. You can find my code in data_parsing.py. In particular, I created a "movie_genres.csv", sourced from "movies.csv", which uses a one hot encoding for the genre-types on each movie. i.e. there is now a boolean column for each genre (e.g. comedy, action, animation) for every movie-row in the table. Note that a movie can belong to multiple genres.

In addition, I split the "ratings.csv" data into two partitions. One before 1/1/2018 ("ratings_batch_input.csv") and the other post 1/1/2018 ("ratings_stream.csv"). The former populates the batch layer while the latter will be run as part of the speed-layer.

For brevity, the raw datasets are not included in the repo. Though if downloaded from the listed sources, one could easily rerun the attached scripts to generate the data.

Batch and serving layers

One can find all my hql scripts in the hql folder. My strategy for creating hive tables was to first upload the csvs in their raw form as external-csv tables, and then to map each to an orc-equivalent table. The hql files and corresponding hive/hbase tables are:

Next I created some derived tables based on the above tables. Some are pure hive tables, some are pure hbase, and some create both hive and hbase. I encourage looking at each hql file and/or using "describe" on each table in the hive shell to learn more about them. The hql files and corresponding tables are:

Web app

All web app code is found in the movie_app directory. It's a simple Node/Express app that uses mustache templating. The most interesting web-app code is found in ./movie_app/app.js where I render each web-page and query the serving layer for the requested data.

The first webpage movie_ratings.html queries first from the "nprabhu_movie_titles_to_ids" table, in order to resolve an imdbid from the title query. From there, I query the "nprabhu_movie_ratings_count_hbase" table in which each row stores the count per rating-type of the given movie. Here is an example query for Toy Story, which has ImdbId 114709. The "value" of each rating count appears as bytecode because their data-types are BigInt.

Ratings

The second webpage movie_rankings.html queries data from the "nprabhu_movie_ratings_average_hbase" table. In order to effectively "sort" the movies by ranking, I made my key for this table "genre:::avg_rating:::imdbid". Since the table stores the keys in lexographic order, the table is inherently stored in sorted order by genre because of the avg_rating component in the key. There is also an "ALL" genre that ranks movies irrespective of genre. Extracting the N worst movies for a genre is as simple as pulling the first N results of a scan; retrieving the N best movies just requires a reversed scan.

Here some example entries from this table for Toy Story: RankingsGet

And here is a sample scan. Note how the entries are effectively sorted by avg rating. RankingsScan

Speed layer

The speed layer (found in the speed-layer directory) allows the app to ingest real-time updates to the movie ratings data, and then update the hbase tables for immediate use of the new data. This layer consists of two parts; the producer and consumer of a kafka queue.

The producer is generated by ./speed-layer/producer.py; this simple python script reads in a data file "ratings_stream.csv", and acts as a producer that pipes in the raw data into the "nprabhu_movie_ratings" kafka queue. It inserts a record into the queue every 5 sec. Note that the stream csv is not included in the repo for brevity (see Data for how it is generated).

The consumer is generated by the StreamRating scala project. Within the src/main/scala directory, we find two files: StreamRatings.scala and RatingRecord.scala. The latter is a class representation of a row from "ratings_stream.csv" and input into the kafka queue. The former has all logic for consuming the records.

The main method in StreamRatings.scala mapped onto the record is "incrementRatings". Here we do the following:

  1. map the record's MovieId to an ImdbId (primary key in most hbase tables)
  2. Increment the appropriate rating in "movieRatingsCount" for this ImdbId
  3. Computes a new average rating for the movie
  4. Updates the "nprabhu_movie_ratings_average_by_genre_hbase" table with this newest avg rating,so all rankings have the latest result.
    • to find the old key, we must reference "nprabhu_movie_ratings_average_secondary_index" to find the old average as well as "nprabhu_movie_genres_hbase" to find all genres associated with the movie. Here are example queries from these tables: SecIndex SecIndex
  5. Updates "nprabhu_movie_ratings_average_secondary_index" to store new avg rating, so it's retreivable for future streaming updates.

Here are some screenshots of the running producer and consumer programs as evidence to the hbase tables being updated with the streamed data.

Producer demo:

Producer

Consumer demo:

Consumer

big-data-movie-ratings's People

Contributors

nprab428 avatar

Watchers

 avatar  avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.