The analysis is aimed at studying and finding out


Assignment - Big Data Management

Domain: Airline Industry

Project: U.S. Airline/AirtrafficAnalysis

Analysis and Data Synopsis:

The data being considered for our project is on U.S. Air-Traffic data of 2016 from U.S. Department of Transportation (DOT), Bureau of Transportation Statistics (BTS) consisting of 16 U.S. air carriers that have at least 1 percent of total domestic scheduled-service passenger revenues, plus two other carriers that report voluntarily. The data cover nonstop scheduled-service flights between points within the United States (including territories).

The analysis is aimed at studying and finding out air-traffic patterns in 2016 across various airports, airlines and sectors over US. The study also aims to analyze the performance of U.S. airliners across various parameters and study the factors affecting the performance as well as patterns in performance across various time periods. The dataset size is around 465 MB and covers around 465k records of data.

Analysis Aimed at Covering:
1. Comprehending/Overall Summarization of the Air-Traffic Dataset

2. U.S. Air-Traffic/Airport Operation Analysis 2016 -
a. Airports with maximum air-traffic flows - Volume of air-traffic
b. Airports serving maximum no. of airlines
c. Airports service pattern across various time periods

3. U.S.Airline Functionality Analysis 2016 -
a. Aircrafts with maximum flight services
b. Airport coverage density
c. Weekday air-travel density analysis

4. Aircraft flight performance analysis -
a. Aircraft delays - causes, volume, frequency
b. Aircraft Cancellation - count, airports, airlines, period, causal effects

5. Diverted Air-lines analysis -
a. Summary
b. Cause and density

Assignment contains four questions and will ask you to get familiar with aspects of Apache Spark. While first three questions require you to get familiar with Spark programming, the last question will ask you to understand an existing code and explain it in simple terms.

Q1. Consider the two data files (users.csv, transactions.csv). Users file has the following fields:
a) UserID
b) EmailID
c) NativeLanguage
d) Location

Transactions file has the following fields:
a) Transaction_ID
b) Product_ID
c) UserID
d) Price
e) Product_Description

By making use of Spark Core (i.e. without using Spark SQL) find out:
a) Count of unique locations where each product is sold.
b) Find out products bought by each user.
c) Total spending done by each user on each product.

You cannot make use of Spark SQL for this.

Q2. For this question, please make use of the attached JSON file (tweets.json). Make use of Spark SQL library to answer the following questions:
a) Save the dataset as a DataFrame, and print the schema.
b) Get all of the tweets made by a user (any user would work. We should be able to replace user names to get tweets by that particular user).
c) Find count of all tweets by each user user.
d) Get a list of all of the people who are mentioned in tweets.
e) Count the number of time each person is mentioned in the entire dataset of tweets.
f) Give top 50 users who are mentioned the most.
g) Get a list of all hashtags mentioned in the dataset.
h) Find how many times each hashtag is mentioned in the dataset.
i) Get a list of all of the people who are located in a particular city (e.g. Paris)
j) Get country wise distribution of users, and find out which country ranks highest in terms of number of tweets, and number of users.
k) Find out number of tweets where a user is from France and mentions Paris in their tweets.

Q3. For this question, you would need to use the concepts learnt in Graph analytics session, and use datasets trip.csv and station.csv. The two files contain bike sharing data provided by SF Bay Area Portal. Trip.csv file contains following fields:
a) tripId
b) Duration
c) StartDate
d) EndDate
e) StartStation
f) StartTerminal
g) EndDate
h) EndStation
i) EndTerminal
j) BikeID
k) SubscriberType
l) ZipCode
Station.csv file contains following fields:

a) stationId
b) Name
c) Lat (Latitude)
d) Long (Longitude)
e) Dockcount
f) Landmark
g) Installation

Using the two files, please perform the following:

a) Import the data and create a graph using GraphFrames (Hint: Your graph will have nodes and edges. Nodes here would be individual stations so id field would be name field in station.csv file. Edges would have src and dst so it would Start Station and End Station fields in trip.csv file respectively. You can make use of other fields as properties of nodes and edges).
b) Find out number of incoming connections and outgoing connections for each node and print the top 10 nodes.
c) Find out which are the most common direct routes that people take and print top 10.
d) From the analysis in b, see which are the stations where people most frequently start their trips but do not come back. (Hint: You might have to think of incoming connections as a ratio of outgoing connections). Print top 10 such stations.
e) Find all such patterns where any station a is connected to station b, b is connected to c, but c is not directly connected to a.
f) Run a PageRank algorithm to figure out which is the most important station in the entire graph.

Q4. Consider the Movie Similarities code and problem that was discussed during the class (Session 4). Please provide a brief write-up on the problem, steps needed to arrive at the solution (recommendation system), and how exactly those steps are implemented in the code. While you are doing so, please also mention what each line of code does (It is not sufficient to mention what each block of code does, you would have to provide explanation for each line).

Solution Preview :

Prepared by a verified Expert
Programming Languages: The analysis is aimed at studying and finding out
Reference No:- TGS02588312

Now Priced at $65 (50% Discount)

Recommended (90%)

Rated (4.3/5)