0% found this document useful (0 votes)
103 views

Trainity Project 3

Uploaded by

Vivek Dhande
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
103 views

Trainity Project 3

Uploaded by

Vivek Dhande
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 18

Operation Analytics and Investigating

Metric Spike

By- Abhishek Salvi

Description:
Operational Analytics is a crucial process that involves analyzing
a company's end-to-end operations. This analysis helps identify
areas for improvement within the company. As a Data Analyst,
you'll work closely with various teams, such as operations,
support, and marketing, helping them derive valuable insights
from the data they collect.
Case Study 1: Job Data Analysis

A] Jobs Reviewed Over Time:

Objective: Calculate the number of jobs reviewed per hour for


each day in November 2020

QUERY:

SELECT ds as date, COUNT(*) as jobs_count , COUNT(*)/24


as JobsPerHourPerDayFROM job_dataGROUP BY date;

Conclusion:

Here we can have number of jobs reviewed per hour per day.
B] Throughput Analysis:

Objective: Calculate the 7-day rolling average of throughput


(number of events per second).

# 7 days rolling # Day by day event per second

QUERY: QUERY:

select total_events, total_time, total_events / total_time as select ds, count(event), count(event) /


7_dayRolling_events_per_secondfrom sum(time_spent) as DaybyDayRollingAvg
( select COUNT(event) AS total_events, from job_datagroup by ds;
SUM(time_spent) AS total_time
from job_data)as countandsum;
Conclusion:

here we can have 7 days rolling average of throughput and also having day by day throughput.
As above results we can see clearly that daily throughput gives us broader picture about each day
throughput.

C] Language Share Analysis:

Objective: Calculate the percentage share of each language in


the last 30 days.

QUERY:

select language,(countOfLanguage / sumOfLanguages) * 100


AS PercentageOfLanguage from ( select language,
count(language) as countOfLanguage, sum(count(language))
over () as sumOfLanguages from job_data group by
language ) as subquery;
Conclusion:

here we have the result of percentage of each language has in last 30 days.

D] Duplicate Rows Detection:

Objective: Identify duplicate rows in the data

QUERY:

select job_id, actor_id, event, count(*) AS


duplicate_countfrom job_data
group by job_id, actor_id, event
having count(*) > 1;

Conclusion:

As we see currently we do not have any duplicate records in our data.


Case Study 2: Investigating Metric Spike

A]Weekly User Engagement:

Objective: Measure the activeness of users on a weekly basis.

QUERY:

Conclusion:

As a result we were having number of week and engagement count of that week.
This gives us activeness of user on weekly basis.
B] User Growth Analysis:

Objective: Analyze the growth of users over time for a product.

QUERY:
Conclusion:

As a result we were having growth of users count of each product on month on month
basis.
C] Weekly Retention Analysis:

Objective: Analyze the retention of users on a weekly basis after


signing up for a product.
D] Weekly Engagement Per Device:

Objective: Measure the activeness of users on a weekly basis


per device.

QUERY:
Conclusion:

As a result we were having average weekly engagement of users per device.

E] Email Engagement Analysis:

Objective: Analyze how users are engaging with the email


service

QUERY:

Conclusion:

As a result we were having the total count of users engagement by email services.

You might also like