Yammer SQL Project
Yammer SQL Project
user_id: A unique ID per user. Can be joined to user_id in either of the other tables.
created_at: The time the user was created (first signed up)
state: The state of the user (active or pending)
activated_at: The time the user was activated, if they are active
company_id: The ID of the user's company
language: The chosen language of the user
Digging in – Data Available (2)
Table 2: Events
This table includes one row per event, where an event is an action that a user has taken on Yammer. These
events include login events, messaging events, search events, events logged as users progress through a signup
funnel, events around received emails
user_id: The ID of the user logging the event. Can be joined to user\_id in either of the other tables.
user_id: The ID of the user to whom the event relates. Can be joined to user_id in either of
the other tables.
occurred_at: The time the event occurred.
action: The name of the event that occurred. "sent_weekly_digest" means that the user
was delivered a digest email showing relevant conversations from the previous
day. "email_open" means that the user opened the email. "email_clickthrough"
means that the user clicked a link in the email.
Digging in – Data Available (4)
Table 4: Rollup Periods
The final table is a lookup table that is used to create rolling time periods.
period_id: This identifies the type of rollup period. The above dashboard uses period 1007, which is rolling 7-day
periods.
time_id: This is the identifier for any given data point — it's what you would put on a chart axis. If time_id is 2014-
08-01, that means that is represents the rolling 7-day period leading up to 2014-08-01.
pst_start: The start time of the period in PST. For 2014-08-01, you'll notice that this is 2014-07-25 — one week prior.
Use this to join events to the table.
pst_end: The start time of the period in PST. For 2014-08-01, the end time is 2014-08-01. You can see how this is
used in conjunction with pst_start to join events to this table in the query that produces the above chart.
All users
Active users
Digging in –
Existing user engagement
Next we will see if the change is brought up by
existing users. By grouping the users into cohort
based on when they signed up at Yammer.
• This chart shows a decrease in engagement
among the existing users who signed up more
than 10 weeks prior.
• Although a deeper pattern is visible that user
engagement decreases as the account age of the
user goes up.
• Hence there probably isn’t any
restriction affecting new traffic
to the site or change in
the rank on search engines.
Digging in – Engagement per user
With this code we can see if the drop
was due to a drop in users or a drop in
the number of engagements/user too.
• As established earlier, now we can
clearly note a drop of 8.6% in the
number of total users from July 2014
to August 2014
• The number of events per user also
fell from 30 to 26, a decrease of
13.3%, which is also fairly significant
• We will now shift our focus to event
related data, and investigate whether
a particular type of event is triggering
the drop off.
Digging in – Engagement event details
There are 17 event names that are
classified as an engagement generating
event. I calculated the month on month
change to find whether there is a
particular feature that is not working or
not used by the users anymore.
• The monthly aggregate of all these 17
event names shows a sharp decline in
August 2014
• The dip in engagement was largely
attributed to home_page,
like_message_ view_inbox,
send_message, and login.
• Hence, It seems like the drop in all of
these events is related. Again pointing
to the fact that users are logging in less
or dropping off.
Digging in – Regional variation
Lets quickly rule out the possibility that
the drop in engagement is accrued to
regional variation in users.
• The regional composition shows that
the drop in engagement has been
observed consistently across
countries.
• Hence it is not a regional problem
and may be related to the
engagement events itself.
Digging in –
Email action
I wanted to take an aggregated look at
the email activity to see if there was a
change in emails sent, click-through
rates (CTRs), or something else that
may have caused a reduction in active
users.
• Notice that there was a steady
increase in the action taken for re-
engagement emails but there was a
decrease in clickthrough rates, weekly
digest and emails opened.
• But compared to July 2014, only
clickthrough rate shows a decline of
27%, other activity has remained
consistent.
Digging in –
Email action
To gather more information, I want to
see if we can narrow the problem even
further by email opening issue or CTR
issue for email type
• It's evident that the drop in clicks is
due to the weekly digest email and
not re-engagement email.
• While it is particularly
attributed to the fall in
click through rate instead
of email opening rate.
Digging in – Device
I wanted to take a deeper look into the
clickthrough rates to see if the decline
had anything to do with the devices. It
could have had to do with the type of
operating system (IOS vs Android) or
mobile vs desktop.
• We see that the clickthrough rates on
laptops and computers were stable
from July to August, but not the
tablets and cellphones.
Digging in – Device
By categorizing the device names into
'mobile', 'tablet', and 'laptop’, it is
possible to determine that this is truly
the case
• The drop in clickthrough rates was
attributed specifically to mobile
devices and tablets
Summary of the analysis
• The drop in engagement was mainly attributed to a drop in five engagement
events (home_page, like_message_ view_inbox, send_message, and login).
• The decrease in events was caused by a reduction in total active users MoM,
as well as a decrease in engagement per user.
• After an aggregated look at the emails table, there was a significant decrease
in MoM click-through rates for weekly digest from July to August. The decline
is not caused by the re-engagement email.
• By segmenting the clickthrough rates by device type (mobile, tablet, laptop),
the drop in clickthrough rates was attributed to mobile and tablet devices.
Recommendation
• Immediately take a deeper look into the weekly digest emails
specifically for mobile devices and tablets.
• It's possible that there's a technical problem, making it difficult for
users to click the email or simply a UX problem, where the content
and layout of the email are not enticing users to click.
• A good first step would be to see what changes have been made
from July to August and working backward.
• Make a quality check on the code that logs data related to digest
emails to verify if the data pipelines are working fine.
• The engagement dips were more pronounced among longer-term
users, indicating a potential need for re-engagement strategies.
Further Analysis
Some other things that could've been looked at include the
following:
• The engagement for July was quite higher than usual, identify
what caused this.
• Cohort analysis revealed that there was a short user lifecycle,
further detailed investigation on it is necessary.
• Detailed Analysis by language
• Detailed Analysis by geography