Other

Analytics for SharePoint on-premises

SharePoint empowers 80 per cent of Fortune 500 companies. With a huge portion of businesses relying on SharePoint for a constantly increasing amount of content, many administrators find that they need to perform some sort of site analytics inside SharePoint.

SharePoint empowers tens of thousands of companies around the world, including 80 percent of Fortune 500 companies. With a huge portion of businesses relaying on SharePoint for a constantly increasing amount of content, many administrators find that they need to perform some sort of site analytics inside SharePoint.

Often, they need to answer simple questions, such as:

  • How many active users are on my farm?
  • Which sites are the most active ones?
  • Which sites have been inactive for a long time?
  • What are the visitors’ trends for a specific site?

Out-of-the-box SharePoint Analytics and usage

Out of the box, SharePoint does provide some help with these questions. SharePoint 2010 had a dedicated Web Analytics service application that provided some insights into the behavior of users on your SharePoint sites. It did a pretty decent job, but primarily for performance reasons, it was discontinued in SharePoint 2013. SharePoint 2013 added a new SharePoint Analytics Processing Component, which is now a part of the Search service. It provides you with some basic reports:

  • Popularity Trends shows usage event statistics for a site collection, a site, or an individual item in a library or list.
  • Most Popular Items shows usage event statistics for all items in a library or list.

You can find detailed instructions on how to use this report here.

sharepoint popularity trends

The main problem with these reports is that they will not give you a big picture of what is happening on your farm. You need to generate an Excel report site collection by site collection, and even when you try to narrow your report down to a single site collection, you still must go site by site, and list by list, to see what is happening. That requires too much work and is too time-consuming to be useful. Another pitfall is that there is no API to automate this work.

Doing your own SharePoint Analytics

Configure usage and health data collection in SharePoint

With so many limitations to the out of the box options, you might need to resort to other options in your search for answers to the questions that started all of this. One option is to use the Page Request events collected by the Usage and Health Data Collection Service, which is easy to configure: You just need to navigate to Central Administration > Monitoring > Configure usage and health data collection in SharePoint Server and ensure that “Enable usage data collection” and “Page requests” are checked. Once configured, SharePoint will immediately start to collect usage logs.

configure analytics

First, the usage logs are saved to the disk on each WFE server in your farm in the default location C:Program FilesCommon FilesMicrosoft SharedWeb Server Extensions16LOGS. There is a timer job, “Microsoft SharePoint Foundation Usage Data Import,” which will move them to the Logging Database Server every five minutes. Depending on the number of active users and events you wish to log, this database can grow very fast, so it is important to check the data retention settings for events that have a default value of 14 days. You can check these with the following PowerShell cmdlet Get-SPUsageDefinition and change them with Set-SPUsageDefinition.

Querying the logging database for Sharepoint Analytics logs

Now that you have set everything up, you can generate reports by querying the logging database. Page Request events are stored in a total of 31 partitioned tables ranging from RequestUsage_Partition0 to RequestUsage_Partition31. To avoid having to UNION across all partitions it is best not to use the tables directly but to use the RequestUsage view instead. 

Once you examine the RequestUsage view, you will notice the most important columns: LogTime, UserLogin, WebApplicationId, SiteId, SiteUrl, and WebId. So, knowing all this, let’s try to determine how many active users were on my farm last week. Here is the sample query:

SELECT  COUNT(RowId) AS FarmHits,
       COUNT(DISTINCT UserLogin) AS FarmUniqueUsers
FROM [dbo].[RequestUsage]
WHERE PartitionId in (SELECT PartitionId from fn_PartitionIdRangeMonthly('20181001', '20181007'))
       AND UserLogin IS NOT NULL
       AND DataLength(UserLogin) > 0
       AND WebApplicationId <> '00000000-0000-0000-0000-000000000000'
       AND SiteId <> '00000000-0000-0000-0000-000000000000'
       AND WebId <> '00000000-0000-0000-0000-000000000000'

From the query, you will notice that we are trying to filter out things that are not a real user interaction with a page on your farm. To further improve the reliability of this query, it would be a good idea to filter out a few more things:

  • Filter UserLogin for Managed accounts and the Farm account.
  • Filter RequestType to only look at the http GET requests.
  • Filter DocumentPath to only look at the .aspx pages and ignore a lot of unnecessary API calls.

Knowing all this, let’s try to solve another problem by finding the most active site collections last week:

SELECT TOP 1000
    SiteId AS SiteId,
       SiteUrl,
    COUNT(RowId) AS Hits,
    COUNT(DISTINCT UserLogin) AS UniqueVisits
FROM [dbo].[RequestUsage]
WHERE PartitionId in (SELECT PartitionId from fn_PartitionIdRangeMonthly('20181001', '20181009'))
       AND UserLogin IS NOT NULL
       AND DataLength(UserLogin)>0
       AND WebApplicationId <> '00000000-0000-0000-0000-000000000000'
       AND SiteId <> '00000000-0000-0000-0000-000000000000'
       AND WebId <> '00000000-0000-0000-0000-000000000000'
       AND DocumentPath LIKE '%.aspx'
       AND RequestType = 'GET'
GROUP BY SiteId, SiteUrl
ORDER BY COUNT(RowId) DESC

You can see that it is simple to modify the query to get information about site collections. You could easily go down to the subsite level by adding a GROUP BY on the WebId column.

Although this approach gives you some answers, you need to be aware of its limitations:

  • The query can become very slow on larger farms if you query a time period longer than a few days.
  • The max history you can set for the data retention period is 31 days, and there is no way to go further backward in your reports.
  • It is very hard to write a query to get information about visitors at a document library or at the document level.
  • The database can use a lot of disk space.

Using SPDockit for SharePoint Analytics

SPDockit’s Analytics and Audit reports can help you get answers to your questions without losing a lot of sleep. They are easy to configure and overcome most of the limitations mentioned before.

You can get a clear picture of how your users are interacting with the farm on the SharePoint Analytics report introduced in version 8. It includes information and summaries about the number of page hits, unique visitors, storage usage, and the number of subsites for each site collection, as well as detailed trends showing how these metrics changed in the past and how they are behaving now. For those who are looking for some additional insight into how the site collections are performing, you can quickly narrow down to the subsite level and get a report that answers questions about site activity.

sharepoint analytics navigation

With SPDockit, there is no need to stop at the subsite level with the number of views and visitors shown by the Analytics reports. You can use our new Audit reports to get information about who is visiting SharePoint sites and opening files. Some of the events you can track here are: 

  • Opened and downloaded documents, viewed items in lists, and viewed item properties;  
  • Edited items; 
  • Checked-out and checked-in items; 
  • Items that have been moved and copied to other locations in the site collection; 
  • Deleted and restored items; and 
  • Searched queries and terms. 

New SPDockit is now available, so download 30-day free trial and try it yourself.

Subscribe to our Newsletter

Related Posts