What Microsoft gives you
The Microsoft 365 admin centre provides a set of Copilot usage reports that are genuinely useful as a starting point. You can see, at a tenant level, how many users have been assigned a Copilot licence, how many have used any Copilot feature in the past 7, 30, or 90 days, and which specific Copilot features (in Word, Teams, Outlook, Excel, and so on) are being used. Microsoft Viva Insights adds a layer of aggregated, privacy-protected data on collaboration patterns and meeting time that can help contextualise Copilot usage within broader working patterns.
These tools tell you who is using Copilot and roughly how often. They do not tell you whether using Copilot is changing how people work, whether the outputs people are generating are useful, or whether the habit is sustainable. For a deployment review conversation with leadership, the dashboard numbers are a reasonable starting point. For a genuine adoption assessment, they are insufficient.
Metrics to ignore (or at least not to lead with)
Seat activation rate
Seat activation (the number of users who have logged into Copilot at least once) is almost meaningless as an adoption indicator. A user who opened Copilot once in week one, produced a mediocre output, and never returned will show as "activated" in perpetuity. For many organisations, seat activation sits above 60 or 70% while genuine ongoing use sits below 20%. Reporting seat activation as an adoption metric overstates progress significantly.
Daily active users (DAU) without context
Daily active users is a standard product metric borrowed from consumer software, where it makes sense as a proxy for engagement. In an enterprise Copilot context, it is misleading without qualification. A user who opens Copilot every day to generate one short output is counted identically to a user who uses it deeply across multiple tasks. Volume does not indicate depth, and depth is what drives productivity change. DAU without a corresponding measure of task complexity or time saving tells you very little.
Feature usage breadth
Knowing that 40 users have used Copilot in Teams and 25 have used it in Word sounds like progress. But feature breadth says nothing about frequency, quality, or whether the feature was used for a meaningful task or out of curiosity. A user who opened Copilot in Excel once to see what happened is not a productive Excel Copilot user. Feature breadth data creates a misleadingly comprehensive picture of adoption depth.
Metrics that actually matter
Spontaneous use rate
The most meaningful adoption indicator is whether users reach for Copilot without being prompted, outside of any programme, challenge, or instruction. For a detailed picture of what this looks like in practice, see our article on what good Microsoft 365 Copilot usage actually looks like. Spontaneous use is the clearest evidence that a habit has formed. It is best measured through self-report: in a week-nine programme survey, ask participants what percentage of their Copilot use in the past week was for tasks outside the formal programme challenge. A high spontaneous use rate signals genuine behaviour change; a low one signals that adoption remains prompted and fragile.
Confidence delta
"The gap between pre-programme and post-programme confidence is a stronger predictor of sustained use than any dashboard metric."
Self-reported confidence in using Copilot, measured before the programme begins and again at week nine, gives you a direct measure of capability change. Programmes that produce a large confidence delta (typically 50–70% increase in self-rated confidence across a cohort) also produce higher spontaneous use and better time saving results. Programmes that produce a small delta are not changing capability meaningfully, regardless of what the usage dashboard shows. Capturing baseline confidence before week one costs nothing and provides essential context for interpreting all other data.
Time saving self-report
Asking users to estimate how much time Copilot saved them in a given week is subjective, but it is directionally reliable and financially translatable. At week nine, ask participants: across all the tasks where you used Copilot this week, how much total time do you estimate you saved? Average the responses across the cohort and you have a figure you can put into an ROI calculation. The number will not be audit-standard, but it will be directionally honest and far more useful than knowing how many times users clicked the Copilot button.
Task completion rate in programme challenges
During a structured programme, challenge completion rate is a leading indicator of final adoption outcomes. Cohorts where 80% or more of participants complete each weekly challenge consistently produce higher spontaneous use rates at week nine than cohorts where completion drops below 60%. If completion is falling, something needs to change, the challenge is too hard, the facilitator support is insufficient, or the leaderboard dynamic is not generating enough accountability. Completion rate gives you a real-time signal to act on, not a retrospective result to report.
Building a simple measurement dashboard
You do not need a sophisticated analytics setup to track the metrics that matter. A structured adoption programme can be adequately measured with four data points collected twice, before and after:
- Active adoption rate from the Microsoft 365 admin centre (monthly active users ÷ licensed users)
- Self-reported Copilot confidence on a 1–10 scale from a pre- and post-programme survey
- Weekly time saving estimate from a post-programme survey question
- Spontaneous use rate from a post-programme survey question
These four metrics, tracked across a cohort of 20–50 participants, give you everything you need to make a credible case to leadership about whether the adoption programme worked, and what to do differently for the next cohort. To translate those figures into a financial return, use our framework for measuring Copilot ROI.
Not sure where your adoption currently sits across these metrics? The free Copilot diagnostic gives you a baseline score in under five minutes, covering adoption structure, measurement, and behaviour change readiness, the three factors that determine whether your current approach is likely to produce lasting results.
Take the free diagnostic