Analysis, Reporting

Adding Funnels in Google’s Looker Studio – NATIVELY!

Back in 2017, I lamented the lack of any option to create funnel visualizations in Data Studio (now known as Looker Studio.) 

So many clients needed a way to visualize their customer’s behavior through key conversion paths on their site, that I found some clever workarounds to bring funnel-like visualizations to life. 

In addition to the methods outlined in my old blog post (and the great posts of others), there were several Community Visualizations available. 

I’m so excited to see that now, funnel visualizations are available natively in Looker Studio! So let’s check them out. 

Under Add a chart, you’ll now see an option for funnel visualizations: 

They are essentially the same three charts (same setup, etc) but just three different ways of viewing it: 

  1. Sloped bar 
  2. Stepped bar
  3. Inverted triangle (note that while this funnel style may be visually appealing, its size doesn’t really tell you about the actual conversion rate, meaning that your users will still need to read and digest the numbers to understand how users convert. Aka… it’s a data visualization, that doesn’t actually visualize the data…) 

My personal favorite is probably the Stepped Bar, so I’ll use that for the following examples. 

The setup is surprisingly simple (certainly, much simpler than the hoops I used to jump through to create these visualizations in 2017!) 

You just need to specify one dimension and one metric

For a dimension, you could use: 

  • Page Path and Query String
  • Event Name 
  • A calculated field that takes some mix of different dimensions (based on a case statement.) 

Obviously if you included every page, or every event, that “funnel” chart would not be terribly useful, as it would include every page/event, and not narrow it down to those that you actually consider to be a part of the funnel: 

You’ll therefore want to use filters to narrow down to just the events or pages that actually form your funnel. For example, you could filter to just the specific events of view_item, add_to_cart, begin_checkout and purchase. 

Another option would be to create a specific dimension for use in your funnels, that uses a combination of events and pages (and/or, collapses various values of a dimension into just those you want included.) 

For example, let’s say you want to analyze a funnel including: 

  • Session on the site (tracked via an event)
  • Viewed a page of your blog (tracked via a page_view event, but might have many different possible values, so we want to collapse them all into one)
  • Submitted a lead form (tracked via an event) 

You could create a CASE statement to combine all of those into one dimension, for easy use in a funnel: 

CASE WHEN Event name="session_start" THEN "session_start"
WHEN REGEXP_CONTAINS(Page path + query string, r"/blog") THEN "blog_view"
WHEN Event name = "generate_lead" THEN "generate_lead"
ELSE NULL END 

(You would then exclude “dimension IS NULL” from your funnel.) 

For your metrics, you could use something like Total Users, Sessions, etc. 

Formatting options: 

  • You can choose to show the dimension value (or not) 
  • You can choose to show the funnel numbers as the raw number, the conversion percentage (from the very first step) or the conversion rate from the previous step. Warning: If you show the conversion rate from the previous step, the funnel visualization still shows the conversion rate from the start of the funnel, so this might be confusing for some users (unless you show both, via two charts.) 

You can choose to “Color by” a single color (my recommendation, because this is garish and awful – I said what I said.) 

Your funnel can include up to 10 steps (which is on par with funnel in Explore, and definitely better than the “create a blended data source” hack we used to use, that only allowed for 5 steps.) 

Have you had a chance to play with the new funnel visualizations in Looker Studio yet? Share what you think in Measure Chat’s Looker Studio channel! 

Analysis, Conferences/Community, Featured, google analytics, Reporting

Go From Zero to Analytics Hero using Data Studio

Over the past few years, I’ve had the opportunity to spend a lot of time in Google’s Data Studio product. It has allowed me to build intuitive, easy-to-use reporting, from a wide variety of data sources, that are highly interactive and empower my end-users to easily explore the data themselves… for FREE. (What?!) Needless to say, I’m a fan!

So when I had the chance to partner with the CXL Institute to teach an in-depth course on getting started with Data Studio, I was excited to help others draw the same value from the product that I have.

Perhaps you’re trying to do more with less time… Maybe you’re tearing your hair out with manual analysis work… Perhaps you’re trying to better communicate your data… Or maybe you set yourself a resolution to add a new tool to your analytics “toolbox” for 2020. Whatever your reasons, I hope these resources will get you started!

So without further adieu, check out my free 30 minute webinar with the CXL Institute team here, which will give you a 10-step guide to getting started with Data Studio.

And if you’re ready to really dive in, check out the entire hour online course here:

 

Analysis, Conferences/Community, Featured, google analytics

That’s So Meta: Tracking Data Studio, in Data Studio

That’s So Meta: Tracking Data Studio, in Data Studio

In my eternal desire to track and analyze all.the.things, I’ve recently found it useful to track the usage of my Data Studio reports.

Viewing data about Data Studio, in Data Studio? So meta!

Step 1: Create a property

Create a new Google Analytics property, to house this data. (If you work with multiple clients, sites or business units, where you may want to be able to isolate data, then you may want to consider one property for each client/site/etc. You can always combine them in Data Studio to view all the info together, but it gives you more control over permissions, without messing around with View filters.)

Step 2: Add GA Tracking Code to your Data Studio reports

Data Studio makes this really easy. Under Report Settings, you can add a GA property ID. You can add Universal Analytics, or GA4.

You’ll need to add this to every report, and remember to add it when you create new reports, if you’d like them to be included in your report.

Step 3: Clean Up Dimension Values

Note: This blog post is based on Universal Analytics, but the same principles apply if you’re using GA4. 

Once you have tracked some data, you’ll notice that the Page dimension in Google Analytics is a gibberish, useless URL. I suppose you could create a CASE formula and rewrite the URLs in to the title of the report…Hmmm… Wait, why would you do that, when there’s already an easier way?!

You’ll want to use the Page Title for the bulk of your reporting, as it has nice, readable, user-friendly values:

However, you’ll need to do some further transformation of Page Title. This is because reports with one page, versus multiple pages, will look different.

Reports with only one page have a page title of:

Report Name

Reports with more than one page have a page title of:

Report Name > Page Name

If you want to report on the popularity at a report level, we need to extract just the report name. Unfortunately, we can’t simply extract “everything before the ‘>’ sign” as the Report Name, since not all Page Titles will contain a “>” (if the report only has one page.)

I therefore use a formula to manipulate the Page Title:

REGEXP_EXTRACT(

(CASE 
WHEN REGEXP_MATCH(Page Title,".*›.*") 
THEN Page Title 
ELSE CONCAT(Page Title," ›")
END)

,'(.*).*›.*')

Step 4: A quick “gotcha”

Please note that, on top of Google Analytics tracking when users actually view your report, Google Analytics will also fire and track a view when:

  1. Someone is loading the report in Edit mode. In the Page dimension, you will see these with /edit in the URL.
  2. If you have a report scheduled to send on a regular cadence via email, the process of rendering the PDF to attach to the email also counts as a load in Google Analytics. In the Page dimension, you will see these loads with /appview in the URL.

This means that if you or your team spend a lot of time in the report editing it, your tracking may be “inflated” as a result of all of those loads.

Similarly, if you schedule a report for email send, it will track in Google Analytics for every send (even if no one actually clicks through and views the report.)

If you want to exclude these from your data, you will want to filter out from your dashboard Pages that contain /edit and /appview.

 

Step 5: Build your report

Here’s an example of one I have created:

Which metrics should I use?

My general recommendation is to use either Users or Pageviews, not Sessions or Unique Pageviews.

Why? Sessions will only count if the report page was the first page viewed (aka, it’s basically “landing page”), and Unique Pageviews will consider two pages in one report “unique”, since they have different URLs and Page Titles. (It’s just confusing to call something “Unique” when there are so many caveats on how “unique” is defined, in this instance.) So, Users will be the best for de-duping, and Pageviews will be the best for a totals count.

What can I use these reports for?

I find it helpful to see which reports people are looking at the most, when they typically look at them (for example, at the end of the month, or quarter?) Perhaps you’re having a lot of ad hoc questions coming to your team, that are covered in your reports? You can check if people are even using them, and if not, direct them there before spending a bunch of ad hoc time! Or perhaps it’s time to hold another lunch & learn, to introduce people to the various reports available? 

You can also include data filters in the report, to filter for a specific report, or other dimensions, such as device type, geolocation, date, etc. Perhaps a certain office location typically views your reports more than another?

Of course, you will not know which users are viewing your reports (since we definitely can’t track PII in Google Analytics) but you can at least understand if they’re being viewed at all!

Analysis, Presentation

10 Tips for Presenting Data

Big data. Analytics. Data science. Businesses are clamoring to use data to get a competitive edge, but all the data in the world won’t help if your stakeholders can’t understand, or if their eyes glaze over as you present your incredibly insightful analysis. This post outlines my top ten tips for presenting data.

It’s worth noting that these tips are tool agnostic—whether you use Data Studio, Domo, Tableau or another data viz tool, the principles are the same. However, don’t assume your vendors are in lock-step with data visualization best practices! Vendor defaults frequently violate key principles of data visualization, so it’s up to the analyst to put these principles in practice.

Tip #1: Recognize That Presentation Matters

The first step to presenting data is to understand that how you present data matters. It’s common for analysts to feel they’re not being heard by stakeholders, or that their analysis or recommendations never generate action. The problem is, if you’re not communicating data clearly for business users, it’s really easy for them to tune out.

Analysts may ask, “But I’m so busy with the actual work of putting together these reports. Why should I take the time to ‘make it pretty’?”

Because it’s not about “making things pretty.” It’s about making your data understandable.

My very first boss in Analytics told me, “As an analyst, you are an information architect.” It’s so true. Our job is to take a mass of information and architect it in such a way that people can easily comprehend it.

… Keep reading on ObservePoint‘s blog …

Analysis, Conferences/Community

Oct. 25th: 1-Day Workshop in San Francisco – Intro to R for the Digital Analyst

I’ll be conducting a small (up to 8 students) hands-on workshop that is an introduction to R for the digital analyst in San Francisco on Wednesday, October 25th.

If you are a digital analyst who is looking to dive into R, this 1-day intensive hands-on training is for you. This class is intended for digital analysts who are just getting started with R or who have tried to use R but have not successfully put it to use on a regular basis.

Course Overview

The course is a combination of lecture and hands-on examples, with the goal being that every attendee leaves the class with a basic understanding of:

  • The syntax and structure of the R platform — packages, basic operations, data types, etc.
  • How to navigate the RStudio interface — the script editor, the console, the environment pane, and the viewer
  • How to pull data from web analytics platforms (Google Analytics and Adobe Analytics) using R and the platforms’ APIs
  • The basics of transforming and manipulating data using R (base R vs. dplyr, with an emphasis on the latter — you don’t need to understand what that means to take the course; we’ll cover it!)
  • The “grammar of graphics” for data visualization (the paradigm for visualizing data in R using the most popular package for doing so — ggplot2)
  • Tips for troubleshooting R scripts (and writing code that can be readily troubleshot!)
  • The various options for producing deliverables directly from R

All of the material presented and applied during the class, as well as more advanced topics that cannot be covered in a one-day course, will be available to the students for reference as they put the material in to practice following the class.

Course Requirements

Students are expected to bring their own laptops. There will be communication prior to the class to ensure the required software (all free/open source) is installed and working.

Other Details

  • Date: Wednesday, October 25th
  • Time: 9:00 AM to 5:00 PM
  • Location: Elite SEM, 100 Bush St. #845, San Francisco, CA 94104
  • Cost: $895
  • Registration: click here to register

Questions?

Contact tim at analyticsdemystified dot com with any questions you have regarding the course.

Adobe Analytics, Analysis, Featured, google analytics

Did that KPI Move Enough for Me to Care?

This post really… is just the setup for an embedded 6-minute video. But, it actually hits on quite a number of topics.

At the core:

  • Using a statistical method to objectively determine if movement in a KPI looks “real” or, rather, if it’s likely just due to noise
  • Providing a name for said statistical method: Holt-Winters forecasting
  • Illustrating time-series decomposition, which I have yet to find an analyst who, when first exposed to it, doesn’t feel like their mind is blown just a bit
  • Demonstrating that “moving enough to care” is also another way of saying “anomaly detection”
  • Calling out that this is actually what Adobe Analytics uses for anomaly detection and intelligent alerts.
  • (Conceptually, this is also a serviceable approach for pre/post analysis…but that’s not called out explicitly in the video.)

On top of the core, there’s a whole other level of somewhat intriguing aspects of the mechanics and tools that went into the making of the video:

  • It’s real data that was pulled and processed and visualized using R
  • The slides were actually generated with R, too… using RMarkdown
  • The video was generated using an R package called ari (Automated R Instructor)
  • That package, in turn, relies on Amazon Polly, a text-to-speech service from Amazon Web Services (AWS)
  • Thus… rather than my dopey-sounding voice, I used “Brian”… who is British!

Neat, right? Give it a watch!

https://youtu.be/eGB5x77qnco

If you want to see the code behind all of this — and maybe even download it and give it a go with your data — it’s available on Github.

Analysis

A Mobile Analytics Comparative: Insurance Apps – Part 1

Throughout this mobile analytics comparative, I was looking for a few specific things to determine data collected by a group of high-profile apps and how their respective analytics tools were architected to facilitate analysis. Part 1 of this blog series focuses on the SDKs installed within each app and Part 2 will dive into the specific events and variables requested by each apps’ analytics tool.

My comparative focuses on insurance companies in the US, because they undoubtedly spend more money battling it out on advertising than any other industry around. I can testify that it’s working because my TV-watching-kids walk around the house humming insurance jingles (and craftily changing the words) on a daily basis.

According to Statista, the Top 5 Big Spenders on advertising include: Geico, State Farm, Progressive, Liberty Mutual, and Allstate (in that order). I was curious to know how these ad spending giants track their mobile apps. So after downloading each app, I ran my iOS device through my computer using manual HTTP proxy settings and observed calls using Charles to determine what data was being collected and passed through each application. Since I don’t have accounts at each of these insurance providers, my assessment involved three steps: 1) launching the app, 2) agreeing to their user acceptance rules, and 3) swiping my way to the auto quote section. Each app eventually made a hand-off to their responsive websites to complete the quote, which is where my simulated scripts ended.

My findings revealed a good deal about each of these organizations just by digging into their apps from the outside. I should mention that none of the companies in this comparative knew they were being evaluated and I have not validated my observed data requests against any of their internal analytics solutions to determine if data is actually populating in their analytics solutions as designed. If you’re reading this and happen to work at one of these firms, I welcome your feedback and please let me know if I’ve missed the mark on anything critical.

Here’s a few of the notable things I uncovered during this study:

Every app evaluated is using one or more analytics tracking tools. Not surprisingly, we found that all of the insurance companies in our evaluation have analytics tracking tools installed on their mobile apps. Either Adobe Analytics and Google Analytics was present in each of these apps, mirroring the industry dominance that these companies have in the web analytics world. Adobe Analytics was present in three out of five iOS applications and in four out of five Android applications. Google Analytics was present in in just two of the iOS apps, but Google Play Analytics and Firebase Analytics (or both) were present in all but one of the Android apps. See the matrix below for more details.   

The average number of SDKs installed within each app is 19. To determine the total number of SDKs installed within each app, I used a free tool offered by SafeDK called the App X-Ray which allows you to scan any Android app to determine the SDKs installed within that app. This clever little tool revealed a whole lot about each app and how 3rd party services are embedded within each one to deliver different services and solutions. If you’re not familiar with the world of SDKs, these “Software Developer Kits” are used to deploy tools and services within apps such as analytics, advertising, payment solutions, location-based services, crash reporting, attribution, and more. According to SafeDK, as of the 2nd Quarter in 2017, the average app has 17.8 SDKs installed. This fits extraordinarily well into our small sample of insurance apps that have on average 18.6 SDKs embedded. I found instances of crash reporting, location, messaging, voice of customer, advertising, payment, and numerous tools installed within the insurance app sample I evaluated. According to SafeDK: analytics, advertising, and social are the three most commonly used SDKs within apps (with payment as a close 4th).

 

 

For app developers, SDKs are essential. Whether it’s adding a tool in the rush to market for a new app, or accessing a library that wouldn’t be available otherwise, SDKs are critical to app development. SDKs provide add-on services and functions that save time and money and there are hundreds available. But just because they exist doesn’t mean you have to use them. I’ll let my bias shine and state that you’d be silly not to use analytics SDKs, but during my assessment, I found that four out of five Android insurance apps had SDKs installed, but not in use. For Allstate, they had a total of 31 SDKs installed in their Android app and a whopping 14 were not actually being used. This presents a whole host of considerations about waste, app bloat, and app size that are considerations for mobile developers and product owners.

 

Messaging and Location SDKs were prevalent. Within my small sample of insurance apps, I found that three out of five used messaging services to (presumably) deliver in-app messaging to customers. Additionally, all but one of the providers used a dedicated location services SDK within their apps, which again makes a lot of sense for the insurance app, where you might find a customer in need of roadside assistance at any given time. Both of these categories are up-and-coming in the world of mobile apps and can tie nicely into your integrated marketing efforts if executed correctly.  

 

Very little user experience testing is going on within these apps. Whereas messaging and location SDKs were apparent, few of the insurance apps we evaluated offered any testing tools within their apps. For testing here we’re talking about analytics testing similar to the Adobe Target or Optimizely tests that you’d find on a website. Not QA testing. Yet, as with websites, testing often reflects a higher level of maturity among companies using analytics. For this group, Geico was the only company to employ Adobe Target for testing and optimization of its app. One of the things that we do know about mobile is that apps are constantly going through new releases and updates, so there could be new features and functions rolled out with each new release. However, this is no substitute for actual user testing with A/B or multivariate combinations of creative to get your install base using your app and coming back for more.

 

Tag management tools are still being shoehorned into mobile apps. Each of the five insurance apps we evaluated included calls to Tag Management Solutions which are increasingly common (and indispensable) in traditional web analytics. Yet, as we all know, mobile is an entirely different beast than the web and data collection requires a different methodology. The event-based tracking method of mobile, coupled with conditional execution and in some cases batch uploading create challenges for web-designed Tag Management Solutions. Lee Isensee of #MeasureSlack riffed on this topic a while ago, but his premise still holds true in that for native apps (which most of my examples are) and a large portion of hybrid apps, Tag Management technology simply doesn’t work well. While I’m not chastising any of these insurance providers for embedding TMS within their apps (mainly because I’m not entirely sure how they’re using them without looking from the inside), I caution you to carefully consider how you utilize TMS within your native apps.

 

Behavioral, Context, and Navigational data collection varies widely across these apps. So far, I’ve spent a lot of time writing about the SDKs installed within each app, but this still tells us very little about what actual data is collected by each of these applications. Since this post is already getting long in the tooth, I will save the nitty gritty details for Part 2, but I can tell you that there are a lot of variances when it comes to behavioral, context, and navigational data collected within each app analytics solution.

Tune into the next post for more details, but in the meantime, write me a comment below or shoot me an email at john@analyticsdemystified.com if you have thoughts, questions or ideas about app analytics.

 

Analysis, Featured, General, Presentation

Foundational Social Psychology Experiments (And Why Analysts Should Know Them) – Part 5 of 5

Digital Analytics is a relatively new field, and as such, we can learn a lot from other disciplines. This post continues exploring classic studies from social psychology, and what we analysts can learn from them.

Jump to an individual topic:

False Consensus

Experiments have revealed that we tend to believe in a false consensus: that others would respond similarly to the way that we would. For example, Ross, Greene & House (1977) provided participants with a scenario, with two different possible ways of responding. Participants were asked to explain which option they would choose, and guess what other people would choose. Regardless of which option they actually chose, participants believed that other people would choose the same one.

Why this matters for analysts: As you are analyzing data, you are looking at the behaviour of real people. It’s easy to make assumptions about how they will react, or why they did what they did, based on what you would do. But our analysis will be far more valuable if we can be aware of those assumptions, and actively seek to understand why our actual customers did these things – without relying on assumptions.

Homogeneity of the Outgroup

There is a related effect here: the Homogeneity of the Outgroup. (Quattrone & Jones, 1980.) In short, we tend to view those who are different to us (the “outgroup”) as all being very similar, while those who are like us (the “ingroup”) are more diverse. For example, all women are chatty, but some men are talkative, some are quiet, some are stoic, some are more emotional, some are cautious, others are more risky… etc.

Why this matters for analysts: Similar to the False Consensus Effect, where we may analyse user behaviour assuming everyone thinks as we do, the Homogeneity of the Outgroup suggests that we may oversimplify the behaviour of customers who are different to us, and fail to fully appreciate the nuance of varied behaviour. This may seriously bias our analyses! For example, if we are a large global company, an analysis of customers in another region may be seriously flawed if we are assuming customers in the region are “all the same.” To overcome this tendency, we might consider leveraging local teams or local analysts to conduct or vet such analyses.

The Hawthorne Effect

In 1955, Henry Landsberger analyzed several studies conducted between 1924 and 1932 at the Hawthorne Works factory. These studies were examining the factors related to worker productivity, including whether the level of light within a building changed the productivity of workers. They found that, while the level of light changing appeared to be related to increased productivity, it was actually the fact that something changed that mattered. (For example, they saw an increase in productivity even in low light conditions, which should make work more difficult…) 

However, this study has been the source of much criticism, and was referred to by Dr. Richard Nisbett as a “glorified anecdote.” Alternative explanations include that Orne’s “Demand Characteristics” were in fact at work (that the changes were due to the workers knowing they were a part of the experiment), or the fact that the changes were always made on a Sunday, and Mondays normally show increased productivity, due to employee’s having a day off. (Levitt & List, 2011.)

Why this matters for analysts: “Demand Characteristics” could mean that your data is subject to influence, if people know they are being observed. For example, in user testing, participants are very aware they are being studied, and may act differently. Your digital analytics data however, may be less impacted. (While people may technically know their website activity is being tracked, it may not be “top of mind” enough during the browsing experience to trigger this effect.) The Sunday vs. Monday explanation reminds us to consider other explanations or variables that may be at play, and be aware of when we are not fully in control of all the variables influencing our data, or our A/B test. However, the Hawthorne studies are also a good example where interpretations of the data may vary! There may be multiple explanations for what you’re seeing in the data, so it’s important to vet your findings with others. 

Conclusion

What are your thoughts? Do these pivotal social psychology experiments help to explain some of the challenges you face with analyzing and presenting data? Are there any interesting studies you have heard of, that hold important lessons for analysts? Please share them in the comments!

Analysis, Featured, General, Presentation

Foundational Social Psychology Experiments (And Why Analysts Should Know Them) – Part 4 of 5

Digital Analytics is a relatively new field, and as such, we can learn a lot from other disciplines. This post continues exploring classic studies from social psychology, and what we analysts can learn from them.

Jump to an individual topic:

The Bystander Effect (or “Diffusion of Responsibility”)

In 1964 in New York City, a woman name Kitty Genovese was murdered. A newspaper report at the time claimed that 38 people had witnessed the attack (which lasted an hour) yet no one called the police. (Later reports suggested this was an exaggeration – that there had been fewer witnesses, and that some had, in fact, called the police.)

However, this event fascinated psychologists, and triggered several experiments. Darley & Latane (1968) manufactured a medical emergency, where one participant was allegedly having an epileptic seizure, and measured how long it took for participants to help. They found that the more participants, the longer it took to respond to the emergency.

This became known as the “Bystander Effect”, which proposes that the more bystanders that are present, the less likely it is that an individual will step in and help. (Based on this research, CPR training started instructing participants to tell a specific individual, “You! Go call 911” – because if they generally tell a group to call 911, there’s a good chance no one will do it.)

Why this matters for analysts: Think about how you present your analyses and recommendations. If you offer them to a large group, without specific responsibility to any individual to act upon them, you decrease the likelihood of any action being taken at all. So when you make a recommendation, be specific. Who should be taking action on this? If your recommendation is a generic “we should do X”, it’s far less likely to happen.

Selective Attention

Before you read the next part, watch this video and follow the instructions. Go ahead – I’ll wait here.

In 1999, Simons and Chabris conducted an experiment in awareness at Harvard University. Participants were asked to watch a video of basketball players, where one team was wearing white shirts, and the other team was wearing black shirts. In the video, the white team and black team respectively were passing the ball to each other. Participants were asked to count the number of passes between players of the white team. During the video, a man dressed as a gorilla walked into the middle of the court, faced the camera and thumps his chest, then leaves (spending a total of 9 seconds on the screen.) Amazingly? Half of the participants missed the gorilla entirely! Since then, this has been termed “the Invisible Gorilla” experiment. 

Why this matters for analysts: As you are analyzing data, there can be huge, gaping issues that you may not even notice. When we focus on a particular task (for example, counting passes by the white-shirt players only, or analyzing one subset of our customers) we may overlook something significant. Take time before you finalize or present your analysis to think of what other possible explanations or variables there could be (what could you be missing?) or invite a colleague to poke holes in your work.

Stay tuned

More to come!

What are your thoughts? Do these pivotal social psychology experiments help to explain some of the challenges you face with analyzing and presenting data?

Analysis, Featured, General, Presentation

Foundational Social Psychology Experiments (And Why Analysts Should Know Them) – Part 3 of 5

Digital Analytics is a relatively new field, and as such, we can learn a lot from other disciplines. This post continues exploring classic studies from social psychology, and what we analysts can learn from them.

Primacy and Recency Effects

The serial position effect (so named by Ebbinghaus in 1913) finds that we are most likely to recall the first and last items in a list, and least likely to recall those in the middle. For example, let’s say you are asked to recall apple, orange, banana, watermelon and pear. The serial position effect suggests that individuals are more likely to remember apple (the first item; primacy effect) and pear (the final item; recency effect) and less likely to remember orange, banana and watermelon.

The explanation cited is that the first item/s in a list are the most likely to have made it to long-term memory, and benefit from being repeated multiple times. (For example, we may think to ourselves, “Okay, remember apple. Now, apple and orange. Now, apple, orange and banana.”) The primacy effect is reduced when items are presented in quick succession (probably because we don’t have time to do that rehearsal!) and is more prominent when items are presented more slowly. Longer lists tend to see a decrease in the primacy effect (Murdock, 1962.)

The recency effect, that we’re more likely to remember the last items, is explained because the most recent item/s are recalled, since they are still contained within our short-term memory (remember, 7 +/- 2!) However, the items in the middle of the list benefit from neither long, nor short, term memory, and therefore are forgotten.

This doesn’t just affect your recall of random lists of items. When participants are given a list of attributes of a person, their order appears to matter. For example, Asch (1964) found participants told “Steve is smart, diligent, critical, impulsive, and jealous” had a positive evaluation of Steve, whereas participants told “Steve is jealous, impulsive, critical, diligent, and smart” had a negative evaluation of Steve. Even though the adjectives are the exact same – only the order is different!

Why this matters for analysts: When you present information, your audience is unlikely to remember everything you tell them. So choose wisely. What do you lead with? What do you end with? And what do you prioritize lower, and save for the middle?

These findings may also affect the amount of information you provide at one time, and the cadence with which you do so. If you want more retained, you may wish to present smaller amounts of data more slowly, rather than rapid-firing with constant information. For example, rather than presenting twelve different “optimisation opportunities” at once, focusing on one may increase the likelihood that action is taken.

This is also an excellent argument against a 50-slide PowerPoint presentation – while you may have mentioned something in it, if it was 22 slides ago, the chance of your audience remembering are slim.

The Halo Effect

Psychologists have found that our positive impressions in one area (for example, looks) can “bleed over” to our perceptions in another, unrelated area (for example, intelligence.) This has been termed the “halo effect.”

In 1977, Nisbet and Wilson conducted an experiment with university students. The two students watched a video of the same lecturer deliver the same material, but one group saw a warm and friendly “version” of the lecturer, while the other saw the lecturer present in a cold and distant way. The group who saw the friendly version rated the lecturer as more attractive and likeable.

There are plenty of other examples of this. For example, “physically attractive” students have been found to receive higher grades and/or test scores than “unattractive” students at a variety of ages, including elementary school (Salvia, Algozzine, & Sheare, 1977; Zahr, 1985), high school (Felson, 1980) and college (Singer, 1964.) Thorndike (1920) found similar effects within the military, where a perception of a subordinate’s intelligence tended to lead to a perception of other positive characteristics such as loyalty or bravery.

Why this matters for analysts: The appearance of your reports/dashboards/analyses, the way you present to a group, your presentation style, even your appearance may affect how others judge your credibility and intelligence.

The Halo Effect can also influence the data you are analysing! It is common with surveys (especially in the case of lengthy surveys) that happy customers will simply respond “10/10” for everything, and unhappy customers will rate “1/10” for everything – even if parts of the experience differed from their overall perception. For example, if a customer had a poor shipping experience, they may extend that negative feeling about the interaction with the brand to all aspects of the interaction – even if only the last part was bad! (And note here: There’s a definite interplay between the Halo Effect and the Recency Effect!)

Stay tuned

More to come soon!

What are your thoughts? Do these pivotal social psychology experiments help to explain some of the challenges you face with analyzing and presenting data?

Analysis, Conferences/Community, Presentation, Reporting

Ten Tips For Presenting Data from MeasureCamp SF #1

Yesterday I got to attend my first MeasureCamp in San Francisco. The “Unconference” format was a lot of fun, and there were some fantastic presentations and discussions.

For those who requested it, my presentation on Data Visualization is now up on SlideShare. Please leave any questions or comments below! Thanks to those who attended.

Analysis, Featured, Presentation

Foundational Social Psychology Experiments (And Why Analysts Should Know Them) – Part 2 of 5

Digital Analytics is a relatively new field, and as such, we can learn a lot from other disciplines. This post continues exploring classic studies from social psychology, and what we analysts can learn from them.

Jump to an individual topic:

Confirmation Bias

We know now that “the facts” may not persuade us, even when brought to our attention. However, Confirmation Bias tells us that we intentionally seek out information that continually reinforces our beliefs, rather than searching for all evidence and fully evaluating the possible explanations.

Wason (1960) conducted a study where participants were presented with a math problem: find the pattern in a series of numbers, such as “2-4-6.” Participants could create three subsequent sets of numbers to “test” their theory, and the researcher would confirm whether these sets followed the pattern or not. Rather than collecting a list of possible patterns, and using their three “guesses” to prove or disprove each possible pattern, Wason found that participants would come up with a single hypothesis, then seek to prove it. (For example, they might hypothesize that “the pattern is even numbers” and check whether “8-10-12”, “6-8-10” and “20-30-40” correctly matched the pattern. When it was confirmed their guesses matched the pattern, they simply stopped. However, the actual pattern was “increasing numbers” – their hypothesis was not correct at all!

Why this matters for analysts: When you start analyzing data, where do you start? With a hunch, that you seek to prove, then stop your analysis there? (For example, “I think our website traffic is down because our paid search spend decreased.”) Or with multiple hypotheses, which you seek to disprove one by one? A great approach used in government, and outlined by Moe Kiss for its applicability to digital analytics, is the Analysis of Competing Hypotheses.

Conformity to the Norm

In 1951, Asch found that we conform to the views of others, even when they are flat-out wrong, surprisingly often! He conducted an experiment where participants were seated in a group of eight others who were “in” on the experiment (“confederates.”) Participants were asked to judge whether a line was most similar in length to three other lines. The task was not particularly “grey area” – there was an obvious right and wrong answer.

(Image Credit)

Each person in the group gave their answer verbally, in turn. The confederates were instructed to give the incorrect answer, and the participant was the sixth of the group to answer.

Asch was surprised to find that 76% of people conformed to others’ (incorrect) conclusions at least once. 5% always conformed to the incorrect answer. Only 25% never once agreed with the group’s incorrect answers. (The overall conformity rate was 33%.)

In follow up experiments, Asch found that if participants wrote down their answers, instead of saying them aloud, the conformity rate was only 12.5%. However, Deutsch and Gerard (1955) found a 23% conformity rate, even in situations of anonymity.

Why this matters for analysts: As mentioned previously, if new findings contradict existing beliefs, it may take more than just presenting new data. However, these conformity studies suggest that efforts to do so may be further hampered if you are presenting information to a group. It is less likely that people will stand up for your new findings against the norm of the group. In this case, you may be better to discuss your findings slowly to individuals, and avoid putting people on the spot to agree/disagree within a group setting. Similarly, this argues against jumping straight to a “group brainstorming” session. Once in a group, Asch demonstrated that 76% of us will agree with the group (even if they’re wrong!) so we stand the best chance of getting more varied ideas and minimising “group think” by allowing for individual, uninhibited brainstorming and collection of all ideas first.

Stay tuned!

More to come next week. 

What are your thoughts? Do these pivotal social psychology experiments help to explain some of the challenges you face with analyzing and presenting data?

Analysis, Featured, Presentation

Foundational Social Psychology Experiments (And Why Analysts Should Know Them) – Part 1 of 5

Digital Analytics is a relatively new field, and as such, we can learn a lot from other disciplines. This series of posts looks at some classic studies from social psychology, and what we analysts can learn from them.

Jump to an individual topic:

The Magic Number 7 (or, 7 +/- 2)

In 1956, George A. Miller conducted an experiment that found that the number of items a person can hold in working memory is seven, plus or minus two. However, all “items” are not created equal – our brain is able to “chunk” information to retain more. For example, if asked to remember seven words or even seven quotes, we can do so (we’re not limited to seven letters) because each word is an individual item or “chunk” of information. Similarly, we may be able to remember seven two-digit numbers, because each digit is not considered its own item.

Why this matters for analysts: This is critical to keep in mind as we are presenting data. Stephen Few argues that a dashboard must be confined to one page or screen. This is due to this limitation of working memory. You can’t expect people to look at a dashboard and draw conclusions about relationships between separate charts, tables, or numbers, while flipping back and forth constantly between pages, because this requires they retain too much information in working memory. Similarly, expecting stakeholders to recall and connect the dots between what you presented eleven slides ago is putting too great a pressure on working memory. We must work with people’s natural capabilities, and not against them.

When The Facts Don’t Matter

In 1957, Leon Festinger studied a Doomsday cult who believed that aliens would rescue them from a coming flood. Unsurprisingly, no flood (nor aliens) eventuated. In their book, When Prophecy Fails, Festinger et al commented, “A man with a conviction is a hard man to change. Tell him you disagree and he turns away. Show him facts or figures and he questions your sources. Appeal to logic and he fails to see your point … Suppose that he is presented with evidence, unequivocal and undeniable evidence, that his belief is wrong: what will happen? The individual will frequently emerge, not only unshaken, but even more convinced of the truth of his beliefs than ever before.”

In a 1967 study by Brock & Balloun, subjects listened to several messages, but the recording was staticky. However, the subjects could press a button to clear up the static. They found that people selectively chose to listen to the message that affirmed their existing beliefs. For example, smokers chose to listen more closely when the content disputed a smoking-cancer link.

However, Chanel, Luchini, Massoni, Vergnaud (2010) found that if we are given an opportunity to discuss the evidence and exchange arguments with someone (rather than just reading the evidence and pondering it alone) we are more likely to change our minds in the face of opposing facts.

Why this matters for analysts: Even if your data seems self-evident, if it goes against what the business has known, thought, or believed for some time, you may need more data to support your contrary viewpoint. You may also want to allow for plenty of time for discussion, rather than simply sending out your findings, as those discussions are critical to getting buy-in for this new viewpoint.

Stay tuned!

More to come tomorrow.

What are your thoughts? Do these pivotal social psychology experiments help to explain some of the challenges you face with analyzing and presenting data?

Analysis, Conferences/Community, Digital Analytics Community

Evolution of the Analysis Exchange

When we created the Analysis Exchange years ago my Partners and I all knew the industry needed better gateways into the field. Being “old school” each of us had more or less found our way to web analytics, and while we all ended up being lucky, we all recognized that the industry wouldn’t be able to grow or scale if that was the only way in. The idea of giving folks interested in the field “hands on” access to data, projects, and guidance was a no-brainer really … but wow did we not see how it would blow up!

In the subsequent years Analysis Exchange has ebbed and flowed, primarily based on our internal ability to focus on finding groups willing to bring projects to the table. The one thing I didn’t really imagine was how difficult it would be to find non-profits that A) had questions that could B) be answered using Google Analytics and C) could spend the time required to participate in a project. And while we had some great partners over time, finding projects ended up being the biggest gateway to the success of the Exchange.

Ironically, since we put Analysis Exchange on the back-burner a year ago … student interest has more or less exploded. We now get an average of 30 new students signing up from around the world every week! This is great and is a really interesting view in how analytics is changing from a global perspective … but what a disappointment for those new students to not have projects to work on.

So we are going to fix that.

For the time being we have taken Analysis Exchange offline and are looking into new ways to scale the effort and serve the needs of nearly 5,000 individuals around the world who want to join the digital analytics industry. We don’t have a timeline for these changes but we are working on them actively and as part of a few other innovative ideas we are planning to roll out. We appreciate your patience while we work.

As always I welcome your comments …

Analysis, Featured

The Trouble (My Troubles) with Statistics

Okay. I admit it. That’s a linkbait-y title. In my defense, though, the only audience that would be successfully baited by it, I think, are digital analysts, statisticians, and data scientists. And, that’s who I’m targeting, albeit for different reasons:

  • Digital analysts — if you’re reading this then, hopefully, it may help you get over an initial hump on the topic that I’ve been struggling mightily to clear myself.
  • Statisticians and data scientists — if you’re reading this, then, hopefully, it will help you understand why you often run into blank stares when trying to explain a t-test to a digital analyst.

If you are comfortably bridging both worlds, then you are a rare bird, and I beg you to weigh in in the comments as to whether what I describe rings true.

The Premise

I took a college-level class in statistics in 2001 and another one in 2010. Neither class was particularly difficult. They both covered similar ground. And, yet, I wasn’t able to apply a lick of content from either one to my work as a web/digital analyst.

Since early last year, as I’ve been learning R, I’ve also been trying to “become more data science-y,” and that’s involved taking another run at the world of statistics. That. Has. Been. HARD!

From many, many discussions with others in the field — on both the digital analytics side of things and the more data science and statistics side of things — I think I’ve started to identify why and where it’s easy to get tripped up. This post is an enumeration of those items!

As an aside, my eldest child, when applying for college, was told that the fact that he “didn’t take any math” his junior year in high school might raise a small red flag in the admissions department of the engineering school he’d applied to. He’d taken statistics that year (because the differential equations class he’d intended to take had fallen through). THAT was the first time I learned that, in most circles, statistics is not considered “math.” See how little I knew?!

Terminology: Dimensions and Metrics? Meet Variables!

Historically, web analysts have lived in a world of dimensions. We combine multiple dimensions (channel + device type, for instance) and then put one or more metrics against those dimensions (visits, page views, orders, revenue, etc.)

Statistical methods, on the other hand, work with “variables.” What is a variable? I’m not being facetious. It turns out it can be a bit a mind-bender if you come at it from a web analytics perspective:

  • Is device type a variable?
  • Or, is the number of visits by device type a variable?
  • OR, is the number of visits from mobile devices a variable?

The answer… is “Yes.” Depending on what question you are asking and what statistical method is being applied, defining what your variable(s) are, well, varies. Statisticians think of variables as having different types of scales: nominal, ordinal, interval, or ratio. And, in a related way, they think of data as being either “metric data” or “nonmetric data.” There’s a good write-up on the different types — with a digital analytics slant — in this post on dartistics.com.

It may seem like semantic navel-gazing, but it really isn’t: different statistical methods work with specific types of variables, so data has to be transformed appropriately before statistical operations are performed. Some day, I’ll write that magical post that provides a perfect link between these two fundamentally different lenses through which we think about our data… but today is not that day.

Atomic Data vs. Aggregated Counts

In R, when using ggplot to create a bar chart that uses underlying data that looks similar to how data would look in Excel, I have to include a parameter that is stat="identity". As it turns out, that is a symptom of the next mental jump required to move from the world of digital analytics to the world of statistics.

To illustrate, let’s think about how we view traffic by channel:

  • In web analytics, we think: “this is how many (a count) visitors to the site came from each of referring sites, paid search, organic search, etc.”
  • In statistics, typically, the framing would be: “here is a list (row) for each visitor to the site, and each visitor is identified as being visiting from referring sites, paid search, organic search, etc.” (or, possibly, “each visitor is flagged as being yes/no for each of: referring sites, paid search, organic search, etc.”… but that’s back to the discussion of “variables” covered above).

So, in my bar chart example above, R defaults to thinking that it’s making a bar chart out of a sea of data, where it’s aggregating a bunch of atomic observations into a summarized set of bars. The stat="identity" argument has to be included to tell R, “No, no. Not this time. I’ve already counted up the totals for you. I’m telling you the height of each bar with the data I’m sending you!”

When researching statistical methods, this comes up time and time again: statistical techniques often expect a data set to be a collection of atomic observations. Web analysts typically work with aggregated counts. Two things to call out on this front:

  • There are statistical methods (a cross tabulation with a Chi square test for independence is one good example) that work with aggregated counts. I realize that. But, there are many more that actually expect greater fidelity in the data.
  • Both Adobe Analytics (via data feeds, and, to a clunkier extent, Data Warehouse) and Google Analytics (via the GA360 integration with Google BigQuery) offer much more atomic level data than the data they provided historically through their primary interfaces; this is one reason data scientists are starting to dig into digital analytics data more!

The big, “Aha!” for me in this area is that we often want to introduce pseudo-granularity into our data. For instance, if we look at orders by channel for the last quarter, we may have 8-10 rows of data. But, if we pull orders by day for the last quarter, we have a much larger set of data. And, by introducing granularity, we can start looking at the variability of orders within each channel. That is useful! When performing a 1-way ANOVA, for instance, we need to compare the variability within channels to the variability across channels to draw conclusions about where the “real” differences are.

This actually starts to get a bit messy. We can’t just add dimensions to our data willy-nilly to artificially introduce granularity. That can be dangerous! But, in the absence of truly atomic data, some degree of added dimensionality is required to apply some types of statistical methods. <sigh>

Samples vs. Populations

The first definition for “statistics” I get from Google (emphasis added) is:

“the practice or science of collecting and analyzing numerical data in large quantities, especially for the purpose of inferring proportions in a whole from those in a representative sample.”

Web analysts often work with “the whole.” Unless we consider historical data the sample and the “whole” including future web traffic. But, if we view the world that way — by using time to determine our “sample” — then we’re not exactly getting a random (independent) sample!

We’ve also been conditioned to believe that sampling is bad! For years, Adobe/Omniture was able to beat up on Google Analytics because of GA’s “sampled data” conditions. And, Google has made any number of changes and product offerings (GA Premium -> GA 360) to allow their customers to avoid sampling. So, Google, too, has conditioned us to treat the word “sampled” as having a negative connotation.

To be clear: GA’s sampling is an issue. But, it turns out that working with “the entire population” with statistics can be an issue, too. If you’ve ever heard of the dangers of “overfitting the model,” or if you’ve heard, “if you have enough traffic, you’ll always find statistical significance,” then you’re at least vaguely aware of this!

So, on the one hand, we tend to drool over how much data we have (thank you, digital!). But, as web analysts, we’re conditioned to think “always use all the data!” Statisticians, when presented with a sufficiently large data set, like to pull a sample of that data, build a model, and then test the model with another sample of the data. As far as I know, neither Adobe nor Google have an, “Export a sample of the data” option available natively. And, frankly, I have yet to come across a data scientist working with digital analytics data who is doing this, either. But, several people have acknowledged this is something that should be done in some cases.

I think this is going to have to get addressed at some point. Maybe it already has been, and I just haven’t crossed paths with the folks who have done it!

Decision Under Uncertainty

I’ve saved the messiest (I think) for last. Everything on my list to this point has been, to some extent, mechanical. We should be able to just “figure it out” — make a few cheat sheets, draw a few diagrams, reach a conclusion, and be done with it.

But, this one… is different. This is an issue of fundamental understanding — a fundamental perspective on both data and the role of the analyst.

Several statistically-savvy analysts I have chatted with have said something along the lines of, “You know, really, to ‘get’ statistics, you have to start with probability theory.” One published illustration of this stance can be found in The Cartoon Guide to Statistics, which devotes an early chapter to the subject. It actually goes all the way back to the 1600s and an exchange between Blaise Pascal and Pierre de Fermat and proceeds to walk through a dice-throwing example of probability theory. Alas! This is where the book lost me (although I still have it and may give it another go).

Possibly related — although quite different — is something that Matt Gershoff of Conductrics and I have chatted about on multiple occasions across multiple continents. Matt posits that, really, one of the biggest challenges he sees traditional digital analysts facing when they try to dive into a more statistically-oriented mindset is understanding the scope (and limits!) of their role. As he put it to me once in a series of direct messages really boils down to:

  1. It’s about decision-making under uncertainty
  2. It’s about assessing how much uncertainty is reduced with additional data
  3. It must consider, “What is the value in that reduction of uncertainty?”
  4. And it must consider, “Is that value greater than the cost of the data/time/opportunity costs?”

The list looks pretty simple, but I think there is a deeper mindset/mentality-shift that it points to. And, it gets to a related challenge: even if the digital analyst views her role through this lens, do her stakeholders think this way? Methinks…almost certainly not! So, it opens up a whole new world of communication/education/relationship-management between the analyst and stakeholders!

For this area, I’ll just leave it at, “There are some deeper fundamentals that are either critical to understand or something that can be kicked down the road a bit.” I don’t know which it is!

What Do You Think?

It’s taken me over a year to slowly recognize that this list exists. Hopefully, whether you’re a digital analyst dipping your toe more deeply into statistics or a data scientist who is wondering why you garner blank stares from your digital analytics colleagues, there is a point or two in this post that made you think, “Ohhhhh! Yeah. THAT’s where the confusion is.”

If you’ve been trying to bridge this divide in some way yourself, I’d love to hear what of this post resonates, what doesn’t, and, perhaps, what’s missing!

Analysis, Featured

“What will you do with that?” = :-(

Remember back when folks wrote blog posts that were blah-blah-blah “best practice”-type posts? I think this is going to be one of those – a bit of a throwback, perhaps. But, hopefully, mildly entertaining and, hell, maybe even useful!

Let’s Start with Three Facts

  • Fact #1: Business users sometimes (often?) ask for data that they’re not actually going to be able to act on.
  • Fact #2: Analysts’ time is valuable.
  • Fact #3: Analysts need to prioritize their time pulling data, compiling reports, and conducting analyses with a bias towards results that will drive action.

None of the above are earth-shattering or particularly insightful observations.

And Yet…

…I am regularly dismayed by the application of these facts by analysts I watch or chat with. (Despite being an analytics curmudgeon, I don’t actually enjoy being dismayed.)

The following questions are all variations of the same thing, and they all make the hair on the back of my neck stand up when I hear an analyst ask them (or proudly tell me they ask them as part of their intake process ):

“What are you going to do with that information (or data or report) if I provide it?”

“What decision will you make based on that information?”

“What action will you take if I provide that information?”

I abhor these questions (and various variations of them).

Do you share my abhorrence?

Pause for a few seconds and ask yourself if you see these types of questions as counterproductive.

If you do see a problem with these questions, then read on and see if it’s for the same reason that I do.

If you do not  see a problem, then read on and see if I can change your mind.

If you’re not sure…well, then, get off the damn fence and form an opinion!

Some More Facts

We have to add to our fact base a bit to explain why these questions elevate my dander:

  • Fact #4: Analysts must build and maintain a positive relationship with their stakeholders.
  • Fact #5: Analysts hold the keys to the data (even if business users have some direct data access, they don’t have the expertise or depth of access that analysts do).

How Those Questions Can Be Heard

When an analyst says, “What decision will you make based on that information?” what they can (rightly!) be heard saying is any (or all) of the following:

“You (the business user) must convince me (the analyst) that it is worth my time to support you.”

“I don’t believe that information would be valuable to you, so you must convince me that it would be.”

“I would rather not add anything to my plate, so I’m going to make you jump through a few more hoops before I agree to assist you. (I’m kinda’ lazy.)”

Do you see the problem here? By asking a well-intended question, the analyst can easily come across as adversarial: as someone who holds the “power of the data” such that the business user must (metaphorically) grovel/justify/beg for assistance.

This is not a good way to build and grow strong relationships with the business! And, we established with Fact #4 that this was important.

But…What About Fact #3?

Do we have an intractable conflict here? Am I saying that we can’t say, “No” or, at least, “Why?” to a business user? There are only so many hours in the day!

I’m not actually saying that at all.

Let’s shift from facts to two assumptions that I (try to) live by:

  • Assumption #1: No business user wants to waste their own or the analyst’s time.
  • Assumption #2: Stakeholders have reasonably deep knowledge of their business areas, and they want to drive positive results.

“Aren’t assumptions dangerous?” you may ask. “Aren’t they the cousins of ‘opinions,’ which we’ve been properly conditioned to eschew?”

Yes… except not really in this case. These are useful assumptions to work from and to only discard if and only if they are thoroughly and conclusively invalidated in a specific situation.

Have You Figured Out Where I Am Heading?

As soon as a business user approaches me with any sort of request:

  • I start with an assumption that the request is based on a meaningful and actionable need.
  • I put the onus on myself to take the next step to articulate what that need is.

Is that a subtle pivot? Perhaps. But, with both of the above in mind, the questions I listed at the beginning of this post should start to appear as clearly inappropriate.

The Savvy Analyst’s Approach

I hope you’re not expecting anything particularly magic here, as it’s not. But, no matter the form of the question or request, I always try to work through the following basic process by myself:

  1. Is the requestor trying to simply measure results or are they looking to validate a hypothesis? (There is no room for “they just want some numbers” – given my own knowledge of the business and any contextual clues I picked up in the request, I will put it into one bucket or the other.)
  2. If I determine the stakeholder is trying to measure results, then I try to articulate (on the fly in conversation or in writing as a follow-up) what I think their objective is for the thing they’re trying to measure. And then I skip to step 4.
  3. If I determine the stakeholder is trying to validate a hypothesis (or “wants some analysis”), then I try to articulate one or more of the most likely and actionable hypotheses that I can using the structure:
    • The requestor believes… <something>.
    • If that belief is right, then we will… <some action>.
  4. I then play back what I’ve come up with to the stakeholder. I’ll couch it as though I’ve just completed a master class in active listening: “I want to make sure I’m getting you information that is as useful as possible. What I think you’re looking for is…(play back of what came out of step 2 or 3).”
  5. Then — after a little (or a lot) of discussion — I’ll dive into actually doing the work.

If you’re more of a graphical thinker, then the above words can be represented as a flowchart:

This approach has several (hopefully obvious) benefits:

  • It immediately makes the request a collaboration rather than a negotiation.
  • It sneakily demonstrates that, as an analyst, I’m focused on business results and on providing useful information.
  • It prevents me from spending time (hours or days) pulling and crunching data that is wildly off the mark for what the stakeholder actually wants.
  • It provides me with a space to outline several different approaches that require various levels of effort (or, often, provides the opportunity to say, “Let’s just check this one thing very quickly before we head too far down this path.”).

Are You With Me?

What do you think? Have you been guilty of guiding a stakeholder to put up her dukes every time she comes to you with a request, or do you take a more collaborative approach right out of the chute?

Analysis, Featured

3-Day Training: R & Statistics for the Digital Analyst – June 13-15 (Columbus, OH)

One challenge I found over the course of last year as I worked to learn R and learn how to apply statistics in a meaningful way to digital analytics data was that, while there is a wealth of information on both subjects, there is limited information available that speaks directly to working with digital analytics data. The data isn’t necessarily all that special, but even something as (theoretically) simple as translating web analytics “dimensions and metrics” to “variables” (multi-level factors, continuous vs. categorical variables, etc.) sent me into multiple mental circles.

In an effort to shorten that learning curve for other digital analysts, Mark Edmondson from IIH Nordic and I recruited Dr. Michael Levin from Otterbein University and have put together a 3-day training class:

  • Dates: June 13, 2017 – June 15, 2017
  • Location: 95 Liberty Street, Columbus, OH, 43215
  • Early Bird Price (through March 15, 2017): $1,695
  • Full Registration (after March 15, 2017): $1,995
  • Event Website

Course Description

The course is a combination of lectures and hands-on examples. The goal is that every attendee will leave with a clear understanding of:

  • The syntax and structure of the R language, as well as the RStudio interface
  • How to automatically pull data from web analytics and other platforms
  • How to transform and manipulate data using R
  • How to visualize data with R
  • How to troubleshoot R scripts
  • Various options for producing deliverables directly from R
  • The application of core statistics concepts and methods to digital analytics data

The course is broken down into three core units, with each day being devoted to a specific unit, and the third day bringing together the material taught on the first two days:

The first and third days have a heavy hands-on component to them.

Who Should Attend?

This training is primarily for digital analysts who have hit the limits of what can be done effectively with Microsoft Excel, the native interfaces of digital analytics platforms, and third party platforms like Tableau. Specifically, it is for digital analysts who are looking to:

  • Improve their efficiency and effectiveness when it comes to accessing and manipulating data from digital/social/mobile/internal platforms
  • Increase the analytical rigor they are able to apply to their work – applying statistical techniques like correlation, regression, standardization, and chi square so they can increase the value they deliver to their organizations

Attendees should be relatively well-versed in digital analytics data. We will primarily be working with Google Analytics data sets in the course, but the material itself is not platform-specific, and the class discussion will include other platforms as warranted based on the make-up of the attendees.

Attendees who currently work (or have dabbled with) R or statistics are welcome. The material goes “beyond the basics” on both subjects. But, attendees who have not used R at all will be fine. We start with the basics, and those basics are reinforced throughout the course.

Oh… and Columbus, Ohio, in June is a great place to be. The class includes meals and evening activities!

Head over to the event website for additional details and to register!

 

Adobe Analytics, Analysis, Featured

R and Adobe Analytics: Did the Metric Move Significantly? Part 3 of 3

This is the third post in a three-post series. The earlier posts build up to this one, so you may want to go back and check them out before diving in here if you haven’t been following along:

  • Part 1 of 3: The overall approach, and a visualization of metrics in a heatmap format across two dimensions
  • Part 2 of 3: Recreating — and refining — the use of Adobe’s anomaly detection to get an at-a-glance view of which metrics moved “significantly” recently

The R scripts used for both of these, as well as what’s covered in this post, are posted on Github and available for download and re-use (open source FTW!).

Let’s Mash Parts 1 and 2 Together!

This final episode in the series answers the question:

Which of the metrics changed significantly over the past week within specific combinations of two different dimensions?

The visualization I used to answer this question is this one:

This, clearly, is not a business stakeholder-facing visualization. And, it’s not a color-blind friendly visualization (although the script can easily be updated to use a non-red/green palette).

Hopefully, even without reading the detailed description, the visualization above jumps out as saying, “Wow. Something pretty good looks to have happened for Segment E overall last week, and, specifically, Segment E traffic arriving from Channel #4.” That would be an accurate interpretation.

But, What Does It Really Mean?

If you followed the explanation in the last post, then, hopefully, the explanation is really simple. In the last post, the example I showed was this:

This example had three “good anomalies” (the three dots that are outside — and above — the prediction interval) in the last week. And, it had two “bad anomalies” (the two dots at the beginning of the week that are outside — and below — the prediction interval).

In addition to counting and showing “good” and “bad” anomalies, I can do one more simple calculation to get “net positive anomalies:”

[Good Anomalies] – [Bad Anomalies] = [Net Positive Anomalies]

In the example above, this would be:

[3 Good Anomalies] – [2 Bad Anomalies] = [1 Net Positive Anomaly]

If the script is set to look at the previous week, and if weekends are ignored (which is a configuration within the script), then that means the total possible range for net positive anomalies is -5 to +5. That’s a nice range to provide a spectrum for a heatmap!

A Heatmap, Though?

This is where the first two posts really get mashed together:

  • The heatmap structure lets me visualize results across two different dimensions (plus an overall filter to the data set, if desired)
  • The anomaly detection — the “outside the prediction interval of the forecast of the past” — lets me get a count of how many times in the period a metric looked “not as expected”

The heatmap represents the two dimensions pretty obviously. For each cell — each intersection of a value from each of the two dimensions — there are three pieces of information:

  • The number of good anomalies in the period (the top number)
  • The number of bad anomalies in the period (the bottom number)
  • The number of net positive anomalies (the color of the cell)

You can think of each cell as having a trendline with a forecast and prediction confidence band for the last period, but actually displaying all of those charts would be a lot of charts! With the heatmap shown above, there are 42 different slices represented for each metric (there is then one slide for each metric), and it’s quick to interpret the results once you know what they’re showing.

What Do You Think?

This whole exercise grew out of some very specific questions that I was finding myself asking each time I reviewed a weekly performance measurement dashboard. I realize that “counting anomalies by day,” is somewhat arbitrary. But, by putting some degree of rigor behind identifying anomalies (which, so far, relies heavily on Adobe to do the heavy lifting, but, as covered in the second post, I’ve got a pretty good understanding of how they’re doing that lifting, and it seems fairly replicable to do this directly in R), it seems useful to me. If and when a specific channel, customer segment, or combination of channel/segment takes a big spike or dip in a metric, I should be able to hone in on it with very little manual effort. And, I can then start asking, “Why? And, is this something we can or should act on?”

Almost equally importantly, the building blocks I’ve put in place, I think, provide a foundation that I (or anyone) can springboard off of to extend the capabilities in a number of different directions.

What do you think?

Adobe Analytics, Analysis, Featured

R and Adobe Analytics: Did the Metric Move Significantly? Part 2 of 3

In my last post, I laid out that I had been working on a bit of R code to answer three different questions in a way that was repeatable and extensible. This post covers the second question:

Did any of my key metrics change significantly over the past week (overall)?

One of the banes of the analyst’s existence, I think, is that business users rush to judge (any) “up” as “good” and (any) “down” as “bad.” This ignores the fact that, even in a strictly controlled manufacturing environment, it is an extreme rarity for any metric to stay perfectly flat from day to day or week to week.

So, how do we determine if a metric moved enough to know whether it warrants any deeper investigation as to the “why” it moved (up or down)? In the absence of an actual change to the site or promotions or environmental factors, most of the time (I contend), metrics don’t move enough in a short time period to actually matter. They move due to noise.

But, how do we say with some degree of certainty that, while visits (or any metric) were up over the previous week, they were or were not up enough to matter? If a metric increases 20%, it likely is not from noise. If it’s up 0.1%, it likely is just ordinary fluctuation (it’s essentially flat). But, where between 0.1% and 20% does it actually matter?

This is a question that has bothered me for years, and I’ve come at answering it from many different directions — most of them probably better than not making any attempt at all, but also likely an abomination in the eyes of a statistician.

My latest effort uses an approach that is illustrated in the visualization below:

In this case, something went a bit squirrely with conversion rate, and it warrants digging in farther.

Let’s dive in to the approach and rationale for this visualization as an at-a-glance way to determine whether the metric moved enough to matter.

Anomaly Detection = Forecasting the Past

The chart above uses Adobe’s anomaly detection algorithm. I’m pretty sure I could largely recreate the algorithm directly using R. As a matter of fact, that’s exactly what is outlined on the time-series page on dartistics.com. And, eventually, I’m going to give that a shot, as that would make it more easily repurposable across Google Analytics (and other time-series data platforms). And it will help me plug a couple of small holes in Adobe’s approach (although Adobe may plug those holes on their own, for all I know, if I read between the lines in some of their documentation).

But, let’s back up and talk about what I mean by “forecasting the past.” It’s one of those concepts that made me figuratively fall out of my chair when it clicked and, yet, I’ve struggled to explain it. A picture is worth a thousand words (and is less likely to put you to sleep), so let’s go with the equivalent of 6,000 words.

Typically, we think of forecasting as being “from now to the future:”

But, what if, instead, we’re actually not looking to the future, but are at today and are looking at the past? Let’s say our data looks like this:

Hmmm. My metric dropped in the last period. But, did it drop enough for me to care? It didn’t drop as much as it’s dropped in the past, but it’s definitely down? Is it down enough for me to freak out? Or, was that more likely a simple blip — the stars of “noise” aligning such that we dropped a bit? That’s where “forecasting the past” comes in.

Let’s start by chopping off the most recent data and pretend that the entirety of the data we have stops a few periods before today:

Now, from the last data we have (in this pretend world), let’s forecast what we’d expect to see from that point to now (we’ll get into how we’re doing that forecast in a bit — that’s key!):

This is a forecast, so we know it’s not going to be perfect. So, let’s make sure we calculated a prediction interval, and let’s add upper and lower bounds around that forecast value to represent that prediction interval:

Now, let’s add our actuals back into the chart:

Voila! What does this say? the next-to-last reporting period was below our forecast, but it was still inside our prediction interval. The most recent period, thought, was actually outside the prediction interval, which means it moved “enough” to likely be more than just noise. We should dig further.

Make sense? That’s  what I call “forecasting the past.” There may be a better term for this concept, but I’m not sure what it is! Leave a comment if I’m just being muddle-brained on that front.

Anomaly Detection in Adobe Analytics…Is This

Analysis Workspace has anomaly detection as an option in its visualizations and, given the explanation above, how they’re detecting “anomalies” may start to make more sense:

Now, in the case of Analysis Workspace, the forecast is created for the entire period that is selected, and then any anomalies that are detected are highlighted with a larger circle.

But, if you set up an Intelligent Alert, you’re actually doing the same thing as their Analysis Workspace anomaly visualization, with two tweaks:

  • Intelligent Alerts only look at the most recent time period — this makes sense, as you don’t want to be alerted about changes that occurred weeks or months ago!
  • Intelligent Alerts give you some control over how wide the prediction interval band is — in Analysis Workspace, it’s the 95% prediction interval that is represented; when setting up an alert, though, you can specify whether you want the band to be 90% (narrower), 95%, or 99% (wider)

Are you with me so far? What I’ve built in R is more like an Intelligent Alert than it is like the Analysis Workspace  representation. Or, really, it’s something of a hybrid. We’ll get to that in a bit.

Yeah…But ‘Splain Where the Forecast Came From!

The forecast methodology used is actually what’s called Holt-Winters. Adobe provides a bit more detail in their documentation. I started to get a little excited when I found this, because I’d come across Holt-Winters when working with some Google Analytics data with Mark Edmondson of IIH Nordic. It’s what Mark used in this forecasting example on dartistics.com. When I see the same thing cropping up from multiple different smart sources, I have a tendency to think there’s something there.

But, that doesn’t really explain how Holt-Winters works. At a super-high level, part of what Holt-Winters does is break down a time-series of data into a few components:

  • Seasonality — this can be the weekly cycle of “high during the week, low on the weekends,” monthly seasonality, both, or something else
  • Trend — with seasonality removed, how the data is trending (think rolling average, although that’s a bit of an oversimplification)
  • Base Level — the component that, if you add in the trend and seasonality to it will get you to the actual value

By breaking up the historical data, you get the ability to forecast with much more precision than simply dropping a trendline. This is worth digging into more to get a deeper understanding (IMHO), and it turns out there is a fantastic post by John Foreman that does just that: “Projecting Meth Demand Using Exponential Smoothing.” It’s tongue-in-cheek, but it’s worth downloading the spreadsheet at the beginning of the post and and walking through the forecasting exercise step-by-step. (Hat tip to Jules Stuifbergen for pointing me to that post!)

I don’t think the approach in Foreman’s post is exactly what Adobe has implemented, but it absolutely hits the key pieces. Analysis Workspace anomaly detection also factors in holidays (somehow, and not always very well, but it’s a tall order), which the Adobe Analytics API doesn’t yet do. And, Foreman winds up having Excel do some crunching with Solver to figure out the best weighting, while Adobe applies three different variations of Holt-Winters and then uses the one that fits the historical data the best.

I’m not equipped to pass any sort of judgment as to whether either approach is definitively “better.” Since Foreman’s post was purely pedagogical, and Adobe has some extremely sharp folks focused on digital analytics data, I’m inclined to think that Adobe’s approach is a great one.

Yet…You Still Built Something in R?!

Still reading? Good on ya’!

Yes. I wasn’t getting quite what I wanted from Adobe, so I got a lot from Adobe…but then tweaked it to be exactly what I wanted using R. The limitations I ran into with Analysis Workspace and Intelligent Alerts were:

  • I don’t care about anomalies on weekends (in this case — in my R script, it can be set to include weekends or not)
  • I only care about the most recent week…but I want to use the data up through the prior week for that; as I read Adobe’s documentation, their forecast is always based on the 35 days preceding the reporting period
  • do want to see a historical trend, though; I just want much of that data to be included in the data used to build the forecast
  • I want to extend this anomaly detection to an entirely different type of visualization…which is the third and final part in this series
  • Ultimately, I want to be able to apply this same approach to Google Analytics and other time-series data

Let’s take another look at what the script posted on Github generates:

Given the simplistic explanation provided earlier in this post, is this visual starting to make more sense? The nuances are:

  • The only “forecasted past” is the last week (this can be configured to be any period)
  • The data used to pull that forecast is the 35 days immediately preceding the period of interest — this is done by making two API calls: 1 to pull the period of interest, and another to pull “actuals only” data; the script then stitches the results together to show one continuous line of actuals
  • Anomalies are identified as “good” (above the 95% prediction interval) or “bad” (below the 95% prediction interval)

I had to play around a bit with time periods and metrics to show a period with anomalies, which is good! Most of the time, for most metrics, I wouldn’t expect to see anomalies.

There is an entirely separate weekly report — not shown here — that shows the total for each metric for the week, as well as a weekly line chart, how the metric changed week-over-week, and how it compared to the same week in the prior year. That’s the report that gets broadly disseminated. But, as an analyst, I have this separate report — the one I’ve described in this post — that I can quickly flip through to see if any metrics had anomalies on one or more days for the week.

Currently, the chart takes up a lot of real estate. Once the analysts (myself included) get comfortable with what the anomalies are, I expect to have a streamlined version that only lists the metrics that had an anomaly, and then provides a bit more detail.

Which may start to sound a lot like Adobe Analytics Intelligent Alerts! Except, so far, when Adobe’s alerts are triggered, it’s hard for me to actually get to a deeper view get more context. That may be coming, but, for now, I’ve got a base that I understand and can extend to other data sources and for other uses.

For details on how the script is structured and how to set it up for your own use, see the last post.

In the next post, I’ll take this “anomaly counting” concept and apply it to the heatmap concept that drills down into two dimensions. Sound intriguing? I hope so!

The Rest of the Series

If you’re feeling ambitious and want to go back or ahead and dive into the rest of the series:

Adobe Analytics, Analysis, Featured

R and Adobe Analytics: Two Dimensions, Many Metrics – Part 1 of 3

This is the first of three posts that all use the same base set of configuration to answer three different questions:

  1. How do my key metrics break out across two different dimensions?
  2. Did any of these metrics change significantly over the past week (overall)?
  3. Which of these metrics changed significantly over the past week within specific combinations of those two different dimensions?

Answering the first question looks something like this (one heatmap for each metric):

Answering the second question looks something like this (one chart for each metric):

Answering the third question — which uses the visualization from the first question and the logic from the second question — looks like this:

These were all created using R, and the code that was used to create them is available on Github. It’s one overall code set, but it’s set up so that any of these questions can be answered independently. They just share enough common ground on the configuration front that it made sense to build them in the same project (we’ll get to that in a bit).

This post goes into detail on the first question. The next one goes into detail on the second question. And, I own a T-shirt that says, “There are two types of people in this world: those who know how to extrapolate from incomplete information.” So, I’ll let you guess what the third post will cover.

The remainder of this post is almost certainly TL;DR for many folks. It gets into the details of the what, wherefore, and why of the actual rationale and methods employed. Bail now if you’re not interested!

Key Metrics? Two Dimensions?

Raise your hand if you’ve ever been asked a question like, “How does our traffic break down by channel? Oh…and how does it break down by device type?” That question-that-is-really-two-questions is easy enough to answer, right? But, when I get asked it, I often feel like it’s really one question, and answering it as two questions is actually a missed opportunity.

Recently, while working with a client, a version of this question came up regarding their last touch channels and their customer segments. So, that’s what the examples shown here are built around. But, it could just as easily have been device category and last touch channel, or device category and customer segment, or new/returning and device category, or… you get the idea.

When it comes to which metrics were of interest, it’s an eCommerce site, and revenue is the #1 metric. But, of course, revenue can be decomposed into its component parts:

[Visits] x [Conversion Rate] x [Average Order Value]

Or, since there are multiple lines per order, AOV can actually be broken down:

[Visits] x [Conversion Rate] x [Lines per Order] x [Revenue per Line]

Again, the specific metrics can and should vary based on the business, but I got to a pretty handy list in my example case simply by breaking down revenue into the sub-metrics that, mathematically, drive it.

The Flexibility of Scripting the Answer

Certainly, one way to tackle answering the question would be to use Ad Hoc Analysis or Analysis Workspace. But, the former doesn’t visualize heatmaps at all, and the latter…doesn’t visualize this sort of heatmap all that well. Report Builder was another option, and probably would have been the route I went…except there were other questions I wanted to explore along this two-dimensional construct that are not available through Report Builder.

So, I built “the answer” using R. That means I can continue to extend the basic work as needed:

  • Exploring additional metrics
  • Exploring different dimensions
  • Using the basic approach with other sites (or with specific segments for the current site — such as “just mobile traffic”)
  • Extending the code to do other explorations of the data itself (which I’ll get into with the next two posts)
  • Extending the approach to work with Google Analytics data

Key Aspects of R Put to Use

The first key to doing this work, of course, is to get the data out. This is done using the RSiteCatalyst package.

The second key was to break up the code into a handful of different files. Ultimately, the output was generated using RMarkdown, but I didn’t put all of the code in a single file. Rather, I had one script (.R) that was just for configurations (this is what you will do most of the work in if you download the code and put it to use for your own purposes), one script (.R) that had a few functions that were used in answering multiple questions, and then one actual RMarkdown file (.Rmd) for each question. The .Rmd files use read_chunk() to selectively pull in the configuration settings and functions needed. So, the actual individual files break down something like this:

This probably still isn’t as clean as it could be, but it gave me the flexibility (and, perhaps more importantly, the extensibility) that I was looking for, and it allowed me to universally tweak the style and formatting of the multi-slide presentations that each question generated.

The .Renviron file is a very simple text file with my credentials for Adobe Analytics. It’s handy, in that it only sits on my local machine; it never gets uploaded to Github.

How It Works (How You Can Put It to Use)

There is a moderate level of configuration required to run this, but I’ve done my best to thoroughly document those in the scripts themselves (primarily in config.R). But, summarizing those:

  • Date Range — you need to specify the start and end date. This can be statically defined, or it can be dynamically defined to be “the most recent full week,”  for instance. The one wrinkle on the date range is that I don’t think the script will work well if the start and end date cross a year boundary. The reason is documented in the script comments, so I won’t go into that here.
  • Metrics — for each metric you want to include, you need to include the metric ID (which can be something like “revenue” for the standard metrics or “event32” for events, but can also be something like “cm300000270_56cb944821d4775bd8841bdb” if it’s a calculated metric; you may have to use the GetMetrics() function to get the specific values here. Then, so that the visualization comes out nicely, for each metric, you have to give it a label (a “pretty name”), specify the type of metric it is (simple number, currency, percentage), and how many places after the decimal should be included (visits is a simple number that needs 0 places after the decimal, but, “Lines per Order” may be a simple number where 2 places after the decimal make sense).
  • One or more “master segments” — it seems reasonably common, in my experience, that there are one or two segments that almost always get applied to a site (excluding some ‘bad’ data that crept in, excluding a particular sub-site, etc.), and the script accommodates this. This can also be used to introduce a third layer to the results. If, for instance, you wanted to look at last touch channel and device category just for new visitors, then you can apply a master segment for new visitors, and that will then be applied to the entire report.
  • One Segment for Each Dimension Value — I went back and forth on this and, ultimately, went with the segments approach. In the example above, this was 13 total segments (one each for the seven channels, which included the “All Others” channel, and one each for each of the six customer segments, which was five customer segment values plus one “none specified” customer segment). I could have also simple pulled the “Top X” values for specific dimensions (which would have had me using a different RSiteCatalyst function), but this didn’t give me as much control as I wanted to ensure I was covering all of the traffic and was able to make an “All Others” catch-all for the low-volume noise areas (which I made with an Exclude segment). And, these were very simple segments (in this case, although many use cases would likely be equally simple). Using segments meant that each “cell” in the heatmap was a separate query to the Adobe Analytics API. On the one hand, that meant the script can take a while to run (~20 minutes for this site, which has a pretty high volume of traffic). But, it also means the queries are much less likely to time out. Below is what one of these segments looks like. Very simple, right?

  • Segment Meta Data — each segment needs to have a label (a “pretty name”) specified, just like the metrics. That’s a “feature!” It let me easily obfuscate the data in these examples a bit by renaming the segments “Channel #1,” “Channel #2,” etc. and “Segment A,” “Segment B,” etc. before generating the examples included here!
  • A logo — this isn’t in the configuration, but, rather, just means replacing the logo.png file in the images subdirectory.

Getting the segment IDs is a mild hassle, too, in that you likely will need to use the GetSegments() function to get the specific values.

This may seem like a lot of setup overall, but it’s largely a one-time deal (until you want to go back in and use other segments or other metrics, at which point you’re just doing minor adjustments).

Once this setup is done, the script just:

  • Cycles through each combination of the segments from each of the segment lists and pulls the totals for each of the specified metrics
  • For each [segment 1] + [segment 2] + [metric] combination, adds a row to a data frame. This results in a “tidy” data frame with all of the data needed for all of the heatmaps
  • For each metric, generates a heatmap using ggplot()
  • Generates an ioslides presentation that can then be shared as is or PDF’d for email distribution

Easy as pie, right?

What about Google Analytics?

This code would be fairly straightforward to repurpose to use googleAnalyticsR rather than RSiteCatalyst. That’s not the case when it comes to answering the questions covered in the next two posts (although it’s still absolutely doable for those, too — I just took a pretty big shortcut that I’ll get into in the next two posts). And, I may actually do that next. Leave a comment if you’d find that useful, and I’ll bump it up my list (it may happen anyway based on my client work).

The Rest of the Series

If you’re feeling ambitious and want to go ahead and dive into the rest of the series:

Analysis, Reporting, Team Demystified

Disseminating Digital Data: Why A One Size Fits All Model Doesn’t Work

[Shared by Nancy Koons, Digital Analytics Consultant, Team Demystified …]

One of the things I love about working with the folks at Demystified are the conversations about analytics that often spring up in our Slack group. Whether it’s a discussion around tool capabilities, proper use of metrics, or how to deliver insights effectively, I’m always learning new things and appreciating the many perspectives brought to the table.

Today a discussion unfolded around Data Studio and the sharing of data within organizations. Data Studio is Google’s newest data visualization tool. It has been built to encourage users to interact directly with the dashboards. You can apply filters, manipulate date ranges – all great features designed to facilitate analysis and engage users. Today, the topic of NOT currently being able to save a version of the dashboard as a PDF came up, with some energized discussion around whether or not this is still a needed piece of functionality in today’s world. One perspective was that Google is trying to shift the way organizations consume analytics and drive innovation – which is a very interesting concept. Getting people more engaged and interacting directly with their data is a worthy goal indeed.

For many organizations, however, I think there is still a need to be able to share snapshot “reports” or dashboards as static docs and I am going to outline those reasons in this post:

1) Executive Consumption:  While there are many tools out there that support pulling in multiple, disparate data sources, in a large or complex organization I still see many companies struggle to pull everything together into one, cohesive dashboarding tool or system. If you are able to do this, then (kudos!) and it could be perfectly reasonable to ask an executive to log on to view dashboards. (They probably approved a decent chunk of change to get the system implemented, after all.) My experience with larger, complex organizations is that the C-Suite is often monitoring things like: offline and online sales, cancelled/return merchandise reports, sales team quotas and leads, operations reports, inventory systems, and getting all of that into one system is still more of a dream than a reality. And when that is the case, I think asking an Exec to log into a one system to view one set of reports, and another tool to access other data is not reasonable. In some cases, sure, they may be open to it, but I know a lot of companies where the expectation is that the business units provide reports in the format the exec asks – not the other way around.

2) Technology Norms and Preferences: One of the clients I work with uses Google Analytics for their websites, and could be a good candidate to build out dashboards using Data Studio. Unfortunately, they are more of a Windows/Microsoft organization, where most end-users within the company do not have Google Accounts, so viewing a dashboard in Data Studio would require an extra hurdle in setting up that type of account just to view a report (hat tip to Michele Kiss for pointing that out!). While not necessarily advanced or ideal, analytics reports and insights are typically distributed via email (slides or PDF format). When data is discussed, it tends to be in meetings in conference rooms- where internet speed can sometimes be a challenge- not to mention you may end up relying on your vendor’s ability to refresh/display data at a critical moment. (Something Elizabeth “Smalls” Eckels encountered with a client while we were discussing this very topic!) Some executives or managers may also prefer to catch up on performance reports while traveling, and the ability to connect to the internet on a plane, in an airport or in a hotel can still prove to be a challenge at times.

3) Resource Knowledge: One of my continual concerns with non-analytics people accessing digital analytics data is the ability to pull invalid metrics or data into a report, or interpret the data incorrectly. There are still many non-digital marketing managers who want to understand their digital data, but need help understanding the terminology, what a metric truly represents, and how to take the information from a report or dashboard and make a good decision.

4) Ease of Use and Advancing Analytics Internally: Finally, if you want to elevate the role of analytics within an organization, making it as easy as possible for people to consume the right information goes a long way. Don’t make an executive hop through hoops (and get irritated or frustrated). Don’t set up a non-analyst to struggle. Evaluate the tech savviness, the appetite, and ability for your end user to consume an interactive dashboard before rolling it out to a team of marketers and executives who are not prepared to use it. While I think it should be much, much easier for anyone to work with digital data, it’s my view that digital analytics tools still have work to do to make it easier for your average marketing or non-analyst end user to pull the right info quickly and easily.

Analysis

How Often Should You Revisit Your KPIs?

The need to define your Key Performance Indicators is an old, tired concept. Been there, done that, we get it.*

But how many of us think about the need to revisit those KPIs? Perhaps you went through an extensive definition exercise. . . three years ago. Since then, you have redesigned your website, optimized via A/B testing, introduced a mobile app, and created the ability to sign up via phone. Do those thoroughly discussed KPIs still apply?

On the one hand, your true “Single Performance Indicator” should stand the test of time, since it measures your overall business success – which ultimately doesn’t change much. An ecommerce or B2B company still wants to drive sales, a content site wants ad revenue, a non-profit wants donations.

However, for the majority of businesses, digital does not encompass their entire customer interaction. It is common for digital KPIs be a leading indicators of overall business success. For example, submitting a lead or form, signing up for a trial, viewing pricing, downloading a coupon or comparing products can be online KPIs. They may not involve money directly changing hands (yet), but they are a crucial first step. These digital KPIs are even more likely to change as you add new site or mobile features.

So how often should you revisit your digital KPIs?

This can depend on two things:

  1. How fast your development is. If it takes you two years to redesign a website or launch an app, you can probably revisit your KPIs less frequently, since your action is generally slow moving. (If nothing has changed, then it’s unlikely your KPIs need updating.)
  1. Your customer purchase cycle. If you have a short, frequent buying cycle, you may want to revisit your KPIs more often. If you are a B2B company with a two-year buying cycle, this may be less critical.

My recommendation is to conduct a review of your KPIs and leading indicators at minimum, once a year. If nothing else, a review where nothing changes at least assures both the business and the analytics team that their efforts are appropriately directed.

If you have a short purchase cycle, and/or typically iterate quickly as a company, this might be better done every six months, and/or while planning for big projects that will may materially change your KPIs. For example, let’s say your signup process was previously all online. Now, a new project is introducing the ability for customer service representatives to help customers sign up via phone. A project like that could have a huge impact on your existing KPIs, and a reassessment is critical.

Now, keep in mind this applies to your overall business KPIs. Individual campaigns should still begin with consideration of what KPIs will be used to measure their success, every time.

Analytics is continually evolving, and performance measurement must assess your business against up-to-date objectives. However, KPIs are not only important for reporting, but also for analysis and optimization, which ultimately seeks to drive performance. By revisiting your KPIs, you can ensure that both the business and your analysts are focused on the right things.

* And still, it’s not always done. But that’s a topic for another day.

Analysis, Conferences/Community

Pairing Analytics with Qualitative Methods to Understand the WHY

Rudimentary analytics can be valuable to understand what your customers and prospects do. However, the true value from analytics comes from marrying that with the why – and more importantly, understanding how to overcome the why not.

At Analytics that Excite in Cincinnati this month, I spoke about qualitative methods that analysts can use to better understand people’s behavior and motivations. Check out the presentation on SlideShare.

analytics-and-qualitative-to-understand-why

Analysis, Conferences/Community, google analytics, Presentation

Advanced Training for the Digital Analyst

In today’s competitive business environments, the expectations placed on the digital analysts are extremely high. Not only do they need to be masters of the web analytics tools necessary for slicing data, creating segments, and extracting insights from fragmented bits of information…but they’re also expected to have fabulous relationships with their business stakeholders; to interpret poorly articulated business needs; to become expert storytellers; and to use the latest data visualization techniques to communicate complex data in simple business terms. It’s no short order and most businesses are challenged to find the staff with the broad set of skills required to deliver insights and recommendations at the speed of business today.

In response to these challenges, Analytics Demystified has developed specific training courses and workshops designed to educate and inform the digital analyst on how to manage the high expectations placed on their job roles. Starting with Requirements Gathering the Demystified Way, we’ll teach you how to work with business stakeholders to establish measurement plans that answer burning business questions with clear and actionable data. Then in Advanced Google Analytics & Google Tag Manager, we’ll teach you or your teams how to get the most from your digital analytics tools. And finally in our workshops for digital analysts, attendees can learn about Data Visualization and Expert Presentation to put all their skills together and communicate data in a visually compelling way. Each of these courses is offered in our two day training session on October 13th & 14th. If any of these courses are of interest…read on:

 

Requirements Gathering the Demystified Way

Every business with a website goes through changes. Sometimes, it’s a wholesale website redesign, other times a new microsite emerges, or maybe it’s small tweaks to navigation, but features change, and sites evolve always. This workshop led by Analytics Demystified Senior Partner, John Lovett will teach you how to strategically measure new efforts coming from your digital teams. The workshop helps analysts to collaborate with stakeholders, agencies, and other partners using our proven method to understand the goals and objectives of any new initiative. Once we understand the purpose, audience and intent, we teach analysts how to develop a measurement plan capable of quantifying success. Backed with process and documentation templates analysts will learn how to translate business questions into events and variables that produce data. But we don’t stop there…gaining user acceptance is critical to our methodology so that requirements are done right. During this workshop, we’ll not only teach analysts how to collect requirements and what to expect from stakeholders, we we also have exercises to jumpstart the process and send analyst’s back to their desk with a gameplan for improving the requirements gathering process.  

 

Advanced Google Analytics & Google Tag Manager

Getting the most out of Google Analytics isn’t just about a quick copy-paste of JavaScript. In this half-day training, you will learn how to leverage Google Analytics as a powerful enterprise tool. This session sets the foundation with basic implementation, but delves deeper into more advanced features in both Google Analytics and Google Tag Manager. We will also cover reporting and analysis capabilities and new features, including discussion of some exclusive Premium features. This session is suitable for users of both Classic and Universal Analytics, both Standard and Premium.

 

Data Visualization and Expert Presentation

The best digital analysis in the world is ineffective without successful communication of the results. In this half-day class, Web Analytics Demystified Senior Partners Michele Kiss and Tim Wilson share their advice for successfully presenting data to all audiences, including communication of numbers, data visualization, dashboard best practices and effective storytelling and presentation.

 

At Analytics Demystified we believe that people are the single most valuable asset in any digital analytics program. While process and technology are essential ingredients in the mix as well, without people your program will not function. This is why we encourage our clients, colleagues, and peers to invest in digital analytics education. We believe that the program we’re offering will help any Digital Analyst become a more valuable member of their team. Reach out to us at partners@analyticsdemystified.com to learn more, or if we’ve already convinced you, sign up to attend this year’s training on October 13th & 14th in San Francisco today!

Analysis, Featured

The Most Important ‘Benchmarks’ Are Your Own

Companies typically expend unnecessary energy worrying about comparing themselves to others. What analyst hasn’t been asked: “What’s a typical conversion rate?” Unfortunately, conversion rates can vary so dramatically—by vertical, by product, by purchase cycle, by site design—not to mention, the benchmark data available is typically so generic that it is essentially useless.

To explain how benchmarks fail to provide insight, let’s pretend instead of conversion rate we’re talking about a new runner.* Sarah is starting a “Couch to 5K.” In planning her training, Sarah might wonder, “What’s the average running pace?” It’s an understandable question – she wants to understand her progress in context. However, a runner’s pace depends on their level of fitness, physical attributes, terrain, distance, weather and more. In theory, there could be some value in a benchmark pace for a runner just like Sarah: age, size, fitness level, training schedule, terrain, even climate. Ideally, this data would be trended, so she could understand that in Week 4 of her training plan, most people “just like her” are running a 11:30 min/mile. By Week 8, she should be at 10:30.
2015-07-20_15-08-31

However, that’s not how benchmarks work. Benchmark data is typically highly aggregated, including professional runners through newbies, each running in such a wide variety of conditions that it’s essentially rendered useless. Why useless? Because the benchmark is only helpful if it’s truly comparable.

But let’s say Sarah had this (highly generic) ‘average pace’ available to her. How would it make any difference to her progress towards a 5K? If her starting pace was slower than the average, she would set goals to slowly increase it. But even if her starting pace was faster than the average, she wouldn’t pat herself on the back and stop training. She would still set goals to slowly increase it. Thus, her actions would be the same regardless of what the data told her.

That’s the issue with benchmarks. Knowing them makes no difference to the actions the business takes, and data is only useful if it drives action. Back to a business context: Let’s say your conversion rate is 3%, and the industry average is 2.5%. No business is going to stop trying to optimize just because they’re above a generic average.

Ultimately, the goal of a business is to drive maximum profit. This means continually optimizing for your KPIs, guided by your historic performance, and taking in to account your business model and users. So instead of getting sidetracked by benchmarks, help your stakeholders by focusing on progress against internal measures.

First set up a discussion to review the historical trends for your KPIs. Come armed with not only the historical data, but also explanations for any spikes or troughs. Be sure to call out:

  • Any major traffic acquisition efforts, especially lower-qualified paid efforts, because they might have negatively affect historical conversion rates.
  • Any previous site projects aimed at increasing conversion, and their impact.

With this information, you can work with stakeholders to set a tangible goal to track toward, and brainstorm marketing campaigns and optimization opportunities to get you there. For example, you might aim to increase the conversion rate by 0.5% each quarter, or set an end of year goal of 2.6%. Your historical review is critical to keep you honest in your goal setting. After all, doubling your conversion rate is a pretty unrealistic goal if you have barely moved the needle in two years!

Be sure to test and measure (and document!) your attempts to optimize this rate, to quantify the impact of changes. (For example, “Removing the phone number from our form increased the conversion rate by 5%.”) This will help you track what actually moves the needle as you progress toward your goal.

* Awesome fitness analogy courtesy of Tim Patten.

Analysis, Analytics Strategy, Featured

Two (Only Two!) Reasons for Analysis: Opportunities and Problem Areas

A common — and seemingly innocuous — question that analysts get asked all the time, in one form or another:

“Can you do some analysis tell us where you think we can improve our results?”

Seemingly innocuous…but what does it really mean? All too often, it seems like we have a tendency to just analyze for the sake of analyzing — without really having a clear purpose in mind. We tell ourselves that we should be doing better, without really thinking about the type of “better” that we’re trying too achieve.

I was having this discussion with a client recently who was challenging me to explain how to approach analysis work. I found myself pointing out that there are really only two scenarios where analysis (or optimization) makes sense:

  • When there is a problem
  • When there is a potential opportunity

It really breaks down – conceptually – pretty simply:

Problems vs. Opportunities

Some examples:

  • I send an email newsletter once a month, which accounts for a pretty small percentage of traffic to my site (Level of Activity = Low), but that channel delivers the highest conversion rate of any channel (Results Being Delivered = High). On the one hand, that’s expected. On the other hand, is this an OPPORTUNITY? Can I send email more frequently and increase the level of activity without killing the results being delivered? Basically…can I move it into the NO ANALYSIS REQUIRED zone with some analysis and action?
  • Or, flip it around to another classic: I have a high volume of traffic (Level of Activity = High) from Display going to a campaign landing page, and that traffic is converting at a very low rate (Results Being Delivered = Low). That’s a PROBLEM AREA that warrants some analysis. Should media spend be scaled back while I try to figure out what’s going on? Is it the page (should I optimize the landing page experience with A/B testing?) or is it the traffic quality (should the media targeting and/or banner ad creative be adjusted)? Again, the goal is to get that segment of traffic into the NO ANALYSIS REQUIRED zone.
  • Finally, I’ve dug into my mobile traffic from new visitors from organic search. It’s performing dramatically below other segments (Results Being Delivered = Low). But, it also represents a tiny fraction of traffic to my site (Level of Activity = Low). How much effort should I put into trying to figure out why this traffic is performing poorly? “But, maybe, if you figure out why it’s performing poorly with the existing traffic, you’ll also get more traffic from it!!! You can’t ignore it. You need to try to make it better!” you exclaim. To which I respond: “Maybe.” What is the opportunity cost of chasing this particular set of traffic? What traffic is already in the PROBLEM AREA or OPPORTUNITY zone? Isn’t it more likely that I’ll be able to address one of these dimensions rather than hoping my analysis addresses both of them simultaneously?

This diagram is nothing more than a mental construct – a way to assess a request for analysis to try to hone in on why you’re doing it and what you’re trying to achieve.

What do you think?

Analysis, General

I Believe in Data*

* [Caveats to follow]

Articles about analytics tend to take two forms. One style exalts data as a cure-all panacea.

Another style implies that people put too much faith in the power of their data. That there are limitations to data. That data can’t tell you everything about how to build, run and optimize your business. I agree.

My name is Michele Kiss, I am an analytics professional, and I don’t believe that data solves everything.

I believe in the appropriate use of data. I don’t believe that clickstream data after you redesign your site can tell you if people “like” it. I believe that there are many forms of data, that ultimately it is all just information, and that you need to use the right information for the right purpose. I do not believe in torturing a data set to extract an answer it is simply not equipped to provide.

I believe in the informed use of data. Data is only valuable when its i) definitions and its ii) context are clearly understood. I believe there is no such thing as pointing an analyst to a mountain of data and magically extracting insights. (Sorry, companies out there hiring data scientists with that overblown expectation.)

When a nurse tells your doctor, “150” and “95” those numbers are only helpful because your doctor knows that i) That’s your blood pressure reading and ii) Has a discussion with you about your diet, exercise, lifestyle, stress, family/genetic history and more. That data is helpful because it’s definition is clear, and your doctor has the right contextual information to interpret it.

I believe that the people and processes around your data determine success more than the data itself. A limited data set used appropriately in an organization will be far more successful than a massive data set with no structure around its use.

I believe data doesn’t have to be perfect to be useful, but I also believe you must understand why imperfections exist, and their effect. Perfect data can’t tell you everything, but outright bad data can absolutely lead you astray!

I believe that data is powerful, when the right data is used in the right way. I believe it is dangerously powerful when misused, either due to inexperience, misunderstanding or malice. But I don’t believe data is all powerful. I believe data is a critical part of how businesses should make decisions. But it is one part. If used correctly, it should guide, not drive.

Data can be incredibly valuable. Use it wisely and appropriately, along with all the tools available to your business.

“There are more things in heaven and earth, Horatio,
Than are dreamt of in your [big data set].”
– Analytics Shakepeare

Analysis, Featured, General

The Curse of Bounce Rate and ‘Easy’ Metrics (And Why We Can Do Better)

One of the benefits of having a number of friends in the analytics industry is the spirited (read: nerdy) debates we get in to. In one such recent discussion, we went back and forth over the merits of “bounce rate.”

I am (often vehemently) against the use of “bounce rate.” However, when I stepped back, I realised you could summarise my argument against bounce rate quite simply:

Using metrics like ‘bounce rate’ is taking the easy way out, and we can do better

Bounce rate (at its simplest) is the percentage of visits that land on your site that don’t take a second action. (Don’t see a second page, don’t click the video on your home page, etc.)

My frustration: bounce rate is heavily dependent upon the way your site is built, and your analytics implementation (do you have events tracking that video? how are the events configured? is your next page a true page load?) Thus, bounce rate varies as to what exactly it represents. (Coupled with the fact that most business users don’t, of course, understand the nuances, so may misuse a metric they don’t understand.)

So let’s take a step back. What are we trying to answer with “bounce rate”?

Acquisition analysis (where bounce rate is commonly used) compares different traffic sources or landing pages by which do the best job of “hooking” the user and getting them to take some next action. You are ultimately trying to decide what does the best job of driving the next step towards business success.

Let’s use that!

Instead of bounce rate, what are the conversion goals for your website? What do you want users to do? Did they do it? Instead of stopping at “bounce rate”, compare your channels or landing pages on how they drive to actual business conversions. These can be early-stage (micro) conversions like viewing pricing, or more information, or final conversions like a lead or a purchase.

So, what is better than bounce rate?

  • Did they view more information or pricing? Download a brochure?
  • Did they navigate to the form? Submit the form?
  • Did they view product details? Add to cart? Add the item to a wish list?
  • Did they click related content? Share the article?

Using any of these will give you better insight into the quality of traffic or the landing pages you’re using, but in a way that truly considers your business goals.

But let’s not stop there…. what other “easy metrics” do we analysts fall back on?

What about impressions?

I frequently see impressions used as a measure of “awareness.” My partner, Tim Wilson, has already detailed a pretty persuasive rant on awareness and impressions that is well worth reading! I’m not going to rehash it here. However the crux of this is:

Impressions aren’t necessarily ‘bad’ – we can just do a better job of measuring awareness.

So what’s better than impressions?

  • At least narrow down to viewable impressions. If we are honest with ourselves, an impression below the fold that the user doesn’t even see does not affect their awareness of your brand!

  • Even measuring clicks or click-through is a step up, since the user at least took some action that tells you they truly “saw” your ad – enough to engage with it.

  • A number of vendors provide measures of true awareness lift based on exposure to ad impressions, by withholding and measuring against a control group. This is what you truly want to understand!

What about about page views per visit or time on site?

Page views per visit or time on site is commonly used in content sites as a measure of “engagement.” However (as I have ranted before) page views or time can be high if a user is highly engaged – but they can also be high if they’re lost on the site!

So what’s better than just measuring time or page views?

So why do we do this?

In short: Because it’s easy. Metrics like bounce rate, page views, time on site and impressions are basic, readily available data points provided in standard reports. They’re right there when you load a report! They are not inherently ‘bad’. They do have some appropriate use cases, and are certainly better than nothing, in the absence of richer data.

However, analysis is most valuable when it addresses how your actions affect your business goals. To do that, you want to focus on those business goals – not some generic data point that vendors include by default.

Thoughts? Leave them in the comments!

Analysis

Every Analyst Should Follow fivethirtyeight.com

I’ll admit it: I’m a Nate Silver fanboy. That fandom is rooted in my political junky-ism and dates back to the first iteration of fivethirtyeight.com back in 2008. Since then, Silver joined the New York Times, so fivethirtyeight.com migrated to be part of that media behemoth, and, more recently, Silver left the New York Times for ESPN — another media behemoth. This bouncing around has been driven by Silver’s passion for various places where data is abundant and underutilized: starting with online poker, then baseball analytics, and then a sharp turn to political polling (the original fivethirtyeight.com), which then went even more deeply into politics (the Times iteration of fivethirtyeight.com), which then went broadly into data across many subjects (his book), and which then stayed fairly broad…but with a return to some heavier sports (with the ESPN iteration of fivethirtyeight.com).

Silver talks a lot about what data can and cannot do and how it gets mis-used, and he often dives into details of statistical analysis that I really can’t quite follow. But, he also has a whole other aspect of what he (and his team) does really, really well that I haven’t seen him talking about much. These are twofold:

  1. Picking the questions that are worth answering
  2. Effectively visualizing that data

These are both key to his success, but they’re also key to any analyst’s ability to deliver value within their organizations.

Picking Questions Worth Answering

Silver originally picked questions that simply intrigued him (winning at online poker, better analyzing baseball players, predicting election outcomes), and those wound up getting him to questions that had mass appeal. Now, as a media site, the questions his team picks, I assume, have a heavy component of “will this drive traffic?” The questions have a pretty diverse range:

  • In the wake of the Sandy Hook shootings, what happened with media coverage and public opinion about gun control? [Article]
  • Will the recent moves to add calorie counts to fast food menus actually change consumer consumption behavior? [Article]
  • Would lifting the ban that prevents gay men from donating blood meaningfully move the needle on blood donations? [Article]

These questions  are often driven by current events and, clearly, would be of interest to a sufficiently large number of potential readers.

“But I’m not trying to drive impressions with my analyses! I just want to drive my business forward!” you exclaim! “How does this relate to me?!”

I’ll claim that it does, but I’ll admit it’s a somewhat meta argument. The dream for most analysts is to find something that gets widely shared internally, because the work reveals something that is surprising and actionable. It’s sooooo easy to lose sight day-in and day-out of the need to be tackling questions that will be most likely to lead to dissemination and action. fivethirtyeight.com — any media site, really — has to focus on content that will be “popular” (in some definition of the word). As an analyst, shouldn’t we constantly be going beyond reacting to the questions that fall in our lap and seeking out meaningful questions to answer? 

For me, every time I read an article on fivethirtyeight.com and think, “Aw, man! That author is so lucky to have gotten to dig into that!” I try to remind myself that I do have some control over what I dig into with most of my clients, and I should constantly be seeking questions that would have broad and actionable appeal (and pushing them to identify those questions themselves).

Effectively Visualizing that Data

This second aspect of the content on fivethirtyeight.com is more tangible and directly applicable. It’s not that every article nails it, but most of the articles include visualized data, and most of those visualizations are very well thought through — neither picking a “standard” visualization, nor getting fancy for fanciness’s sake.

I’m a casual college football fan, at best, but it’s been interesting to watch Silver struggle with predicting who would be in the first “final four” with the change to the championship system that went into place this year. One of his approaches was to run simulations based on what clues he could find about how the selection committee would act, combined with predictions for the results of as-yet-unplayed games. This resulted in a chart like the one below.

Although the one below didn’t actually get the final four “right,” in that TCU dropped out and Ohio State was in…this was something that was almost impossible to accurately predict (between the wildcard of the selection committee’s process, and the fact that Ohio State surprised everyone by blowing out Wisconsin in the Big Ten championship game that occurred several days after he ran this simulation). But, the visualization works on two levels: 1) at a glance, it’s clear which teams his analysis show as being in contention for a final four spot, and 2) the use of the heatmap and dividing lines provides a second level of detail as to the skewing and variability that the model predicted for each team:

College Football Predictions

 Are you not a “sportsball” (<– Michele Kiss hat tip) fan? Let’s look at an example from politics!

When Jeb Bush took an offical pre-pre-pre-pre-“I’m running for U.S. President” step, Silver asked the question: “Is Jeb Bush Too Liberal To Win the Republican Nomination in 2016?”  To tackle this, he pulled third party data from three different sources that all used different techniques to quantify where various political figures fall on the liberal-conservative spectrum. The result? Another exceedingly well-presented visualization!

Again, the visualization works on two levels: 1) at a glance, it shows that Bush appears to skew to the left side of the conservative spectrum, but he’s not extremely so, and 2) the second layer of detail shows where current (potential) and past Republican candidates fall relative to each other, how consistent each of 2 or 3 different measurement systems aligned  when making that assessment (see Rand Paul!), and even how the times they have a’ changed as to the “average” for the party (for Congress):

Political Conservatives Relative Conservatism

The great visualizations aren’t limited to sports and politics, nor are they limited to Silver’s posts. One final example is, in one sense, “just” a simple histogram, but it’s a histogram that has had some real care put into by Mona Chalabi. She tried to answer the question: “How Common Is It For A Man To Be Shorter Than His Partner?” She was limited to secondary data (which was quite limiting!), and she noted at the outset that, for a range of reasons, the results weren’t all that surprising. But, in the histogram below, look at how much care was put into adding clear labels (“Woman taller.” “Man taller”), using color to emphasize the “answer to the original question,” and even the addition of a simple vertical line to represent “equal height.”

How Common Is It for a Man to Be Shorter Than His Partner?

I absolutely love the level of care that fivethirtyeight.com puts into their visualizations. They clearly have a well-defined style guide when it comes to the palette, fonts, and font size. But, as with any good style guide, those constraints enable a high level of creativity to then determine what the truly best way to visualize the information is.

fivethirtyeight.com is my newest most favorite site. As I opened with, much of the underlying content is actually on topics I care about, but I’m going to justify my on-going consumption of that content by claiming that it is also a source of inspiration and motivation for improving my work as an analyst!

Analysis, Reporting

I’ve Become Aware that Awareness Is a #measure Bugaboo

A Big Question that social and digital media marketers grapple with constantly, whether they realize it or not:

Is “awareness” a valid objective for marketing activity?

I’ve gotten into more than a few heated debates that, at their core, center around this question. Some of those debates have been with myself (those are the ones where I most need a skilled moderator!).

The Arguments For/Against Awareness

Here’s the absolutist argument against awareness:

“There is no direct business value in driving ‘Awareness.’ It’s a hope and a prayer that increasing awareness of your brand/product will eventually lead to increased sales, but, if you’re not actually making that link with data, then you might as well admit that you’re trying to live in the Mad Men era of Marketing.”

Here’s the absolutist argument for awareness:

“While ‘the funnel’ has been completely blown up by the introduction of digital and the increasingly fragmented consumer experience, it’s impossible for a consumer to make a purchase of a consumer brand without being aware that the brand exists. Logically, then, if and until we know that 100% of our target consumers are aware that we exist (and even what we stand for — awareness is more than just ‘recognize the brand’ and, when I [the absolutist] say ‘awareness’ I mean that consumers have some knowledge of the brand, and that knowledge gives them a favorable impression!). But, between that fragmented experience and the fact that it’s totally reasonable to expect a time delay between achieving ‘awareness’ and a consumer actually making a purchase, we just have to accept that we won’t reasonably be able to tie directly to sales as easily as direct response activity can!”

Obviously, any time an argument gets framed with “absolutist” viewpoints, the blogger thinks the reality is somewhere in between the two extremes.

And I do.

But I’m much closer to the absolutist-for-awareness position. I wouldn’t possibly be considering pre-ordering a WhistleGPS if I wasn’t at least aware that the product exists. At the same time, I am only vaguely aware of when it crept into my consciousness as existing. Now, many of the impressions that led to my awareness are trackable, and, if and when I pre-order, those impressions (the digital ones, at least, but I think all of my exposure has been digital) can be linked to me as a purchaser. But, the conversion lag will be several months at that time — even when trackable, that’s not “real-time” conversion data that could have been used to optimize their sponsored posts or remarketing campaigns. So, whether I’m being included in a media mix model or an attribution management exercise, I’m posing some big challenges.

But That Doesn’t Mean I’m Happy with Awareness

The against-awareness absolutists have a valid point, in that “hope and a prayer” is really not a valid measurement approach. And, neither is “impressions,” which is what marketers often use as their KPI for awareness. Impressions is a readily available and easily understood measure, but it’s a measure of exposure rather than awareness.

IMPRESSIONS = EXPOSURE <> AWARENESS

So, the question for marketers is: “Is your goal to just increase brand exposure, or do you really care about increasing brand awareness?”

“Well, gee, Tim. You have to increase exposure of the brand — impressions! — in order to increase awareness. And, you can’t truly measure ‘awareness,’ can you?”

Oh, how I would kill to actually have that discussion. Because you can measure awareness in many cases. And, that can be extended to be both unaided or aided awareness, as well as brand affinity and even purchase intent!

I’m actually appalled at how often digital media agencies don’t more effectively measure the impact “awareness-driving” campaigns! It’s easy to resort to “impressions.” Is it laziness, or is it that they’re terrified that measuring awareness may be a much less compelling story than a “millions of impressions!!!” story?

Measuring Awareness

There is one more nuance here. We don’t actually want to measure awareness in absolute terms. Rather, we want to measure the increase or lift in awareness resulting from a particular campaign. And that is doable. Even macro-level — quarterly or annual — brand awareness surveys are more interested in if they have increased awareness since the prior study and, if so, by how much.

This is in not an endorsement of a specific product or service, but it would be disingenuous for me to describe one such methodology without crediting where I first saw and learned about it, which was through Vizu (the image below is from their home page):

vizulift

This is for measuring the lift in awareness for a display ad campaign. The concept is fairly simple:

  1. Track which users have been exposed to display ads and which ones haven’t.
  2. Use a small portion of the ad buy to actually serve an in-banner survey to both groups to gauge awareness (or preference or intent or whatever attitudinal data you want).
  3. Compare the “not exposed” group’s responses (your control) to the “exposed” group’s responses. The delta is the lift that the display campaign delivered.

This can — and generally needs to — measure the lift from multiple exposures to an ad. Repetition does matter. But, a technique like this can help you find the sweet spot for when you start reaching diminishing returns for incremental repeated impressions.

And, depending on the size of the media buy, a simple lift study like this can often be included as a value-add service. And it can be used to optimize the creative and placements against something much closer to “business impact” than clickthroughs or viewthroughs.

Vizu is actually part of Nielsen, which has other services for measuring awareness, and Dynamic Logic (part of Millward Brown) also offers solutions for measuring “brand” rather than simply measuring exposure.

My Advice? Be Precise.

At the end of the day, if you’re fine with measuring impressions then be clear that you really care more about exposure than actual awareness, affinity, or purchase intent. If you do care about true brand impact, then do some research and find a tool or service that enables you to measure that impact more appropriately.

Analysis

Hello. I’m a Radical Analytics Pragmatist

I was reading a post last week by one of the Big Names in web analytics…and it royally pissed me off. I started to comment and then thought, “Why pick a fight?” We’ve had more than enough of those for our little industry over the past few years. So I let it go.

Except I didn’t let it go.

Source: Flickr / Adrian Tombu

I was still fuming about it later that day.

And the next day.

And now…almost a week later. Still fuming.

I went back to the post to see if any of the commenters had called out the blather for what it was (I’d initially read the post shortly after it was published, so there were no comments at the time). Several dozen comments…and they were all fawning over the content: “This is brilliant! I’m totally going to start doing what this recommends.”

Here’s the issue: what the post recommended, in my mind, was wrong. And, thus…

I am radical

There it is — in all of it digital pixel clarity from Merriam-by-gawd-Webster:

radical

If a Big Name writes a post that I read as some of the worst possible advice ever gets glowingly praised by his acolytes (many of whom are my industry peers)…am I radical or an idiot? Is a radical just an idiot with delusions of grandeur (see Act Three from this recently rebroadcast This American Life episode)?

Sometimes, certainly. But, as I get older, spend more and more time in the analytics profession (I passed the decade mark several years ago), and continue to work with client after client that has approached their analytics work by trying to apply clichés that have been (mis)interpreted as best practices, I’m becoming increasingly entrenched in my belief that, while I may have some radical and contrarian views…these views are right.*

Some of my favorites (most hated) of these clichés:

  • “Good dashboards don’t just show what happened. They show why it happened.” Wrong!
  • “Dashboards are only useful if they include insights and recommendations.” Poppycock!!!
  • “Good analysts dig into the data after a campaign and find insights from that data.” I hate this one because it’s insidious — good analysts do do exactly this, but, when an analyst or marketer makes this statement, they’re generally making it as a way to ignore what needs to happen before the analyst digs in in order to make this an efficient reality.
  • “Google Analytics Intelligence Events and Adobe Analytics Anomaly Detection are the wave of the future — ‘the technology’ is finally telling analysts where they should start their analysis!” For the love of all things dimensional and metrical, please remove your tool-centric cranium from your rectum. NO!!!
  • “If the marketer hasn’t articulated what questions they want to answer, the analyst should know the business well enough to come up with those questions — hypotheses — on their own and should dig into the data and answer them.” This is another insidious one — analysts wayyyy too willing to ignore the marketer’s experience and brain as a critical part of the team.
  • “Designers shy away from analytics. They know it will stunt their ability to be creative.” I’m not a violent man, but I’ve wanted to punch more than one analyst in the face when I’ve heard this “statement of fact.”

Occasionally, I’ve taken some of these clichés head-on and ripped out a blog post, like this one about why I don’t include text-based commentary on dashboards. In other cases, I’ve bitten my tongue (I have permanent teethmarks on it to prove it. I’ve had to bite pretty hard).

But, these supposed truisms get my goat every time I come across them in a post or in my work. Just because a lot of people have said something, and it seems to make sense and be easy to comprehend, doesn’t make it true.

So, I’ll give myself the radical label — it’s easier to spell than contrarian.

That’s the easy part. What is the internet if not a forum for individuals to criticize and complain? (Well, it’s a place for people to post and view cat videos…but criticizing and complaining is easily in the Top 5.) That brings me to…

I am a pragmatist

I am a firm believer in the maxim:

“If you aren’t trying to change it, don’t complain about it.”

I can document that, in one way or another, I’ve spent well over half of my career trying to “change it.” I’ve tried out different approaches when I’ve seen “truisms” not work. I’ve refined how I approach different situations, and I’ve spent countless hours (well, in theory, they’re countable…but I haven’t always tracked my time at that level — poor data capture, I guess) developing and refining different ways to counter these industry myths.

At the core of that work is one word: pragmatism. I’ve never proposed an approach that isn’t workable in practice:

  • I’ve distilled each “radical” approach to its essence — to the point where I’m terrified that what I’m stating is so clear and obvious that my audience won’t realize that it’s a radical departure from how they have actually been operating. I tried to summarize the essence of these ideas as aphorisms in a post last year (not a complete list, but a start).
  • I’ve built tools and templates that take these pragmatic concepts and make them directly applicable. Many of those tools are posted here, but I also regularly post downloadable templates in individual blog posts.

And I still fail. But I’m working on it. It’s why I’m now a consultant — I want to spend as much time with as many different analysts and marketers who realize “something isn’t working with this analytics stuff” at as many companies as possible to try to “change it.”

 

* I am an analyst to the core and am wired with the fairly common trait therein of rampant insecurity. I’ve spent most of my professional life assuming that everyone else knows a lot more than me, and it’s only a matter of time until I’m found out and revealed as an utter fraud. It takes a lot for me to make the bold statement: “I am right.”

Photo credit: Flickr / Adrian Tombu

Analysis, Analytics Strategy

How to Deliver Better Recommendations: Forecast the Impact!

One of the most valuable ways to be sure your recommendations are heard is to forecast the impact of your proposal.

Consider what is more likely to be heard:

“I think we should do X…”

vs

“I think we should do X, and with a 2% increase in conversion, that would drive a $1MM increase in revenue”

The benefits of modeling out the impact of your recommendations include:

  1. It forces you to think through your recommendation. Is this really going to drive revenue? If so, how? What are the behaviours that will change that will drive the growth?
  2. A solid revenue estimate will help you “sell” your idea
  3. Comparing the revenue impact estimate of a number of initiatives can help the business to prioritise

There are a few basic steps to putting together an impact estimate:

  • Clarify your idea
  • Detail how it will have an impact
  • Collect any existing data that will help you model that impact
  • Build your model, with the ability to adjust assumptions
  • Using your current data, and assumed impact, calculate your revenue estimate
  • Discuss your proposal with stakeholders and fine-tune the model and its assumptions

Example 1: Adding videos to an ecommerce product page

Sample Revenue Model: Videos on the Product Page

View model

This model forecasts the revenue impact of adding videos to an ecommerce site’s product pages. This model makes a few assumptions about how this project will drive revenue:

  1. It assumes some product page visits will view a video, where those visits would not have previously engaged with photo details
  2. It assumes that conversion from product page to cart page will be improved because of users who were viewing photos being further convinced by video
    • Note: This assumption could be more general, or more specific. In the model we have assumed that conversion will be better for users who view photos or videos. The model could also simplify, and assume a generic lift, without taking in to account whether users view the video or click photos.

It does not assume there will be an impact on:

  1. Migration to the product pages (since users won’t even know there are videos until they get there)
  2. Conversion from cart to purchase
  3. Average Order Value

However, for #2 and #3, placeholders are there to allow the business to adjust those if there is a good reason to.

There are a lot of other levers that could be added, if appropriate:

  • Increase in order size
  • Cross-sell
  • Increase in migration to the product page, if videos were widely advertised elsewhere on the site

So you will see it’s a matter of thinking through the project and how it’s expected to affect behaviour (and subsequently, revenue) in choosing what assumptions to adjust.

Example 2: Adding a new ad unit to the home page

Sample Revenue Model: Ad Unit on Home Page

View model

This is a non-ecommerce example, for a website monetised via advertising. The recommendation is to add a third advertising unit to the home page, a large expanding unit above the nav.

The assumptions made are:

  1. The new ad unit will have high sell through and high CPM. This is because we are proposing a “high visibility” unit that we think can sell well.
  2. The existing Ad Unit 1 will suffer a small decrease in sell through, but retain its CPM
  3. The existing Ad Unit 2, as the cheaper ad unit, will not be affected as those advertisers would not invest in the new, expensive unit

There are of course other levers that could be adjusted:

  • We could factor in the impact to click-through rate for the existing ads, and assume a decrease in CPM for both ads due to lower performance.
  • We could take into account the impact on down-stream ad impressions, as the new ad unit generates clicks off site. For users to click the ad, we would lose revenue from the ads they would have otherwise seen later in their visit.
  • We could, as a business, consider only selling the new ad unit half the time (to avoid such a high-visibility ad being “in user’s faces” all the time), and adjust the sell through rate down accordingly.

Five Tips to Success

  1. Keep the model as simple as possible, while accounting for necessary assumptions and adjustments. The simpler the model, the easier it will be for stakeholders to follow your logic (a critical ingredient for their support!)
  2. Be clear on your assumptions. Why did you assume certain things? And why didn’t you assume others?
  3. Encourage stakeholder collaboration. You want your stakeholders to weigh in on what they think the impact can be. Getting them involved is key to getting them on board. Make it easy for them to adjust assumptions and have the model re-calculate. (A user experience tip: On the example models, you’ll see that I used colour coding: yellow fill with blue text means this is an “adjustable assumption.” Using that same formatting repeatedly will help train stakeholders how to easily adjust assumptions.)
  4. Be cautious. If in doubt, be conservative in your assumptions. If you’re not sure, consider providing a range – a conservative estimate and an aggressive estimate. E.g. With a 1% lift in conversion, we’ll see X, with a 10% lift we’ll see Y.
  5. Track your success. If a project gets implemented, compare the revenue generated to your model, and consider why the model was / was not in line with the final results. This will help fine-tune future models.

Bonus tip: Remember this an estimate. While the model may calculate “$1,927,382.11”, don’t confuse being precise with being accurate. When going back to the business, consider presenting “$1.8-2.0MM” as the final estimate.

What tips would you add?

Share your experiences in the comments!

Analysis, General

7 Tips For Delivering Better Analytics Recommendations

As an analyst, your value is not just in the data you deliver, but in the insight and recommendations you can provide. But what is an analyst to do when those recommendations seem to fall on deaf ears?

1. Make sure they’re good

Too often, analysts’ “recommendations” are essentially just half-baked (or even downright meaningless) ideas. This is commonly due to poor process, and expectation setting between the business and analytics team. If the business expects “monthly recommendations”, analysts will feel pressured to just come up with something. But what ends up getting delivered is typically low value.

The best way to overcome this is to work with the business to clearly set expectations. Recommendations will not be delivered “on schedule”, but when there is a valuable recommendation to be made.

2. Make sure you mean it

Are you just throwing out ideas? Or do you truly believe they are an opportunity to drive revenue, or improve the experience? Product Managers have to stand by their recommendations, and be willing to answer to initiatives that don’t work. Make sure you would be willing to do the same!

3. Involve the business

Another good way to have your recommendations fall on deaf ears is if they 1) Aren’t in line with current goals / objectives; or 2) Have already been proposed (and possibly even discussed and dismissed!)

Before you just throw an idea out there, discuss it with the business. (We analysts are quick to fault the business for not involving us, but should remember this applies both ways!) This should be a collaborative process, involving both the business perspective and data to validate and quantify.

4. Let the data drive the recommendation

It is typically more powerful (and less political…!) to use language like, “The data suggests that…” rather than “I propose that…”

5. Consider your distribution method

The comments section of a dashboard or report is not the place for solid, well thought out recommendations. If the recommendation is valuable, reconsider your delivery methods. A short (or even informal) meeting or proposal is likely to get more attention than the footnote of a report.

6. Find the right receiver

Think strategically about who you present your idea to. The appropriate receiver depends on the idea, the organisation (and its politics…) and the personalities of the individuals! But don’t assume the only “right” person is at the top. Sometimes your manager, or a more hands-on, tactical stakeholder, may be better able to bring the recommendation into reality. Critically evaluate the appropriate audience is, before proposing it to just anyone.

Keep in mind too: Depending on who your idea gets presented to, you should vary the method and level of detail you present. You wouldn’t expect your CMO to walk through every data point and assumption of your model! But your hands-on marketer might want to go through and discuss (and adjust!) each and every assumption.

7. Provide an estimate of the potential revenue impact

In terms of importance, this could easily be Tip #1. However, it’s also important enough for an entire post! Stay tuned …

What about you?

What have you found effective in delivering recommendations? Share your tips in the comments!

Analysis

Overcoming The Analyst Curse: DON’T Show Your Math!

If I could give one piece of advice to an aspiring analyst, it would be this: Stop showing your ‘math’. A tendency towards ‘TMI deliverables’ is common, especially in newer analysts. However, while analysts typically do this in an attempt to demonstrate credibility (“See? I used all the right data and methods!”) they do so at the expense of actually being heard.

The Cardinal Rule: Less is More

Digital measurement is not ninth grade math class. So by default, analysts should refrain from providing too much detail about how they arrived at their conclusions and calculations.

What should you show?

  • Your conclusions / findings

  • Any assumptions you had to make along the way (where these impact the conclusions)

  • Your recommendations

  • An estimated revenue impact

  • A very brief overview of your methodology (think: abstract in a science journal.) This should be enough for people to have necessary context, but not so much they could repeat your entire analysis step by step!

What shouldn’t you show?

  • Detailed methodology

  • Calculations or formulas used

  • Data sources

  • Props / vars / custom variables / events used

  • That’s not to say that this stuff isn’t important to document. But don’t present it to stakeholders.

But Of Course: “It Depends”

Because there is never one rule that applies to all situations, there are additional considerations.

How much ‘math’ you show will also depend on:

  1. The level of your audience

    • Your manager should get more detail. After all, if your work is wrong, they are ultimately responsible.

    • Executives will typically want far less detail.

  1. The individual

    • Some individuals need to see more detail to be confident in relying upon your work. For example, your CFO may need to see more math than your Creative Director. Get to know your stakeholders over time, and take note of who may need a little extra background to be persuaded.

    • Note: We commonly hear the argument “But my stakeholders really want more details!” Keep in mind there is a difference between them hearing you out, and truly wanting it. To test this, try presenting without the minutiae (though, have it handy) and see whether they actually ask for it.

  1. Your findings

    • Are you confirming or refuting existing beliefs?

    • If what you are presenting comes as a surprise, you should be prepared to give more detail as to how you got to those findings.

In Your Back Pocket

Keep additional information handy, both for those who might want it, as well as to remind yourself of the details later. A tab in the back of a spreadsheet, an appendix in a presentation or a definitions/details section in a dashboard can all be a handy reference if the need arises now, or later.

[Credit: Thanks to Tim Wilson, Christopher Berry, Peter O’Neill and Tristan Bailey for the discussion on the topic!]

Analysis

Attribution Management Takes More than Technology

I’ll admit it: I’m something of a late joiner to the attribution management bandwagon. Over the last year or so, though, I’ve come around, and I attribute that (pun totally intended) to some great clients and some vigorous discussions with both platform providers and practitioners.

What I’ve seen — and was a victim of myself — was confusion about what true and meaningful attribution actually is. While it’s widely understood (and pined for!) that attribution management is intended to get beyond the “last click” — factoring in the contribution of each of multiple consumer touchpoints that lead up to a conversion — there are fundamentally two different approaches to actually achieving that goal…and one of them has some pretty major flaws.

There seems to be limited awareness and very little acknowledgment or discussion of these two different approaches, which is a problem. Marketers know what they want — accurate and meaningful attribution of credit to each of their channels and initiatives — but the conversation gets murky in a hurry when it comes to how to best achieve that goal.

Basic Attribution Is Flawed

Basic attribution is the approach that web analytics vendors tend to promote as “attribution management.” It’s more about cross-session visitor tracking — tracking which channels drove a visitor to the web site over time, and then enabling the marketer or analyst to choose how they want to distribute the credit among those channels (e.g., 40% to the first touchpoint, 40% to the last touchpoint, and the remaining 20% distributed between all touches that occurred in between). This is attribution, but it’s attribution with a couple of non-trivial shortcomings:

  1. In most cases, this attribution is limited to “clickthrough” traffic — it tends to ignore the impact of impressions, because impression data is not something that is readily available within many web analytics implementations
  2. The marketer has to choose how to distribute credit among the tracked channels based on their own opinion and instincts. This is reminiscent of medieval medicine: “We believe this disease is caused by an excess of blood, so we’re going to bleed the patient. If he still dies, we either didn’t bleed him early enough, or we didn’t bleed him enough” is really not all that different from, “We believe that the last touchpoint contributes 3 times as much to the ultimate conversion as all prior touchpoints combined (a ‘last touch = 75%’ model).” In modern times, marketers don’t inadvertently kill consumers with misguided attribution, but the scenarios are similar in that they start with non-fact-based assumptions.

Both of these gaps are troubling. In some cases, various platforms — including web analytics platforms — tackle the first issue through integration with DSPs or other media-based data sources, but the second issue almost always remains.

Advanced Attribution

Advanced attribution, almost by definition, addresses the first issue above. Advanced attribution providers use various techniques for data capture beyond the data available when a visitor clicks through to the site. It’s the second issue, and how advanced attribution can address it, that gets really interesting. True advanced attribution removes “assumption” and “instinct” (i.e., “picking an attribution model”) as the starting point.

One approach to do this is, in a sense, multivariate testing in the absence of the ability to define a control group at the outset. As a simplistic example, think of a series of marketing touchpoints (impressions and clickthroughs) you are tracking at the user level: A, B, C, and D. If you have a group of users for whom you tracked their touches as “A –> B –> C –> D –> <conversion>,” and another group of users for whom you tracked their touches as “A –> B –> D –> <conversion>” then, without choosing how much proportional credit any individual step should get, you can assign (attribute) the value delivered by “C” in this sequence: it’s the incremental lift between the first sequence and the second.

Obviously, this has to be done for hundreds (or thousands) of touchpoint-series. But, even without fully understanding the math and modeling that goes into that, this is obviously a more robust approach to the problem.

And…It’s Not Just Technology

Shifting gears a bit, there’s another wrinkle when it comes to successfully implementing an advanced attribution management program: the technology is just one part of the equation. In this sense, attribution management is no different than web analytics, a data warehouse, or a CRM platform: if you overly rely on the technical implementation on its own to deliver results, you’re liable to stumble on multiple process, communication, and organizational hurdles along the way. With luck, you will be able to recover, but why not eliminate those hurdles altogether? It’s doable, but it requires flexing your soft skills throughout the course of the implementation: establishing realistic and clear measures of program success up front, ensuring your stakeholders understand what advanced attribution is (and isn’t), involving IT early and often to minimize last-minute technical surprises, and developing and rehearsing your process for converting results from the program into action.

I’ve been fortunate to get to partner with Adometry (now Google) to write a white paper on this very subject: “10 Tactics for Building an Effective Attribution Management Program.”

 

Analysis

The Dirtiest Word in Analytics Is Interesting

There are few sentence openers that set off analytics alarm bells for me more than:

“It would be interesting to see…”

That phrase gives me a 15-minute cardio workout without needing to get up from my chair. Really. (Well. Almost. I definitely have developed a physiological reaction to the phrase — elevated heart rate, burning sensation in my ears, etc.)

The curse of the analyst is how often we find ourselves producing results that fall into the dastardly void of “interesting but not actionable.” The real kicker is that, if we let ourselves be guided by “interesting”-based requests, and we repeatedly deliver analyses that scratch those curiosity itches, then, over time, we wind up with the worst possible feedback:

“The reports you deliver have a lot of interesting information…but they don’t have insights. They don’t make recommendations. They don’t tell me what I should do.”

In other words, the reports and analyses we’ve been delivering are exactly what was requested, but seldom produce actionable insights and, ultimately, lead to nothing more than a big, fat, depressing, “So, what?”

The root cause of this vicious and soul-sucking ritual goes back to the initial request: a lack of discipline regarding the intake of the request itself. The “So what?” test can be applied much more cheaply at the point of intake rather than waiting until a full-blown analysis has been conducted and presented!

In a perfect world, where egos do not have to be protected and where organizational pecking orders are not part of the identification and qualification of requests for analysis, it’s simple (the request below is essentially verbatim from an email, just with masked/bracketed details — the subsequent exchange is idealized and fictitious):

Executive: I woke this morning thinking about an interesting impact metric for our website.  To what extent does a blip in twitter result in a blip on our website.  For example, I suspect that something related to [cultural topic] was trending on [date].  Did we also see an increase in traffic on our website?  We could play around with the metric and use  our website analytics to see where people were going.

Analyst: That’s a pretty broad ask. Can you help me understand what we would do with that data?

Executive: We’d have a better understanding of, when a relevant topic spikes in social media, how people behave on our site.

Analyst: Okay. But, wouldn’t you expect that to change depending on the topic? And, I’m trying to envision what the results of such an analysis might look like where you would easily be able to act on the information.

Executive: Well…er…hmmm. That’s a good question. I guess it would…well…no…maybe not. Um. Wow. I really hadn’t thought this through. Now that you’re asking me to…I don’t think this would actually be all that useful. And, you would probably have to spend a lot of time to respond to the request. Thanks for probing a bit rather than just running off and spinning your wheels on my behalf!

We don’t live in a perfect world. Vague requests are going to get floated. Organizational hierarchies and relationship-building realities mean that, as analysts, we do have to chase “interesting” requests that lead nowhere. But, that doesn’t mean we shouldn’t recognize and strive to minimize how often that happens. Here’s how you can do that:

  • Condition yourself to go to full alert whenever the word “interesting” is used (as well as “understand,” as in, “I want to understand…<some sort of behavior>;” and, heck, let’s throw in, from the example above: “play around!”)
  • When you’re on full alert, probe for clarification as much as you can without being an ass — be delicate!
  • Even if you are not able to probe very much, ask yourself the “So what?” question repeatedly — frame the specific questions that you might try to answer based on the request, envision a possible answer, and then apply the litmus test: “So what? If that was the answer, what would we do differently than we’re doing now?”

These tips are all pre-analysis, but they focus your analysis and the way the results get delivered.

Analysis, General

What Marketing/Analytics Can Learn from Mythbusters

Earlier this month, I gave a presentation at the Columbus Web Group meetup that I titled Mythbusters: Analytics Edition. The more I worked on the presentation — beating the same drums and mounting the same soapboxes I’ve mounted for years — the more I realized that the Discovery Channel show is actually a pretty useful analog for effective digital analytics. And, since I’m always on the lookout for new and better ways to talk to analysts and marketers about how to break out of the soul-sucking and money-wasting approaches that businesses have developed for barfing data and gnashing teeth about the dearth of “actionable insights,” this one seemed worth trying to write down.

Note: If you’re not familiar with the show…you can just bail on this post now. It’s written with the assumption that the reader actually knows the basic structure and format of the program.

Mythbusters - Analytics Edition

First, a Mythbusters Episode Produced by a Typical Business

When I do a thought experiment of putting an all-too-typical digital marketer and their analytics team in charge of producing a Mythbusters episode, here’s what happens:

mythbusters_jamie_armorThe show’s opening credits roll. Jamie and Adam stand in their workshop and survey the tools they have: welding equipment, explosives, old cars, Buster, ruggedized laptops, high-speed cameras, heavy ropes and chain, sheet metal, plexiglass, remote control triggers, and so on. They chat about which ones seem would be the most fun to do stuff with, and then they head their separate ways to build something cool and interesting.

[Commercial break]

Jamie and Adam are now out in a big open space. They have a crane with an old car suspended above it. They have an explosive device constructed with dynamite, wire, and a bunch of welded metal. They have a pole near the apparatus with measurements marked on it. They have a makeshift bomb shelter. They have high-speed cameras pointed at the whole apparatus. They get behind the bomb shelter, trigger the crane to drop the car and, right as it lands on the explosive device, the device goes off and blows the car up into the air.

[Commercial break]

Jamie and Adam are now reviewing the footage of the whole exercise. They play and replay videos in slow motion from different angles. They freeze-frame the video at the peak of the old car’s trajectory and note how high it went. Then, the following dialogue ensues:

Adam: “That was soooooo cool.”

Jamie: “Yeah. It was. What did we learn?”

Adam: “Well, the car was raised 7’2″ into the air.”

Jamie: “Right. So, how are we going to judge this myth? Busted, plausible, or confirmed?”

Adam: “Um… what was the myth we were trying to bust?”

Jamie: “Oh. I guess we didn’t actually identify one. We just came up with something cool and did it.”

Adam: “And we measured it!”

Jamie: “That’s right! We measured it! So… busted, plausible, or confirmed?!”

Adam: “Hmmm. I don’t know. I don’t think how high the car went really tells us anything. How loud do you think the explosion was?”

Jamie: “It was pretty loud. Did we measure the sound?”

Adam: “No. We probably should have done that. But…man…that was a bright flash when it blew up! I had to shield my eyes!”

Jamie: “Aha! We have software that will measure the brightness of the flashes from the video footage! Let’s do that!”

[They measure the brightness.]

Adam: “Wow. That’s pretty bright.”

Jamie: “Yeah. So, have we now done enough analysis to call the myth busted, plausible, or confirmed?”

Adam: “Well…we still don’t know what ‘it’ is. What’s the myth?”

Jamie: “Oh, yeah. I forgot about that.” [turns to the camera] “Well, we’re about out of time. We’ll be back next week! You know the format, folks! We’ll do this again next week — although we’ll come up with something else we think is cool to build and blow up. Hopefully, we’ll be able to make a busted, plausible, or confirmed call on that episode!”

[Credits roll]

This is how we’ve somehow managed to train ourselves to treat digital analytics!!!

We produce weekly or monthly reports and expect them to include “analysis and insights.” Yet, like the wrongheaded Mythbusters thought experiment above, we don’t actually ask questions that we want answered.

Sure, We Can Find Stuff Just by Looking

Keeping with the Mythbusters theme and, actually, lifting a slide straight from the presentation I did, what happens — in reality — when we simply point a web analyst to the web analytics platform and tell them to do some analysis and provide some insights for a monthly report? Poking around, clicking into reports, correlating data, even automatically detecting anomalies, we can turn up all sorts of things that don’t help the marketer one whit:

mythbusters_anomaly

To be clear, the marketer (Jamie) is complicit here. He is the one who expects the analyst to simply dig into the data and “find insights.” But, week in and week out, month in and month out, he gets the report, the report includes “analysis” of the anomalies in the data and other scattershot true-but-not-immediately-relevant findings, but he doesn’t get information that he can immediately and directly act on. (At which point we invoke Einstein’s definition of insanity: “doing the same thing over and over again and expecting different results.”)

“Insights” that are found this way, more often than not, have a perfectly logical and non-actionable explanation. This is what analysis becomes when the analyst is told to simply dig into the data and produce a monthly report with “analysis and insights.”

The Real Mythbusters Actually Gets It Right

Let’s look at how the real Mythbusters show runs:

  1. A well-known (or obscure) belief, urban legend, or myth is identified.
  2. The Mythbusters team develops a plan for testing that myth in a safe, yet scientifically valid, way.
  3. They experiment/construct/iterate as they implement the plan.
  4. They conclude with a one-word and unequivocal assessment of the result: “Confirmed,” “Plausible,” or “Busted.”

Granted, the myths they’re testing aren’t ones that lead to future action (just because they demonstrate that a lawn chair with a person on it can be lifted by balloons if you tie enough of them on doesn’t mean they’re going to start promoting a new form of air travel). But, aside from that, the structure of their approach is exactly where marketers could get the most value. It is nothing more and nothing less than a basic application of the scientific method.

Sadly, it’s not an approach that marketers intuitively follow (they’re conditioned not to by the legacy of bloated recurring reports). And, even worse, it’s not an approach that many analysts embrace and push themselves.

Outlining those same exact steps, but in marketing analytics terms:

  1. A marketer has an idea about some aspect of their site that, if they’re right, would lead them to make a change. (This is a hypothesis, but without the fancy label.)
  2. The analyst assesses the idea and figures out the best option for testing it, either through digging into historical web analytics or voice of the customer data or by conducting an A/B test.
  3. The analyst does the analysis or conducts the test
  4. The analyst clearly and concisely communicates the results of the analysis back to the marketer, who then takes action (or doesn’t, as appropriate)

So clear. So obvious. Yet…so NOT the mainstream reality that I see. I have a lot of theories as to why this is, and it’s becoming a personal mission to change that reality. Are you on board to help? It will be the most mundane revolution, ever…but, who knows? Maybe we’ll at least come up with a cool T-shirt.

Jamie armor photo courtesy of TenSafeFrogs

Analysis

An Aspirational Report Structure

The life of an analyst invariably includes responsibility for some set of recurring reports: daily, weekly, monthly, and even quarterly. I hate reports. Or, to be more precise, I have come to hate the term “report.” (I’ve also developed something of an aversion to the term “analysis,” but that’s a topic for another day.)

To make a bold claim: corporate cultures force employees to develop cognitive dissonance when it comes to recurring reports:

  • We believe we need to get them, that they’re supposed to be lengthy, that emailing them as PowerPoint presentations makes sense, and that we’re supposed to use them to drive the business forward
  • We often don’t actually look at them when they arrive in our inboxes, and, when we do, we simply scroll through each slide, look at it long enough to know what it’s saying, and then close the file and get on with our lives

It’s cognitive dissonance because we believe we need something…even though, in practice, it’s not something we get much value out of.

Occasionally, this cognitive dissonance actually bubbles up into our consciousness — unrecognized for what it is — and results in a conversation with the analyst or the analytics manager that is, essentially, a single statement:

“I get this report each month, and it has the data I asked for in it, but it doesn’t include any meaningful insights or recommendations that actually help me drive the business.”

I’ve seen or heard a statement to this effect more times than I can count. I’ve also seen the typical reaction: the report gets longer! Analysts are pleasers by nature. If we hear “the report isn’t working,” and that’s articulated as a, “it’s missing something” (insights, recommendations) critique, then the “obvious” way to fix it is to “add the missing stuff.”

In my mind, I look like this when this happens (apologies to any kiddos reading this who don’t get the Susan Powter reference):

susan-powter

Performance Measurement vs. Hypothesis Validation

I’ve been beating a fairly persistent drum over the past few years that is a pretty simple rhythm. But, it’s also an, apparently, somewhat radical departure from long-established corporate norms that have not evolved, even though the tools we have for analytics have advanced dramatically.

This tune I’ve been pounding out has an easy, if not particularly melodic, claim as its chorus: there are fundamentally two (and only two) ways that analytics can actually be used to drive a business forward: performance measurement and hypothesis validation:

Two Uses of Data

There’s a temptation to look at the above image and make the leap that “performance measurement” is “reporting” and that “hypothesis validation” is “analysis.” I’d be fine with that leap…if that’s how most businesspeople actually defined the words. And they don’t.

Consider the following requests or statements made every day in businesses around the world:

  • “Be sure to include insights and recommendations in the monthly report!”
  • “The report must include analysis of what happened.”
  • “Can you analyze the results of the recent campaign and tell me if it was successful?”
  • “I need a weekly insights report for the web site.”

Ick. Ugh. And phooey! These all represent a conflation of performance measurement and hypothesis validation. I hate the fact that both “hypothesis” and “validation” are four-syllable, fancy-pants-sounding words, but it’s hard to argue that they’re not more descriptive and precise than “analysis.”

My Dreamy-Dream State of Information Delivery

Let me wave my magic wand to how I believe analytics stuff should be delivered. It’s pretty simple:

  • The only recurring reports are: single-screen dashboards with KPIs (with targets) organized under clearly worded business goals. They are automated or semi-automated and are pushed out to stakeholders with minimal latency on whatever schedule makes sense. If it’s a weekly report that runs Sunday through Saturday, and the recipient actively uses email, it hits their inbox (embedded in the email body — NOT as an attachment) on Sunday morning at 12:01 AM. It’s got clear indicators for each KPI as to whether the metric is on track or not. It does not include any qualitative commentary, insights, or recommendations (see my last post).
  • Additional information is delivered in an ad hoc fashion and is driven entirely by what hypotheses were prioritized for validation, when they were validated, and who actually cares about the results. The core “results” are delivered on no more than five slides or pages (and, ideally, fit on 1 or 2). When appropriate, there is a supplemental appendix, a separate detailed writeup that goes into the methodology and detailed results, and/or a reasonably cleanly formatted spreadsheet with the deeper data.

“OMG, Tim! Are you actually saying that the only regularly scheduled reporting we should be doing are 1-page dashboards? And, are you saying that we should never be presenting more than 5 slides of information when we present analysis results?”

Yup. That’s my dreamy dream world.

But, sadly, although common sense says that makes sense — both from saving analysts time and preventing narcolepsy-by-PowerPoint — that is a too-radical shift for most companies.

So…

A Recurring Report Structure that Deviously Moves in This Direction

Let’s get a little pragmatic. Harry Potter had less trouble taking out Voldemort than he would have had if he’d tried to drastically curtail the volume and length of corporate monthly reports.

Drowning in Paper

Photo by Christian Guthier

So, let’s accept that monthly reports are here to stay for the foreseeable future. How can we make them better? Well, we have to combine performance measurement and hypothesis validation. So, first, I still say we do a low-latency, one-page, automated or quasi-automated distribution of the performance measurement dashboard. The intro to that email is pretty simple:

“Below is the performance measurement dashboard for X for <timeframe>. We are in the process of developing the full report. Let us know if you have any specific questions or comments based on the dashboard itself.”

dashboardexample

That’s it. It’s in the stakeholders’ inboxes on the first day of the month. Everyone knows at a glance if anything has gone awry performance-wise. If something in that limited set of data is unexpected, the recipients are asked to quickly respond. In many cases, someone on the distribution list will already know what the root cause is, (“Oh. Yeah. We turned off paid search 3 weeks into the month. We forgot to let everyone know we’d done that.”) In other cases, a KPI’s miss of its target will be cause for immediate concern, and the root cause is not immediately known. That’s good! Better to have the alarm raised on Day 1 rather than on Day 9 when a “full report” is finally made available! The more people hypothesizing early about the root cause, the better!

So, then, what does the “full report” that goes out several days or a week later look like? Like this:

  1. Title slide
  2. Dashboard — the exact same one that was sent out on the first of the month
  3. A slide listing all of the questions (hypotheses — in plain English!) that were tackled during the previous month and an indicator as to whether each question was definitively answered or not (some of these may have been spawned by the initial dashboard — trying to get to the root cause of a problem that manifested itself there), and an indicator if there is action that should be taken for each question. It’s a simple table.
  4. One slide for each question that was definitively answered — the title of the slide is the question. Big, bold text underneath provides the answer. Big, bold text that summarizes the recommended action (“No action warranted” is an acceptable recommended action, as long as there is a different potential answer to the question that would have led to actual action). The body of the slide is a clean and clear set of context (including a chart or two as warranted) providing the essence of why the answer is what it is.
  5. A single slide of “(Preliminary) questions to be addressed this month.” Some of the questions that were not definitively answered (from slide 3) may be included here if additional work will likely support answering them.

That’s it. It’s still pretty radical, but it can safely be labeled a “monthly report” because it has a defined structure and has multiple pages!

The last slide is “Preliminary” because, if the report is presented to stakeholders, this is the opportunity to look ahead from an analytics perspective. You have your audience’s attention (because you’ve delivered such useful information in an easy-to-consume way), and you want to now collaborate with them to make sure that you are spending your time as efficiently as possible. It quickly becomes clear that the last slide in this month’s report is the basis for the third slide in next month’s report (other questions will come in over the course of the month that will adjust the final list of questions answered), and it will help stakeholders learn that analysis isn’t merely an after-the-fact (after the end of the month) exercise — it starts by looking ahead to what the business wants to learn!

A note on one-time reports

Sometimes, a campaign runs for a short enough period of time that there is only a single “report” rather than a recurring weekly or monthly report. In these situations, the structure is still very similar. But, the “questions” in the last slide are limited to “questions to be answered through future campaigns.” Otherwise, the structure is identical.

Still too radical?

What do you think? Would this report structure work in your organization (it’s not a theoretical construct — I’ve used it with multiple clients and it’s a direction I try to subtly evolve any report where I inherit some other report structure)?

Analysis, General

Better ways to measure content engagement than time metrics

I spent five years responsible for web analytics for a major ad-monetised content site, so I’m not immune to the unique challenges of measuring a “content consumption” website. Unlike an eCommerce site (where there is a more clear “conversion event”) content sites have to struggle with how to measure nebulous concepts like “engagement.” It can be tempting to just fall back on measures like “time on site”, but these metrics have significant drawbacks. This post outlines those, as well as proposing alternatives to better measure your content site.

So … what’s wrong with relying on time metrics?

1. Most business users don’t understand what they really mean

The majority of business users, and perhaps even newer analysts, may not understand the nuance of time calculations in the typical web analytics tool.

In short, time is calculated from subtracting two time stamps. For example:

Time on Page A = (Time Stamp of Page B) – (Time Stamp of Page A)

So time on page is calculated by subtracting what time you saw the next page from what time you saw the page in question. Time on site works similarly:

Time on Site = (Time Stamp of last call) – (Time Stamp of first call)

A call is often a page view, but could be any kind of call – an event, ecommerce transaction, etc.

Can you spot the issue here? What if a user doesn’t see a Page B, or only sends one call to your web analytics tool? In short: those users do not count in time calculations.

So why does that skew your data?

Let’s take a page, or website, with a 90% bounce rate. Time metrics are only based on 10% of traffic. Aka, time metrics are based on traffic that has already self-selected as “more interested”, by virtue of the fact that they didn’t bounce!

2. They are too heavily influenced by implementation and technical factors unrelated to user behaviour

The way your web analytics solution is implemented can have a significant impact on time metrics.

Consider these two implementations and sets of behaviour:

  • I arrive on a website and click to expand a menu. This click is not tracked as event. I then leave.
  • I arrive on a website and click to expand a menu. This click is tracked as an event. I then leave.

In the first example, I only sent one call to analytics. I therefore count as a “bounce”, and my time on the website does not count in “Time on Site”. In the second example, I have two calls to analytics, one for the page view and one for the event. I no longer count as a bounce, and my time on the website counts as “Time on Site.” My behaviour is the same, but the website’s time metrics are different.

You have to truly understand your implementation, and the impact of changes made to it, before you can use time metrics.

However, it’s not even just your site’s implementation that can affect time metrics. Tabbed browsing – default behaviour for most browsers these days – can skew time, since a user who keeps a tab open will keep “ticking” until the session times out in 30 mins.

Even the time of day your customers choose to browse can also impact time on site, as many web analytics tools end visits automatically at midnight. This isn’t a problem for all demographics, but perhaps the TechCrunches and the Mashables of the world see a bigger impact due to “night owls”!

3. They are misleading

It’s easy to erroneously determine ‘good’ and ‘bad’ based on time on site. However, I may spend a lot of time on a website because I’m really interested in the content, but I can also spend a lot of time on a website because the navigation is terrible and I can’t find what I need. There is nothing about a time metric that tells you if the time spent was successful, yet companies too often consider “more time” to indicate a successful visit. Consider a support site: a short time spent on site, where the user immediately got the help they needed and left, is an incredibly successful visit, but this wouldn’t be reflected by relying on time measures.

So what should you use instead?

Rather than relying on “passive” measures to understand engagement with your website, consider how you can measure engagement via “active” measures: aka, measuring the user’s actions instead of time passing.

Some examples of “active” measures on a content site:

  • Content page views per visit. A lot of my concerns about regarding time measures also apply to “page views per visit” as a measure. (Did I consume lots of page views because I’m interested, or because I couldn’t find what I was looking for?) For a better “page views per visit” measure of engagement, track content page views, and calculate consumption of those per visit. This would therefore exclude navigational and more “administrative” pages and reflect actual content consumption. You can also track what percentage of your traffic actually sees a true content page, vs. just navigational pages.
  • Ad revenue per visit. While this is less a measure of “engagement”, businesses do like to get paid, so this is definitely an important measure for most content sites! It can often be difficult to measure via your analytics tool, since you need to not only take in to account the page views, but what kind of ad the user saw, whether the space was sold or not and what the CPM was. However, it’s okay to use informed estimates. For example:Click-through rate to other articles. A lot of websites will include links to “related articles” or “you also might be interested in….” Track clicks to these links and measure click rate. This will tell you that users not only read an article, but were interested enough to click to read another.
    • I saw 2 financial articles during my visit. We sell financial article pages at an average $10CPM and have an estimated 80% sell through rate. My visit is therefore worth 2/1000*$10*80% = 1.6 cents. This can be a much more helpful measure than “page views per visit” since not all page views are created equal. Having insight in to content consumed and its value can help drive decisions like what to promote or share.
  • Number of shares or share rate. If sharing is considered important to your business, clearly highlight this call to action, and measure whether users share content, and what they share. Sharing is a much stronger indicator of engagement than simply viewing. (You won’t be able to track all shares, for example, copy-and-pasting URLs won’t be tracked, but tracking shares will still give you valuable information about content sharing trends.)
  • Download rate. For example, downloading PDFs.
  • Poll participation rate or other engaging activities.
  • Video Play rate. Even better, track completion rate and drop-off points.
  • Sign up and/or Follow on social.
  • Account creation and sign in.

If you’re already doing a lot of the above, consider taking it a step further and calculating visit scores. For example, you may decide that each view of a content article is 1 point, a share is 5 points, a video start is 2 points and a video complete is 3 points. This allows you to calculate a total visit score, and analyse your traffic by “high” vs “low” scoring visitors. What sources bring high scoring visitors to the site? What content topics do they view more? This is more helpful than “1:32min time on site”!

By using these active measures of user behaviour, you will get better insight than through passive measures like time, which will enable better content optimisation and monetisation.

Is there anything else you would add to the list? What key measures do you use to understand content consumption and behaviour?

Analysis, Reporting

Why I Don’t Put Recommendations on Dashboards

WARNING: Gilligan contrarianism alert! The following post posits a thesis that runs contrary to popular opinion in the analytics community.

Many companies these days rely on some form of internal dashboard(s). That’s a good thing. Even better is when those companies have actually automated these dashboards – pulling data from multiple data sources, structuring it in a way that directly maps to business objectives, and delivering the information in a clean, easy-to-digest format. That’s nirvana.

dashboard

Reality, often, is that the dashboards can only be partially automated. They wind up being something an analyst needs to at least lightly touch to bridge inevitable API gaps before delivering them on some sort of recurring schedule: through email, through an intranet, or even in person in a regularly scheduled meeting.

So, what is the purpose of these dashboards? Here’s where a lack of clarity — clearly communicated — becomes a slippery slope faster than Miley Cyrus can trigger a TV viewer’s gag reflex. Dashboards are, first and foremost, performance measurement tools. They are a mechanism for quickly (at a glance!) answering a single question:

“Are we achieving the goals we set out to achieve?”

They can provide some minimal context around performance, but everything beyond answering that question is a distant second purpose-wise.

It’s easy enough to wax sophomoric on this. It doesn’t change the fact, though, that one of the top complaints dashboard-delivering analysts hear is: “I get the [weekly/monthly/quarterly] dashboard from the analyst, but it doesn’t have recommendations on it. It’s just data!”

I get it. And, my response? When that complaint is leveled, it’s a failure on the part of the analyst to educate (communicate), and a failure of process — a failure to have mechanisms in place to deliver actionable analytical results in a timely and effective manner.

But…here…I’m just going to lay out the various reasons that dashboards are not the place to expect to deliver recommendations, because, in my experience, analysts hear that complaint and respond by trying to introduce recommendations to their dashboards. Why shouldn’t they? I can give four reasons!

Reason No. 1: Dashboards Can’t Wait

Another complaint analysts often hear is that dashboards aren’t delivered quickly enough at the end of the reporting period. Well, no one, as far as I know, has found a way to stop time. It marches on inexorably, with every second taking exactly one second, every minute having a duration of 60 seconds, and every hour having a duration of 60 minutes (crappy Adam Sandler movies — pardon the adjectival redundancy — notwithstanding).

timeflies
Source: aussiegal

Given that, let’s step back and plot out a timeline for what it takes in an “insights and recommendations delivered with the dashboard” scenario for a dashboard that gets delivered monthly:

  1. Pull data (can’t happen until the end of the previous month)
  2. Consolidate data to get it into the dashboard
  3. Review the data — look at KPIs that missed targets and supporting metrics that moved unexpectedly
  4. Dig in to do analysis to try to figure out why those anomalies appeared
  5. IF the root cause is determined, assess whether this is something that needs “fixing” and posit ways that it might be fixable
  6. Summarize the results — the explanation for why those anomalies appeared and what might be done to remedy them going forward (if the root cause was something that requires a near-term change)
  7. Add the results to the dashboard
  8. Deliver the dashboard
  9. [Recipient] Review the dashboard and the results
  10. [Recipient] Decide whether to take action
  11. [Recipient] If action will be taken, then take the action

Seems like a long list, right? I didn’t write it trying to split out separate steps and make it needlessly long. What’s interesting is that steps 1 and 2 can (and should!) be shortened through automation. Aside from systems that are delayed in making their data available, there is no reason that steps 1 and 2 can’t be done within hours (or a day) of the end of the reporting period.

Steps 3 through 7, though, are time-consuming. And, often, they require conversations and discussion — not to mention time to actually conduct analysis. Despite vendor-perpetuated myths that “the tool” can generate recommendations… tools really suck at doing so (outside of highly operationalized processes).

Here’s the other kicker, though: steps 9 through 11 take time, too! So, realistically, let’s say that steps 1 and 2 take a day, steps 3 through 8 take a week, steps 9 and 10 takes 3 days (because the recipient doesn’t drop everything to review the dashboard when it arrives), and then step 11 takes a week (because “action” actually requires marshalling resources and getting something done). That means — best case — we’re 2.5 weeks into the month before action gets taken.

So, what happens at the end of the month? The process repeats, but there was only 1.5 weeks of the change actually being in place… which could easily get dwarfed by the 2.5 weeks of the status quo!!!

Let’s look at how a “dashboard without insights” process can work:

  1. Pull data (can’t happen until the end of the previous month)
  2. Consolidate data to get it into the dashboard
  3. Deliver the dashboard (possibly calling out any anomalies or missed targets)
  4. [Recipient] Review the dashboard and hones in on anything that looks troubling that she cannot immediately explain (more on that in the next section)
  5. The analyst and the recipient identify what, if any, trouble spots require deeper analysis and jointly develop actionable hypotheses to dig in
  6. The analyst conducts a very focused analysis (or, in some cases, proposes an A/B test) and delivers the results.
  7. [Recipient] If action is warranted, takes action

Time doesn’t stop for this process, either. But, it gets the information into the business’s hand inside of 2 days. The analyst doesn’t waste time discovering root causes that the business owner already knows (see the next section). The analysis that gets conducted is focused and actionable, and the business owner is already primed to take action, because she participated in determining what analyses made the most sense.

Reason No. 2: Analysts Aren’t Omniscient

I alluded to this twice in the prior paragraph. Let’s look at a generic and simplistic (but based on oft-observed real-world experience) example:

  1. The analyst compiles the dashboard and sees that traffic is down
  2. The analyst digs into the traffic sources and sees that paid search traffic is down dramatically
  3. The analyst digs in further and sees that paid search traffic went to zero on the 14th of the month and stayed there
  4. The analyst fires off an urgent email to the business that paid search traffic went to zero mid-month and that something must be wrong with the site’s SEM!
  5. The business responds that SEM was halted mid-month due to budget adjustments, and they’ve been meaning to ask what impact that has had

What’s wrong with this picture? Steps 2 through 4 are largely wasted time and effort! There is very real analysis to be done… but it doesn’t come until step 5, when the business provides some context and is ready for a discussion.

This happens all the time. It’s one of the reasons that it is imperative that analysts build strong relationships with their marketing stakeholders, and one of the reasons that a sign of a strong analytics organization is one where members of the team are embedded – literally or virtually – in the teams they support.

But, even with a strong relationship, co-location with the supported team, regular attendance at the team’s recurring meetings, and a spot on the team’s email distribution list, analysts are seldom aware of every activity that might result in an explainable anomaly in the results delivered in a dashboard.

This gets to a data source that gets ignored all too often: the minds and memories of the marketing team. There is nothing at all wrong with an analyst making the statement: “Something unexpected happened here, and, after I did some cursory digging, I’m not sure why. Do you have any ideas as to what might have caused this?” There are three possible responses from the marketer who is asked this question:

  • “I know exactly what’s going on. It’s almost certainly the result of X.”
  • “I’m not sure what might have caused that, but it’s something that we should get to the bottom of. Can you do some more digging to see if you can figure it out?”
  • “I’m not sure what might have caused that, but I don’t really care, either. It’s not important.”

These are quick answers to an easy question that can direct the analyst’s next steps. And, two of the three possible answers lead to a next step of moving onto a value-adding analysis — not pursuing a root cause that will lead to no action! Powerful stuff!

Reason No. 3: Insights Don’t have a Predictable and Consistent Length

I see it all the time: a standard dashboard format that, appropriately, has a consistent set of KPIs and supporting metrics carefully laid out in a very tightly designed structure. Somewhere in that design is a small box – at the top of the dashboard, at the bottom right of the dashboard, somewhere – that has room for a handful of bullet points or a short paragraph. This  area of the dashboard often has an ambitious heading: “Insights,” “Recommendations,” “Executive Summary.”

The idea – conceived either on a whiteboard with the initial design of the dashboard, or, more likely, added the first time the dashboard was produced – is that this is where the analysts real value will be manifested. THIS is where the analyst will place the Golden Nuggets of Wisdom that have been gleaned from the data.

Here’s the problem: some of these nuggets are a flake of dust, and some are full-on gold bars. Expecting insights to fit into a consistent, finite space week in and week out or month in and month out is naïve. Sometimes, the analyst has half a tweet’s worth of prose-worthy material to include, which makes for a largely empty box, leaving the analyst and the recipient to wonder if the analyst is slacking. At other times, the analyst has a handful of useful nuggets to impart…but then has to figure out how to distill a WordPress-sized set of information into a few tweet-sized statements.

Now, if you buy into my first two reasons as to why recommendations shouldn’t be included with the dashboard in the first place, then this whole section becomes moot. But, if not — if you or your stakeholders still insist that performance measurement include recommendations — then don’t constrain the space to include that information to a fixed box on the dashboard.

Reason No. 4: Insights Can’t Be Scheduled

A scene from The Marketer and the Analyst (it’s a gripping — if entirely fictitious — play):

Marketer: “This monthly dashboard is good. It’s showing me how we’re doing. But, it doesn’t include any insights based on the performance for the month. I need insights to take action!”

Analyst: “Well, what did you do differently this month from previous months?”

Marketer: “What do you mean?”

Analyst: “Did you make any changes to the site?”

Marketer: “Not really.”

Analyst: “Did you change your SEM investment or strategy?”

Marketer: “No.”

Analyst: “Did you launch any new campaigns?”

Marketer: “No.”

Analyst: “Were there any specific questions you were trying to answer about the site this month?”

Marketer: “No.”

Analyst: ???!

Raise your hand if this approximates an exchange you’ve had. It’s symptomatic of a completely ass-backward perception of analytics: that the data is a vast reserve of dirt and rock with various veins of golden insights threaded throughout. And, that the analyst merely needs to find one or more of those veins, tap into it, and then produce a monthly basket of new and valuable ingots from the effort.

The fact is, insights come from analyses, and analyses come from hypotheses. Some analyses are small and quick. Some are large and require gathering data – through an A/B or multivariate test, for instance, or through a new custom question on a site survey. Confusing “regularly scheduled performance measurement” with “hypothesis-driven analysis” has become the norm, and that is a mistake.

While it is absolutely fine to measure the volume and value of analyses completed, it is a recipe for failure to expect a fixed number of insights to be driven from and delivered with a scheduled dashboard.

A Final Word: Dashboards vs. Reports

Throughout this post, I’ve discussed “dashboards.” I’ve steered clear of the word “report,” because it’s a word that has become pretty ambiguous. Should a report include insights? It depends on how you define a report:

  • If the report is the means by which, on a regularly scheduled basis, the performance of a [site/campaign/channel/initiative] is performing, then my answer is: “No.” Reasons 1, 2, and 4 explain why.
  • If the report is the term used to deliver the results of a hypothesis-driven analysis (or set of hypothesis-driven analyses), then my answer is, “Perhaps.” But…why not call it “Analysis Results” to remove the ambiguity in what it is?
  • If the report is intended to be a combination of both of the above, then you will likely be delivering a 25+ deck of rambling slides that — despite your adoration for the content within — is going to struggle to hold your audience’s attention and is going to do a poor job of both measuring performance and of delivering clearly actionable analysis results.

We live in a real-time world. Consumers — all marketers have come to accept — have short attention spans and consume content in bite-sized chunks. An effective analyst delivers information that is super-timely and is easily digestible.

So. Please. Don’t spend 3 weeks developing insights and recommendations to include on a 20-page document labeled “dashboard.”

Analysis, Reporting

Analytics Aphorisms — Gilligan-Style

Last week, I had the pleasure of presenting at a SEER Interactive conference titled “Marketing Analytics: Proving and Improving Online Performance.” The conference was at SEER’s main office, which is an old church (“old” as in “built in 1850” and “church” as in “yes…a church”) in the Northern Liberties part of Philadelphia. The space itself is, possibly, the most unique that I’ve presented in to date (photo courtesy of @mgcandelori — click to view the larger version…that’s real stained glass!):

IMG_0965

As luck would have it, Michele Kiss attended the conference, which meant all of the speakers got a pretty nice set of 140-character notes on the highlights of what they’d said.

Reviewing the stream of tweets afterwards, I realized I’ve developed quite a list of aphorisms that I tend to employ time and again in analytics-oriented conversations. I’m sufficiently self-aware that I’ll often preface them with, “So, this is soapbox #23,” but, perhaps, not self-aware enough to not actually spout them!

The occasion of standing on an actual altar (SEER maintained much of the the space’s original layout) seemed like a good time to put together a partial catalog of my go-to one-liners. Enjoy!

Being data-driven requires People AND Process AND Technology

I beat this drum fairly often. It’s not enough to have a pristine and robust technology stack. Nor is it sufficient to have great data platforms and great analysts. I believe — firmly — that successful companies have to have a solid analytics process, too. Otherwise, those wildly-in-demand analysts sifting through exponentially growing volumes of data don’t have a prayer. Effective analytics has to be efficient analytics, and efficiency comes from a properly managed process for identifying what to test and analyze.

peopleprocesstech

Identifying KPIs is nothing more than answering two question: 1) What are we trying to achieve? and 2) How will we know if we’ve done that?

I’ve got to credit former colleague Matt Coen for the clarity of these. I like to think I’ve done a little more than just brand them “the two magic questions,” but it’s possible that I haven’t! The point here is that business-speak is, possibly, more vile than Newspeak. “Goals” vs. “strategies” vs. “objectives” vs. “tactics” — these are all words that different people define in different ways. I actually have witnessed — on multiple occasions — debates between smart people as to whether something is an “objective” or a “strategy.”

As soon as we use the phrase “key performance indicator,” the acronym “KPI,” or the phrase “success measure,” we’re asking for trouble. So, whether I verbally articulate the two questions above, or whether I simply ask them of myself and try to answer them, I try to avoid business-speak:

  1. What are we trying to achieve? Answer that question without data. It’s nothing more than the elevator pitch for an executive who, while making conversation while traveling from the ground floor to the 8th floor, asks, “What’s the purpose of <insert whatever you’re working on>?”
  2. How will we know if we’ve done that? This question sometimes gets asked…but it skips the first question and invites spouting of a lengthy list of data points. As a plain English follow-on to the first question, though, it invites focus!

The K in KPI stands for “Key” — not for 1,000.

This one is a newer one for me, but I’ll be using it for a lonnnng time. All too often, “KPI” gets treated as a fancy-pants way to say “data” or “metrics” or “measures.” Sure, we feel like we’re business sophisticates when we can use fancy language…but that doesn’t mean that we should be using fancy language poorly! I covered this one in more detail in my last post…but I’m going to repeat the picture I used there, anyway, because it cracks me up:

Barfing Metrics

 

A KPI is not a KPI if it doesn’t have a target.

“Visits” is not a KPI. Nor is “conversion rate.” Or “customer satisfaction.” A KPI is not a KPI without a target. Setting targets is an inexact science and is often an uncomfortable exercise. But…it’s not as hard as it often gets made out to be.

Human nature is to think, “If I set a target and I miss it…then I will be viewed as having FAILED!” In reality, that’s the wrong view of targets. If you work for a manager, a company, or a client where that is the de facto response…then you need to find a new job.

Targets set up a clear and objective way to: 1) ensure alignment on expectations at the outset of an effort, and 2) objectively determine whether you were able to meet those expectations. If you wildly miss a very-hard-to-set target, then you will have learned a lot more and will be better equipped to set expectations (targets) the next time.

This all leads into another of my favorites…

You’re never more objective about what you *might* accomplish than before you set out to achieve it.

“I have no idea and no expectations!” is almost always an unintentional lie. Somebody decided that time and energy would be spent on the campaign/channel/initiative/project. That means there was some expectation that it would be worthwhile to do so. And “worthwhile” means somethingIt’s really, really hard to, at the end of a 6-month redesign where lots of people pulled lots of long hours to hit the launch date, stand up and say, “This didn’t do as well as we’d hoped.” In the absence of targets, that never happens. The business owner or project manager or analyst automatically starts looking for ways to illustrate to the project team and to the budget owner that the effort paid off.

But, that’s short-sighted. For starters, without a target set up front, just about any reporting of success will carry with it a whiff of disingenuousness (“You’re telling me that’s good…but this is the first I’m hearing that we knew what ‘good’ would look like!”). And, the after-the-fact-looking-for-success means effort is spent looking backwards rather than minimal effort to look backwards so that real effort can go into looking forward: “Based on how we performed against our expectations (targets), what should we do next, and what do we expect to achieve?”

Any meaningful analysis is based on a hypothesis or set of hypotheses.

I’ve had the debate many times over the years as to whether there are cases where “data mining” means “just poking around in the data to see what patterns emerge.” In some cases, the person is truly misguided and believes that, with a sufficiently large data set and sufficiently powerful analytical tools, that is truly all that is needed: data + tools = patterns –> actionable insights. That’s just wrongheaded.

More often, though, the debate is an illustration that a lot of analysts don’t realize that, in reality, they and their stakeholders are actually testing hypotheses. Great analysts may subconsciously be doing that…but it’s happening. The more we recognize that that’s what we’re doing, the more focused and efficient we can be with our analyses!

Actionable hypotheses come from filling in the blanks on two questions: 1) I believe _______, and 2) If I’m right, I will ______.

Having railed against fancy business-speak…it’s really not all that cool of me to be floating the word “hypothesis” now, is it? In day-to-day practice…I don’t! Rather, I try to complete these two statements (in order) before diving into any hypothesis:

  1. I believe [some idea]  — this actually is the hypothesis. A hypothesis is an assumption, an idea, a guess, a hunch, or a belief. Note that this isn’t “I know…” and it’s not even “I strongly believe…” It’s the lowest level of conviction possible, so we should be fine learning (quickly, and with data) when the belief is untrue!
  2. If I am right, I will [take some action] — this isn’t actually part of the hypothesis. Rather, it’s a way to qualify the hypothesis by ensuring that it is sufficiently focused that, if the belief holds up, action could be taken. In my experience, taking one broad belief and breaking it down into multiple focused hypotheses leads to much more efficient and actionable analysis.

Like the two magic questions, I don’t necessarily force my clients to use this terminology. I’ll certainly introduce it when the opportunity arises, but, as an analyst, I always try to put requests into this structure. It helps not only focus the analysis (and, often, promote some probing and clarification before I dig into the time-intensive work of pulling and analyzing the data), but focus the output of the analysis in a way that makes it more actionable.

Do you have favorite analytics aphorisms?

I’d love to grow my list of meaningful analytics one-liners. Do you have any you use or have heard that you really like?

Analysis

Avoiding Analytics Data-Wandering

It’s something of a given that any efficient and meaningful analysis will be driven by a clear hypothesis or set of hypotheses. Yet…it’s rare for analysts or marketers to actually think or speak in terms of hypotheses. That’s a problem, in my mind. It leads to gross inefficiency on the part of the analyst (casting about semi-aimlessly in various analytics platforms), and it leads to often non-actionable results (because the results answer questions that can’t lead to action).

Having said that, the word “hypothesis” itself can be intimidating. And, the fact is that, for marketers, we not only need to be working with clearly articulated hypotheses, but we need to be sure that the results of validating those hypotheses will actually be actionable!

To avoid intimidating language while also ensuring actionability, I’ve started using a pretty simple two-sentence construct when it comes to approaching almost any analysis:

  1. I believe… [some idea about the site or channel]
  2. If I am right, we will… [take some specific action]

The first statement is nothing more than the articulation of a hypothesis. The second statement ensures that the hypothesis is sufficiently specific to be validated, and it ensures that there is the possibility of taking real action based on the results of the analysis.

I wrote up some additional thoughts on this approach in a recent article on Practical eCommerce, and it is a core part of the presentation I will be giving at eMetrics in Boston in early October.

Analysis, Featured

Sequential Segmentation in Adobe Discover

The segment builder for Adobe Discover had some great features added during last week’s release. To celebrate, I thought I would put out a short video explaining the new layout and sequential segmentation. I was planning on a video that would be just a few minutes in length but it turned into a half hour mini-training! I split it up into the three videos below so you can bite off a piece at a time. Hopefully these basic examples give you a good start. I will likely add more examples depending on the response to these videos and the questions I get. If you want more comprehensive training feel free to contact us to take advantage of our full Discover, SiteCatalyst, ReportBuilder, and Testing training courses. If you are attending our ACCELERATE conference in September you can also take a Discover class or one of our other great classes.

Part 1 – Intro to the New Discover Segment Builder

Part 2 – Sequential Segmentation

Part 3 – Sequential Segmentation with Time Intervals

 

Lastly, here is the example data used in the videos for your reference:

Analysis, Analytics Strategy

Some (Practical) eCommerce Google Analytics Tips

A short, partially self-promotional post — two links and one, “Hey…look out for…” note about sampling.

Post No. 1: Feras Alhlou’s 3 Key GA Reports

Feras Alhlou of E-Nor recently wrote an article for Practical eCommerce that describes three Google Analytics reports with which he recommends eCommerce site owners become familiar. The third one in his list — Funnel Segments — is particularly intriguing (breaking down your funnels by Medium).

Post No. 2: (Log Rolling) 5 Custom Events for eCommerce Sites

I also recently published a Practical eCommerce article with some handy (I claim) tips for eCommerce site owners running Google Analytics that describes five of my favorite custom events for eCommerce sites.

“Hey…look out for…sampling (with conversion rates)”

Sampling in Google Analytics is one of those weird things that people either totally freak out about (especially people who currently or previously worked for the green-themed-vendor-that-has-been-red-for-a-few-years-now) or totally poo-poo as not a big deal at all. Once Google Analytics Premium came out, Google actually started talking about sampling more…because its impact diminishes with Premium.

I actually fell in the “poo-poo” camp for years. The fact was, every time I dug into a metric in a sampled report — when I jumped through hoops to get unsampled data — the result was similar enough for the difference to be immaterial. I patted myself on the back for being a sharp enough analyst to know that an appropriately chosen sample of data can provide a pretty accurate estimate of the total population.

And that’s true.

But, if you start segmenting your traffic and have segments that represent a relatively small percentage of your site’s overall traffic, and if you combine that with a metric like Ecommerce conversion rate (which is a fraction that relies on two metrics: Visits and Transactions), things can start to get pretty wonky. Ryan at Blast Analytics wrote a post that I found really helpful when I was digging into this on behalf of a client a couple of months back.

Obviously, if you’re running the free Google Analytics and you never see the yellow “your data is sampled” box, then this isn’t an issue. Even if you do see the box, you may be able to slide the sampling slider all the way to the right and get unsampled data. If that doesn’t work, you may want to pull your data using shorter timeframes to remove sampling (which throws Unique Visitors out the window as a metric you can use, of course).

Be aware of sampling! It can take a nice hunk of meat out of your tush if you blithely disregard it.

Analysis

QA: It's for Analysts, Too (and I'm not talking about tagging)

There is not an analyst on the planet with more than a couple of weeks of experience who has not delivered an analysis that is flawed due to a mistake he made in pulling or analyzing the data. I’m not talking about messy or incomplete data. I’m talking about that sinking feeling when, following your delivery of analysis results, someone-somewhere-somehow points out that you made a mistake.

Now, it’s been a while since I experienced that feeling for something I had produced. <Hold on for a second while I find a piece of wood to knock on… Okay. I’m back.> I think that’s because it’s an ugly enough feeling that I’ve developed techniques to minimize the chance that I experience it!

As a blogger…I now feel compelled to write those down.

I get it. There is a strong urge to skip QA’ing your analysis!

No one truly enjoys quality assurance work. Just look at the number of bugs that QA teams find that would have easily been caught in proper unit testing by the developer. Or, for that matter, look at the number of typos that occur in blog posts (proofreading is a form of QA).

Analysis QA isn’t sexy or exciting work (although it can be mildly stimulating), and, when under the gun to “get an answer,” it can be tempting to hasten to the finish by skipping past a step of QA, but it’s not a wise step to skip.<

I mean it.  Skipping Analysis QA is bad, bad, BAD!

9 times out of 10, QA’ing my own analysis yields “nothing” – the data I pulled and the way I crunched it holds up to a second level of scrutiny. But, that’s a “nothing” in quotes because “9 times everything checked out” is the wrong perspective. That one time in ten when I catch something pays for itself and the other nine analyses many times over.

You see, there are two costs of pushing out the results of an analysis that have errors in them:

doh

  1. It can lead to a bad business decision. And, once an analysis is presented or delivered, it is almost impossible to truly “take it back.” Especially if that (flawed) analysis represents something wonderful and exciting, or if it makes a strong case for a particular viewpoint, it will not go away. It will sit in inboxes, on shared drives, and in printouts just waiting to be erroneously presented as a truth days and weeks after the error was discovered and the analysis was retracted.
  2. It undermines the credibility of the analyst (or, even worse, the entire analytics team). It takes 20 pristine analyses* that hold up to rigorous scrutiny to recover the trust lost when a single erroneous analysis is delivered. This is fair! If the marketer makes a decision  (or advocates for a decision) based on bad data from the analyst, they wind up taking bullets on your behalf.

Analysis QA is important!

With that lengthy preamble, below are my four strategies for QA’ing my own analysis work before it goes out the door.

1. Plausibility Check

Like it or not, most analyses don’t turn up wildly surprising and dramatic insights. When they do – or, when they appear to – my immediate reaction is one of deep suspicion.

My favorite anecdote on this front goes back almost a decade, when a product marcom who had been digging into SEO and making tweaks to his product line’s main landing page, popped his head into my cubicle one day and asked me if I’d seen “what he’d done.” He’d been making minor — and appropriate — updates to his product line’s main landing page to try to improve the SEO. When he looked at a traffic report for the page, he saw a sudden and dramatic increase in visits starting one day in the middle of the prior month. He immediately took a printout of the traffic chart and told everyone he could find — including the VP of marketing — that he’d achieved a massive and dramatic success by updating some meta data and page copy!

Of course…he hadn’t.

I dug into the data and pretty quickly found that a Gomez (uptime/load time monitoring software) user agent was the source of the increased traffic. It turned out that Gomez was pitching my company’s web admins, and they’d turned on a couple of monitors to have data to show to the people in the company to whom they were pitching. (The way their monitors worked, each check of the site recorded a new visit, and none of those monitors were filtered out as bots…until I discovered the issue and updated our bots configuration.)
In other words, “Doh!!!”

That’s a dramatic example, but, to adjust the “if it seems too good to be true…” axiom:

If the data looks too surprising or too counter intuitive to be true…it probably is!

Considering the plausibility of the results is not, in and of itself, actual QA, but it’s a way to get the hairs on your back standing up to help you focus on the other QA strategies!

2. Proofread

Proofreading is tedious in writing, and it’s not much less tedious in analytics. But, it’s valuable!

looklfet

Here’s how I proofread my analyses for QA purposes:

  • I pull up each query and segment in the tool I created it in and literally walk back through what’s included.
  • I re-pull the data using those queries/segments and do a spot-check comparison with wherever I wound up putting the data to do the analysis
  • I actually proofread the analysis report – no need to have poor grammar, typos, or inadvertently backwards labeling.

That’s really all there is to it for proofreading. It takes some conscious thought and focus, but it’s worth the effort.

3. Triangulation

This is one of my favorite – and most reliable – techniques. When it comes to digital data and the increasing flexibility of digital analytics platforms, there are almost always multiple ways to come at any given analysis. Some examples:

  • In Google Analytics, you looked at the Ecommerce tab in an events report to check the Ecommerce conversion rate for visits that fired a specific event. To check the data, build a quick segment for visits based on that event and check the overall Ecommerce conversion rate for that segment. It should be pretty close!
  • In SiteCatalyst, you have a prop and an eVar populated with the same value, and you are looking at products ordered by subrelating the eVar with Products and using Orders as the metric. For a few of the eVar values, build a Visit-container-based segment using the prop value and then look at the Products report. The numbers should be pretty close.
  • If you’ve used the eCommerce conversion rate for a certain timeframe in your analysis, pull the visits by day and the orders by day for that timeframe, add them both up, and divide to see if you get the same conversion rate.
  • Use flow visualization (Google Analytics) or pathing (SiteCatalyst) to compare results that you see in a funnel or fallout report – they won’t match, but you should be able to easily explain why when the steps when they differ.
  • Pull up a clickmap to see what it reports when you’ve got a specific link tracked as an event (GA) or a custom link (SiteCatalyst).
  • If you have a specific internal link tracked as an event or custom link, compare the totals for that event to the value from the Previous Page report for the page it links to.

You get the idea. These are all web analytics examples, but the same approach applies for other types of digital analysis as well (if your Twitter analytics platform says there were 247 tweets yesterday that included a certain keyword, go to search.twitter.com, search for the term, and see how many tweets you get back).

triangulation

Quite often, the initial triangulation will turn up wildly different results. That will force you to stop and think about why, which, most of the time, will result in you realizing why that wasn’t the primary way you chose to access the data. The more ass-backwards of a triangulation that you can come up with to get to a similar result, the more confidence you will have that your data is solid (and, when a business user decides to pull the data themselves to check your work and gets wildly different results, you may already be armed to explain exactly why…because that was your triangulation technique!).

4. Phone a friend

Granted, for this one, you have to tap into other resources. But, a fresh set of eyes is invaluable (there’s a reason that development teams generally split developers out from the QA team, and there’s a reason that even professional writers have an editor review their work).

phone a friend

When phoning a friend, you actually can request any or all of the three prior tips:

  • Ask them if the results you are seeing pass the “sniff test” – do they seem plausible?
  • Ask them to look at the actual segment or query definitions you used – get them to proofread your work.
  • Ask them to spot-check your work by trying to recreate the results – this may or may not be triangulation (even if they approach the question exactly as you did, they’re still checking your work).

To be clear, you’re not asking that they completely replicate your analysis. Rather, you’re handing them a proverbial napkin and asking them to quickly and messily put a pen to that napkin to see if anything emerges that calls your analysis into question.

This Is Not As Time-Consuming As It Sounds

I positively cringe when someone excitedly tells me that they “just looked at the data and saw something really interesting!”

  • If it’s a business user, I shake my head and gently probe for details (“Really? That’s interesting. Let me see if I’m seeing the same thing. How is it that you got this data?…”)
  • If it’s an analyst, I say a silent prayer that they really have found something really interesting that holds up as interesting under deeper scrutiny. The more surprising and powerful the result, the stronger I push for a deep breath and a second look.

So, obviously, there is a lot of judgment involved when it comes to determining the extent of QA to perform. The more complex the project, and the more surprising the results, the more time it’s worth investing in QA. The more you get used to doing QA, the earlier in the analysis you will be thinking about it (and doing it), and the less incremental time it takes.

And it’s worth it.

Photos courtesy of, in order, hobvias sudoneighm, Terry Whalebone, Nate Steiner, and Bùi Linh Ngân.

*Yup. I totally made that number up…but it feels about right based on my own experience.

 

Analysis

Tiger Woods Is Batting .260 Lifetime

Tiger Woods won his 78th career PGA event on Sunday at The Players Championship. The commentators were tireless in their mentions of the fact that his was Woods’s 300th PGA event start.

I’m a bad golfer and a worse baseball player, but I found myself wanting to combine the two sports by calculating Woods’s “batting average” for PGA tour events. This required two major definitional leaps:

  • An “at bat” was a tournament
  • A “hit” was a win

This is a whopper of a stretch, I realize, but stick with me, anyway. 🙂

The batting average math is now simply: with Woods’s win, his career batting average in tour events was 78/300, or .260! In baseball, a “good” hitter bats over .300. Of course, for my definitions to hold up, in real baseball, a player would only get credited with a hit if he hit a game-winning walkoff home run every time he got a hit!

This led me to wonder what Woods’s batting average over his career to date has been. So, using data from Woods’ profile on pgatour.com, I plotted it out (even though Woods was an amateur until 1996, the tournaments he played in before that still counted as PGA tour starts):

Tiger Woods Cumulative Win Percentage

His batting average peaked in 2009, just a couple of months before he had his worst Thanksgiving ever.

As the end of the chart shows, it does look like he is on his way back. Keep in mind that, like a real batting average, the fewer tournaments he’d played in, the more a win would increase his cumulative average and the less a non-win would drop it. That’s one reason that, in baseball, there is more focus on the batting average for the season than on the career batting average.

So, that got me wondering how this tour season compares to Woods’s past seasons. The gray in the chart below shows his average as of the end of each season:

Tiger Woods Cumulative and Yearly Win Percentage

To date, this is his highest win percentage of any year other than 2008, which was severely shortened by a knee injury. In 2008, he won 4 out of 6 PGA events before his season ended. In 2013, he has won 4 out of 7 so far!

Idle fun with Excel and online data!

 

 

Analysis, Analytics Strategy

#eMetrics Reflection: Self-Service Analysis in 2 Minutes or Less

I’m chunking up my reflections on last week’s eMetrics conference in San Francisco into several posts. I’ve got a list of eight possible topics, but I seriously doubt I’ll managed to cover all of them.

The closing keynote at eMetrics was Matt Wilson and Andrew Janis talking about how they’ve been evolving the role of digital (including social) analytics at General Mills.

Almost as a throwaway aside, Matt noted that one of the ways he has gone about increasing the use of their web analytics platform by internal users is with video:

  1. He keeps a running list of common use cases (types of data requests)
  2. He periodically makes 2-minute (or less) videos of how to complete these use cases

Specifically:

  • He uses Snagit Pro to do a video capture of his screen while he records a voiceover
  • If a video lasts more than 120 seconds, he scraps it and starts over

Outside of basic screen caps with annotations, the “video with a voiceover” is my favorite use of Snagit. When I need to “show several people what is happening,” it’s a lot more efficient than trying to find a time for everyone to jump into GoToMeeting or a Google Hangout. I just record my screen with my voiceover, push the resulting video to YouTube (in a non-public way — usually “anyone with the link” mode), and shoot off an email.

I’ve never tried this with analytics demos — as a way to efficiently build a catalog of accessible tutorials — but I suspect I’m going to start!

Analysis, Analytics Strategy

#eMetrics Reflection: Visits / Visitors / Cohorts / Lifetime Value

I’m chunking up my reflections on last week’s eMetrics conference in San Francisco into several posts. I’ve got a list of eight possible topics, but I seriously doubt I’ll managed to cover all of them.

One of the first sessions I attended at last week’s eMetrics was Jim Novo’s session titled “The Evolution of an Attribution Resolution.” We’ll (maybe) get to the “attribution” piece in a separate post (because Jim turned on a light bulb for me there), but, for now, we’ll set that aside and focus on a sub-theme of his talk.

Later at the conference, Jennifer Veesenmeyer from Merkle hooked me up with a teaser copy of an upcoming book that she co-authored with others at Merkle called It Only Looks Like Magic: The Power of Big Data and Customer-Centric Digital Analytics. (It wasn’t like I got some sort of super-special hookup. They had a table set up in the exhibit hall and were handing copies out to anyone who was interested. But I still made Jennifer sign my copy!) Due to timing and (lack of) internet availability on one of the legs of my trip, I managed to read the book before landing back in Columbus.

A Long-Coming Shift Is About to Hit

We’ve been talking about being “customer-centric” for years. It seems like eons, really. But, almost always, when I’ve hear marketers bandy about the phrase, they mean, “We need to stop thinking about ‘our campaigns’ and ‘our site’ and ‘our content’ and, instead, start focusing on the customer’s needs, interests, and experiences.” That’s all well and good. Lots of marketers still struggle to actually do this, but it’s a good start.

What I took away from Jim’s points, the book, and a number of experiences with clients over the past couple of years is this:

Customer-centricity can be made much more tangible…and much more tactically applicable when it comes to effective and business-impacting analytics.

This post covers a lot of concepts that, I think, are all different sides of the same coin.

Visitors Trump Visits

Cross-session tracking matters. A visitor who did nothing of apparent importance on their first visit to the site may do nothing of apparent importance across multiple visits over multiple weeks or months. But…that doesn’t mean what they do and when they do it isn’t leading to something of high value to the company.

Caveat (defended) to that:

Visitors Trump Visits

Does this means visits are dead? No. Really, unless you’re prepared to answer every new analytics question with, “I’ll have an answer in 3-6 months once I see how visitors play out,” you still need to look at intra-session results.

When I asked Jim about this, his response totally made sense. Paraphrasing heavily: “Answering a question with a visit-driven response is fine. But, if there’s a chance that things may play out differently from a visitor view, make sure you check back in later and see if your analysis still holds over the longer term.”

Cohort Analysis

Cohort analysis is nothing more than a visitor-based segment. Now, a crap-ton of marketers have been smoking the Lean Startup Hookah Pipe, and, in the feel-good haze that filled the room, have gotten pretty enamored with the concept. Many analysts, myself included, have asked, “Isn’t that just a cross-session segment?” But “cross-session segment” isn’t nearly as fun to say.

Cohort Analysis Tweet

Here’s the deal with cohort analysis:

  • It is nothing more than an analysis based around segments that span multiple sessions
  • It’s a visitor-based concept
  • It’s something that we should be doing more (because it’s more customer-centric!)

The problem? Mainstream web analytics tools capture visitors cross-session, and they report cross-session “unique visitors,” but this is only in aggregate. You can dig into Adobe Discover to get cross-session detail, or, I imagine, into Adobe Insight, but that is unsatisfactory. Google has been hinting that this is a fundamental pivot they’re making — to get more foundationally visitor-based in their interface. But, Jim asked the same question many analysts are:

Visitor Value Prediction

Having started using and recommending visitor-scope custom variables more and more often, I’m starting to salivate at the prospect of “visitor” criteria coming to GA segments!

Surely, You’ve Heard of “Customer Lifetime Value?”

“Customer Lifetime Value” is another topic that gets tossed around with reckless abandon. Successful retailers, actually, have tackled the data challenges behind this for years. Both Jim and the Merkle book brought the concept back to the forefront of my brain.

It’s part and parcel to everything else in this post: getting beyond, “What value did you (the customer) deliver to me today?” to “What value have you (or will you) deliver to me over the entire duration of our relationship” (with an eye to the time value of money so that we’re not just “hoping for a payoff wayyyy down the road” and congratulating ourselves on a win every time we get an eyeball).

Digital data is actually becoming more “lifetime-capable:”

  • Web traffic — web analytics platforms are evolving to be more visitor-based than visit-based, enabling cross-session tracking and analysis
  • Social media — we may not know much about a user (see the next section), but, on Twitter, we can watch a username’s activity over time, and even the most locked down Facebook account still exposes a Facebook ID (and, I think, a name)…which also allows tracking (available/public) behavior over time
  • Mobile — mobile devices have a fixed ID. There are privacy concerns (and regulations) with using this to actually track a user over time, but the data is there. So, with appropriate permissions, the trick is just handling the handoff when a user replaces their device

Intriguing, no?

And…Finally…Customer Data Integration

Another “something old is new again” is customer data integration — the “customer” angle of of the world of Master Data Management. In the Merkle book, the authors pointed out that the illusive “master key” that is the Achilles heel of many customer data integration efforts is getting both easier and more complicated to work around.

One obvious-once-I-read-it concept was that there are fundamentally two different classes of “user IDs:”

  • strong identifier is “specifically identifiable to a customer and is easily available for matching within the marketing database.”
  • weak identifier is “critical in linking online activity to the same user, although they cannot be used to directly identify the user.”

Cookie IDs are a great example of a weak identifier. As is a Twitter username. And a Facebook user ID.

The idea here is that a sophisticated map of IDs — strong identifiers augmented with a slew of weak identifiers — starts to get us to a much richer view of “the customer.” It holds the promise of enabling us to be more customer-centric. As an example:

  • An email or marketing automation system has a strong identifier for each user
  • Those platforms can attach a subscriber ID to every link back to the site in the emails they send
  • That subscriber ID can be picked up by the web analytics platform (as a weak identifier) and linked to the visitor ID (cookie-based — also a weak identifier)
  • Now, you have the ability to link the email database to on-site visitor behavior

This example is not a new concept by any means. But, in  my experience, the way each of the platforms involved in a scenario like this has preferred to work is that they set their own strong and weak identifiers. What I took away from the Merkle book is that we’re getting a lot closer to being able to have those identifiers flow between systems.

Again…privacy concerns cannot be ignored. They have to be faced head on, and permission has to be granted where permission would be expected.

Lotta’ Buzzwords…All the Same Thing?

Nothing in this post is really “new.” They’re not even “new to me.” The dots I hadn’t connected was that they are all largely the same thing.

That, I think, is exciting!

 

Analysis

A.D.A.P.T. to Act and Learn

I keep posting things elsewhere and forgetting to get a post here to reference them.

Last fall, I pitched a session topic to Jim Sterne for the eMetrics conference that occurred last week. At the time, I was just a few weeks into my job at Clearhead, and I figured that, by April 2013, I’d easily have a fully baked, deliverable-supporting process that I could use as the basis for the session.

You’re expecting this sentence — the one following that last paragraph — to say, “Boy…was I wrong!” The fact is…I was mostly right!

A handful of articles, posts, and content all came out of my effort to get spit and polish on the material in time for the session:

  • The eMetrics session itself, as well as the various downloadable templates that accompanied it, are posted at clearhead.me/emetrics
  • I did a high-level summary of the content and approach in a Practical eCommerce article that was published last week
  • My unified theory of analytics (requests) was an operational umbrella for the ADAPT to Act and Learn thinking
  • Thanks to Avinash, I sorta’ rediscovered Lean Analytics right as I was wrapping up the presentation

Lots of content. You be the judge if it’s good content. Or, if you’re reading this shortly after it got posted and you’re in central Ohio, come get an abbreviated version at this month’s Columbus Web Analytics Wednesday.

Analysis, Reporting

Gilligan's Unified Theory of Analytics (Requests)

The bane of many analysts’ existence is that they find themselves in a world where the majority of their day is spent on the receiving end of a steady flow of vague, unfocused, and misguided requests:

“I don’t know what I don’t know, so can you just analyze the traffic to the site and summarize your insights?”

“Can I get a weekly report showing top pages?”

“I need a report from Google Analytics that tells me the gender breakdown for the site.”

“Can you break down all of our metrics by: new vs. returning visitors, weekend vs. weekday visitors, working hours vs. non-working hours visitors, and affiliate vs. display vs. paid search vs. organic search vs. email visitors? I think there might be something interesting there.”

“Can you do an analysis that tells me why the numbers I looked at were worse this month than last?”

“Can you pull some data to prove that we need to add cross-selling to our cart?”

“We rolled out a new campaign last week. Can you do some analysis to show the ROI we delivered with it?”

“What was traffic last month?”

“I need to get a weekly report with all of the data so I can do an analysis each week to find insights.”

The list goes on and on. And, in various ways, they’re all examples of well-intended requests that lead us down the Nefarious Path to Reporting Monkeydom. It’s not that the requests are inherently bad. The issue is that, while they are simple to state, they often lack context and lack focus as to what value fulfilling the request will deliver. That leads to the analyst spending time on requests that never should have been worked on at all, making risky assumptions as to the underlying need, and over-analyzing in an effort to cover all possible bases.

I’ve given this a lot of thought for a lot of years (I’m not exaggerating — see the first real post I wrote on this blog almost six years ago…and then look at the number of navel-gazing pingbacks to it in the comments). And, I’ve become increasingly convinced that there are two root causes for not-good requests being lobbed to the analytics team:

  • A misperception that “getting the data” is the first step in any analysis — a belief that surprising and actionable insights will pretty much emerge automagically once the raw data is obtained.
  • A lack of clarity on the different types and purposes of analytics requests — this is an education issue (and an education that has to be 80% “show” and 20% “tell”)

I think I’m getting close to some useful ways to address both of these issues in a consistent, process-driven way (meaning analysts spend more time applying their brainpower to delivering business value!).

Before You Say I’m Missing the Point Entirely…

The content in this post is, I hope, what this blog has apparently gotten a reputation for — it’s aimed at articulating ideas and thoughts that are directly applicable in practice. So, I’m not going to touch on any of the truths (which are true!) that are more philosophical than directly actionable:

  • Analysts need to build strong partnerships with their business stakeholders
  • Analysts have to focus on delivering business value rather than just delivering analysis
  • Analysts have to stop “presenting data” and, instead “effectively communicate actionable data-informed stories.”

All of these are 100% true! But, that’s a focus on how the analyst should develop their own skills, and this post is more of a process-oriented one.

With that, I’ll move on to the three types of analytics requests.

Hypothesis Testing: High Value and SEXY!

Hands-down, testing and validation of hypotheses is the sexiest and, if done well, highest value way for an analyst to contribute to their organization. Any analysis — regardless of whether it uses A/B or multivariate testing, web analytics, voice of the customer data, or even secondary research — is most effective when it is framed as an effort to disprove or fail to disprove a specific hypothesis. This is actually a topic I’m going to go into a lot of detail (with templates and tools) on during one of the eMetrics San Francisco sessions I’m presenting in a couple of weeks.

The bitch when it comes to getting really good hypotheses is that “hypothesis” is not a word that marketers jump up and down with excitement over. Here’s how I’m starting to work around that: by asking business users to frame their testing and analysis requests in two parts:

Part 1: “I believe…[some idea]”

Part 2: “If I am right, we will…[take some action]”

This construct does a couple of things:

  • It forces some clarity around the idea or question. Even if the requestor says, “Look. I really have NO IDEA if it’s ‘A’ or ‘B’!” you can respond with, “It doesn’t really matter. Pick one and articulate what you will do if that one is true. If you wouldn’t do anything different if that one is true, then pick the other one.”
  • It forces a little bit of thought on the part of the requestor as to the actionability of the analysis.

And…it does this in plain, non-scary English.

So, great. It’s a hypothesis. But, how do you decide which hypotheses to tackle first? Prioritization is messy. It always is and it always will be. Rather than falling back on the simplistic theory of “effort and expected impact” for the analysis, how about tackling it with a bit more sophistication:

  • What is the best approach to testing this hypothesis (web analytics, social media analysis, A/B testing, site survey data analysis, usability testing, …)? That will inform who in your organization would be best suited to conduct the analysis, and it will inform the level of effort required. 
  • What is the likelihood that the hypothesis will be shown to be true? Frankly, if someone is on a fishing expedition and has a hypothesis that making the background of the home page flash in contrasting colors…common sense would say, “That’s a dumb idea. Maybe we don’t need to prove it if we have hypotheses that our experience says are probably better ones to validate.”
  • What is the likelihood that we actually will take action if we validate the hypothesis? You’ve got a great hypothesis about shortening the length of your registration form…but the registration system is so ancient and fragile that any time a developer even tries to check the code out to work on it, the production code breaks. Or…political winds are blowing such that, even if you prove that always having an intrusive splash page pop up when someone comes to your home page is hurting the site…it’s not going to change.
  • What will be the effort (time and resources) to validate the hypothesis? Now, you damn well better have nailed down a basic approach before answering this. But, if it’s going to take an hour to test the hypothesis, even if it’s a bit of a flier, it may be worth doing. If it’s going to take 40 hours, it might not be.
  • What is the business value if this hypothesis gets validated (and acted upon)? This is the “impact” one, but I like “value” over “impact” because it’s a little looser.

I’ve had good results when taking criteria along these lines and building a simple scoring system — assigning High, Medium, Low, or Unknown for each one, and then plugging in some weighted scores for each value for each criteria. The formula won’t automatically prioritize the hypotheses, but it does give you a list that is sortable in a logical way, It, at least, reveals the “top candidates” and the “stinkers.”

Performance Measurement (think “Reporting”)

Analysts can provide a lot of value by setting up automated (or near-automated) performance measurement dashboards and reports. These are recurring (hypothesis testing is not — once you test a hypothesis, you don’t need to keep retesting it unless you make some change that makes sense to do so).

Any recurring report* should be goal- and KPI-oriented. KPIs and some basic contextual/supporting metrics should go on the dashboard, targets need to be set (and set up such that alerts are triggered when a KPI slips). Figuring out what should go on a well-designed dashboard comes down to answering two questions:

  1. What are we trying to achieve? (What are our business goals for this thing we will be reporting on?)
  2. How will we know that we’re doing that? (What are our KPIs?)

They need to get asked and answered in order, and that’s a messier exercise oftentimes than we’d like it to be. Analysts can play a strong role in getting these questions appropriately answered…but that’s a topic for another time.

Every other recurring report that is requested should be linkable back to a dashboard (“I have KPIs for my paid search performance, so I’d like to always get a list of the keywords and their individual performance so I have that as a quick reference if a KPI changes drastically.”)

Having said that, a lot of tools can be set up to automatically spit out all sorts of data on a recurring basis. I resist the temptation to say, “Hey…if it’s only going to take me 5 minutes to set it up, I shouldn’t waste my time trying to validate its value.” But, it can be hard to not appear obstructionist in those situations, so, sometimes, the fastest route is the best. Even if, deep down, you know you’re delivering something that will get looked at the first 2-3 times it goes out…and will never be viewed again.

Quick Data Requests — Very Risky Territory (but needed)

So, what’s left? That would be requests of the,. “What was traffic to the site last month?” ilk. There’s a gross misperception when it comes to “quick” requests that there is a strong correlation between the amount of time required to make the request and the amount of time required to fulfill the request. Whenever someone tells me they have a “quick question,” I playfully warn them that the length of the question tends to be inversely correlated to the time and effort required to provide an answer.

Here’s something I’ve only loosely tested when it comes to these sorts of requests. But, I’ve got evidence that I’m going to be embarking on a journey to formalize the intake and management of these in the very near future, so I’m going to go ahead and write them down here (please leave a comment with feedback!).

First, there is how the request should be structured — the information I try to grab as the request comes in:

  • The basics — who is making the request and when the data is needed; you can even include a “priority” field…the rest of the request info should help vet out if that priority is accurate.
  • A brief (255 characters or so) articulation of the request — if it can’t be articulated briefly, it probably falls into one of the other two categories above. OR…it’s actually a dozen “quick requests” trying to be lumped together into a single one. (Wag your finger. Say “Tsk, tsk!”
  • An identification of what the request will be used forthere are basically three options, and, behind the scenes, those options are an indication as to the value and priority of the request:
    • General information — Low Value (“I’m curious,” “It would be be interesting — but not necessarily actionable — to know…”)
    • To aid with hypothesis development — Medium Value (“I have an idea about SEO-driven visitors who reach our shopping cart, but I want to know how many visits fall into that segment before I flesh it out.”)
    • To make a specific decision — High Value
  • The timeframe to be included in the data — it’s funny how often requests come in that want some simple metric…but don’t say for when!
  • The actual data details — this can be a longer field; ideally, it would be in “dimensions and metrics” terminology…but that’s a bit much to ask for many requestors to understand.
  • Desired delivery format — a multi-select with several options:
    • Raw data in Excel
    • Visualized summary in Excel
    • Presentation-ready slides
    • Documentation on how to self-service similar data pulls in the future

The more options selected for the delivery format, obviously, the higher the effort required to fulfill the request.

All of this information can be collected with a pretty simple, clean, non-intimidating intake form. The goal isn’t to make it hard to make requests, but there is some value in forcing a little bit of thought rather than the requestor being able to simply dash off a quickly-written email and then wait for the analyst to fill in the many blanks in the request.

But that’s just the first step.

The next step is to actually assess the request. This is the sort of thing, generally, an analyst needs to do, and it covers two main areas:

  • Is the request clear? If not, then some follow-up with the requestor is required (ideally, a system that allows this to happen as comments or a discussion linked to the original request is ideal — Jira, Sharepoint, Lotus Notes, etc.)
  • What will the effort be to pull the data? This can be a simple High/Medium/Low with hours ranges assigned as they make sense to each classification.

At that point, there is still some level of traffic management. SLAs based on the priority and effort, perhaps, and a part of the organization oriented to cranking out those requests as efficiently as possible.

The key here is to be pretty clear that these are not analysis requests. Generally speaking, it’s a request for data for a valid reason, but, in order to conduct an analysis, a hypothesis is required, and that doesn’t fit in this bucket.

So, THEN…Your Analytics Program Investment

If the analytics and optimization organization is framed across these three main types of services, then conscious investment decisions can be made:

  • What is the maximum % of the analytics program cost that should be devoted to Quick Data Requests? Hopefully, not much (20-25%?).
  • How much to performance measurement? Also, hopefully, not much — this may require some investment in automation tools, but once smart analysts are involved in defining and designing the main dashboards and reports, that is work that should be automated. Analysts are too scarce for them to be doing weekly or monthly data exports and formatting.
  • How much investment will be made in hypothesis testing? This is the highest value

With a process in place to capture all three types of efforts in a discrete and trackable way enables reporting back out on the value delivered by the organization:

  • Hypothesis testing — reporting is the number of hypotheses tested and the business value delivered from what was learned
  • Performance measurement — reporting is the level of investment; this needs to be done…and it needs to be done efficiently
  • Quick data requests — reporting is output-based: number of requests received, average turnaround time. In a way, this reporting is highlighting that this work is “just pulling data” — accountability for that data delivering business value really falls to the requestors. Of course, you have to gently communicate that or you won’t look like much of a team player, now, will you?

Over time, shifting an organization to think it terms of actionable and testable hypotheses is the goal — more hypotheses, fewer quick data requests!

And, of course, this approach sets up the potentially to truly close the loop and follow through on any analysis/report/request delivered through a Digital Insight Management program (and, possibly, platform — like Sweetspot, which I haven’t used, personally, but which I love the concept of).

What Do You Think?

Does this make sense? It’s not exactly my opus, but, as I’ve hastily banged it out this evening, I realize that it includes many of the ways that I’ve had the most success in my analytics career, and it includes many of the structures that have helped me head off the many ways I’ve screwed up and had failures in my analytics career.

I’d love your thoughts!

 

*Of course, there are always valid exceptions.

Analysis, General

Three Things You Need To Move From Reporting To Analysis

Reporting is necessary but not sufficient. Don’t get me wrong – there will always be some need to see on-going status of key metrics with an organisation, and for business people to see numbers that trigger them to ask analysis questions. But if your analysts spend 40 hours a week providing you with report after report after report, you are failing to get value from analytics. Why? Because you’re not doing it.

So, what factors are critical to an increased focus on analysis?

1. Understand the difference

Reporting should be standardised, regular and raise alerts.

Standardised: You should be looking at the same key metrics each time.

Regular: Occurring on an agreed-upon schedule. For example, daily, weekly, monthly.

Alerts: If something does not change much over time, or dramatic shifts in “key” metrics are no big deal, you shouldn’t be monitoring them. It’s the “promoted” or “fired” test – if a KPI shifts dramatically and no one could be fired or promoted as a result, was it really that important? Okay, most of the time it’s not as dire as promoted/fired, but dramatic shifts should trigger action. A report may not answer every question, however it should alert you to changes that warrant further investigation. Reporting can inspire deeper analysis.

Analysis is ad-hoc and investigative, an exploration of the data. It may come from something observed in a report, an analyst’s curiosity or a business question, but it should aim to figure out something new.

Unfortunately, it’s far too common for what should be a one-time analysis to turn into an on-going report. After all, if it was useful once, it “must” be useful again, right?

2. The right (minded) people

Having the right analysts is critical to doing more than just reporting. Do you have an analyst who is bored to tears running reports? Good! That is a sign that you hired well. Ideally, reporting should be a “rite of passage” that new analysts go through, to teach them the basic concepts, how to use the tools, how to present data, what key metrics are important to the business and how to spot shifts that require extra investigation.

The right analysts are intellectually curious, interested in understanding the root cause of things. They are puzzle solvers who enjoy the process of discovery. The right analysts therefore thrive on, not surprisingly, analysis, not reporting.

That’s not to say that more seasoned analysts should have no role in reporting. They should be monitoring key reports and fully informed about trends and changes in the data. They just should be able to step back from the heavy lifting.

3. Trust

Analytics requires trust. The business needs to trust that the analytics team are monitoring trends in key metrics, that they know the business well enough and that they are focusing analyses on what really matters. This requires open dialogue and collaboration. Analytics has a far better success rate when tightly integrated with the business.

It’s easy to feel you’re getting “value for money” when you get tons of reports delivered to you, because you’re seeing output. But it’s also a sign that you don’t trust your analysts to find you the insights. And sadly, it’s the business that misses out on opportunities.

The first steps to action

If you are ready to start seeing the value of analytics, here are a few ways you can start:

  1. Limit reporting to what is necessary for the business. This may mean discontinuing reports of little or no value. This can be difficult! Perhaps propose “temporarily” discontinuing a number of reports. Once the “temporary” pause is over, and people realise they didn’t really miss those reports, it should be clear that they are no longer necessary.
  2. Review your resources. Make sure you have the right people focused on analysis and that they are suited to, and ready for, that kind of work.
  3. Allocate a certain percentage of analysts’ time to exploration of data and diving into ad hoc business questions. Don’t allow this to be “when they have time” work. (Hint: They’ll never “have time.”) It needs to be an integral part of their job that gets prioritised. The key to ensuring prioritisation is for analysis to be aligned with critical business questions, so stakeholders are anxiously awaiting the results.
  4. Introduce regular sharing and brainstorming sessions, to present and develop analyses. You don’t have to limit this to your analytics team! Invite your business stakeholders, to help collaboration between teams.

The hardest part will be getting started. Once you start seeing the findings of analysis, and getting much deeper insight that some standard report would provide you, it will be easy to see the benefits and continue to build this practice.

Analysis, General, Industry Analysis

Getting comfortable with data sampling in the growing data world

“Big data” is today’s buzz word, just the latest of many. However, I think analytics professionals will agree: “Big data” is not necessarily better than “small data” unless you use it to make better decisions.

At a recent conference, I heard it proposed that “Real data is better than a representative sample.” With all due respect, I disagree. That kind of logic assumes that a “representative sample” is not, in fact, representative.

If the use of “representative” data would not accurately reflect the complete data set, and its use would lead to different conclusions, using “real” data is absolutely better. However, it’s not actually because “real” data is somehow superior, but rather because the representative sample itself is not serving its intended purpose.

On the flip side, let’s assume the representative sample does actually represent the complete data set, and would reveal the same results and lead to the same decisions. In this case, what are the benefits of leveraging the sample?

  • Speed – sampling is typically used to speed up the process, since the analytics process doesn’t need to evaluate every collected record.
  • Accuracy – if the sample is representative (the assumption we are making here) using the full or sampled data set should make no difference. Results will be just as accurate.
  • Cost-savings – a smaller, sampled data set requires less effort to clean and analyse than the entire data set.
  • Agility – by gaining time and freeing resources, digital analytics teams can become more agile and responsive to acting wisely on small (sampled) data.

There is no doubt that technology continues to develop rapidly. Storage and computing power that used to require floors of space now fits into my iPhone 5. However, the volume of data we leverage is growing at the same rate (or faster!) The bigger data gets, and the quicker we demand answers, the more sampling will become an accepted best practice. After all, statisticians and researchers in the scientific community have been satisfied with sampling for decades. Digital too will reach this level of comfort in time, and focus on decisions instead of data volume.

What do you think?

Analysis, Analytics Strategy, Presentation

Effectively Communicating Analysis Results

I was fortunate enough to not only get to attend the Austin DAA Symposium this week, but to get to deliver one of the keynotes. The event itself was fantastic — a half day that seemed to end pretty much as soon as it started, but in which I felt like I had a number of great conversations, learned a few things, and got to catch up with some great people whom I haven’t seen in a while.

The topic of my keynote was “Effectively Communicating Analysis Results,” and, as sometimes tends to happen between the writing of the description and the actual creation of the content, the scope morphed a bit by the time the symposium arrived.

My theme, ultimately, was that, as analysts, we have to play a lot of roles that aren’t “the person who does analysis” if we really want to be effective. I illustrated why that is the case…in a pie chart (I compensated by explaining that pie charts are evil later in the presentation). The pie chart was showing, figuratively, a breakdown of all of the factors that actually contribute to an analysis driving meaningful and positive action by the business:

What Goes Into Effective Analysis

 

The roles? Well:

  • Translator
  • Cartographer
  • Process Manager
  • Communicator
  • Neuroscientist
  • Knowledge Manager

I recorded one of my dry runs, which is available as a 38 minute video, and the slides themselves are available as well, over on the Clearhead blog.

It was a fun presentation to develop and deliver, and a fantastic event!

Analysis

Quotable Quotes from Nate Silver

It’s hard to be an analyst and not be a fan of Nate Silver. Actually, I think it might actually be the law — one of those “natural law” things, like gravity (“Obey gravity! It’s the law!”), rather than one of those legislated ones.

Not too long ago, I wrote a post that gave my take on one aspect of the post-election commentary about Silver’s work. In some of the Twitter exchanges around that post, Jim Cain suggested that I really should read Silver’s book, as the content of the post lined up well with some of the topics Silver covered. I’d planned to read the book over the holidays, anyway, but his nudge convinced me to go ahead and buy the Kindle edition and bump it up to the top of my list.

THAT was a great move (thank you, Twitter and Jim!). I’ve still got some digesting (and rereading) to do, but I thought I’d throw out some of my favorite quotes from the book as a blog post. In order of appearance…

The most elegant description of the why and what of having massive amounts of data at your disposal to tell whatever story you want:

The instinctual shortcut that we take when we have “too much information” is to engage with it selectively, picking out the parts we like and ignoring the remainder, making allies with those who have made the same choices and enemies with the rest.

On the role of the analyst and the necessity for thought behind the data:

The numbers have no way of speaking for themselves. We speak for them. We imbue them with meaning…Data-driven predictions can succeed–and they can fail. It is when we deny our role in the process that the odds of failure rise. Before we demand more of our data, we need to demand more of ourselves.

But…also recognizing that people are not machines (he spends a lot of time breaking down the evolution of chess-playing computers to articulate the strengths and weaknesses of computers and humans when it comes to prediction):

We can never make perfectly objective predictions. They will always be tainted by our subjective point of view.

Silver actually quotes John P. A. Ioannidis, author of a paper called “Why Most Published Research Findings Are False,” at length and then explains in simple terms the mathematical realities of digging into Big Data:

“In the last twenty years, with the exponential growth in the availability of information, genomics, and other technologies, we can measure millions and millions of potentially interesting variables,” Ioannidis told me. “The expectation is that we can use that information to make predictions work for us. I’m not saying that we haven’t made any progress. Taking into account that there are a couple of million papers, it would be a shame if there wasn’t. But there are obviously not a couple of million discoveries. Most are not really contributing much to generating knowledge.”

This is why our predictions may be more prone to failure in the era of Big Data. As there is an exponential increase in the amount of available information, there is likewise an exponential increase in the number of hypotheses to investigate. For instance, the U.S. government now publishes data on about 45,000 economic statistics. If you want to test for relationships between all combinations of two pairs of these statistics–is there a causal relationship between the bank prime loan rate and the unemployment rate in Alabama?–that gives you literally one billion hypotheses to test.

But the number of meaningful relationships in the data–those that speak to causality rather than correlation and testify to how the world really works–is orders of magnitude smaller. Nor is it likely to be increasing at nearly so fast a rate as the information itself; there isn’t any more truth in the world than there was before the internet or the printing press. Most of the data is just noise, as most of the universe is filled with empty space.

I have a whole slew of reading and understanding-deepening to do around Bayesian reasoning, Fisher’s statistical method, Frequentists, and all sorts of other data science-y topics spawned by the middle part of the book (so, my original plan to read this book has now been replaced by a plan to dig into Matt Gershoff’s list of data science and machine learning resources). Silver is a strong believer in what he calls “The Bayesian Path to Less Wrongness:”

…I’m of the view that we can never achieve perfect objectivity, rationality, or accuracy in our beliefs. Instead, we can strive to be less subjective, less irrational, and less wrong. Making predictions based on our beliefs is the best (and perhaps even only) way to test ourselves. If objectivity is the concern for a greater truth beyond our personal circumstances, and prediction is the best way to examine how closely aligned our personal perceptions are with that greater truth, the most objective among us are those who make the most accurate predictions.

And, more on the “art and science” of analytics — the need to not simply expect the numbers to give the right answer on their own:

It would be nice if we could just plug data into a statistical model, crunch the numbers, and take for granted that it was a good representation of the real world. Under some conditions, especially in data-rich fields like baseball, that assumption is fairly close to being correct. In many other cases, a failure to think carefully about causality will lead us up blind alleys.

As analysts, how often are we faced with stakeholders who have unrealistic expectations of getting a black-and-white answer? The reality:

In science, one rarely sees all the data point toward one precise conclusion. Real data is noisy–even if the theory is perfect, the strength of the signal will vary.

And, finally, a conclusion that wraps with a clever turn on Reinhold Niebuhr’s Serenity Prayer:

Prediction is difficult for us for the same reason that it is so important: it is where objective and subjective reality intersect. Distinguishing the signal from the noise requires both scientific knowledge and self-knowledge: the serenity to accept the things we cannot predict, the courage to predict the things we can, and the wisdom to know the difference.

These were a sample of some of the passages I highlighted throughout the book. They capture a degree of the concepts and ideas that the book covers. They don’t — at all — cover the deeply researched examples that Silver uses to illustrate these ideas. From poker to election prediction to weather forecasting (which has gotten much better in the past few decades) to earthquake and financial market prediction (that have barely improved at all when compared to weather forecasting) to predicting terrorism, the depth and breadth of his research is impressive!

I suspect I will be returning to specific aspects of his book in greater detail in the future, but this was a fun re-skim to remind me that the writing and ideas were both outstanding!

Analysis, Testing and Optimization

Big Data without Digital Insight Management Is a Big Hot Mess

One of the many exciting aspects of joining a new company is the opportunity for reflection. The lead-up to the job change forced some introspection — what was it I really most enjoyed about my profession and what would a dream job look like that allowed me to spend as much of each day doing that as possible? And, as a new company, everyone has had to put their heads together to build out the processes needed to bring the vision for the company to life, which has required a different flavor of reflection: reflecting on what has and has not worked in our collective experience when it comes to enabling brands to be as data-informed as possible in their daily processes.

Shortly after joining Clearhead, I attended eMetrics in Boston. The conference, as always, was a great time. And, as often is the case, one of the conversations that stuck with me the most occurred where I didn’t expect it — in the exhibit hall during the sessions with a vendor I’d never heard of before the conference: Sweetspot Intelligence. Sergio Maldonado (@sergiomaldo) explained the vision for Sweetspot, gave me a brief product tour, and handed me a copy of the paper they sponsored Eric Peterson to write: Digital Insight Management: Ten Tips to Better Leverage Your Existing Investment in Digital Analytics and Optimization. The concept of “Digital Insight Management” is intriguing. And, luckily, it’s much more than an abstract idea — it’s real and, I believe, something that all analysts should be striving to implement.

Let’s Start with the Basics — Demystified’s Hierarchy of Analytical Needs

Early in the paper, Eric included Analytics Demystified‘s Hierarchy of Analytical Needs:

Experienced analysts look at this diagram and think, “Well…yeah. That’s a good depiction of the battle we fight every day.” Any sort of ho-hum response to the diagram is because we’ve been fighting the battle to move “up the pyramid” for a while, and we often feel undermined by the business environment in which we work. This is one of the more succinct and elegant depictions (not just the labels on the left — the assessment in the boxes on the right) that I’ve seen.

One Step Back Adds Another Element

When viewed through the lens of “what an analyst can do,” the hierarchy is complete. In some respects, the analyst can only lead the proverbial horse to water (clearly communicate a data-informed recommendation). The analyst can’t necessarily make the horse drink (take action). But, still, it’s worth recognizing that, if we take just one step back from this pyramid, we want to see one more level on the hierarchy:

Again, this is somewhat obvious. Yet, it’s where “we” (businesses) seem to so often stumble. There is so much “Data” now that marketers are now conditioned to prepend any mention of the word “data” with the word “big!” Few reports rely on data from a single source as analysts, and marketers work hard to place the data into meaningful context.  But, of course, the further up the pyramid we go, the easier and easier it is to get derailed. Ultimately…limited action.

Pivoting the Process

While the hierarchies above are unequivocally true, the actual process for meaningful analytics — analysis that drives relevant action — actually looks quite different:

Let’s break this down a bit:

  • Everything hinges on having clear objectives and measures of success — it’s scary how often marketers stumble on this, and, as analysts, it behooves us to be skilled in helping marketers get these nailed down (these are soft skills!)
  • Performance measurement is key…but it’s not the source of insights — performance measurement is the alerting system; it tracks the KPIs against targets, as well as some supporting and contextual metrics. But, the reports and dashboards themselves don’t yield insights — they surface problems that then need to be further explored.
  • All analysis starts with a business problem, business question, or business idea — the lefthand column is where th magic happens (or, all too often, doesn’t!).

It is impossible to attend any analytics-oriented conference these days without being hit over the head with how critical it is to develop and foster strong relationships with your business partners: regularly communicate, listen for the problems they’re having that your analytical skills can help with, learn how to communicate effectively, etc. That is a recurring theme because actually teasing out the right business questions and problems can be tricky!

Conversely, the back end of the process can be tricky, too. We’ve all had cases where we completed the right analysis and got actionable results…but action never occurred. As I understand it, that is where Sweetspot comes in: technology that supports communication and workflow related to getting actionable information to the people who can take action:

So…Will Tag Management Solve This?

(Blog authors get to crack themselves up with their headings…)

What Eric’s paper, and Sweetspot’s product, got me thinking about are a couple of gaps that, hopefully, I’ve covered in this post:

  • As analysts, we need to develop, implement, and own workable processes within our companies to make analytics truly gain and sustain traction
  • There is an opportunity for better technology to support these processes…and that is analytics technology that has nothing to do with the mechanics of capturing customer data

Is “Digital Insight Management” the next big thing? I think it is. Big Data is just a big hot mess without it.

Analysis, Featured, Technical/Implementation

The T&T Plugin – Integrate T&T with Google Analytics

When Test&Target was being built back in the day and doing business as Offermatica, it was designed to be an open platform so that its data can be made available to any analytics platform.  While the integration with SiteCatalyst has since been productized, a very similar approach approach can be used to integrate your T&T test data with Google Analytics.  Let me explain how here.

The integration of SiteCatalyst leverages a feature of Test&Target called a “Plug-in”.  This plug-in concept allows you to specify code snippets that will be brought to the page upon certain conditions.  The SiteCatalyst integration is simply a push of a code snippet or plug-in to the page that tells SiteCatalyst key T&T info.

Having something like this can be incredibly helpful for all sorts of reasons such as integrating your optimization program with third party tools, or by allowing you to deliver code to the page via T&T which saves you from having IT make changes to the page code on the site.

To push your campaign or test data over to SiteCatalyst, you create a HTML offer in T&T that looks like this:

<script type=”text/javascript”>
if (typeof(s_tnt) == ‘undefined’) {
var s_tnt = ”;
}
s_tnt += ‘${campaign.id}:${campaign.recipe.id}:​${campaign.recipe.trafficType},’;
</script>

This code is simply taking the T&T profile values in red, which represent your test name and test experience names, and passes them to a variable called s_tnt for SiteCatalyst to pick up.  There is a back end classification process that takes place where these numerical values are translated into what you named them in T&T.  This is helpful to shorten the call being made to SiteCatalyst but not required unless the call to your SiteCatalyst has a relatively high character count.

After you save this HTML offer in your T&T account, you then have to create the “Plug-in”.  You can do so by accessing the configuration area as seen here:

T&T plugin, SiteCatalyst, Google AnalyticsThen we simply configure the plug-in here:

T&T Plug-in ConfiguratorThe area surrounded by a red box is where you select the previously created HTML offer with your plug-in code.  You also have the option to specify when the code gets fired.  Typically you want it to only fire when a visitor becomes a member of a test or when test content (T&T offers) are being displayed and to do so, simply select, Display mbox requests only.   If you wanted to, you can have your code fire on all mbox requests as that can be need sometimes.  Additionally, you can limit the code firings to a particular mbox or even by certain date periods.

Pretty straightforward.  To do this for Google Analytics you use the code right below to create a HTML offer and configure the plug-in in the exact same manner.  Note that we are not passing Campaign or Recipe (Experience) ID’s but rather profile tokens that represent the exact name of the Campaign name and Experience name specified in your test setup.

<script type=”text/javascript”>
_gaq.push([‘_trackEvent’, ‘Test&Target’,’${campaign.name}’,’${campaign.recipe.name}’]);
</script>

And that is it.  Once that is in place, your T&T test data is being pushed to your Google Analytics account.

Before I show you what it looks like in Google Analytics, it is important to understand a key concept in Google Analytics.

Test&Target is using the Custom Events capability of Google Analytics to populate the data.  Each Event has a Category, an Action, and a Label.  In this integration, the Google Analytics Event Category is simply Test&Target because that is our categorization of these Events.  The Google Analytics Action Event represents the Test&Target Test name.  And finally, the Event Label in Google Analytics represents the Test&Target Test Experience.  Here is a mapping to hopefully relate this easier:

Google Analytics EventsNow that we understand that, lets see what the integration gets you:

Google Analytics Test&TargetWhat we have here is a report of a specific Google Analytics Event Category, in this case the Test&Target Event.  Most of my clients have many Event Categories so it’s important to classify Test&Target as a separate Event and this plug-in code does that for you.

This is a very helpful report as we can get a macro view of the optimization efforts.  This report allows you to look at how ALL of your tests impact success events being tracked in Google Analytics at the SAME time.  Instead of looking at just a unique test as you might be used to when looking at test results in T&T, here we can see if Test A was more impactful then Test B – essentially comparing any and all tests against each other.  This is great if organizations have many groups running tests or if you want to see what particular test types impact a particular metric or combination of metrics.

Typically though, one likes to drill into a specific test and that is available by changing the Primary Dimension to Event Label which, as you know, represents the T&T Test Experience.  Here we are looking at Event Labels (Experiences) for a unique Event Action (Test):

Google Analytics Test ExperiencesHere we can look at how a unique test and its experiences impacted given success events captured in Google Analytics. Typically, most organizations include their key success events for analysis in T&T but this integration is helpful if you want to look at success events not included in your T&T account or if you want to see how your test experiences impacted engagement metrics like time on site, page views, etc….

So there you have it.  A quick and easy way to integrate your T&T account with Google Analytics.  While this can be incredibly helpful and FREE, it is important to also understand that statistical confidence is not communicated here in Google Analytics or any analytics platform that I know of, including SiteCatalyst.  It is important to leverage your testing platform for these calculations or offline calculators of statistical confidence before making any key decisions based on test data.

While this was fun to walk you through how to leverage the T&T plug-in to push data into Google Analytics please know that you can use the plug-in for a wide array of things.  I’ve helped clients leverage the plug-in capability to integrate T&T with MixPanel, CoreMetrics, and Webtrends.  You can also use this plug-in capability to integrate with other toolsets other then analytics.  For example, I have helped clients integrate T&T data into SFDC, ExactTarget, Responsys, Causata, internal CRM databases, Eloqua/Aprimo/Unica , Demdex (now DBA Audience Manager), and display retargeting toolsets.  Any platform that can accept a javascript call or pick up a javascript variable can make use of this plug-in concept.

I’ve also helped customers over the years leverage the plug-in to publish tags to the site.  Years before the abundance of Tag Management Platforms became available, there were T&T customers using the plug-in to publish Atlas, DoubleClick, and Analytic tags to the site.  In fact, if Adobe wanted to, they could make this plug-in capability into a pretty nice Tag Management Platform and one that would work much more efficiently with T&T then the current Tag Management tool they have on the market today.

Analysis, Reporting, Social Media

Analysts as Community Managers' Best Friends

I had a great time in Boston last week at eMetrics. The unintentional theme, according to my own general perception and the group messaging backchannel that I was on, was that tag management SOLVES ALL!!!.

My session…had nothing to do with tag management, but it seemed worth sharing nonetheless: “The Community Manager’s Best Friend: You.” The premise of the presentation was twofold:

  • Community managers plates are overly full as it is without them needing to spend extensive time digging into data and tools
  • Analysts have a slew of talents that are complementary to community managers’, and they can apply those talents to make for a fantastic partnership

Due to an unfortunate mishap with the power plug on my mixing board while I was out of town a few month ago, my audio recording options are a bit limited, so the audio quality in the 50-minute video (slides with voiceover) below isn’t great. But, it’s passable (put on some music in the background, and the “from the bottom of a deep well” audio effect in the recording won’t bug you too much):

I’ve also posted the slides on Slideshare, so you can quickly flip through them that way as well, if you’d rather:

As always, I’d love any any and all feedback! With luck, I’ll reprise the session at future conferences, and a reprise without refinement would be a damn shame!

Analysis

A Pragmatic Approach to "Test and Learn"

“We’re going to use a ‘test and learn’ approach” has become as common a buzzphrase as, “We’re going to be data-driven,” and “We’re going to derive actionable insights.” I’m not a fan of buzzphrases. Buzzphrases tend to originate as statements of an aspirational goal that then quickly morph to be treated as accepted reality. When someone like Eric Peterson steps up and delves into one of these buzzphrases, I do figurative backflips of joy.

The devil is in the details (which is not only a buzzphrase, but a full-on cliché…but it’s true!). And, the details come down to the right people with a valid process using capable tools. When it comes to “test and learn,” the gap between concept and actual implementation often seems to be a true chasm.

The concept: use a combination of A/B (and multivariate) testing and the analysis of historical data to test hypotheses. Based on the disproving or failure to disprove each hypothesis, take appropriate action to drive continuous improvement.

The actual implementation: HiPPOs, lack of clarity on what the KPIs are (without KPIs to optimize against, there can be no optimization), limited resources, over-focus on a specific technique or tool as “the answer,” inability to coordinate/align between marketers/designers/strategists/analysts, analyses resulting in “light gray or dark gray” conclusions rather than “black or white” ones, and so on.

On top of the challenges that have always existed, even in the “simple” world of a brand’s digital presence being primarily limited to their web site and the drivers of traffic to the site (SEO, SEM, banner ads, affiliate programs), we now operate in a world that includes social media. And, most of a brand’s social media activity cannot be A/B tested in a classical sense, so that tried-and-trued (but, alas, still too rare) technique is not available.

None of these challenges mean that “test and learn” is an unattainable ideal. But, it does mean that a strong process with a diligent steward (read: an analyst who is willing to expend some bandwidth as a project manager) is in order. For reasons we’ll cover at a later date, I’m working on codifying such a process, based on what has (and hasn’t) worked for me in past and current roles. Here we go!

Step 1: Develop a Structured, Living Learning List

Step 1 is key. When we talk about learning in a digital data context, we’re talking about a never-ending process. This isn’t “Algebra I,” where a syllabus can be developed once, locked down, and then pulled out semester after semester to each new set of incoming students. Rather, we’re talking about a list that will grow over time. Use Excel. Use MS Access. Use a spreadsheet in Google Drive. Or, get fancy, and use Sharepoint or Jive or any of a gazillion knowledge management platforms. It doesn’t really matter. But, having a centralized, living, taggable and trackable list of “learning possibilities” is critical. Otherwise, great ideas can be fleeting and temporal — lost to a tragedy of poor timing and imperfect human memory.

This list is a list of learning opportunities that any stakeholder (core or extended) proposed as being a useful target for testing and analysis. Here’s a start for what should be captured for each item on the list:

  • A title for the learning opportunity
  • The name of the person who submitted it
  • The date it was submitted
  • A description of the question being asked
  • The potential business impact (High/Medium/Low) that answering the question would enable

Those are “core” pieces of information that the person submitting the opportunity needs to provide. In addition, the list needs to include some other fields to be populated over time:

  • The status of the question (open, in work, cancelled/rejected, completed, etc.)
  • The date the question was cancelled or completed
  • What sort of testing or analysis would be required to answer the question (historical data analysis, secondary research, primary research, A/B or multivariate testing, in-market experimentation, etc.)
  • The level of effort / time estimated to answer the question
  • A summary of the results of the analysis (the “answer” to the question) and where the full analysis can be found

Once you’ve settled on where this list will live, who will maintain it, and exactly what fields it will contain, it’s time to move on to…

Step 2: Capture Ideas from Stakeholders

A fairly common delusion in business is that analysts have access to all of the data and have tools at their disposal that will crunch that data in a way such that insights will magically emerge. It doesn’t work that way. Analysis is an exercise in asking smart and valid business questions and then using the specifics of each question to frame and execute an analysis to get an answer.

Analysts are only ONE source for smart and valid business questions!

The reason we set up the list in Step 1 the way we did was so that we can capture more questions than we could possibly ever answer. That’s a fantastic situation in which to be, because it means you have a vast pool of learning opportunities to draw from, rather than scrambling around to find one or two things worth analyzing.

The idea is to make it very easy for any stakeholder who has an idea or a question to quickly and easily get it recorded and available for consideration for analysis: the developer who read a Mashable article that inspired a thought about the current web site, the designer who was torn between the treatment of the global navigation on the site, the marketer who knows an upcoming campaign will be a litmus test as to whether a particular channel is worth the company’s investment or not, etc.

This doesn’t mean this process is all about volume. The “input form” (the first bulleted list above) should force some basic consideration on the part of the submitter to qualify the idea. The fact is, it’s the people in the organization who are most interested in being data-informed, and the people who are most interested in moving the business forward, will be the people who engage in this quasi-crowdsourced learning process.

Step 3: Develop a Means of Prioritizing the Ideas

While Step 2 is intended to be as egalitarian as possible, the act of prioritizing the ideas is, by necessity, much less so. The HiPP  is the HiPP for a reason (I say HiPP here rather than HiPPO because we’re talking about the Highest Paid Person rather than her specific Opinions), and the person who is accountable for the budget that pays for the analyst deserves a greater say in where the analyst spends her time.

First, of course, there needs to be some set of agreed-to criteria for prioritizing the ideas. The specifics will vary, but these will likely include:

  • The likely impact to the business if the question is successfully answered (including how quickly the organization will be able to act on the results and the cost to the organization to do so)
  • The long-term applicability of the results of the exercise
  • The expected time and cost to conduct the analysis

An assessment of these criteria should be captured and recorded in the list developed in Step 1. But, they are inherently subjective assessments, so it still comes back to people-driven decision making as to what gets tested and analyzed and when. There are several ways to tackle this — the ones listed below are all ones that I have used with success in one form or another over the years, but I’m sure there are others:

  • Have a core team of stakeholders regularly review the list of questions and decide which ones to tackle and when (see Step 5)
  • Set up a way for everyone who submitted ideas to also vote on ideas — a “thumbs-up” for an idea moves it up the list such that the more people who give it a boost, the more likely it is to be a question that gets tackled
  • A modified combination of the above is to vary the voting weight of each person; I’ve been through exercises where everyone is given an amount of (fake) money that they get to “invest” in the questions they would like to see answered. They can invest in as few or as many questions as they like. With this approach, the key decision makers / budget owners can be given more money to invest

This is one of the trickier aspects of managing the program, because it requires that the right people are actively engaged with the process and available to participate on a recurring basis (see Step 5).

Step 4: Have a Clear Start/End for Each Analysis

Step 4 and Step 5 are all about process and rigor. Each accepted/approved learning opportunity should be treated as a mini-project and managed as such. In most cases, the analyst can also be the project manager. But, in some cases, the project management may be something that a professional project manager should take on. Key elements of project managing the analysis include:

  • Identifying all of the people who will be involved or impacted in some way (a RASCI matrix is handy for this)
  • Development of a work breakdown structure — a list of all of the steps that will be required to complete the test or analysis, as well as the sequence of that effort
  • Development of a project schedule — when each task on the work breakdown structure will be tackled (in some cases, there will be a “wait and see” period that needs to be built into the schedule; with social media, especially, it’s common to need to explicitly change a tactic for a period of time — a week or two — and then evaluate the results of that change; in these cases, there are periods of time when there is no actual analysis “work” being performed on the project)
  • Establishment and publication of key milestones
  • A plan for communication — who will be updated, when, and how
  • A commitment to regularly checking the project schedule against the actual work completed

“Egad!” you exclaim. “All of THAT is supposed to be done by the analyst?” Well, yes. It doesn’t have to be a huge deal. For many analyses, this can all be done in an Excel spreadsheet with a few recurring Outlook or Google Calendar reminders. In many cases, the test or analysis may be so small that the “schedule” is a single day. But, having the diligence to think through the what, the who, the how, and the when — and then managing the work to the results of that thinking — brings rigor to the work and credibility to the process. It prevents wheel-spinning, and, by feeding back to the list from Step 1, helps build an inventory of discrete, completed work over time that can then be used to assess the overall effectiveness of the analytics program and ensure that what has been learned in the past gets applied in the future.

Step 5: Develop a Fixed Cadence for Updates

As I noted at the beginning of Step 4, that step and this one are complementary. And, in some ways, they are in tension:

  • Step 4 — recognize that each analysis is different and unique and has it’s own schedule; that schedule may be a single day, it may be a week, it may be several months. Treat it as such and manage each effort as a small project
  • Step 5 — establish a fixed cadence for providing communication, updates, and assessment of the overall program

The fixed cadence may be daily (although I’ve never had a case where that is warranted, the Agile development methodology dictates daily stand-ups, so it may be that there are analytics programs where that is warranted), weekly, or even monthly. Having been at an agency for the last three years, monthly was often the most frequent cadence I could manage. This fixed cadence can include the delivery of recurring performance measurement results (KPI-driven dashboards) if those require an in-person review. But, the focus of these communications should be on the learning plan:

  • What questions from the list have been answered since the last update
  • What questions are currently being worked on and how (historical data analysis, A/B test, adjustment of digital tactics for a fixed period of time to measure results, etc.)
  • What new questions have come in for consideration

If this cadence includes a formal meeting with the stakeholders, which is ideal, then a discussion that generates new questions, as well as the prioritization of new questions, can also be part of this meeting.

“Test and Learn” Is the Core of Analysis

In addition to laying out a practical process for effectively driving continuous learning, I hope I have also illustrated that “testing” is inextricably bound with “analysis” and vice versa. We can’t treat testing as being limited to A/B and multivariate testing and analysis as being limited to historical data. To truly learn in a way that delivers business value, the focus has to be on the business questions. Depending on the question, the best way to answer the question should be selected from a comprehensive arsenal at the analyst’s disposal, and the overall process has to be rigorously managed.

Analysis, Social Media

Smart Analytics (Sometimes) = Upset Consumers

The recent media coverage of Orbitz’s OS-based content targeting was intriguing, if not surprising. The Wall Street Journal broke the story about how Mac users were presented with higher priced hotel options than PC users on Orbitz’s site. This was NOT, mind you, a case of any actual difference in prices, but, rather, simply pricier hotels presented in search results. Within a few days, there was a minor social media backlash, with the facts of the situation being misrepresented and Orbitz being accused of nefarious behavior.

Orbitz did something that was smart, was right, and benefited consumers. Yet, because they did it by sniffing out easily detectable information — the operating system of visitors to their site — they were (unfairly) accused of bad behavior.

This is a case where, sadly, perception is reality, and sound bites are the full story. To read my complete thoughts on the subject, check out my post on the Resource Interactive blog.

Analysis, Reporting, Social Media

Imperfect Options: Social Media Impact for eCommerce Sites

I’m now writing a monthly piece for Practical eCommerce, and the experience has been refreshing. At ACCELERATE in Chicago earlier this year, April Wilson‘s winning Super ACCELERATE session focused on digital analytics for smaller companies. Her point was that a lot of the online conversation about “#measure” (or “#msure”) focuses on multi-billion dollar companies and the challenges they have with their Hadoop clusters, while there are millions of small- to medium-sized businesses who have very little time and very limited budgets who need some love from the digital analytics community. To that end, she proposed an #occupyanalytics movement — the “99%” of business owners who can get real value from digital analytics, but who can’t push work to a team of analysts they employ.

Practical eCommerce aims to provide useful information to small- to medium-sized businesses that have an eCommerce site. It’s refreshing to focus on that analytics for that target group!

My latest piece was an exploration of the different ways that managers of eCommerce sites running Google Analytics can start to get a sense of how much of their business can be linked to social media. It touches on some of the very basics — campaign tracking, referral traffic, and the like — but also dips into some of the new social media-oriented reporting in Google Analytics, as well as some of the basics of multi-channel funnels as they related to social media.And, of course, a nod to the value of voice of the customer data. Interested in more? You can read the full article on the Practical eCommerce site.

Analysis, Analytics Strategy, Reporting, Social Media

Four Dimensions of Value from Measurement and Analytics

When I describe to someone how and where analytics delivers value, I break it down into four different areas. They’re each distinct, but they are also interrelated. A Venn diagram isn’t the perfect representation, but it’s as close as I can get: Earlier this year, I wrote about the three-legged stool of effective analytics: Plan, Measure, Analyze. The value areas covered in this post can be linked to that process, but this post is about the why, while that post was about the how.

Alignment

Properly conducted measurement adds value long before a single data point is captured. The process of identifying KPIs and targets is a fantastic tool for identifying when the appearance of alignment among the stakeholders hides an actual misalignment beneath the surface. “We are all in agreement that we should be investing in social media,” may be a true statement, but it lacks the specificity and clarity to ensure that the “all” who are in agreement are truly on the same page as to the goals and objectives for that investment. Collaboratively establishing KPIs and targets may require some uncomfortable and difficult discussions, but it’s a worthwhile exercise, because it forces the stakeholders to articulate and agree on quantifiable measures of success. For any of our client engagements, we spend time up front really nailing down what success looks like from a hard data perspective for this very reason. As a team begins to execute an initiative, being able to hold up a concise set of measures and targets helps everyone, regardless of their role, focus their efforts. And, of course, Alignment is a foundation for Performance Measurement.

Performance Measurement

The value of performance measurement is twofold:

  • During the execution of an initiative, it clearly identifies whether the initiative is delivering the intended results or not. It separates the metrics that matter from the metrics that do not (or the metrics that may be needed for deeper analysis, but which are not direct measures of performance). It signifies both when changes must be made to fix a problem, and it complements Optimization efforts by being the judge as to whether a change is delivering improved results.
  • Performance Measurement also quantifies the results and the degree to which an initiative added value to the business. It is a key tool in driving Internal Learning by answering the questions: “Did this work? Should we do something like this again? How well were we able to project the final results before we started the work?”

Performance Measurement is a foundational component of a solid analytics process, but it’s Optimization and Learning that really start to deliver incremental business value.

Optimization

Optimization is all about continuous improvement (when things are going well) and addressing identified issues (when KPIs are not hitting their targets). Obviously, it is linked to Performance Measurement, as described above, but it’s an analytics value area unto itself. Optimization includes A/B and multivariate testing, certainly, but it also includes straight-up analysis of historical data. In the case of social media, where A/B testing is often not possible and historical data may not be sufficiently available, optimization can be driven by focused experimentation. This is a broad area indeed! But, while reporting squirrels can operate with at least some success when it comes to Performance Measurement, they will fail miserably when it comes to delivering Optimization value, as this is an area that requires curiousity, creativity, and rigor rather than rote report repetition. Optimization is a “during the on-going execution of the initiative” value area, which is quite different (but, again, related) to Internal Learning.

Learning

While Optimization is focused on tuning the current process, Internal Learning is about identifying truths (which may change over time), best practices, and, “For the love of Pete, let’s not make the mistake of doing that again!” tactics. It pulls together the value from all three of the other analytics value areas in a more deliberative, forward-looking fashion. This is why it sits at the nexxus of the other three areas in the diagram at the beginning of this post. While, on the one hand, Learning seems like a, “No, duh!” thing to do, it actually can be challenging to do effectively:

  • Every initiative is different, so it can be tricky to tease out information that can be applied going forward from information that would only be useful if Doc Brown appeared with his Delorean
  • Capturing this sort of information is, ideally, managed through some sort of formal knowledge management process or program, and such programs are quite rare (consultancies excluded)
  • Even with a beautifully executed Performance Management process that demonstrates that an initiative had suboptimal results, it is still very tempting to start a subsequent initiative based on the skeleton of a previous one. Meaning, it can be very difficult to break the, “that’s how we’ve always done it” barrier to change (remember how long it took to get us to stop putting insanely long registration forms on our sites?)

Despite these challenges, it is absolutely worth finding ways to ensure that ongoing learning is part of the analytics program:

  • As part of the Performance Measurement post mortem for a project, formally ask (and document), what aspects, specifically, of the initiative’s results contain broader truths that can be carried forward.
  • As part of the Alignment exercise for any new initiative, consciously ask, “What have we done in the past that is relevant, and what did we learn that should be applied here?” (Ideally, this occurs simply by tapping into an exquisite knowledge management platform, but, in the real world, it requires reviewing the results of past projects and even reaching out and talking to people who were involved with those projects)
  • When Optimization work is successfully performed, do more than simply make the appropriate change for the current initiative — capture what change was made and why in a format that can be easily referenced in the future

This is a tough area that is often assumed to be something that just automatically occurs. To a certain extent, it does, but only at an individual level: I’m going to learn from every project I work on, and I will apply that learning to subsequent projects that I work on. But, the experience of “I” has no value to the guy who sits 10′ away if he is currently working on a project where my past experiences could be of use if he doesn’t: 1) know I’ve had those experiences, or 2) have a centralized mechanism or process for leveraging that knowledge.

What Else?

What do you say when someone asks you, “How does analytics add value?” Do you focus on one or more of the areas above, or do you approach the question from an entirely different perspective? I’d love to hear!

Analysis, Analytics Strategy, Reporting

Digital Analytics: From Data to Stories and Communication

This will be a quick little post as I try to pull together what seems to be an emerging theme in the digital analytics space. In a post late last year, I wrote:

I haven’t attended a single conference in the last 18 months where one of the sub-themes of the conference wasn’t, “As analysts, we’ve got to get better at telling stories rather than simply presenting data.

Lately, though, it seems that the emphasis on “stories” has shifted to a more fundamental focus on “communication.” As evidence, I present the following:

A 4-Part Blog Series

Michele Kiss published a 4-part blog series over the course of last week titled “The Most Undervalued Analytics Tool: Communication.” The series covered communication within your analytics teamcommunication across departments, communication with executives and stakeholders, and communication with partners. Whether intentionally or not, the series highlighted how varied and intricate the many facets of “communication” really are (and she makes some excellent tips for addressing those different facets!).

A Data Scientist’s “Day to Day” Advice

Christopher Berry, VP of Marketing Science at Syncapsealso published a post last week that touched on the importance of communication. Paraphrasing (a bit), he advised:

  • Recognize that you’re going to have to repeat yourself — not because the people your communicating with are stupid, but because they’re not as wired to the world of data as you are
  • Communicate to both the visual and auditory senses — different people learn better through different channels (and neuroscience has shown that ideas stick better when they’re received through multiple sensory registers)
  • Use bullet points (be concise)

Christopher is one of those guys who could talk about the intricacies of shoe leather and have an audience spellbound…so his credibility on the communication front comes more from the fact that he’s a great communicator than from his position as a top brain in the world of data scientistry.

Repetition at ACCELERATE

During last Wednesday’s ACCELERATE conference in Chicago, I tweeted the following:

The tweet was mid-afternoon, and it was after a run of sessions — all very good — where the presenters directly spoke to the importance of communication when it come to a range of analytics responsibilities and challenges.

A Chat with Jim Sterne

At the Web Analytics Wednesday that followed the conference, I got my first chance (ever!) to have more than a 2-sentence conversation with Jim Sterne (I’m pretty sure the smile on his face all day was the smile of a man who was attending a conference as a mere attendee than as a host and organizer, and the plethora of attendant stresses of that role!).

During that discussion, Jim asked me the question, “What is it that you are doing now that is moving towards [where you want to be with your career].” We’ll leave the details of the bracketed part of my quote aside and focus on my answer, which I’d never really thought of in such explicit terms. My answer was that, being a digital analyst at an agency that was built over the course of 3 decades on a foundation of great design work and outstanding consumer research (as in: NOT on measurement and analytics), I have to keep honing my communication skills. In many, many ways I have a conversation every day where I am trying to communicate the same basics about digital analytics that I’ve been communicating for the past decade in different environments. But, I’m not just repeating myself. If I look back over my 2.5 years at the agency, I’ve added a new “tool” to my analytics communication toolbox every 2-3 months, be it a new diagram, a new analogy, a new picture, or a new anecdote. I’ve been working really hard (albeit not explicitly or even consciously) to become the most effective communicator I can be on the subject of digital analytics. Not every new tool sticks, and I try to discard them readily when I realize they’re not resonating.

It’s a work in progress. Are you consciously working on how you communicate as an analyst? What’s your best tip?

Analysis, Analytics Strategy, Social Media

The Many Dimensions of Social Media Data

I’ve been thinking a bit of late about the different aspects of social media data. This was triggered by a few different things:

  • Paul Phillips of Causata spoke at eMetrics in San Francisco, and his talk was about leveraging data from customer touchpoints across multiple channels to provide better customer relationship management
  • I’ve been re-reading John Lovett’s Social Media Metrics Secrets book as part of an internal book group at Resource Interactive
  • We’ve had clients approaching us with some new and unique questions related to their social media efforts

What’s become clear is that “social media analytics” is a broad and deep topic, and discussions quickly run amok when there isn’t some clarity as to which aspect of social media analytics is being explored.

As I see it, there are four broad buckets of ways that social media data can be put to use by companies:

No company that is remotely serious about social media in 2012 can afford to ignore the top two boxes. The bottom two are much more complex and, therefore, require a substantial investment, both in people and technology.

Now, I could stop here and actually have a succinct post. But, why break a near-perfect (or consistently imperfect) streak? Let’s take a slightly deeper look at each bucket.

Operational Execution

(I almost labeled this bucket “Community Management,” but the variety of viewpoints in the industry on the scope of that role convinced me to leave that can of worms happily sealed for the purposes of this post.)

Social media requires a much more constant intake and rapid response/action based on data than web sites typically do. Having the appropriate tools, processes, and people in place to respond to conversations with appropriate (minimal) latency is key.

Key challenges to effectively managing this aspect of social media data include: determining a reasonable scope, being realistic about the available on-going people who will manage the process, and, to a lesser extent, selecting the appropriate set of tools. Tool selection is challenging because this is the area where the majority of social media platforms are choosing to play — from online listening platforms like Radian6, Sysomos, Alterian, and Syncapse; to “social relationship management” platforms like Vitrue, Buddy Media, Wildfire, (Adobe) Context Optional, and Shoutlet; and even to the low-cost platforms such as Hootsuite and TweetDeck. These platforms have a range of capabilities, and their pricing models vary dramatically.

Performance Measurement

Ahhh, performance measurement. When it comes to social media, it definitely falls in the “simple, but not easy” bucket. And, it’s an area where marketers are perpetually dissatisfied when they discover that there is no “value of a fan” formula, nor is there “the ROI of a tweet.” But, any marketer who has the patience to step back and consider where social media plays in his/her business can absolutely do effect performance measurement and report on meaningful business results!

Chapters 4 and 5 of John Lovett’s book, Social Media Metrics Secrets, get to the heart of social media performance measurement by laying out possible social media objectives and appropriate KPIs therein. High on my list is to make it through Olivier Blanchard’s Social Media ROI: Managing and Measuring Social Media Efforts in Your Organization, as I’m confident that his book is equally full of usable gems when it comes to quantifying the business value delivered from social media initiatives.

When it comes to technologies for social media performance measurement, we generally find ourselves stuck trying to make use of the Operational Execution platforms. They all tout their “powerful analytics,” but their product roadmaps have typically been driven more by “listenting” and “publishing” features than they have been driven by “metrics” capabilities. With Google’s recent announcement ofGoogle Analytics Social Reports, and with Adobe’s recent announcement of Adobe Social, this may be starting to change.

(Social-Enhanced) CRM

Leveraging social media data to improve customer relationship management is something that there has been lots of talk about…but that very few companies have successfully implemented. At its most intriguing, this means companies identifying — through explicit user permission or through mining the social web — which Twitter users, Facebook fans, Pinterest users, Google+ users, and so on can be linked to their internal systems. Then, by listening to the public conversations of those users and combining that information with internally-captured transactional data (online purchases, in-store purchases, loyalty program membership, email clickthroughs, etc.), getting a much more comprehensive view of their customers and prospects. That “more comprehensive view,” in theory, can be used to build much more robust predictive models that can let the brand know how, when, and with what content to engage individual customers to maximize the value of that relationship for the brand.

The challenges are twofold:

  • Consumer privacy concerns — even if a brand doesn’t do anything illegal, consumers and the press have a tendency to get alarmed when they realize how non-anonymous their relationship with the brand is (as Target learned…and they weren’t even using social media data!)
  • Complexity and cost — there is a grave tendency for marketers to confuse “freely available data” with “data that costs very little to gather and put to good use.” Companies’ customer data is data they have collected through controllable interactions with consumers — through a form they filled out on the web, through a credit card being run as part of a purchase, through a call into the service center, etc. Data that is pulled from social media platforms is at the whim of the platforms and the whim of the consumer who set up the account. No company (except Twitter) can go out to a Twitter account and, in an automated fashion, bring back the user’s email address, real name, gender, or even country of residence. It takes much more sophisticated data crawling, combined with probabilistic matching engines, to get this data.

Despite these challenges, this is an exciting opportunity for brands. And, the technology platforms are starting to emerge, with the three that spring the most quickly to my mind being Causata, iJento, and Quantivo.

Trend / Opportunity Prediction

This is another area that is really tough to pull off, but it’s an area that, admittedly, has great potential. It’s a “Big Data” play if ever there was one — along the lines of how the Department of Homeland Security supposedly harnesses the data in millions of communications streams to identify terrorist hot spots. It’s sifting through a haystack and not knowing whether your’re looking for a needle, a twig, a small piece of wire, or a paperclip, but knowing that, if you find any of them, you’ll be able to put it to good use.

The wistfully optimistic marketing strategist describes this area something like this: “I want to pick up on patterns and trends in the psychographic and attitudinal profile of my target consumers that emerge in a way that I can reasonably shift my activities. I want an ‘alert’ that tells me, ‘There’s something of interest here!'”

It’s a damn vague dream…but that doesn’t mean it’s unrealistic. It’s a multi-faceted challenge, though, because it requires the convergence of some rather sticky wickets:

  • Identifying conversations that are occurring amongst people who meet the profile of a brand’s target consumers (demographic, psychographic, or otherwise) — yet, social media profiles don’t come with a publicly available list of the user’s attitudes, beliefs, purchasing behavior, age, family income, educational level, etc.
  • Identifying topics within those conversations that might be relevant for the brand — we’re talking well beyond “they’re talking about what the brand sells” and are looking for content with a much, much fuzzier topical definition
  • Identifying a change in these topics — generally, what marketers want most is to pick up on an emerging trend rather than simply a long-held truism

To pull this off will require a significant investment in technology and infrastructure, a significant investment in a team of people with specialized skills, and a significant amount of patience. I chuckle every time I hear an anecdote about how a brand managed to pick up on some unexpected opportunity in real time and then quickly respond…without a recognition that the brand was spending an awful lot of time listening in real-time and picking up nothing of note!

This area, I think, is what a lot of the current buzz around Big Data is focused on. I’m hoping there are enough companies investing in trying to pull it off that we get there in the next few years, because it will be pretty damn cool. Maybe IBM can set Watson up with a Digital Marketing Optimization Suite login and see what he can do!

Analysis, Reporting

3-Legged Stool of Effective Analytics: Plan, Measure, Analyze

Several weeks ago, Stéphane Hamel wrote a post that got me all re-smitten with his thought process. In the post, he postulated that there are three heads of online analytics. He covered three different skillsets needed to effectively conduct online analytics: business acumen, technical (tools) knowledge, and analysis. And, he made the claim that no one person will ever excel at all three, which led to his case for building out teams of “analysts” who have complementary strengths.

I’ve had several unrelated experiences with different clients and internal teams of late that have led me to try to capture, in a similar fashion, the three-legged stool of an online analytics program. Just as others have started tacking on additional components to Stéphane’s three skillsets, I’m sure my three-legged stool will quickly become a traditional chair…then some sort of six-legged oddity. But, I’d be thrilled if I could consistently communicate the basics to my non-analyst co-workers and clients:

I hold to a pretty strict distinction between “measurement and reporting” and “analysis,” and I firmly believe there is value in “reporting,” as long as that reporting is appropriately set up and applied.

Just as I believe that reporting should generally occur either as a one-time event (campaign wrap-up, for instance) or at regular intervals, I firmly believe that testing and analysis should not be forced into a recurring schedule. It’s fine (desirable) to be always conducting analysis, but the world of “present the results of your analysis — and your insights and recommendations therein — once/month on the first Wednesday of the month” is utterly asinine. Yet…it’s a mindset with which a depressing majority of companies operate.

Reporting Done Poorly…Which Is an Unfortunately Ubiquitous Habit

I’ve been client side. I’ve been agency side. I’ve done a decent amount of reading on human nature as it relates to organizational change. My sad conclusion:

The business world has conditioned itself to confuse “cumbersome decks of data” with “reporting done well.”

It happens again and again. And again. And…again! It goes like this:

  1. Someone asks for some data in a report
  2. Someone else pulls the data
  3. The data raises some additional questions, so the first person asks for more data.
  4. The analyst pulls more data
  5. The initial requestor finds this data useful, so he/she requests that the same data be pulled on a recurring schedule
  6. The analyst starts pulling and compiling the data on a regular schedule
  7. The requestor starts sharing the report with colleagues. The colleagues see that the report certainly should be useful, but they’re not quite sure that it’s telling them anything they can act on. They assume that it’s because there is not enough data, so they ask the analyst to add in yet more data to the report
  8. The report begins to grow.
  9. The recipients now have a very large report to flip through, and, frankly, they don’t have time month in and month out to go through it. They assume their colleagues are, though, so they keep their mouths shut so as to not advertise that the report isn’t actually helping them make decisions. Occasionally, they leaf through it until they see something that spikes or dips, and they casually comment on it. It shows that they’re reading the report!
  10. No one tells the analyst that the report has grown too cumbersome, because they all assume that the report must be driving action somewhere. After all, it takes two weeks of every month to produce, and no one else is speaking up that it is too much to manage or act on!
  11. The analyst (now a team of analysts) and the recipients gradually move on to other jobs at other companies. At this point, they’re conditioned that part of their job is to produce or receive cumbersome piles of data on a regular basis. Over time, it actually seems odd to not be receiving a large report. So, if someone steps up and asks the naked emperor question: “How are you using this report to actually make decisions and drive the business?”…well…that’s a threatening question indeed!

In the services industry, there is the concept of a “facilitated good.” If you’re selling brainpower and thought, the theory goes, and you’re billing out smart people at a hefty rate, then you damn well better leave behind a thick binder of something to demonstrate that all of that knowledge and consultation was more than mere ephemera!

And, on the client side, if the last 6 consultancies and agencies that you worked with all diligently delivered 40-slide PowerPoint decks or 80-page reports, then, by golly, you’re going to look askance at the consultant who shows up and aims for actionable concision!

Nonetheless, I will continue my quixotic quest to bring sanity to the world. So, onto the three legs of my analytics stool…

First, Plan (Dammit!!!)

Get a room full of experienced analysts together and ask them where any good analytics program or initiative starts, and you’ll get a unanimous response that it starts: 1) at the beginning of the initiative, and 2) with some form of rigorous planning.

The most critical question to answer during analytics planning is: “How are we going to know if we’re successful?” Of course, you can’t answer that question if you haven’t also answered the question: “What are we trying to accomplish?” Those are the two questions that I wrote about in this Getting to Great KPIs post.

Of course, there are other components of analytics planning:

  • Where will the data come from that we’ll use?
  • What other metrics — beyond the KPIs — will we need to capture?
  • What additional data considerations need to be factored into the effort to ensure that we are positioned for effective analysis and optimization down the road?
  • What (if any) special tagging, tracking, or monitoring do we need to put into place (and who/how will that happen)?
  • What are the known limitations of the data?
  • What are our assumptions about the effort?
  • …and more

In my experience both agency-side and client-side, this step regularly gets skipped like it’s a smooth round, rock in the hand of an adolescent male standing on the shore of a lake on a windless day.

An offshoot of the planning is the actual tagging/tracking/monitoring configuration…but I consider that an extension of the planning, as it may or may not be required, depending on the nature of the initiative.

Next, Measure and Report

Yup. Measurement’s important. That’s how you know if you’re performing at, above, or below your KPIs:

Here’s where I start to get into debates, both inside the analytics industry and outside. I strongly believe that it is perfectly acceptable to deliver reports without accompanying insights and analysis. Ideally, reports are automated. If they’re not automated, they’re produced damn quickly and efficiently.

Dashboards — the most popular form of reports — have a pretty simple purpose: provide an at-a-glance view of what has happened since the last update, and ensure that, at a glance, any anomalies jump out. More often than not, there won’t be anomalies, so there is nothing that needs to be analyzed based on the report! That’s okay!

I was discussing this concept with a co-worker recently, and, in response to my claim that reports should simply get delivered with minimal latency and, at best, a note that says, “Hey, I noticed this apparent anomaly that might be important. I’m going to look into it, but if you (recipient) have any ideas as to what might be going on, I’d love to get your thoughts,” she responded:

I think this makes sense, but wouldn’t we provide some analysis as to the “why” on the monthly reports?

I immediately went to the “dashboard in your car” analogy (I know — it breaks down on a lot of fronts, but it works here) with my response:

You don’t look at your fuel gauge when you get in the car every day and ask, “Why is the needle pointing where it is?” You take a quick look, make sure it’s not pegged on empty, and then go about your day.

That’s measurement. It may spawn analysis, but, often, it does not. And that’s to be expected!

Which Brings Us to Testing and Analysis

Analysis requires (or, at least, is much more likely to yield value in an efficient manner) having conducted some solid planning and having KPI-centric measurement in place. But, the timing of analysis shouldn’t be forced into a fixed schedule.

The bottom part of the figure above gets to the crux of the biscuit when it comes to timing: sometimes, the best way to answer a business question is through analyzing historical data. Sometimes, the best way to answer a question is through go-forward testing. Sometimes, it’s a combination of the two (develop a theory based on the historical data, but then test it by making a change in the future and monitoring the results). Sometimes the analysis can be conducted very quickly. Other times, the analysis requires a large chunk of analyst time and may take days or weeks to complete.

Facilitating the collaboration with the various stakeholders, managing the analysis projects (multiple analyses in flight at once — starting and concluding asynchronously based on each effort’s unique nature), can absolutely fall under the purview of the analyst (again referencing Stéphane’s post, this should be an analyst with a strong “head” for business acumen).

In Conclusion…(I promise!)

There is a fundamental flaw in any approach to using data that attempts to bundle scheduled reporting with analysis. It forces efforts to find “actionable insights” in a context where there may very well be none. And, it perpetuates an assumption that it’s simply a matter of pointing an analyst at data and waiting for him/her to find insights and make recommendations.

I’ve certainly run into business users who flee from any effort to engage directly when it comes to analytics. They hide behind their inboxes lobbing notes like, “You’re the analyst. YOU tell me what my business problem is and make recommendations from your analysis!” I’m sure some of these users had one too many (and one is “too many”) interactions with an analyst who wanted to explain the difference between a page view and a visit, or who wanted to collaboratively sift through a 50-page deck of charts and tables. That’s not good, and that analyst should be flogged (unless he/she is less than two years out of college and can claim to have not known any better). But, using data to effectively inform decisions is a collaborative effort. It needs to start early (planning), it needs to have clear, concise performance measurement (KPI-driven dashboards), and it needs to have flexibility to drive the timing and approach of analyses that deliver meaningful results.

Analysis, Reporting, Social Media

Digital and Social Measurement Based on Causal Models

Working for an agency that does exclusively digital marketing work, with a heavy emphasis on emerging channels such as mobile and social media, I’m constantly trying to figure out the best way to measure the effectiveness of the work we do in a way that is sufficiently meaningful that we can analyze and optimize our efforts.

Fairly regularly, I’m drawn into work where the team has unrealistic expectations of the degree to which I can accurately quantify the impact of their initiatives on their top (or bottom) line. I’ve come at these discussions from a variety of angles:

This post is largely an evolution of the last link above. It’s something I’ve been exploring over the past six months, and which was strongly reinforced when I read John Lovett’s recent book. As I’ve been doing measurement planning (measurement strategy? marketing optimization planning?) with clients, it’s turned out to be quite useful when I have the opportunity to apply it.

Initially, I referred to this approach as developing a “logical model” (that’s even what I called it towards the end of my second post that referenced John’s book), but that was a bit bothersome, since “logical model” has a very specific meaning in the world of database design. Then, a couple of months ago, I stumbled on an old Harvard Business Review paper about using non-financial measures for performance measurement, and that paper introduced the same concept, but referred to it as a “causal model.” I like it!

How It Works

The concept is straightforward, it’s not particularly time-consuming, it’s a great exercise for ensuring everyone involved is aligned on why a particular initiative is being kicked off, it sets up meaningful optimization work as individual tactics and campaigns are implemented, and it positions you to be able to demonstrate a link (correlation) between marketing activities and business results.

This approach acknowledges that there is no existing master model that shows exactly how a brand’s target consumers interact and respond to brand activity. The process starts with more “art” than “science” — knowledge of the brand’s target consumers and their behaviors, knowledge of emerging channels and where they’re most suited (e.g., a QR code on a billboard on a busy highway…not typically a good match), and a hefty dose of strategic thought.

The exact structure of this sort of model varies widely from situation from situation, but I like to have my measurable objectives — what we think we’re going achieve through the initiative or program that we believe has underlying business value — listed on the left side of the page, and then build linkages from that to a more definitive business outcome on the right:

It should fit on a single page, and it requires input from multiple stakeholders. Ultimately, it can be a simple illustration of “why we’re doing this” for anyone to review and critique. If there are some pretty big leaps required, or if there are numerous steps along the way to get to tangible business value, then it begs the question: “Is this really worth doing?” It’s an easy litmus test as to whether an initiative makes sense.

What I’ve found is that this exercise can actually alter the original objectives in the planning stage, which is a much better time and place to alter them than once execution is well under way!

Once the model is agreed to, then you can focus on measuring and optimizing to the outputs from the base objectives — using KPIs that are appropriate for both the objective and the “next step” in the causal model.

And, over time, the performance of those KPIs can be correlated with the downstream components of the causal model to validate (and adjust) the model itself.

This all gets back to the key that measurement and analytics is a combination of art and science. Initially, it’s more art than science — the science is used to refine, validate, and inform the art.

Analysis, Analytics Strategy, Reporting

The Analyst Skills Gap: It's NOT Lack of Stats and Econometrics

I wrote the draft of this post back in August, but I never published it. With the upcoming #ACCELERATE event in San Francisco, and with what I hope is a Super Accelerate presentation by Michael Healy that will cover this topic (see his most recent blog post), it seemed like a good time to dust off the content and publish this. If it gives Michael fodder for a stronger takedown in his presentation, all the better! I’m looking forward to having my perspective challenged (and changed)!

A recent Wall Street Journal article titled Business Schools Plan Leap Into Data covered the recognition by business schools that they are sending their students out into the world ill-equipped to handle the data side of their roles:

Data analytics was once considered the purview of math, science and information-technology specialists. Now barraged with data from the Web and other sources, companies want employees who can both sift through the information and help solve business problems or strategize.

That article spawned a somewhat cranky line of thought. It’s been a standard part of presentations and training I’ve given for years that there is a gap in our business schools when it comes to teaching students how to actually use data. And, the article includes a quote from an administrator at the Fordham business school: “Historically, students go into marketing because they ‘don’t do numbers.'” That’s an accurate observation. But, what is “doing numbers?” In the world of digital analytics, it’s a broad swath of activities:

  • Consulting on the establishment of clear objectives and success measures (…and then developing appropriate dashboards and reports)
  • Providing regular performance measurement (okay, this should be fully automated through integrated dashboards…but that’s easier said than done)
  • Testing hypotheses that drive decisions and action using a range of analysis techniques
  • Building predictive models to enable testing of different potential courses of action to maximize business results
  • Managing on-going testing and optimization of campaigns and channels to maximize business results
  • Selecting/implementing/maintaining/governing data collection platforms and processes (web analytics, social analytics, customer data, etc.)
  • Assisting with the interpretation/explanation of “the data” — supporting well-intended marketers who have found “something interesting” that needs to be vetted

This list is neither comprehensive nor a set of discrete, non-overlapping activities. But, hopefully, it illustrates the point:

The “practice of data analytics” is an almost impossibly broad topic to be covered in a single college course.

What bothered me about the WSJ article are two things:

  • The total conflation of “statistics” with “understanding the numbers”
  • The lack of any recognition of how important it is to actually be planning the collection of the data — it doesn’t just automatically show up in a data warehouse

On the first issue, there is something of an on-going discussion as to what extent statistics and predictive modeling should be a core capability and a constantly applied tool in the analyst’s toolset. Michael Healy made a pretty compelling case on this front in a blog post earlier this year — making a case for statistics, econometrics, and linear algebra as must-have skills for the web analyst. As he put it:

If the most advanced procedure you are regularly using is the CORREL function in Excel, that isn’t enough.

I’ve…never used the CORREL function in Excel. It’s certainly possible that I’m a total, non-value-add reporting squirrel. Obviously, I’m not going to recognize myself as such if that’s the case. I’ve worked with (and had work for me) various analysts who have heavy statistics and modeling skills. And, I relied on those analysts when conditions warranted. Generally, this was when we were sifting through a slew of customer data — profile and behavioral — and looking for patterns that would inform the business. But this work accounted for a very small percentage of all of the work that analysts did.

I’m a performance measurement guy because, time and again, I come across companies and brands that are falling down on that front. They wait until after a new campaign has launched to start thinking about measurement. They expect someone to deliver an ROI formula after the fact that will demonstrate the value they delivered. They don’t have processes in place to monitor the right measures to trigger alarms if their efforts aren’t delivering the intended results.

Without the basics of performance measurement — clear objectives, KPIs, and regular reporting — there cannot be effective testing and optimization. In my experience, companies that have a well-functioning and on-going testing and optimization program in place are the exception rather than the rule. And, companies that lack the fundamentals of performance management that try to jump directly to testing and optimization find themselves bogged down when they realize they’re not entirely clear what it is they’re optimizing to.

Diving into statistics, econometrics, and predictive modeling in the absence of the fundamentals is a dangerous place to be. I get it — part of performance measurement and basic analysis is understanding that just because a number went “up” doesn’t mean that this wasn’t the result of noise in the system. Understanding that correlation is not causation is important — that’s an easy concept to overlook, but it doesn’t require a deep knowledge of statistics to sound an appropriately cautionary note on that front. 9 times out of 10, it simply requires critical thinking.

None of this is to say that these advanced skills aren’t important. They absolutely have their place. And the demand for people with these skills will continue to grow. But, implying that this is the sort of skill that business schools need to be imparting to their students is misguided. Marketers are failing to add value at a much more basic level, and that’s where business schools need to start.

Analysis, Reporting, Social Media

"Demystifying" the Formula for Social Media ROI (there isn't one)

I raved about John Lovett’s new book, Social Media Metrics Secrets in an earlier post, and, while I make my way through Marshall Sponder’s Social Media Analytics book that arrived on bookshelves at almost exactly the same time, I’ve also been working on putting some of Lovett’s ideas into action.

One of the more directly usable sections of the book is in Chapter 5, where Lovett lays out pseudo formulas for KPIs for various possible (probable) social media business objectives. This post started out to be about my experiences drilling down into some of those formulas…but then the content took a turn, and one of Lovett’s partners at Analytics Demystified wrote a provocative blog post…so I’ll save the formula exploration for a subsequent post.

Instead…Social Media ROI

Lovett explicitly notes in his book that there is no secret formula for social media ROI. In my mind, there never will be — just as there will never be unicorns, world peace, or delicious chocolate ice cream that is as healthy as a sprig of raw broccoli, no matter how much little girls and boys, rationale adults, or my waistline wish for them.

Yes, the breadth of social media data available is getting better by the day, but, at best, it’s barely keeping pace with the constant changes in consumer behavior and social media platforms. It’s not really gaining ground.

What Lovett proposes, instead of a universally standard social media ROI calculation, is that marketers be very clear as to what their business objectives are – a level down from “increase revenue,” “lower costs,”and “increase customer satisfaction” – and then work to measure against those business objectives.

The way I’ve described this basic approach over the past few years is using the phrase “logical model,” – as in, “You need to build a logical link from the activity you’re doing all the way to ultimate business benefit, even if you’re not able to track those links all the way along that chain. Then…measure progress on the activity.”

Unfortunately, “logical model” is a tricky term, as it already has a very specific meaning in the world of database design. But, if you squint and tilt you’re head just a bit, that’s okay. Just as a database logical model is a representation of how the data is linked and interrelated from a business perspective (as opposed to the “physical model,” which is how the data actually gets structured under the hood), building a logical model of how you expect your brand’s digital/social activities to ladder up to meaningful business outcomes is a perfectly valid  way to set up effective performance measurement in a messy, messy digital marketing world.

No Wonder These Guys Work Together

Right along the lines of Lovett’s approach comes one of the other partners at Analytics Demystified with, in my mind, highly complementary thinking. Eric Peterson’s post about The Myth of the “Data-Driven Business” postulates that there are pitfalls a-looming if the digital analytics industry continues to espouse “being totally data-driven” as the penultimate goal. He notes:

…I simply have not seen nearly enough evidence that eschewing the type of business acumen, experience, and awareness that is the very heart-and-soul of every successful business in favor of a “by the numbers” approach creates the type of result that the “data-driven” school seems to be evangelizing for.

What I do see in our best clients and those rare, transcendent organizations that truly understand the relationship between people, process, and technology — and are able to leverage that knowledge to inform their overarching business strategy — is a very healthy blend of data and business knowledge, each applied judiciously based on the challenge at hand. Smart business leaders leveraging insights and recommendations made by a trusted analytics organization — not automatons pulling levers based on a hit count, p-value, or conversion rate.

I agree 100% with his post, and he effectively counters the dissenting commenters (partial dissent, generally – no one has chimed in yet fully disagreeing with him). Peterson himself questions whether he is simply making a mountain out of a semantic molehill. He’s not. We’ve painted ourselves into corners semantically before (“web analyst” is too confining a label, anyone…?). The sooner we try to get out of this one, the better — it’s over-promising / over-selling / over-simplifying the realities of what data can do and what it can’t.

Which Gets Back to “Is It Easy?”

Both Lovett’s and Peterson’s ideas ultimately go back to the need for effective analysts to have a healthy blend of data-crunching skills and business acumen. And…storytelling! Let’s not forget that! It means we will have to be communicators and educators — figuring out the sound bites that get at the larger truths about the most effective ways to approach digital and social media measurement and analysis. Here’s my quick list of regularly (in the past…or going forward!) phrases:

  • There is no silver bullet for calculating social media ROI — the increasing fragmentation of the consumer experience and the increasing proliferation of communication channels makes it so
  • We’re talking about measuring people and their behavior and attitudes — not a manufacturing process; people are much, much messier than widgets on a production line in a controlled environment
  • While it’s certainly advisable to use data in business, it’s more about using that data to be “data-informed” rather than aiming to be “data-driven” — experience and smart thinking count!
  • Rather than looking to link each marketing activity all the way to the bottom line, focus on working through a logical model that fits each activity into the larger business context, and then find the measurement and analysis points that balance “nearness to the activity” with “nearness to the ultimate business outcome.”
  • Measurement and analytics really is a mix of art and science, and whether more “art” is required or more “science” is required varies based on the specific analytics problem you’re trying to solve

There’s my list — cobbled from my own experience and from the words of others!

Analysis, Analytics Strategy, Social Media

What I Learned About #measure & Google+ from a Single Blog Post

Quite unintentionally, I stirred up a lengthy discussion last week with a blog post where I claimed that web analytics platforms were fundamentally broken. In hindsight, the title of the post was a bit flame-y (not by design — I dashed off a new title at the last minute after splitting up what was one really long post into two posts; I’m stashing the second post away for a rainy day at this point).

To give credit where credit is due, the discussion really took off when Eric Peterson posted an excerpt and a link in Google+ and solicited thoughts from the Google+/#measure community. That turned into the longest thread I’ve participated in to date on Google+, and subsequently led to a Google+ hangout that Eric set up and then moderated yesterday.

This post is an attempt to summarize the highlights of what I saw/heard/learned over the past week.

What I Learned about the #measure Community

Overall, the discussion brought back memories of some of the threads that would occasionally get started on the webanalytics Yahoo! group back in the day. That’s something we’ve lost a bit with Twitter…but more on that later.

What I took away about the group of people who make up the community was pretty gratifying:

  • A pretty united “we” — everyone who participated in the discussions was contributing with the goal of trying to move the discussion forward; as a community, everyone agrees that we’re at some sort of juncture where “web analytics” is an overly limiting label, where the evolution of consumer behavior (read: social media and mobile) and consumer attitudes (read: privacy) are impacting the way we will do our job in the future, and where the world of business is desperately trying to be more data-driven…and floundering more often than succeeding. There are a lot of sharp minds who are perfectly happy to share every smart thought they’ve got on the subject if it helps our industry out — the ol’ “a rising tide lifts all boats” scenario. That’s a fun community with whom to engage.
  • Strong opinions but small egos — throughout the discussion that occurred both on Google+ and on Twitter (as well as in several blog posts that the discussion spawned, like this one by Evan LaPointe and Nancy Koon’s inaugural one and Eric’s post), there were certainly differing points of view, but things never got ugly; I actually had a few people reach out to me directly to make sure that their thoughts hadn’t been taken the wrong way (they hadn’t been)
  • 100s of years of experience — we have a lot of experience from a range of backgrounds when it comes to trying to figure out the stickiest of the wickets that we’re facing. That is going to serve us well.
  • (Maybe) Agencies and vendors leading the way? — I don’t know that I learned this for sure, but an informal tally of the participants in the discussion showed a heavy skewing towards vendor and agency (both analytics agencies and marketing/creative/advertising agencies) representation with pretty limited “industry” participation. On the one hand, that is a bit concerning. On the other hand, having been in “industry” for more of my analytics career than I’ve been on the agency side, it makes sense that vendors and agencies are exposed to a broader set of companies facing the same challenges, are more equipped to see the patterns in the challenges the analytics industry is facing, and are being challenged from more directions to come up with answers to these challenges sooner rather than later.

These were all good things to learn — the people in the community are one of the reasons I love my job, and this thread demonstrated some of the reasons why that is.

Highlights of the Discussion

Boiling down the discussion is bound to leave some gaps, and, if I started crediting individuals with any of the thoughts, I’d run the serious risk of misrepresenting them, so feel free to read the Google+ thread yourself in its entirety (and the follow-up thread that Eric started a few days later). I’ve called out any highlights that came specifically from the hangout as being from there (participants there were Adam GrecoJohn LovettJoseph StanhopeTim WilsonMichael HelblingJohn RobbinsEmer KirraneLee IsenseeKeith Burtis, and me), since there isn’t a reviewable transcript for that.

Here goes:

  • Everyone recognizes that a “just plug it in and let the technology spit out insights” solution will likely never exist — the question is how much of the technical knowledge (data collection minutia, tool implementation nuances, reporting/analysis interface navigation) can be automated/hidden. A couple of people (severalpublicly, one privately) observed that we want (digital) analytics platforms to be a like a high-performance car — all the complexity as needed under the hood, but high reliability and straightforward to operate. Pushing that analogy — how far and fast it runs will still be highly dependent on the person behind the wheel (the analyst).
  • Adobe/Omniture and Google Analytics had near-simultaneous releases of their latest versions; both companies touted the new features being rolled out…but both companies have stressed that there was a lot more about the releases that were under-the-hood changes that were positioning the products for greater advances in subsequent releases; time will tell, no? And, several people who have actually been working  with SC15 (I’ve only seen a couple of demos, watched some videos, and read some blog posts — the main Omniture clients I support are over a year out from seeing SC15 in production), have pointed out that some of the new features (Processing Rules and Context Data, specifically) will really make our lives better
  • There was general consensus that Omniture has gotten much, much better over the years about listening to customer feedback and incorporating changes based on that feedback; there is still a Big Question as to whether customer-driven incremental improvements (even improvements that require significant updates on the back end) will get to true innovation — the “last big innovations” in web analytics were pointed out as being a decade ago (I would claim that the shift from server logs and image beacons to Javascript-based page tags was innovative and wasn’t much older) — or whether “something else” will have to happen was a question that did not get resolved
  • Getting beyond “the web site” is one major direction the industry is heading — integrating cross-channel data and then getting value from it — introduces a whole other level of complexity…but the train is barrelling along on a track that has clearly been laid in that direction
  • We all get sucked into “solving the technical problem” over “focusing on the business results” — the tools have enough complexity that we count it a “win” when we solve the technical issues…but we’re not really serving anyone well when we stop there; this is one of those things, I suspect, that we all know and we constantly try to remind ourselves…and yet still get sucked into the weeds of the technology and forget to periodically lift our heads up and make sure we’re actually adding value; John Lovett has been preaching about this conundrum for years (and he hits on it again in his new book)
  • Marketing/business are getting increasingly complex, which means the underlying data is getting more complex (and much more plentiful — another topic John touches on in his book), which means getting the data into a format that supports meaningful analysis is getting tougher; trying to keep up with that trend is hard enough without trying to get ahead!
  • Tag management — is it an innovation, or is it simply a very robust band-aid? Or is it both? No real consensus there.
  • Possible areas where innovation may occur: cross-channel integration, optimization, improved conversion tracking (which could encompass both of the prior two areas), integration of behaviora/attitudinal/demographic data
  • [From the hangout] “Innovation” is a pretty loaded term. Are we even clear on what outcome we’re hoping to drive from innovation?
  • [From the hangout] Privacy, privacy, privacy! Is it possible to educate the consumer and/or shift the consumer’s mindset such that they are informed about why that “tracking” them isn’t evil? Can we kill the words “tracking” and “targeting,” which both freak people out? Why are consumers fine with allowing the mobile or Facebook application access to their private data…but freak out about no-PII behavioral tracking (we know why, but it still sucks)?
  • [From the hangout] How did a conversation about where and how innovation will occur devolve into the nuts and bolts of privacy? Why does that happen so often with us? Is that a problem, or is it a symptom of something else?

Yikes! That’s my attempt to summarize the discussion! And it’s still pretty lengthy!

What I Learned about Google+

I certainly didn’t expect to learn anything about Google+ when I wrote the post — it was focusing on plain ol’ web (site) analytics, for Pete’s sake! But, I learned a few things nonetheless:

The good:

  • Longer-form (than 140 characters) discussions, triggered by circles, with the ability to quickly tag people, are pretty cool; Twitter sort of forced us over to blog posts (and then comments on the posts) to have discussions…and Google+ has the potential to bring back richer, more linear dialogue
  • Google+ hangouts…are pretty cool and fairly robust; we had a few hiccups here and there, but I was able to participate reasonably well from inside a minivan traveling down the highway that had the other four members of my family in it (Verizon 4G aircard, in case you’re wondering); and, as the system detects who is speaking, that person’s video jumps to the “main screen” pretty smoothly. It’s not perfect (see below), but we had a pretty meaty conversation in a one-hour slot (and credit, again, to Eric Peterson for his mad moderation skills — that helped!)

The not-so-good:

  • Discussions aren’t threaded, and the “+1” doesn’t really drive the organization of the discussion — multiple logical threads were spawned as the discussion continued, but the platform didn’t really reflect that, which many discussion forums have supported for years
  • Linking the blog post to the discussion was a bit clunky. Who knows what long tail search down the road would benefit from seeing the original post and the ensuing conversation? I added a link to the Google+ discussion to the post after the fact…but it’s not the same as having a string of comments immediately following a post (and if Google+ fizzles…that discussion will be lost; I’ve made a PDF of the thread, but that feels awfully 2007)
  • Google+ hangouts could use some sort of “hand-raising” or “me next” feature; everyone who participated in the hangout worked hard to not speak over anyone else, but we still had a number of awkward transitions

So, that’s what I took away. It was a busy week, especially considering I was knocking out the first half of John Lovett’s new book book (great stuff there) at the same time!

Analysis, Analytics Strategy, Reporting

In Defense of "Web Reporting"

Avinash’s last post attempted to describe The Difference Between Web Reporting and Web Analysis. While I have some quibbles with the core content of the post — the difference between reporting and analysis — I take real issue with the general tone that “reporting = non-value-add data puking.”

I’ve always felt that “web analytics” is a poor label for what most of us who spend a significant amount of our time with web behavioral data do day in and day out. I see three different types of information-providing:

  • Reporting — recurring delivery of the same set of metrics as a critical tool for performance monitoring and performance management
  • Analysis —  hypothesis-driven ad hoc assessment geared towards answering a business question or solving a business problem (testing and optimization falls into this bucket as well)
  • Analytics — the development and application of predictive models in the support of forecasting and planning

My dander gets raised when anyone claims or implies that our goal should be to spend all of our time and effort in only one of these areas.

Reporting <> (Necessarily) Data Puking

I’ll be the first person to decry reporting squirrel-age. I expect to go to my grave in a world where there is still all too much pulling and puking of reams of data. But (or, really, BUT, as this is a biggie), a wise and extremely good-looking man once wrote:

If you don’t have a useful performance measurement report, you have stacked the deck against yourself when it comes to delivering useful analyses.

It bears repeating, and it bears repeating that dashboards are one of the most effective means of reporting. Dashboards done well (and none of the web analytics vendors provide dashboards well enough to use their tools as the dashboarding tool) meet a handful of dos and don’ts:

  • They DO provide an at-a-glance view of the status and trending of key indicators of performance (the so-called “Oh, shit!” metrics)
  • They DO provide that information in the context of overarching business objectives
  • They DO provide some minimal level of contextual data/information as warranted
  • They DON’T exceed a single page (single eyescan) of information
  • They DON’T require the person looking at them to “think” in order to interpret them (no mental math required, no difficult assessment of the areas of circles)
  • They DON’T try to provide “insight” with every updated instance of the dashboard

The last item in this list uses the “i” word (“insight”) and can launch a heated debate. But, it’s true: if you’re looking for your daily, weekly, monthly, or real-time-on-demand dashboard to deliver deep and meaningful insights every time someone looks at it, then either:

  • You’re not clear on the purpose of a dashboard, OR
  • You count, “everything is working as expected” to be a deep insight

Below is a perfectly fine (I’ll pick one nit after the picture) dashboard example. It’s for a microsite whose primary purpose is to drive registrations to an annual user conference for a major manufacturer. It is produced weekly, and it is produced in Excel, using data from Sitecatalyst, Twitalyzer, and Facebook. Is this a case of, as Avinash put it, us being paid “an extra $15 an hour to dump the data into Excel and add a color to the table header?” Well, maybe. But, by using a clunky Sitecatalyst dashboard and a quick glance at Twitalyzer and Facebook, the weekly effort to compile this is: 15 minutes. Is it worth $3.75 per week to get this? The client has said, “Absolutely!”

I said I would pick one nit, and I will. The example above does not do a good job of really calling out the key performance indicators (KPIs). It does, however, focus on the information that matters — how much traffic is coming to the site, how many registrations for the event are occurring, and what the fallout looks like in the registration process. Okay…one more nit — there is no segmentation of the traffic going on here. I’ll accept a slap on the wrist from Avinash or Gary Angel for that — at a minimum, segmenting by new vs. returning visitors would make sense, but that data wasn’t available from the tools and implementation at hand.

An Aside About On-Dashboard Text

I find myself engaged in regular debates as to whether our dashboards should include descriptive text. The “for” argument goes much like Avinash’s implication that “no text” = “limited value.” The main beef I have with any sort of standardized report or dashboard including a text block is that, when baked into a design, it assumes that there is the same basic word count of content to say each time the report is delivered. That isn’t my experience. In some cases, there may be quite a bit of key callouts for a given report…and the text area isn’t large enough to fit it all in. In other cases, in a performance monitoring context, there might not be much to say at all, other than, “All systems are functioning fine.” Invariably, when the latter occurs, in an attempt to fill the space, the analyst is forced to simply describe the information already effectively presented graphically. This doesn’t add value.

If a text-based description is warranted, it can be included as companion material. <forinstance> “Below is this week’s dashboard. If you take a look at it, you will, as I did, say, ‘Oh, shit! we have a problem!’ I am looking into the [apparent calamitous drop] in [KPI] and will provide an update within the next few hours. If you have any hypotheses as to what might be the root cause of [apparent calamitous drop], please let me know” </forinstance> This does two things:

  1. Enables the report to be delivered on a consistent schedule
  2. Engages the recipients in any potential trouble spots the (well-formed) dashboard highlights, and leverages their expertise in understanding the root cause

Which…gets us to…

Analysis

Analysis, by [my] definition, cannot be something that is scheduled/recurring/repeating. Analysis is hypothesis-driven:

  • The dashboard showed an unexpected change in KPIs. “Oh, shit!” occurred, and some root cause work is in order
  • A business question is asked: “How can we drive more Y?” Hypotheses ensue

If you are repeating the same analysis…you’re doing something wrong. By its very nature, analysis is ad hoc and varied from one analysis to another.

When it comes to the delivery of analysis results, the medium and format can vary. But, I try to stick with two key concepts — both of which are violated multiple times over in every example included in Avinash’s post:

  • The principles of effective data visualization (maximize the data-pixel ratio, minimize the use of a rainbow palette, use the best visualization to support the information you’re trying to convey, ensure “the point” really pops, avoid pie charts at all costs, …) still need to be applied
  • Guy Kawasaki’s 10-20-30 rule is widely referenced for a reason — violate it if needed, but do so with extreme bias (aka, slideuments are evil)

While I am extremely wordy on this blog, and my emails sometimes tend in a similar direction, my analyses are not. When it comes to presenting analyses, analysts are well-served to learn from the likes of Garr Reynolds and Nancy Duarte when it comes to how to communicate effectively. It’s sooooo easy to get caught up in our own brilliant writing that we believe that every word we write is being consumed with equal care (you’re on your third reading of this brilliant blog post, are you not? No doubt trying to figure which paragraph most deserves to be immortalized as a tattoo on your forearm, right? You’re not? What?!!!). “Dumb it down” sounds like an insult to the audience, and it’s not. Whittle, hone, remove, repeat. We’re not talking hours and hours of iterations. We’re talking about simplifying the message and breaking it up into bite-sized, consumable, repeatable (to others)  chunks of actionable information.

Analysis Isn’t Reporting

Analysis and reporting are unquestionably two very differing things, but I don’t know that I agree with assertions that analysis requires an entirely different skillset from reporting. Meaningful reporting requires a different mindset and skillset from data puking, for sure. And, reporting and analysis are two different things, but you can’t be successful with the latter without being successful with the former.

Effective reporting requires a laser focus on business needs and business context, and the ability to crisply and effectively determine how to measure and monitor progress towards business objectives. In and of itself, that requires some creativity — there are seldom available metrics that are perfectly and directly aligned with a business objective.

Effective analysis requires creativity as well — developing reasonable hypotheses and approaches for testing them.

Both reporting and analysis require business knowledge, a clear understanding of the objectives for the site/project/campaign/initiative, a better-than-solid understanding of the underlying data being used (and its myriad caveats), and effective presentation of information. These skills make up the core of a good analyst…who will do some reporting and some analysis.

What About Analytics?

I’m a fan of analytics…but see it as pretty far along the data maturity continuum. It’s easy to poo-poo reporting by pointing out that it is “all about looking backwards” or “looking at where you’ve been.” But, hey, those who don’t learn from the past are condemned to repeat it, no? And, “How did that work?” or “How is that working?” are totally normal, human, helpful questions. For instance, say we did a project for a client that, when it came to the results of the campaign from the client’s perspective, was a fantastic success! But, when it came to what it cost us to deliver the campaign, the results were abysmal. Without an appropriate look backwards, we very well might do another project the same way — good for the client, perhaps, but not for us.

In general, I avoid using the term “analytics” in my day-to-day communication. The reason is pretty simple — it’s not something I do in my daily job, and I don’t want to put on airs by applying a fancy word to good, solid reporting and analysis. At a WAW once, I actually heard someone say that they did predictive modeling. When pressed (not by me), it turned out that, to this person, that meant, “putting a trendline on historical data.” That’s not exactly congruent with my use of the term analytics.

Your Thoughts?

Is this a fair breakdown of the work? I scanned through the comments on Avinash’s post as of this writing, and I’m feeling as though I am a bit more contrarian than I would have expected.

Analysis, Social Media

If the Data Looks too Amazing to Be True…

I’ve hauled out this same anecdote off and on for the past decade:

Back in the early aughts [I’m not Canadian, but I know a few of ’em], I was the business owner of the web analytics tool for a high tech B2B company. We were running Netgenesis (remember Netgenesis? I still have nightmares), which was a log file analysis tool that generated 100 or so reports each month and published them as static HTML pages. It took a week for all of the reports to process and publish, but, once published, they were available to anyone in the company via a web interface. One of the product marcoms walked past my cubicle one day early in the month, then stopped, backed up, and stuck his head in: “Did you see what happened to traffic to <the most visited page on our site other than the home page> last month?” I indicated I had not. We pulled up the appropriate report, and he pointed to a step function in the traffic that had occurred mid-month — traffic had jumped 3X and stayed there for the remainder of the month.

“I made a couple of changes to the meta data on the page earlier in the month. This really shows how critical SEO is! I shared it with the weekly product marketing meeting [which the VP of Marketing attended most weeks].”

I got a sinking feeling in my stomach, told him I wanted to look into it a little bit, and sent him on his way. I then pulled up the ad hoc analysis tool and started doing some digging and quickly discovered that a pretty suspicious-looking user-agent seemed to be driving an enormous amount of traffic. It turned out that Gomez was trying to sell into the company and had just set up their agent to ping that page so they could get some ‘real’ data for an upcoming sales demo. Since it was a logfile-based tool, and since the Gomez user agent wasn’t one that we were filtering out, that traffic looked like normal, human-based traffic. When the traffic from that user-agent was filtered out, the actual overall visits to the page had not shown any perceptible change. I explained this to the product marcom, and he then had to do some backtracking on his claims of a wild SEO success (which he had continued to make in the course of the few hours since we’d first chatted and I’d cautioned him that I was skeptical of the data). The moral of the story: If the data looks too dramatic to be true, it probably is!

This anecdote is an example of The Myth of the Step Function (planned to be covered in more detail in Chapter 10 of the book I’ll likely never get around to writing) — the unrealistic expectation that analytics can regularly deliver deep and powerful insights that lead to immediate and drastic business impact. And, the corollary to that myth is the irrational acceptance of data that shows such a step function.

Any time I do training or a presentation on measurement and analytics, I touch on this topic. In an agency environment, I want our client managers and strategists to be comfortable with web analytics and social media analytics data. I even want them to be comfortable exploring the data on their own, when it makes sense. But, (or, really, it’s more like “BUT“), I implore them that, if they see anything that really surprises them, to seek out an analyst to review the data before sharing it with the client. More often than not, the “surprise” will be a case of one of two things:

  • A misunderstanding of the data
  • A data integrity issue

All of this is to say, I know this stuff. I have had multiple experiences where someone has jumped to a wholly erroneous conclusion when looking at data that they did not understand or that was simply bad data. I’d even go so far as to say it’s one of my Top Five Pieces of Personal Data Wisdom!

And yet…

When I did a quick and simple data pull from an online listening tool last week, I had only the slightest of pauses before jumping to a conclusion that was patently erroneous.

Maybe it’s good to get burned every so often. And, I’m much happier to be burned by a frivolous data analysis shared with the web analytics community than to be burned by a data analysis for a paying client. It’s tedious to do data checks — it’s right up there with proof-reading blog posts! — and it’s human nature to want to race to the top of the roof and start hollering when a truly unexpected result (or a more-dramatically-than-expected affirming result) comes out of an analysis.

For me, though, this was a good reminder that taking a breath, slowing down, and validating the data is an unskippable step.

Analysis, Reporting

Reporting: You Can't Analyze or Optimize without It

Three separate observations from three separate co-workers over the past two weeks all resonated with me when it comes to the fundamentals of effective analytics:

  • As we discussed an internal “Analytics 101” class  — the bulk of the class focusses on the ins and outs of establishing clear objectives and valid KPIs — a senior executive observed: “The class may be mislabeled. The subject is really more about effective client service delivery — the students may see this as ‘something analysts do,’ when it’s really a a key component to doing great work by making sure we are 100% aligned with our clients as to what it is we’re trying to achieve.”
  • A note added by another co-worker to the latest updated to the material for that very course said: “If you don’t set targets for success up front, someone else will set them for you after the fact.”
  • Finally, a third co-worker, while working on a client project and grappling with extremely fuzzy objectives, observed: “If you’ve got really loose objectives, you actually have subjectives, and those are damn tough to measure.”

SEO search engine optimization indiaIt struck me that these comments were three sides to the same coin, and it got me to thinking about how often I find myself talking about performance measurement as a critical fundamental building block for conducting meaningful analysis.

“Reporting” is starting to be a dirty word in our industry, which is unfortunate. Reporting in and of itself is extremely valuable, and even necessary, if it is done right.

Before singing the praises of reporting, let’s review some common reporting approaches that give the practice a bad name:

  • Being a “report monkey” (or “reporting squirrel” if you’re an Avinash devotee) — just taking data requests willy-nilly, pulling the numbers, and returning them to the requestor
  • Providing “all the data” — exercises of listing out every possible permutation/slicing of a data set, and then providing a many-worksheeted spreadsheet to end users so that they can “get any data they want”
  • Believing that, if a report costs nothing to generate, then there is no harm in sending it — automation is a double-edged sword, because it can make it very easy to just set up a bad report and have it hit users’ inboxes again and again without adding value (while destroying the analyst’s credibility as a value-adding member of the organization)

None of these, though, are reasons to simply toss reporting aside altogether. My claim?

If you don’t have a useful performance measurement report, you have stacked the deck against yourself when it comes to delivering useful analyses.

Let’s walk through a logic model:

  1. Optimization and analysis are ways to test, learn, and drive better results in the future than you drove in the past
  2. In order to compare the past to the future (an A/B test is a “past vs. future” because the incumbent test represents the “past” and both the incumbent and the challenger represent “potential futures”), you have to be able to quantify “better results”
  3. Quantifying “better results” mean establishing clear and meaningful measures for those results
  4. In order for measures to be meaningful, they have to be linked to meaningful objectives
  5. If you have meaningful objectives and meaningful measures, then you have established a framework for meaningfully monitoring performance over time
  6. In order for the organization to align and stay aligned, it’s incredibly helpful to actually report performance over time using that framework, quod erat demonstrandum (or, Q.E.D., if you want to use the common abbreviation — how in the hell the actual Latin words, including the correct spelling, were not only something I picked up in high school geometry in Sour Lake, TX, but that has actually stuck with me for over two decades is just one of those mysteries of the brain…)

So, let’s not just bash reporting out of hand, okay? Entirely too many marketing organizations, initiatives, and campaigns lack truly crystallized objectives. Without clear objectives, there really can’t be effective measurement. Without effective measurement, there cannot be meaningful analysis. Effective measurement, at it’s best, is a succinct, well-structured, well visualized report.

Photo: Greymatterindia

Analysis, Social Media

Twitter Analytics — Turmoil Abounds, and I'm a Skeptic

Last week was a little crazy on the Twitter front, with two related — but very different —  analytics-oriented announcements hitting the ‘net within 24 hours of each other. Let’s take a look.

Selling Tweet Access

On Wednesday, Twitter announced they would be selling access to varying volumes of tweets, with 50% of all tweets being available for the low, low price </sarcasm> of $360,000/year. It appears there will be a variety of options, with “50%” being the maximum tweet volume, but with other options in the offing to get 5% of all tweets, 10% of all tweets, or all tweets/references/retweets that are tied to a specific user. All of these sound like they’re going to come with some pretty tight usage constraints, including that they can’t be resold and that the actual tweet content can’t be published.

Twitter has made an API available almost from the moment the service was created. That’s one of the reasons the service grew so explosively — developers were able to quickly build a range of interfaces to the tool that were better than what Twitter’s development team was able to create. But, the API came with limitations — a very tight limit on how often an application could get updates, and a tight limit on just how many updates could be pushed/pulled at once.

As various Twitter analytics-type services began to crop up, Twitter opened up a “garden hose” option — developers could contact Twitter, show that they had a legitimate service with a legitimate need, and they could get access to more tweets more often through the API. Services like Twitalyzer, TweetReach, and Klout jumped all over that option and have built out robust and useful solutions over the course of the last 6-12 months. Now it looks like Twitter is looking to coil up the garden hose, which could spell a permanent end to the growing season for these services. This will be a shame if it comes to pass.

For a steep price, these paid options from Twitter will have limited use: limited to some basic monitoring/listening and some basic performance measurement. Even with the $360K/year option, providing half of the tweets seems problematic when you consider Twitter from a social graph perspective — in theory, half of the network ripple from any given tweet will be lost, or, more confusingly, will crop up as a 2nd or 3rd degree effect with no ability to trace it back to its source because the path-to-the-source passes through the “unavailable 50%!”

This data also won’t be of much use as a listen-and-respond tool. Imagine a brand that has a fantastic ability to monitor Twitter and appropriately engage and respond…but appears schizophrenic because they’re operating with one eye closed (and paying a pretty penny to do even that!). To be clear, for any given brand or user, only a tiny fraction of all tweets are actually of interest, but that tiny fraction is going to be spread across 100% of the Twitterverse, so only having access to a 5%, 10%, or even 50% sample means that relevant tweets will be missed.

Online listening platforms — Radian6, SM2, Buzzmetrics, Crimson Hexagon, Sysomos, etc. — may actually have deep enough pockets to pay for these tweets to improve their own underlying data…but they will have to significantly alter the services they provide in order to comply with the usage guidelines for the data.

Ugh.

Twitter Analytics

On Thursday, Mashable reported that Twitter Analytics was being tested by selected users. Unfortunately, I’m not one of those users (<sniff><sob>), so I’m limited to descriptions in the Mashable article. Between that article and Pete Cashmore’s (Mashable CEO) editorial on cnn.com, I’ve got pretty low expectations for Twitter Analytics.

Both pieces seem somewhat naive in that they overplay the value to brands that Facebook has delivered with Facebook Insights, and they confuse “pretty graphs” with “valuable data.” All I can think to do is rattle off a series of reactions from the limited information I’ve been able to dig up:

  • Replies/references over time: um…thanks, but that’s always been something that’s pretty easy to get at, so no real value there.
  • Follows/unfollows: this seems to be taking a page directly from Facebook Insights with it’s new fans/removed fans reporting (which, by the way, never agrees with the “Total Fans” data available in the same report, but I digress…); this has marginal value — in practice, unless a user is really pissing off followers or baiting them to follow with a very specific promotional giveaway (“Follow us and retweet this and you’ll be entered to win a BRAND NEW CAR!!!”), there’s probably not going to be a big spike in unfollows, and it isn’t that hard to trend “total followers” over time, so I can’t get too excited about this, either
  • Unfollows (cont’d.): “tweets that cause people to unfollow” is another apparent feature of Twitter analytics. Really? Was that something that someone living on planet Earth came up with? This sounds nifty initially, but, in practice, isn’t going to be of much use. If a user posts offensive, highly political (for a non-political figure user), or obnoxiously self-promoting tweets…he’s going to lose followers. I don’t think “analytics” will really be needed to figure out the root cause (if it was a single tweet) driving a precipitous follower drop. Common sense should suffice for that.
  • Retweets: this is like references, in that it’s not really that hard to track, and I wouldn’t be surprised at all if Twitter Analytics only counts retweets that use the official Twitter retweet functionality, rather than using a looser definition that includes “RT @<username>” occurrences (which are retweets that are often more valuable, because they can include additional commentary/endorsement by the retweeters)
  • Impressions: I’m expecting a simplistic definition of impressions that is based just on the number of followers, which is misleading, because most users of Twitter see only a fraction of the tweets that cross their stream. Twitalyzer calculates an “effective reach” and Klout calculates a “true reach” — both make an attempt to factor in how receptive followers are to messages from the user. None of these measures is going to be perfect, but I’m happier relying on companies whose sole focus is analytics trying to tinker with a formula than I am with the “owner” of the data coming up with a formula that they think makes sense.

With the screen caps I’ve seen, there is no apparent “export data” button, and that’s a back-breaker. Just as Facebook Insights is woefully devoid of data export capabilities (the “old interface” enables data export…but not of some of the most useful data, and API access to the Facebook Insights data doesn’t exist, as best as I’ve been able to determine), Twitter looks like they may be yet another technology vendor who doesn’t understand that “their” dashboard is destined to be inadequate. I’m always going to want to combine Twitter data with data that Twitter doesn’t have when it comes to evaluating Twitter performance. For instance, I’m going to want to include referrals from Twitter to my web site, as well as short URL click data in my reporting and analysis.

Ikong Fu speculated during an exchange (on Twitter) that Twitter may also, at some point, include their internal calculations of a user’s influence in Twitter Analytics:

I didn’t realize that Twitter was calculating an internal reputation score. It makes sense, though, that that would be included when they make recommendations of who else a user might want to follow. I found a post from Twitter’s blog back in July that announced the rollout of  “follow suggestions,” and that post indicated these were based on “algorithms…built by our user relevance team.” The only detail the post provided was that these suggestions were “based on several factors, including people you follow and the people they follow.” That sounds more like a social graph analysis (“If you’re following 10 people who are all following the same person who you are not following, then we’re going to recommend that you follow that person”) than an analysis of each user’s overall influence/quality. Again…I’m more comfortable with third party companies who are fully focussed on this measurement and who make their algorithms transparent providing me with that information than I am with Twitter in that role.

So, Where Does This Leave Us?

Maybe, for once, I’m just seeing a partially filled glass of data as being half empty rather than half full (okay, so that’s the way I view most things — I’m pessimistic by nature). In the absence of more information, though, I’m forced to think that, just as I was headed towards analytics amour when it came to Twitter data, Twitter is making some unfortunate moves and rapidly smudging the luster right off of that budding relationship.

Or, maybe, I’m unfairly pre-judging. Time will tell.

Analysis, Social Media

Four Ways that Media Mix Modeling (MMM) Is Broken

Many companies rely on some form of media mix modeling (or “marketing mix modeling”) to determine the optimal mix of their advertising spend. With the growth of “digital” media and the explosion of social media, these models are starting to break down. That puts many marketing executives in a tough bind:

  1. Marketing, like all business functions, must be data-driven — more so now than ever
  2. Digital is the “most measurable medium ever” (although their are wild misperceptions as to what this really means)
  3. Ergo, digital media investments must be precisely measured to quantify impact on the bottom line

For companies that have built up a heavy reliance on media mix modeling (MMM), the solution seems easy: simply incorporate digital media into the model! What those of us who live and breathe this world recognize (and lament over drinks at various conferences for data geeks), is that this “simple” solution simply doesn’t work. Publicly, we say, “Well…er…it’s problematic, but we’re working on it, and the modeling techniques are going to catch up soon.”

My take: don’t hold your breath that MMM is going to catch up — even if it catches up to today’s reality, it will already be behind, because digital/social/mobile will have continued its explosive evolution (and complexity to model).

Believe it or not, I’m not saying that MMM should be completely abandoned. It still has it’s place, I think, but there are a lot of things it’s going to really, really struggle to address. I’d actually like to see companies who provide MMM services weigh in on what that is. At eMetrics earlier this month, I attended a session where the speaker did just that. Skip ahead to the last section to find out who!

Geographic Test/Control Data

Both traditional and digital marketing have a mix of geo-specific capabilities. The cost of TV, radio, print, and out-of-home (OOH) marketing provides an imperative to geo-target when appropriate (or simply to minimize the peanut butter effect of spreading a limited investment so thinly that it doesn’t have an impact anywhere). Many digital channels, though, such as web sites and Facebook pages, are geared towards being “available to everyone.” Other channels – SEM, banner ads, and email, for instance – can be geo-targeted, but there often isn’t a cost/benefit reason to do so. Without different geographic placements of marketing, the impact on sales in “exposed areas” vs. “unexposed areas” cannot be teased out:

Cross-Channel Interaction

While marketers have long known that multi-channel campaigns produce a whole that is greater than the sum of the parts, the sheer complexity that digital has introduced into the equation forces MMM to guess at attribution. For example, we know (or, at least, we strongly suspect) that a large TV advertising campaign will not only provide a lift in sales, but it will also produce a lift in searches for a brand. Those increased searches will increase SEM results, which will drive traffic to the brand’s web site. Consumers who visit the site can then be added to a retargeting campaign. Those are four different marketing channels that all require investment…but which one gets the credit when the consumer buys?

This is both data capture and a business rules question. Entire companies (Clearsaleing being the one that I hear the most about) have been built just to address the data capture and application of business rules. While they provide the tools, they’re a long way from really being able to capture data across the entire continuum of a consumer’s experience. The business rules question is just as significant — most marketers’ heads will explode if they’re asked to figure out what the “right” attribution is (and simply trying different attribution models won’t answer the question — different models will show different channels being the “best”). Is this a new career option for Philosophy majors, perhaps?

Fragmentation of Consumer Experiences

This one is related to the cross-channel interaction issue described above, but it’s another lens applied to the same underlying challenge. Consumer behavior is evolving — there are exponentially more channels through which consumers can receive brand exposure (I picked up the phrase “cross-channel consumer” at eMetrics, which is in the running for my favorite three-letter phrase of 2010!). Some of these channels operate both as push and pull, whereas traditional media is almost exclusively “push” (marketers push their messages out to consumers through advertising):

We’re now working with an equation that has wayyyyyyyy more variables, each of which has a lesser effect than the formulas we were trying to solve when MMM first came onto the scene. HAL? Can you help? This is actually beyond a question of simply “more processing power.” It’s more like predicting what the weather will be next week — even with meteoric advancements in processing power and a near limitless ability to collect data, the models are still imprecise.

Self-Fulfilling Mix

Finally, there is a chicken-and-egg problem. While there are reams of secondary research documenting the shifting of consumer behavior from offline to online consumption…many brands still disproportionately invest in offline marketing. It’s understandable — they’re waiting for the data to be able to “prove” that digital marketing works (and prove it with an unrealistic degree of accuracy — digital his held to a higher standard than offline media, and the “confusion of precision with accuracy” syndrome is alive and well). But, when digital marketing investments are overly tentative (and those investments are spread across a multitude of digital channels), the true impact of digital can’t be detected because it’s dwarfed by the impact of the massive — if less efficient — investments in offline marketing:

If I shoot a pumpkin simultaneously with a $1,500 shotgun and a $30 BB gun and ask an observer to tell me how much of an impact the BB gun had…

So, Should We Just Start Operating on Faith and Instinct?

I wrote early in this post that I think MMM has its place. I don’t fully understand what that place is, but the credibility of anyone whose bread is buttered by their MMM book of business who stands up and says, “Folks, MMM has some issues,” immediately skyrockets. That’s exactly what Steve Tobias from Marketing Management Analytics (MMA) did at eMetrics. In his session, “Marketing Mix Modeling: How to Make Digital Work for a True ROI,” he talked at length about many of the same challenges I’ve described in this post (albeit in greater detail and without the use of cartoon-y diagrams). But, he went on to lay out how MMA is using traditional MMM in conjunction with panel-based data (in his examples, he used comScore for the analysis) to get “true ROI” measurement. All I’ve seen is that presentation, so I don’t have direct experience with MMA’s work in action, but I liked what I heard!

Analysis, Analytics Strategy, Reporting, Social Media

Analyzing Twitter — Practical Analysis

In my last post, I grabbed tweets with the “#emetrics” hashtag and did some analysis on them. One of the comments on that post asked what social tools I use for analysis — paid and free. Getting a bit more focussed than that, I thought it might be interesting to write up what free tools I use for Twitter analysis. There are lots of posts on “Twitter tools,” and I’ve spent more time than I like to admit sifting through them and trying to find ones that give me information I can really use. This, in some ways, is another one of those posts, except I’m going to provide a short list of tools I actually do use on a regular basis and how and why I use them.

What Kind of Analysis Are We Talking About?

I’m primarily focussed on the measurement and analysis of consumer brands on Twitter rather than on the measurement of one’s personal brand (e.g., @tgwilson). While there is some overlap, there are some things that make these fundamentally different. With that in mind, there are really three different lenses through which Twitter can be viewed, and they’re all important:

  • The brand’s Twitter account(s) — this is analysis of followers, lists, replies, retweets, and overall tweet reach
  • References of the brand or a campaign on Twitter — not necessarily mentions of @<brand>, but references to the brand in tweet content
  • References to specific topics that are relevant to the brand as a way to connect with consumers — at Resource Interactive, we call this a “shared passion,” and the nature of Twitter makes this particularly messy, but, to whatever level it’s feasible, it’s worth doing

While all three of these areas can also be applied in a competitor analysis, this is the only mention (almost) I’m going to make of that  — some of the techniques described here make sense and some don’t when it comes to analyzing the competition.

And, one final note to qualify the rest of this post: this is not about “online listening” in the sense that it’s not really about identifying specific tweets that need a timely response (or a timely retweet). It’s much more about ways to gain visibility into what is going on in Twitter that is relevant to the brand, as well as whether the time spent investing in Twitter is providing meaningful results. Online listening tools can play a part in that…but we’ll cover that later in this post.

Capturing Tweets?

When it comes to Twitter analysis, it’s hard to get too far without having a nice little repository of tweets themselves.  Unfortunately, Twitter has never made an endless history of tweets available for mining (or available for anything, for that matter). And, while the Library of Congress is archiving tweets, as far as I know, they haven’t opened up an API to allow analysts to mine them. On top of that, there are various limits to how often and how much data can be pulled in at one time through the Twitter API. As a consumer, I suppose I have to like that there are these limitations. As a data guy, it gets a little frustrating.

Two options that I’ve at least looked at or heard about on this front…but haven’t really cracked:

  • Twapper Keeper — this is a free service for setting up a tweet archive based on a hashtag, a search, or a specific user. In theory, it’s great. But, when I used it for my eMetrics tweet analysis, I stumbled into some kinks — the file download format is .tar (which just means you have to have a utility that can uncompress that format), and the date format changed throughout the data, so getting all of the tweets’ dates readable took some heavy string manipulation
  • R — this is an open source statistics package, and I talked to a fellow several months ago who had used it to hook into Twitter data and do some pretty intriguing stuff. I downloaded it and poked around in the documentation a bit…but didn’t make it much farther than that

I also looked into just pulling Tweets directly into Excel or Access through a web query. It looks like I was a little late for that — Chandoo documented how to use Excel as a Twitter client, but then reportd that Twitter made a change that means that approach no longer works as of September 2010.

So, for now, the best way I’ve found to reliably capture tweets for analysis is with RSS and Microsoft Outlook:

  1. Perform a search for the twitter username, a keyword, or a hashtag from http://search.twitter.com (or, if you just want to archive tweets for a specific user, just go to the user’s Twitter page)
  2. Copy the URL for the RSS for the search (or the user)
  3. Add a new RSS feed in MS Outlook and paste in the URL

From that point forward, assuming Outlook is updating periodically, the RSS feeds will all be captured.

There’s one more little trick: customize the view to make it more Excel/export-friendly. In Outlook 2007, go to View » Current View » Customize Current View » Fields. I typically remove everything except From, Subject, and Received. Then go to View » Current View » Format Columns and change the Received column format from Best Fit to the dd-Mmm-yy format. Finally, remove the grouping. This gives you a nice, flat view of the data. You can then simply select all the tweets you’re interested in, press <Ctrl>-<C>, and then paste them straight into Excel.

I haven’t tried this with hundreds of thousands of tweets, but it’s worked great for targeted searches where there are several thousand tweets.

Total Tweets, Replies, Retweets

While replies and retweets certainly aren’t enough to give you the ultimate ROI of your Twitter presence, they’re completely valid measures of whether you are engaging your followers (and, potentially, their followers). Setting up an RSS feed as described above based on a search for the Twitter username (without the “@”) will pick up both all tweets by that account as well as all tweets that reference that account.

It’s then a pretty straightforward exercise to add columns to a spreadsheet to classify tweets any number of ways by some use of the IF, ISERROR, and FIND functions. These can be used to quickly flag each tweet  as a reply, a retweet, a tweet by the brand, or any mix of things:

  • Tweet by the brand — the “From” value is the brand’s Twitter username
  • Retweet — tweet contains the string “RT @<username>
  • Reply — tweet is not a retweet and contains the string “@<username>

Depending on how you’re looking at the data, you can add a column to roll up the date — changing the tweet date to be the tweet week (e.g., all tweets from 10/17/2010 to 10/23/2010 get given a date of 10/17/2010) or the tweet month. To convert a date into the appropriate week (assuming you want the week to start on Sunday):

=C1-WEEKDAY(C1)+1

To convert the date to the appropriate month (the first day of the month):

=DATE(YEAR(C1),MONTH(C1),1)

C1, of course, is the cell with the tweet date.

Then, a pivot table or two later, and you have trendable counts for each of these classifications.

This same basic technique can be used with other RSS feeds and altered formulas to track competitor mentions, mentions of the brand (which may not match the brand’s Twitter username exactly), mention of specific products, etc.

Followers and Lists

Like replies and retweets, simply counting the number of followers you have isn’t a direct measure of business impact, but it is a measure of whether consumers are sufficiently engaged with your brand. Unfortunately, there are not exactly great options for tracking net follower growth over time. The “best” two options I’ve used:

  • Twitter Counter — this site provides historical counts of followers…but the changes in that historical data tend to be suspiciously evenly distributed. It’s better than nothing if you don’t have a time machine handy. (See the Twitalyzer note at the end of this post — I may be changing tools for this soon!)
  • Check the account manually — getting into a rhythm of just checking an account’s total followers is the best way I’ve found to accurately track total followers over time; in theory a script could be written and scheduled that would automatically check this on a recurring basis, but that’s not something I’ve tackled

I also like to check lists and keep track of how many lists the Twitter account is included on. This is a measure, in my mind, of whether followers of the account are sufficiently interested in the brand or the content that they want to carve it off into a subset of their total followers so they are less likely to miss those tweets and/or because they see the Twitter stream as being part of a particular “set of experts.” Twitalyzer looks like it trends list membership over time, but, since I just discovered that it now does that, I can’t stand up and say, “I use that!” I may very well start!

Referrals to the Brand’s Site

This doesn’t always apply, but, if the account represents a brand, and the brand has a web site where the consumer can meaningfully engage with the brand in some way, then measuring referrals from Twitter to the site are a measure of whether Twitter is a meaningful traffic driver. There are fundamentally two types of referrals here:

  • Referrals from tweeted links by the brand’s Twitter account that refer back to the site — these can be tracked by a short URL (such as bit.ly), by adding campaign tracking parameters to the URL so the site’s web analytics tool can identify the traffic as a brand-triggered Twitter referral, or both. The campaign tracking is what is key, because it enables measuring more than simply “clicks:” whether the visitors are first-time visitors to the site or returning visitors, how deeply they engaged with the site, and whether they took any meaningful action (conversions) on the site
  • “Organic” referrals — overall referrals to the site from twitter.com. Depending on which web analytics tool you are using on your site, this may or may not include the clickthroughs from links tweeted by the brand.

By looking at referral traffic, you can measure both the volume of traffic to the site and the relative quality of the traffic when compared to other referral sources for the site.

(If the volume of that traffic is sufficiently high to warrant the effort, you may even consider targeting content on the landing page(s) for Twitter referral traffic to try to engage visitors more effectively– you know the visitor is engaged with social media, so why not test some secondary content on the page to see if you can use that knowledge to deliver more relevant content and CTAs?)

Word Clouds with Wordle

While this isn’t a technique for performance management, it’s hard to resist the opportunity to do a qualitative assessment of the tweets to look for any emerging or hot topics that warrant further investigation. Because all of the tweets have been captured, a word cloud can be interesting (see my eMetrics post for an example). Hands-down, Wordle makes the nicest word clouds out there. I just wish it was easier to save and re-use configuration settings.

One note here: you don’t want to just take all of the tweet content and drop it straight into Wordle, as the search criteria you used for the tweets will dwarf all of the other words. If you first drop the tweets into Word, you can then do a series of search and replaces (which you can record as a macro if you’re going to repeat the analysis over time) — replace the search terms, “RT,” and any other terms that you know will be dominant-but-not-interesting with blanks.

Not Exactly the Holy Grail…

Do all of these techniques, when appropriately combined, provide near-perfect measurement of Twitter? Absolutely not. Not even close. But, they’re cheap, they do have meaning, and they beat the tar out of not measuring at all. If I had to pick one tool that I was going to bet on that I’d be using inside of six months for more comprehensive performance measurement of Twitter, it would be Twitalyzer. It sure looks like it’s come a long way in the 6-9 months since I last gave it a look. What it does now that it didn’t do initially:

  • Offers a much larger set of measures — you can pick and choose which measures make sense for your Twitter strategy
  • Provides clear definitions of how each metric is calculated (less obfuscated than the definitions used by Klout)
  • Allows trending of the metrics (including Lists and Followers).

Twitalyzer, like Klout, and Twitter Counter and countless other tools, is centered on the Twitter account itself. As I’ve described here, there is more going on in Twitter that matters to your brand than just direct engagement with your Twitter account and the social graph of your followers. Online listening tools such as Nielsen Buzzmetrics can provide keyword-based monitoring of Twitter for brand mentions and sentiment — this is not online listening per se, really, but it is using online listening tools for measurement.

For the foreseeable future, “measuring Twitter” is going to require a mix of tools. As long as the mix and metrics are grounded in clear objectives and meaningful measures, that’s okay. Isn’t it?

Analysis, Reporting

Dear Technology Vendor, Your Dashboard Sucks (and it’s not your fault)

Working in measurement and analytics at a digital marketing agency, I find myself working with a seemingly (at times) countless number of of technology platforms – most of them are measurement platforms (web analytics, social media analytics, online listening), but many of them are operational systems that, by their nature, collect data that is needs to be reported and analyzed (email platforms, marketing automation and CRM platforms, gamification systems, social media moderation systems, and so on). And, not only do I get to work with these systems in action, but one of the many fun things about my job is that I constantly get to explore new and emerging platforms as well.

During a recent presentation by one of our technology partners, I had a minor out of body experience where I saw this dopey-voiced Texan turn into something of a crotchety crank. He (I) fairly politely, and with (I hope) a healthy serving of humor poured over the exchange, lit into the CEO. I didn’t know where it came from…except I did (when I pondered the exchange afterwards).

When it comes to reporting, technology vendors fall into the age-old trap of, “When all you have is a hammer, all the world looks like a nail.” The myopia these vendors display varies considerably – some are much more aware of where they fit in the overall marketing ecosystem than others – but they consistently don blinders when it comes to their data and their dashboards.

The most important data to their customers, they assume, is the data within their system. Sure, they know that there are other systems in play that are generating some useful supplemental data, and that’s fantastic! “All” the customer needs to do is use the vendor’s (cumbersome) integration tools to bring the relevant subsets of that data into their system. “Sure, you can bring customer data from your CRM system into our web analytics environment. I’ll just start writing up a statement of work for the professional services you’ll need to do that! What? You want data from our system to be fed into your CRM system, too? I’ll get an SOW rolling for that at the same time! Did I mention that my youngest child just got into an Ivy League school? Up until five minutes ago, I was sweating how we were going to pay for it!”

The vendors – their sales teams – tout their “reporting and analytics” capabilities. They frequently lead off their demos with a view of their “dashboards” and tout how easy and intuitive the dashboard interface is! What they’re really telling their prospective customers, though, is, “You’ll have one more system you’ll have to go to to get the data you need to be an effective marketer.” <groan>

Never mind the fact that these “dashboards” are always data visualization abominations. Never mind the fact that they require new users to climb a steep learning curve. Never mind that they are fundamentally centered around the “unit of analysis” that the stem is built for (a content management system’s dashboard is content-centric, while a CRM system’s dashboard is customer-centric). They only provide access to a fraction of the data that the marketer really cares about most of the time.

Clearly, these platforms need to provide easy access to their data. I’m not really arguing that dashboard and reporting tools shouldn’t be built into these systems. What I am claiming is that vendors need to stop believing (and stop selling) that this is where their customers will glean the bulk of their marketing insights. In most cases, they won’t. Their customers are going to export the data from that system and combine it (or at least look at it side by side) with data from other systems. That’s how they’re going to really get a handle on what is happening.

The CEO with whom I had the out-of-body experience that triggered this post quickly and smartly turned my challenge back on me: “Well, what is it, ideally, that you would want?” I watched myself spout out an answer that, now 24 hours later, still holds up. Here are my requirements, and they apply to any technology vendor who offers a dashboard (including web analytics platforms, which, even though they exist purely as data capture/reporting/analysis systems…still consistently fall short when it comes to providing meaningful dashboards – partly due to lousy flexibility and data visualization, which they can control, and partly due to the lack of integration with all other relevant data sources, which they really can’t):

Within your tool, I want to be able to build a report that I can customize in four ways:

  • Define the specific dimensions in the output
  • Define the specific measures to include in the output
  • Define the time range for the data (including a “user-defined” option – more on that in a minute)
  • Define whether I want detailed data or aggregated data, and, if aggregated, the granularity of the trending of that data over time (daily, weekly, monthly, etc.)

Then, I want that report to give me a URL – an https one, ideally – onto which I can tack login credentials such that that URL will return the data I want any time I refresh it. I want to be able to drop that URL into any standalone reporting environment – my data warehouse ETL process, my MS Access database, or even my MS Excel spreadsheet – to get the data I want returned to me. I’m want to be able to pass a date range in with that request so that I can pull back the range of data I actually need.

Sure, in some situations, I’m going to want to hook into your data more efficiently than through a secure http request – if I’m looking to pull down monstrous data sets on a regular basis – but let’s cross that “API plus professional services” bridge when we get to it, okay?

I’m never going to use your dashboard. I’m going to build my own. And it’s going to have your data and data from multiple other platforms (some of them might even be your competitors), and it’s going to be organized in a way that is meaningful to my business, and it’s going to be useful.

Stop over-hyping your dashboards. You’re just setting yourselves up for frustrated customers.

It’s a fantasy, I realize, but it’s my fantasy.

Analysis, Analytics Strategy, Social Media

Integrated View of Visitors = Multiple Data Sources

I attended the Foresee Results user summit last month, and John Lovett of Analytics Demystified was the keynote speaker. It’s a credit to my general lack of organization that I wasn’t aware he was going to be speaking, much less keynoting!

John showed this diagram when discussing the importance of recognizing your capabilities:

The diagram starts to get at the never-ending quest to obtain a “360 degree customer view.” A persistent misperception among marketers when it comes to web analytics is that behavioral data alone can provide a comprehensive view of the customer. It really can’t — force your customers to behave in convoluted ways and then only focus on behavioral data, and you can draw some crazily erroneous conclusions (“Our customers appear to visit our web site and then call us multiple times to resolve a single issue. They must like to have a lot of interactions with us!”).

Combining multiple data sources — behavioral and attitudinal — is important. As it happened, Larry Freed, the Foresee Results CEO, had a diagram that came at the same idea:

This diagram was titled “Analytics Maturity.” It’s true — slapping Google Analytics on your web site (behavioral data) is cheap and easy. It takes more effort to actually capture voice-of-the-customer (attitudinal) data; even if it’s with a “free” tool like iPerceptions 4Q, there is still more effort required to ensure that the data being captured is valid and to analyze any of the powerful open-ended feedback that such surveys provide. Integrating behavioral and attitudinal data from two sources is tricky enough, not to mention integrating that data with your e-mail, CRM, marketing automation, and ERP systems and third-party data sources that provide demographic data!

It’s a fun and challenging world we live in as analysts, isn’t it?

(On the completely off-topic front: I did snag 45 minutes one afternoon to walk around the University of Michigan campus a bit, as the conference was hosted at the Ross School of Business; a handful of pictures from that moseying is posted over on Flickr.)

Analysis

From Data to Action — The Many Flavors of Latency

I was flipping through the slides from a workshop that Teradata put on at The Ohio State University several months ago, and one of the diagrams jumped out and resonated with me. As I did some digging, it turns out this diagram has been floating around since at least 2004, if not for longer. It was created by Dr. Richard Hackathorn of Bolder Technology Inc. (BTI).

There are a slew of lousy recreations of the diagram (the original diagram wasn’t so hot, either). Rather than recreating it myself, I just snagged one of the cleaner ones, which came from a 4-year-old TDWI article:

The point of the diagram, as well as of most of the derivative works that reference it, is that the value of information has a direct relationship to the speed with which you can react to it. And, there are three distinct things that have to happen between the business event that triggers the information and ation actually being taken.

I don’t know if there is any real math or science behind the shape of the curve. As diagrammed, this says that you’ve already lost most of your value by the time you get to the “decision latency” point in the process. I don’t know that that is necessarily true in most cases. The diagram supports the assertions by all of the various BI/data tool vendors that data needs to be available in near real-time (and, of course, that’s something that all of the vendors claim they are better at than their competition).

But, is the data latency and analysis latency really the big value driver for marketers? In some cases, the data latency is a structural issue — conducting a campaign where the people exposed to it are likely to not convert for anywhere from 1 to 30 days…means you really need to wait for 30 days to see how the campaign played out. Analysis latency is real…but this really can be broken into two pieces: 1) the time to do the analysis and get it packaged for delivery, and 2) the time to schedule/coordinate the information delivery. And, then, certainly the decision latency is real.

In short, the “action time” components totally make sense, and it’s good to understand them. The shape of the curve, though, doesn’t necessarily stand up to scrutiny when looked at through a marketer’s lens.

Analysis, Analytics Strategy, Reporting

Answering the "Why doesn't the data match?" Question

Anyone who has been working with web analytics for more than a week or two has inevitably asked or been asked to explain why two different numbers that “should” match don’t:

  • Banner ad clickthroughs reported by the ad server don’t match the clickthroughs reported by the web analytics tool
  • Visits reported by one web analytics tool don’t match visits reported by another web analytics tool running in parallel
  • Site registrations reported by the web analytics tool don’t match the number or registrations reported in the CRM system
  • Ecommerce revenue reported by the web analytics tool doesn’t match that reported from the enterprise data warehouse

In most cases, the “don’t match” means +/- 10% (or maybe +/- 15%). And, seasoned analysts have been rattling off all the reasons the numbers don’t match for years. Industry guru Brian Clifton has written (and kept current) the most comprehensive of white papers on the subject. It’s 19 pages of goodness, and Clifton notes:

If you are an agency with clients asking the same accuracy questions, or an in-house marketer/analyst struggling to reconcile data sources, this accuracy whitepaper will help you move forward. Feel free to distribute to clients/stakeholders.

It can be frustrating and depressing, though, to watch the eyes of the person who insisted on the “match” explanation glaze over as we try to explain the various nuances of capturing data from the internet. After a lengthy and patient explanation, there is a pause, and then the question: “Uh-huh. But…which number is right?” I mentally flip a coin and then respond either, “Both of them” or “Neither of them” depending on how the coin lands in my head. Clifton’s paper should be required reading for any web analyst. It’s important to understand where the data is coming from and why it’s not simple and perfect. But, that level of detail is more than most marketers can (or want to) digest.

After trying to educate clients on the under-the-hood details…I almost wind up at a point where I’m asked the “Well, which number is right?” question. That leads to a two-point explanation:

  • The differences aren’t really material
  • What matters in many, many cases is more the trend and change over time of the measure — not its perfect accuracy (as Webtrends has said for years: “The trends are more important than the actual numbers. Heck, we put ‘trend’ in our company name!”

This discussion, too, can have frustrating results.

I’ve been trying a different tactic entirely of late in these situations. I can’t say it’s been a slam dunk, but it’s had some level of results. The approach is to list out a handful of familiar situations where we get discrepant measures and are not bothered by it at all, and then use those to map back to the data that is being focussed on.

Here’s my list of examples:

  • Compare your watch to your computer clock to the time on your cell phone. Do they match? The pertinent quote, most often attributed to Mark Twain, is as follows: “A man with one watch knows what time it is; a man with two watches is never quite sure.” Even going to the NIST Official U.S. Time Clock will yield results that differ from your satellite-synched cell phone. Two (or more) measures of the time that seldom match up, and with which we’re comfortable with a 5-10 minute discrepancy.

Photo courtesy of alexkerhead

  • Your bathroom scale. You know you can weigh yourself as you get out of the shower first thing in the morning, but, by the time you get dressed, get to the doctor’s office, and step on the scale there, you will have “gained” 5-10 lbs. Your clothes are now on, you’ve eaten breakfast, and it’s a totally different scale, so you accept the difference. You don’t worry about how much of the difference comes from each of the contributing factors you identify. As long as you haven’t had a 20-lb swing since your last visit to the doctor, it’s immaterial.

Photo courtesy of dno1967

  • For accountants…”revenue.” If the person with whom your speaking has a finance or accounting background, there’s a good chance they’ve been asked to provide a revenue number at some point and had to drill down into the details: bookings or billings? GAAP-recognized revenue? And, within revenue, there are scads of nuances that can alter the numbers slightly…but almost always in non-material ways.

Photo courtesy of alancleaver_2000

  • Voting (recounts). In close elections, it’s common to have a recount. If the recount re-affirms the winner from the original count, then the results is accepted and moved on from. There isn’t a grand hullabaloo about why the recount numbers differed slightly from the original account. In really close races, where several recounts occur, the numbers always come back differently. And, no one knows which one is “right.” But, once there is a convergence as to the results, that is what gets accepted.

Photo courtesy of joebeone

    That’s my list. Do you have examples that you use to explain why there’s more value in picking either number and interpreting it rather than obsessing about reconciling disparate numbers. I’m always looking for other analogies, though. Do you have any?

    Analysis, Analytics Strategy

    A Record-Setting WAW in Columbus with CRM Metrix

    Last week’s Columbus set a new record for the meetup — we had exactly FIFTY attendees, which was a great showing. Part of the large draw was undoubtedly the event sponsor, CRM Metrix (@crm_metrix on Twitter).

    Pre-Meal Networking (and a Friendly Wave from Jonghee!)
    Columbus Web Analytics Wednesday -- Jan 2010

    Hemen Patel, CRM Metrix CTO, facilitated a lively discussion about incorporating the voice of the customer in web site measurement and optimization.

    Hemen Patel Presents
    Columbus Web Analytics Wednesday -- Jan 2010

    Hemen walked through a brief deck (below) that sparked some great back-and-forth with the crowd.

    A Rapt Audience
    Columbus Web Analytics Wednesday -- Jan 2010

    Monish Datta Asks a Question
    Columbus Web Analytics Wednesday -- Jan 2010

    With a crowd of fifty people, not only did I not get to meet the first-time attendees, but I barely had a chance to say, “Hi” to some of the long-time regulars. I guess we’ll just have to have another one in February (I’m working on it!) so I’ll get that chance!

    Analysis, Analytics Strategy, Social Media

    The Spectrum of Data Sources for Marketers Is Wide (& Overwhelming)

    I’ve been using an anecdote of late that Malcolm Gladwell supposedly related at a SAS user conference earlier this year: over the last 30 years, the challenge we face when it comes to using data to drive actions has fundamentally shifted from a challenge of “getting the right data” to “looking at an overwhelming array of data in the right way.” To illustrate, he compared Watergate to Enron — in the former case, the challenge for Woodward and Bernstein was uncovering a relatively small bit of information that, once revealed, led to immediate insight and swift action. In the latter case, the data to show that Enron had built a house of cards was publicly available, but there was so much data that actually figuring out how to extract the underlying chicanery without knowing exactly where to look for it was next to impossible.

    With that in mind, I started thinking about all of the sources of data that marketers now have available to them to drive their decisions. The challenge is that almost all of the data sources out there are good tools — while they all claim competitive advantage and differentiation from other options…I believe in the free markets to the extent that truly bad tools don’t survive (do a Google search for “SPSS Netgenesis” and the first link returned is a 404 page — the prosecution rests!). To avoid getting caught up in the shiny baubles of any given tool, it seems worth organizing the range of available data some way — put every source into a discrete bucket.  It turns out that that’s a pretty tricky thing to do, but one approach would be to put each data source available to us somewhere on a broad spectrum. At one end of the spectrum is data from secondary research — data that someone else has gone out and gathered about an industry, a set of consumers, a trend, or something else. At the other end of the spectrum is the data we collect on our customers in the course of conducting some sort of transaction with them — when someone buys a widget from our web site, we know their name, how they paid, what they bought, and when they bought it!

    For poops and giggles, why not try to fill in that spectrum? Starting from the secondary research end, here we go…!

    Secondary Research (and Journalism…even Journalism 2.0)

    This category has an unlistable number of examples. From analyst firms like Forrester Research and Gartner Group, to trade associations like the AMA or The ARF, to straight-up journalists and trade publications, and even to bloggers. Specialty news aggregators like alltop.com fall into this category as well (even if, technically, they would fit better into a “tertiary research” category, I’m going to just leave them here!).

    I stumbled across iconoculture last week as one interesting company that falls in this category…although things immediately start to get a little messy, because they’ve got some level of primary research as well as some tracking/listening aspects of their offer.

    Listening/Collecting

    Moving along our spectrum of data sources, we get to an area that is positively exploding. These are tools that are almost always built on top of a robust database, because what they do is try to gather and organize what people — consumers — are doing/saying online. As a data source, these are still inherently “secondary” — they’re “what’s happening” and “what’s out there.” But, as our world becomes increasingly digital, this is a powerful source of information.

    One group of tools here are sites like compete.com, Alexa, and even Google’s various “insights” tools: Google Trends, Google Trends for Websites, and Google Insights for Search. These tools tend to not be so much consumer-focussed as site-focussed, but they’re getting their data by collecting what consumers are doing. And they are darn handy.

    “Online listening platforms” are a newer beast, and there seems to be a new player in the space every day. The Forrester Wave report by Suresh Vittal in Q1 2009 seems like it is at least five years old. An incomplete list of companies/tools offering such platforms includes (in no particular order…except Nielsen is first because they’re the source of the registration-free PDF of the Forrester Wave report I just mentioned):

    And the list goes on and on and on… (see Marshall Sponder’s post: 26 Tools for Social Media Monitoring). Each of these tools differentiates itself from their competition in some way, but none of them have truly emerged as a  sustained frontrunner.

    Web Analytics

    I put web analytics next on the spectrum, but recognize that these tools have an internal spectrum all their own. From the “listening/collecting” side of the spectrum, web analytics tools simply “watch” activity on your web site — how many people went where and what they did when they got there. Moving towards the “1:1 transactions” end of the spectrum, web analytics tools collect data on specifically identifiable visitors to your site and provide that user-level specificity for analysis and action.

    Google Analytics pretty much resides at the “watching” end of this list, as does Yahoo! Web Analytics (formerly IndexTools). But, then again, they’re free, and there’s a lot of power in effectively watching activity on your site, so that’s not a knock against them. The other major players — Omniture Sitecatalyst, Webtrends, Coremetrics, and the like — have more robust capabilities and can cover the full range of this mini-spectrum. They all are becoming increasingly open and more able to be integrated with other systems, be that with back-end CRM or marketing automation systems, or be that with the listening/collecting tools described in the prior section.

    The list above covered “traditional web analytics,” but that field is expanding. A/B and multivariate testing tools fall into this category, as they “watch” with a very specific set of options for optimizing a specific aspect of the site. Optimost, Omniture Test&Target, and Google Website Optimizer all fall into this subcategory.

    And, entire companies have popped up to fill specific niches with which traditional web analytics tools have struggled. My favorite example there is Clearsaleing, which uses technology very similar to all of the web analytics tools to capture data, but whose tools are built specifically to provide a meaningful view into campaign performance across multiple touchpoints and multiple channels. The niche their tool fills is improved “attribution management” — there’s even been a Forrester Wave devoted entirely to tools that try to do that (registration required to download the report from Clearsaleing’s site).

    Primary Research

    At this point on the spectrum, we’re talking about tools and techniques for collecting very specific data from consumers — going in with a set of questions that you are trying to get answered. Focus groups, phone surveys, and usability testing all fall in this area, as well as a plethora of online survey tools. Specifically, there are online survey tools designed to work with your web site — Foresee Results and iPerceptions 4Q are two that are solid for different reasons, but the list of tools in that space outnumbers even the list of online listening platforms.

    The challenge with primary research is that you have to make the user aware that you are collecting information for the purpose of research and analysis. That drops a fly in the data ointment, because it is very easy to bias that data by not constructing the questions and the environment correctly. Even with a poorly designed survey, you will collect some powerful data — the problem is that the data may be misleading!

    Transaction Data

    Beyond even primary research is the terminus of the spectrum — it’s customer data that you collect every day as a byproduct of running your business and interacting with customers. Whenever a customer interacts with your call center or makes a purchase on your web site, they are generating data as an artifact. When you send an e-mail to your database, you’ve generated data as to whom you sent the message…and many e-mail tools also track who opened and clicked through on the e-mail. This data can be very useful, but, to be useful, it needs to be captured, cleansed, and stored in a way that sets it up for useful analysis. There’s an entire industry built around customer data management, and most of what the tools and processes in that industry focus on is transaction data.

    What’s Missing?

    As much as I would like to wrap up this post by congratulating myself on providing an all-encompassing framework…I can’t. While there are a lot of specific tools/niches that I haven’t listed here that I could fit somewhere on the spectrum of tools as I’ve described it, there are also sources of valuable data that don’t fit in this framework. One type that jumps out to me is marketing mix-type data and tools (think Analytic Partners, ThinkVine, or MarketShare Partners). I’m sure there are many other types. Nevertheless, it seems like a worthwhile framework to have when it comes to building up a portfolio of data sources. Are you getting data from across the entire spectrum (there are free or near-free tools at every point on the spectrum)? Are you getting redundant data?

    What do you think? Is it possible to organize “all data sources for marketers” in a meaningful way? Is there value in doing so?

    Analysis, Analytics Strategy, Reporting, Social Media

    The Most Meaningful Insights Will Not Come from Web Analytics Alone

    Judah Phillips wrote a post last week laying out why the answer to the question, “Is web analytics hard or easy?” is a resounding “it depends.” It depends, he wrote, on what tools are being used, on how the site being analyzed is built, on the company’s requirements/expectations for analytics, on the skillset of the team doing the analytics, and, finally, on the robustness of the data management processes in place.

    One of the comments on the blog came from John Grono of GAP Research, who, while agreeing with the post, pointed out:

    You refer to this as “web analytics”. I also know that this is what the common parlance is, but truth be known it is actually “website analytics”. “web” is a truncation of “world wide web” which is the aggregation of billions of websites. These tools do not analyse the “web”, but merely individual nominated “websites” that collectively make up the “web”. I know this is semantics … but we as an industry should get it right.

    It’s a valid point. Traditionally, “web analytics” has referred to the analysis of activity that occurs on a company’s web site, rather than on the web as a whole. Increasingly, though, companies are realizing that this is an unduly narrow view:

    • Search engine marketers (SEO and SEM) have, for years, used various keyword research tools to try to determine what words their target customers are using explicitly off-site in a search engine (although the goal of this research has been to use that information to bring these potential customers onto the company’s site)
    • Integration with a company’s CRM and/or marketing automation system — to combine information about a customer’s on-site activity with information about their offline interactions with the company — has been kicked around as a must-do for several years; the major web analytics vendors have made substantial headway in this area over the past few years
    • Of late, analysts and vendors have started looking into the impact of social media and how actions that customers and prospects take online, but not on the company’s web site, play a role in the buying process and generate analyzable data in the process

    The “traditional” web analytics vendors (Omniture, Webtrends, and the like) were, I think, a little late realizing that social media monitoring and measurement was going to turn into a big deal. To their credit, they were just getting to the point where their platforms were opening up enough that CRM and data warehouse integration was practical. I don’t have inside information, but my speculation is that they viewed social media monitoring more as an extension of traditional marketing and media research companies that as an adjacency to their core business that they should consider exploring themselves. In some sense, they were right, as Nielsen, J.D. Power and Associates (through acquisition), Dow Jones, and TNS Media Group all rolled out social media monitoring platforms or services fairly early on. But, the door was also opened for a number of upstarts: Biz360, Radian6, Alterian/Techrigy/SM2, Crimson Hexagon, and others whom I’m sure I’ve left off this quick list. The traditional web analytics vendors have since come to the party through partnerships — leveraging the same integration APIs and capabilities that they developed to integrate with their customers’ internal systems to integrate with these so-called listening platforms.

    Somewhat fortuitously, a minor hashtag snafu hit Twitter in late July when #wa, which had settled in as the hashtag of choice for web analytics tweets was overrun by a spate of tweets about Washington state. Eric Peterson started a thread to kick around alternatives, and the community settled on #measure, which Eric documented on his blog. I like the change for two reasons (notwithstanding those five precious characters that were lost in the process):

    1. As Eric pointed out, measurement is the foundation of analysis — I agree!
    2. “Web analytics,” which really means “website analytics,” is too narrow for what analysts need to be doing

    I had a brief chat with a co-worker on the subject last week, and he told me that he has increasingly been thinking of his work as “digital analytics” rather than “web analytics,” which I liked as well.

    It occurred to me that we’re really now facing two fundamental dimensions when it comes to where our customers (and potential customers) are interacting with our brand:

    • Online or offline — our website, our competitors’ websites, Facebook, blogs, and Twitter are all examples of where relevant digital (online) activities occur, while phone calls, tradeshows, user conferences, and peer discussions are all examples of analog (offline) activities
    • On-site or off-site — this is a bit of a misnomer, but I haven’t figured out the right words yet. But, it really means that customers can interact with the company directly, or, they can have interactions with the company’s brand through non-company channels

    Pictorially, it looks something like this:
    Online / Offline vs. Onsite / Offsite

    I’ve filled in the boxes with broad descriptions of what sort of tools/systems actually collect the data from interactions that happen in each space. My claim is that any analyst who is expecting to deliver meaningful insight for his company needs to understand all four of these quadrants and know how to detect relevant signals that are occuring in them.

    What do you think?

    Analysis, General, Reporting

    Where BI Is Heading (Must Head) to Stay Relevant

    I stumbled across a post by Don Campbell (CTO of BI and Performance Management at IBM — he was at Cognos when they got acquired) today that really got my gears turning. His 10 Red Hot BI Trends provide a lot of food for thought for a single post (for one thing, the post only lists eight trends…huh?). It’s worth clicking over to the post for a read, as I’m not going to repeat the content here.

    BUT…I can’t help but add in my own drool thoughts on some of his ideas:

    1. Green Computing — not much to add here; this is more about next generation mainframes that run on a less power than the processors of yesteryear
    2. Social Networking — it stands to reason that Web 2.0 has a place in BI, and Campbell starts to explain the wherefore and the why. One gap I’ve never seen a BI tool fill effectively is the ability to embed ad hoc comments and explanations within a report. That’s one of the reasons that Excel sticks around — because an Excel based report has to be “produced” in some fashion, there is an opportunity to review, analyze, and provide an assessment within the report. Enterprise BI tools have a much harder time enabling this — when it’s come up with BI tool vendors, it tends to get treated more as a data problem than a tool problem. In other words, “Sure, if you’ve got data about the reports stored somewhere, you can use our tool to display it.” What Campbell starts to touch on in his post is the potential for incorporating social bookmarking (“this view of this data is interesting and here is why”) and commenting/collaboration to truly start blending BI with knowledge management. The challenge is going to be that reports are becoming increasingly dynamic, and users are getting greater control over what they see and how. With roles-based data access, the data that users see on the same report varies from user to user. That’s going to make it challenging to manage “social” collaboration. Challenging…but something that I hope the enterprise BI vendors are trying to overcome.
    3. Data Visualization — I wouldn’t have a category on this blog dedicated to data visualization if I didn’t think this was important. I can’t help but wonder if Campbell is realizing that Cognos was as guilty as the other major BI players of confusing “demo-y neat” with “effective” when it comes to past BI tool feature development. From his post: “The best visualizations do not necessarily involve the most complex graphics or charts, but rather the best representation of the data.” Amen, brother!!! Effective data visualization is finally starting to get some traction — or, at least, a growing list of vocal advocates (side note: Jon Peltier has started up a Chart Busters category on his blog — worth checking out). What I would like to see: BI vendors taking more responsibility for helping their users present data effectively. Maybe a wizard in report builders that ask questions about the type of data being presented? Maybe a blinking red popup warning (preferably with loud sirens) whenever someone selects the 3D effect for a chart? The challenge with data visualization is that soooooo many analysts: 1) are not inherently wired for effective visualization, and 2) wildly underestimate how important it is.
    4. Mobile — I attended a session on mobile BI almost five years ago at a TDWI conference…and I still don’t see this as being a particularly hot topic. Even Campbell, with his mention of RFIDs, seems to think this is as much about new data sources as it is about reporting and analysis in a handheld environment.
    5. Predictive Analytics — this has been the Holy Grail of BI for years. I don’t have enough exposure to enough companies who have successfully operationalized predictive analytics to speak with too much authority here. But, I’d bet good money that every company that is successful in this area has long since mastered the fundamentals of performance measurement. In other words, predictive analytics is the future, but too many businesses are thinking they can run (predictive analytics) before they crawl (performance measurement / KPIs / effective scorecards).
    6. Composite Applications — this seems like a fancy way to say “user-controlled portals.” This really ties into the social networking (or at least Web 2.0), I think, in that a user’s ability to build a custom home page with “widgets” from different data sources that focus on what he/she truly views as important. Taking this a step farther — measuring the usage of those widgets — which ones are turned on, as well as which ones are drilled into — seems like a good way to assess whether what the corporate party line says is important is what line management is really using. There are some intriguing possibilities there as an extension of the “reports on the usage of reports” that gets bandied about any time a company starts coming to terms with report explosion in their BI (or web analytics) environment.
    7. Cloud Computing — I actually had to go and look up the definition of cloud computing a couple of weeks ago after asking a co-worker who used the term if cloud computing and SaaS were the same thing (answer: SaaS is a subset of cloud computing…but probably the most dominant form). This is a must-have for the future of BI — as our lives become increasingly computerized, the days of a locally installed BI client are numbered. I regularly float between three different computers and two Blackberries…and lose patience when what I need to do is tied to only one machine.
    8. Multitouch — think of the zoom in / zoom out capabilities of an iPhone. This, like mobile computing, doesn’t seem so much “hot” to me as somewhat futuristic. The best example of multitouch data exploration that I can think of is John King’s widely-mocked electoral maps on CNN (never did I miss Tim Russert and his handheld whiteboard more than when watching King on election night!). I get the theoretical possibilities…but we’ve got a long ways to go before there is truly a practical application of multitouch.

    As I started with, there are a lot of exciting possibilities to consider here. I hope all of these topics are considered “hot” by BI vendors and BI practicitioners — making headway on just a few of them would get us off the plateau we’ve been on for the past few years.

    Analysis, Reporting

    What is "Analysis?"

    Stephen Few had a recent post, Can Computers Analyze Data?, that started: “Since ‘business analytics’ has come into vogue, like all newly popular technologies, everyone is talking about it but few are defining what it is.” Few’s post was largely a riff off of an article by Merv Adrian on the BeyeNETWORK: Today’s ‘Analytic Applications’ — Misnamed and Mistargeted. Few takes issue (rightly so), with Adrian’s implied definition of the terms “analysis” and “analytics.” Adrian outlines some fair criticisms of BI tool vendors, but Few’s beef regarding his definitions are justified.

    Few defines data analysis as “what we do to make sense of data.” I actually think that is a bit too broad, but I agree with him that analysis, by definition, requires human beings.

    Fancy NancyWith data “coming into vogue,” it’s hard to walk through a Marketing department without hearing references to “data mining” and “analytics.” Given the marketing departments I tend to walk through, and given what I know of their overall data maturity, this is often analogous to someone filling the ice cube trays in their freezer with water and speaking about it in terms of the third law of thermodynamics.

    I’ve got a 3-year-old daughter, and it’s through her that I’ve discovered the Fancy Nancy series of books, in which the main character likes to be elegant and sophisticated well beyond her single-digit age. She regularly uses a word and then qualifies it as “that’s a fancy way to say…” a simpler word. For instance, she notes that “perplexed” is a fancy word for “mixed up.”

    “Analytics” is a Fancy Nancy word. “Web analytics” is a wild misnomer. Most web analysts will tell you there’s a lot of work to do with just basic web site measurement. And, that work is seldom what I would consider “analytics.” As cliché as it is, you can think about data usage as a pyramid, with metrics forming the foundation and analysis (and analytics) being built on top of them.

    Metrics Analysis Pyramid

    There are two main types of data usage:

    • Metrics / Reporting — this is the foundation of using data effectively; it’s the way you assess whether you are meeting your objectives and achieving meaningful outcomes. Key Performance Indicators (KPIs) live squarely in the world of metrics (KPIs are a fancy way to say “meaningful metrics”). Avinash Kaushik defines KPIs brilliantly: “Measures that help you understand how you are doing against your objectives.” Metrics are backward-looking. They answer the question: “Did I achieve what I set out to do?” They are assessed against targets that were set long before the latest report was pulled. Without metrics, analysis is meaningless.
    • Analysis — analysis is all about hypothesis testing. The key with analysis is that you must have a clear objective, you must have clearly articulated hypotheses, and, unless you are simply looking to throw time and money away, you must validate that the analysis will lead to different future actions based on different possible outcomes. Analysis tends to be backward looking as well — asking questions, “Why did that happen?”…but with the expectation that, once you understand why something happened, you will take different future actions using the knowledge.

    So, what about “analytics?” I asked that question of the manager of a very successful business intelligence department some years back. Her take has always resonated with me: “analytics” are forward-looking and are explicitly intended to be predictive. So, in my pyramid view, analytics is at the top of the structure — it’s “advanced analysis,” in many ways. While analysis may be performed by anyone with a spreadsheet, and hypotheses can be tested using basic charts and graphs, analytics gets into a more rigorous statistical world: more complex analysis that requires more sophisticated techniques, often using larger data sets and looking for results that are much more subtle. AND, using those results, in many cases, to build a predictive model that is truly forward-looking.

    The key is that the foundation of your business (whether it’s the entire company, or just your department, or even just your own individual role) is your vision. From your vision comes your strategy. From your strategy come your objectives and your tactics. If you’re looking to use data, the best place to start is with those objectives — how can you measure whether you are meeting them, and, with the measures you settle on, what is the threshold whereby you would consider that you achieved your objective? Attempting to do any analysis (much less analytics!) before really nailing down a solid foundation of objectives-oriented metrics is like trying to build a pyramid from the top down. It won’t work.

    Analysis, Analytics Strategy, Excel Tips, General, Presentation, Reporting

    The Best Little Book on Data

    How’s that for a book title? Would it pique your interest? Would you download it and read it? Do you have friends or co-workers who would be interested in it?

    Why am I asking?

    Because it doesn’t exist. Yet. Call it a working title for a project I’ve been kicking around in my head for a couple of years. In a lot of ways, this blog has been and continues to be a way for me to jot down and try out ideas to include in the book. This is my first stab at trying to capture a real structure, though.

    The Best Little Book on Data

    In my mind, the book will be a quick, easy read — as entertaining as a greased pig loose at a black-tie political fundraiser — but will really hammer home some key concepts around how to use data effectively. If I’m lucky, I’ll talk a cartoonist into some pen-and-ink, one-panel chucklers to sprinkle throughout it. I’ll come up with some sort of theme that will tie the chapter titles together — “myths” would be good…except that means every title is basically a negative of the subject; “Commandments” could work…but I’m too inherently politically correct to really be comfortable with biblical overtones; an “…In which our hero…” style (the “hero” being the reader, I guess?). Obviously, I need to work that out.

    First cut at the structure:

    • Introduction — who this book is for; in a nutshell, it’s targeted at anyone in business who knows they have a lot of data, who knows they need to be using that data…but who wants some practical tips and concepts as to how to actually go about doing just that.
    • Chapter 1: Start with the Data…If You Want to Guarantee Failure — it’s tempting to think that, to use data effectively, the first thing you should do is go out and query/pull the data that you’re interested in. That’s a great way to get lost in spreadsheets and emerge hours (or days!) later with some charts that are, at best, interesting but not actionable, and, at worst, not even interesting.
    • Chapter 2: Metrics vs. Analysis — providing some real clarity regarding the fundamentally different ways to “use data.” Metrics are for performance measurement and monitoring — they are all about the “what” and are tied to objectives and targets. Analysis is all about the “why” — it’s exploratory and needs to be hypothesis driven. Operational data is a third way, but not really covered in the book, so probably described here just to complete the framework.
    • Chapter 3: Objective Clarity — a deeper dive into setting up metrics/performance measurement, and how to start with being clear as to the objectives for what’s being measured, going from there to identifying metrics (direct measures combined with proxy measures), establishing targets for the metrics (and why, “I can’t set one until I’ve tracked it for a while” is a total copout), and validating the framework
    • Chapter 4: When “The Metric Went Up” Doesn’t Mean a Gosh Darn Thing — another chapter on metrics/performance measuremen. A discussion of the temptation to over-interpret time-based performance metrics. If a key metric is higher this month than last month…it doesn’t necessarily mean things are improving. This includes a high-level discussion of “signal vs. noise,” an illustration of how easy it is to get lulled into believing something is “good” or “bad” when it’s really “inconclusive,” and some techniques for avoiding this pitfall (such as using simple, rudimentary control limits to frame trend data).
    • Chapter 5: Remember the Scientific Method? — a deeper dive on analysis and how it needs to be hypothesis-driven…but with the twist that you should validate that the results will be actionable just by assessing the hypothesis before actually pulling data and conducting the analysis
    • Chapter 6: Data Visualization Matters — largely, a summary/highlights of the stellar work that Stephen Few has done (and, since he built on Tufte’s work, I’m sure there would be some level of homage to him as well). This will include a discussion of how graphic designers tend to not be wired to think about data and analysis, while highly data-oriented people tend to fall short when it comes to visual talent. Yet…to really deliver useful information, these have to come together. And, of course, illustrative before/after examples.
    • Chapter 7: Microsoft Excel…and Why BI Vendors Hate It — the BI industry has tried to equate MS Excel with “spreadmarts” and, by extension, deride any company that is relying heavily on Excel for reporting and/or analysis as being wildly early on the maturity curve when it comes to using data. This chapter will blow some holes in that…while also providing guidance on when/where/how BI tools are needed (I don’t know where data warehousing will fit in — this chapter, a new chapter, or not at all). This chapter would also reference some freely downloadable spreadsheets with examples, macros, and instructions for customizing an Excel implementation to do some of the data visualization work that Excel can do…but doesn’t default to. Hmmm… JT? Miriam? I’m seeing myself snooping for some help from the experts on these!
    • Chapter 8: Your Data is Dirty. Get Over It. — CRM data, ERP data, web analytics data, it doesn’t matter what kind of data. It’s always dirtier than the people who haven’t really drilled down into it assume. It’s really easy to get hung up on this when you start digging into it…and that’s a good way to waste a lot of effort. Which isn’t to say that some understanding of data gaps and shortcomings isn’t important.
    • Chapter 9: Web Analytics — I’m not sure exactly where this fits, but it feels like it would be a mistake to not provide at least a basic overview of web analytics, pitfalls (which really go to not applying the core concepts already covered, but web analytics tools make it easy to forget them), and maybe even providing some thoughts on social media measurement.
    • Chapter 10: A Collection of Data Cliches and Myths — This may actually be more of an appendix, but it’s worth sharing the cliches that are wrong and myths that are worth filing away, I think: “the myth of the step function” (unrealistic expectations), “the myth that people are cows” (might put this in the web analytics section), “if you can’t measure it, don’t do it” (and why that’s just plain silliness)
    • Chapter 11: Bringing It All Together — I assume there will be such a chapter, but I’m going to have to rely on nailing the theme and the overall structure before I know how it will shake out.

    What do you think? What’s missing? Which of these remind you of anecdotes in your own experience (haven’t you always dreamed of being included in the Acknowledgments section of a book? Even if it’s a free eBook?)? What topic(s) are you most interested in? Back to the questions I opened this post with — would you be interested in reading this book, and do you have friends or co-workers who would be interested? Or, am I just imagining that this would fill a gap that many businesses are struggling with?

    Analysis, Reporting

    Performance Measurement vs. Analysis

    I’ve picked up some new terminology over the course of the past few weeks thanks to an intermediate statistics class I’m taking. Specifically — what inspired this post — is the distinction between two types of statistical studies, as defined by one of the fathers of statisical process control, W. Edwards Deming. There’s a Wikipedia entry that actually defines them and the point of making the distinction quite well:

    • Enumerative study: A statistical study in which action will be taken on the material in the frame being studied.
    • Analytic study: A statistical study in which action will be taken on the process or cause-system that produced the frame being studied. The aim being to improve practice in the future.

    …In other words, an enumerative study is a statistical study in which the focus is on judgment of results, and an analytic study is one in which the focus is on improvement of the process or system which created the results being evaluated and which will continue creating results in the future. A statistical study can be enumerative or analytic, but it cannot be both.

    I’ve now been at three different schools in three different states where one of the favorite examples used for processes and process control is a process for producing plastic yogurt cups. I don’t know if Yoplait just pumps an insane amount of funding into academia-based research, or if there is some other reason, but I’ll go ahead and perpetuate it by using the same as an example here:

    • Enumerative study — imagine that the yogurt cup manufacturer is contractually bound to provide shipments where less than 0.1% of the cups are defective. Imagine, also, that to fully test a cup requires destroying it in the process of the test. Using statistics, the manufacturer can pull a sample from each shipment, test those cups, and, if the sampling is set up properly, be able to predict with reasonable confidence the proportion of defective cups in the entire shipment. If the prediction exceeds 0.1%, then the entire shipment can be scrapped rather than risking a contract breach. The same test would be conducted on each shipment.
    • Analytic study — now, suppose the yogurt cup manufacturer finds that he is scrapping one shipment in five based on the process described in the enumerative study. This isn’t a financially viable way to continue. So, he decides to conduct a study to try to determine what factors in his process are causing cups to come out defective. In this case, he may set up a very different study — isolating as many factors in the process as he can to see if can identify where the trouble spots in the process itself are and fix them.

    It’s not an either/or scenario. Even if an analytics study (or series of studies) enables him to improve the process, he will likely still need to continue the enumerative studies to identify bad batches when they do occur.

    In the class, we have talked about how, in marketing, we are much more often faced with analytic situations rather than enumerative ones. I don’t think this is the case. As I’ve mulled it over, it seems like enumerative studies are typically about performance measurement, while analytic studies are about diagnostics and continuous improvement. See if the following table makes sense:

    Enumerative Analytic
    Performance management Analysis for continuous improvement
    How did we do in the past? How can we do better in the future?
    Report Analysis

    Achievement tests administered to schoolchildren are more enumerative than analytic — they are not geared towards determining which teaching techniques work better or worse, or even to provide the student with information about what to focus on and how going forward. They are merely an assessment of the student’s knowledge. In aggregate, they can be used as an assessment of a teacher’s effectiveness, or a school’s, or a school district’s, or even a state’s.

    “But…wait!” you cry! “If an achievement test can be used to identify which teachers are performing better than others, then your so-called ‘process’ can be improved by simply getting rid of the lowest performing teachers, and that’s inherently an analytic outcome!” Maybe so…but I don’t think so. It simply assumes that each teacher is either good, bad, or somewhere in between. Achievement tests do nothing to indicate why a bad teacher is a bad teacher and a good teacher is a good teacher. Now, if the results of the achievement tests are used to identify a sample of good and bad teachers, and then they are observed and studied, then we’re back to an analytic scenario.

    Let’s look at a marketing campaign. All too often, we throw out that we want to “measure the results of the campaign.” My claim is that there are two very distinct purposes for doing so…and both the measurement methods and the type of action to be taken are very different:

    • Enumerative/performance measurement — Did the campaign perform as it was planned? Did we achieve the results we expected? Did the people who planned and executed the campaign deliver on what was expected of them?
    • Analytic/analysis — What aspects of the campaign were the most/least effective? What learnings can we take forward to the next campaign so that we will achieve better results the next time?

    In practice, you will want to do both. And, you will have to do both at the same time. I would argue that you need to think about the two different types and purposes as separate animals, though, rather than expecting to “measure the results” and muddle them together.

    Analysis

    The "Right" Talent: an MVT-Meets-Fractional-Factorial-Design Anecdote

    When it comes to business/management books, one of my favorites is First, Break All the Rules. When I first read it, it didn’t strike me as particularly profound. I was relatively new to managing people, and I was being “forced” to read it for an internal class, so my natural reaction was to view it cynically. I’m not proud of it, but that’s how I roll.

    Over time, I’ve found myself quoting and recommending the book again and again (I can’t say the same for the follow-up book — Now, Discover Your Strengths — but that’s a topic for another post). The fundamental premise of First, Break All the Rules, goes something like:

    1. There is a difference between skills and talents — the former can be taught, whereas the latter are more innate characteristics, the combination of which make each person unique
    2. Conventional wisdom has managers focussing on hiring for skills and then focussing on employees’ weaknesses and trying to “fix” them
    3. The “weaknesses” are, all too often, talents the employee simply does not have
    4. It’s much better to identify each employee’s talents/strengths and then help them figure out how to capitalize on those strengths rather than simply managing to their weaknesses

    The book also explains the fallacy of spending a disproportionate amount of time with your weakest employees, which is where day-to-day management tends to pull you. 

    And, I’m possibly butchering the book in my summary — it’s been a few years since I reread it!

    When it comes to data-driven job roles — think most roles where the word “analyst” appears in the job title — the real challenge is finding people who have the right mix of talents. A big part of what we worked on in the Business Intelligence department at National Instruments was building a capability to operate effectively in two different dimensions:

    • As business experts — we recognized that we needed to know our business and the business of sales and marketing as well or better than our internal customers. We had to genuinely want to understand the business problems they were facing — know them well enough that we could articulate them effectively on our own.
    • As data geeks — at the end of the day, we were expected to be able to pull data, analyze it, explain the results, and present them effectively.

    What made our team effective, in my view, is that everyone in the department was very strong in one of these areas, and at least competent in the other. And, we paired up people with complementary skills when it came to tackling any project. On the one hand, this sounds like it was inefficient, but it really wasn’t — it didn’t mean that these teams were joined at the hip and never operated alone. Rather, it meant that they collaborated — both directly with our internal customers as well as offline with each other — to come at each project from multiple angles.

    Now, here I am several years later, taking an Intermediate Statistics class through The Ohio State University. The class is taught on-site at my company, and it’s taught by a professor who spent a big chunk of his career in applied statistics working for Battelle. He’s a good professor, and I particularly like that he beats a pretty hard drum when it comes to the parallel talents needed to effectively use data in a business setting: subject matter expertise, effective problem formulation, and statistical/analytical knowledge. He rails against trained statisticians and even “applied mathematicians” who don’t want to really address the first two requirements head-on — those who jump to crunching the data prematurely, relying on their technical tools and skills to route them to “the answer.”

    Bravo, I say!

    At the same time, even as the professor is eloquently speaking to this issue (and politely patting himself on the back for not falling into the trap), it’s readily apparent that he’s coming at the analysis of data from a heavy background of hardcore statistics. And, while he spent much of his career working on industrial (and defense) processes and problems, he is now mired full-time in the world of marketing and consumer behavior. He is not the first data guru to cross over, by any means, but he is clearly new to the space, and, in many ways, seems hell-bent on retreading ground that has been covered already. 

    As one example, there is the case of MVT, or multivariate testing. MVT has been getting a lot of buzz in marketing over the past five years or so. It’s been touted as a way to accurately test many different variables without having to run a gazillion experiments to test every combination of them. One place that MVT gets used these days is with web landing page design — enabling a marketer to test various color schemes, banner ad taglines, and headline placements to derive the optimal combination without breaking the bank with the number of tests that have to be run to make a valid, data-driven decision. That’s all well and good, and it’s clearly gotten enough traction that it works.

    An old-school process improvement expert I worked with back when I was first starting to hear about MVT pulled me aside one day and said, “You know, Tim, none of this stuff is really knew — MVT’s been around since World War II. It just wasn’t applied to Marketing until recently, so there are a lot of people capitalizing on it.” And…he was right!

    So, back to the present day. In a recent class, this OSU professor was walking through various ways to design experiments and how they could be analyzed, and he kept referencing “fractional factorial design.” We’re only going to touch on the technique in the class, he assured us, but he explained how there were trade-offs you have to make when using that approach. From his explanation, it sounded like fractional factorial design was a lot like MVT, and I asked him about it after class. He had never heard of the “MVT” acronym, but said it sure sounded a lot like fractional factorial design (which was a term that was entirely knew to me). It only took a few seconds on Google to find out that our suspicions were correct.

    I was surprised that MVT was a new term for this professor, as it’s hard to do much of any poking around in marketing circles these days without stumbling into it.

    At the same time, “fractional factorial design” was a new term for me. But, then again, I’m coming at things from a background much more grounded in marketing and business, rather than deep mathematics. I understand the point of MVT, but I’m still wildly fuzzy on the actual mechanics of it.

    And…that’s the short point of this lengthy post: it seems like, in order to truly use data effectively, requires a mix of talents that rarely occur naturally in a single person. This professor has the deep statistical chops and an awareness that he is not a subject matter expert when it comes to marketing and consumer behavior, so he needs to increase that knowledge. Partner him with someone with a deep marketing background (who is almost certainly not also a statistician), and, as long as that person has an awareness of analytics and an interest in applying analysis effectively, you’ve got a winning formula.

    It’s a Venn Diagram, really. If there is no intersection of the talents, then chances are it’s a combination of talents doomed to fail.

    UPDATE: Check the first comment below for a clarification — MVT is not the same thing as fractional factorial design. Rather, fractional factorial is one approach for conducting MVT experiment analysis. The link on the subject earlier in this post makes this same point. I was unclear/ambiguous. I still think that, in Marketing circles, when MVT has gotten recent play, that it is fractional factorial design and analysis that gets the buzz. But, I don’t know for sure.

    Analysis

    "The Axiom of Research" and "The Axiom of Action"

    I attended a one-day seminar today on “The Role of Statistical Concepts and Methods in Research” taught by Dr. Tom Bishop of The Ohio State University. Dr. Bishop heads up a pretty cool collaboration between Nationwide (all areas of the company, including Nationwide: Car Insurance) and OSU, and this seminar was one of the types of minor artifacts of that collaboration.

    Dr. Bishop had me on page 5 of the seminar materials when he introduced “The Fundamental Axioms of Research,” which he stated are twofold:

    • The Axiom of Variation — all research data used for inference and decision making are subject to uncertainty and variation
    • The Axiom of Action — in research, theory is developed, experiments are conducted, and data are collected and analyzed to generate knowledge to form a rational basis for action

    The rest of the seminar was a healthy mix of theory and application, with all of the “theory” being tied directly to how it should be applied correctly. Dr. Bishop is a statistician by training with years of industry experience, so it was pretty cool to hear him emphasize again and again and again that you can get a lot of value from the data without running all sorts of complex, Greek-letter-rich, statistical analyses. The key is to apply a high degree of rigor in understanding and articulating the problem and the approach.

    Lots of good stuff!

    Analysis, Presentation, Reporting

    Techrigy — New Kid on the Social Media Measurement Block

    When Connie Bensen posted that she had formalized a relationship with Techrigy to work on their community, I had to take a look! She gave me a demo of their SM2 product today, and it is very cool. SM2 is pretty clearly competing with radian6, in that their tool is geared around measuring and monitoring a brand/person/company/product’s presence in the world of Web 2.0. I’m not an expert on this space by any means, although I have caught myself describing these sorts of tools as “clip services” for social media. But, hey, I’m not a PR person, either, so I barely know what clip services do!

    I started out by stating how little I know about this area for a reason. It’s because this post is my take on the tool from something of a business intelligence purist perspective. Take it for what it’s worth.

    What I Liked

    The things that impressed me about SM2 — either enough to stick in my head through the rest of the day or because I jotted them down:

    • They brought a community expert (Connie) on board early; on the one hand, Connie is there to help them “build their community,” which, in and of itself, is a pretty brilliant move. But, what they’ve gotten at the same time is someone who is going to use their product heavily to support herself in the role, which means they’ll be eating their own dogfood and getting a lot of great feedback about what does/does not work from a true thought leader in the space. More on what I expect on that on the “Opportunities for Maturity” below…
    • The tool keeps data for all time — it doesn’t truncate after 30 days or, as I understand it, aggregate data over a certain age so that there is less granularity. I’m not entirely sure, but it sort of sounds like the tool is sitting on a Teradata warehouse. If that’s the case, then they’re starting off with some real data storage and retrieval horsepower — it’s likely to scale well
      UPDATE: I got clarification from Techrigy, and it’s not Teradata (too expensive) as the data store. It’s “a massively parallel array of commodity databases/hardware.” That sounds like fun!
    • Users can actually add data and notes in various ways to the tool; a major hurdle for many BI tools is that they are built to allow users to query, report, slice, dice, and, generally pull data…but don’t provide users with a way to annotate the data; I would claim this is one of the reasons that Excel remains so popular — users need to make notes on the data as they’re evaluating it. Some of the ways SM2 allows this sort of thing:
      • On some of there core trending charts, the user can enter “events” — providing color around a spike or dip by noting a particular promotion, related news event, a crisis of some sort, etc. That is cool:
      • The tool allows drilling down all the way to specific blog authors — there is a “Notes” section where the user can actually comment about the author: “tried to contact three times and never heard back,” “is very interested in what we’re doing,” etc. This is by no means a robust workflow, but is seemed like it would have some useful applications
      • The user could override some of the assessments that the tool made — if it included references from “high authority” sources that really weren’t…the user could change the rating of the reference
    • Integration at some level with Technorati, Alexa, and compete.com — it’s great to see third-party data sources included out of the box (although it’s not entirely clear how deep that integration goes); all three of these have their own shortcomings, but they all also have a wealth of data and are good at what they do; SM2 actually has an “SM2 Popularity” calculation that is analogous to Technorati Authority (or Google PageRank, to extend it a bit farther)
    • The overall interface is very clean — much more Google Analytics‘y than WebTrends-y (sorry, WebTrends)

    Overall, the tool looks very promising! But, it’s still got a little growing up to do, from what I could see.

    Opportunities for Maturity

    I need to put in another disclaimer: I got an hour long demo of the tool. I saw it, but haven’t used it.

    With that said, there were a few things that jumped out at me as, “Whoa there, Nellie!” issues. All are fixable and, I suspect, fixable rather easily:

    • I said the interface overall was really clean, and the screen capture above is a good example — Stephen Few would be proud, for the most part. Unfortunately, there are some pretty big no-no’s buried in the application as well from a data visualization perspective:
      • The 3D effect on a bar chart is pointless and evil
      • The tool uses pie charts periodically, which are generally a bad idea; worse, though, is that they frequently represent data where there is a significant “Unknown” percentage — the tool consistently seems to put “Unknown: <number>” under the graph. The problem is that pie charts are deeply rooted in our brains to represent “the whole” — not “the whole…except for the 90% that we’re excluding”

      The good news on this is that, whatever tool SM2 is running under the hood to do the visualization clearly has the flexibility to present the data just about any way they want (see the screen capture earlier in this post; it should be an easy fix

    • The “flexibility” of the tool is currently taken to a bit of an extreme. This is really a bit of an add-on to the prior point — it doesn’t look like any capabilities of the underlying visual display tool have been turned off. There are charting and graphing options that make the data completely nonsensical. This is actually fairly common in technology-driven companies (especially software companies): make the tool infinitely flexible so that the user “can” do anything he wants. The problem? Most of the users are going to simply stick with the defaults…and even more so if clicking on any of the buttons to tweak the defaults brings on a tidal wave of flexibility. Can you say…Microsoft Word?
    • There is some language/labeling inconsistency in the tool, which they’re clearly working to clean up. But, the tool has the concept of “Categories,” which, as far as I could tell, was a flat list of taggability. That meant that a “category” could be “Blogs.” Another category could be “Blogger,” which is a subset of Blogs…presumably. Another category could be “mobile healthcare,” which is really more of a keyword. In some places, these different types of tags/categories were split out, but the “Categories” area, which can be used for filtering and slicing the data, seemed to invite apples-and-oranges comparison. This one, definitely, may just be me not fully understanding the tool

    Overall, Though, I’d Give It a “Strong Buy”

    The company and the product seem to have a really solid foundation — strategy, approach, infrastructure, and so on. There are some little things that jumped out at me as clear areas for improvement…but they’re small and agile, so I suspect they’ll take feedback and incorporate it quickly. And, most of the things I noticed are the same traps that the enterprise BI vendors stumble into release after release after release.

    Mostly, I’m interested to see what Connie comes up with as she gets in and actually road tests the tool for herself and for Techrigy. In one sense, SM2 is “just” an efficiency tool — it’s pulling together and reporting data that is available already through Google Alerts, Twitter Search, Twemes, Technorati, and so on. And, with many of these tools providing information through customized RSS feeds, a little work with Yahoo! Pipes can aggregate that information nicely. The problem is that it takes a lot of digging to get that set up, and the end result is still going to be clunky. SM2 is set up to do a really nice job of knocking out that legwork and presenting the information in a way that is useful and actionable.

    Fun stuff!

    Analysis, Reporting

    VORP, EqA, FIP and Pure "Data" as the Answer

    I’ve written about baseball before, and I’ll do it again. My local paper, The Columbus Dispatch, had a Sunday Sports cover page two weekends ago titled Going Deeper – Baseball traditionalists make way for a new kind of statistician, one who looks beyond batting averages and homers and praises players’ EqA and VORP. The article caught my eye for several reasons:

    • Lance Berkman was pictured embedded in the article — hey, I’ll always be a Texan no matter where I live, and “The Big Puma” has been one of the real feel-good stories for the Astros for the past few years (I’ll overlook that he played his college ball at Rice, the non-conference arch nemesis of my beloved Longhorns)
    • The graphic above the article featured five stats…of which I only recognized one (OPS)
    • The article is written around the Cleveland Indians, who have one of the worst records in major league baseball this year

    With my wife and kids out of town, I got to head to the local bagel shop and actually read beyond the front page of the paper, and the article was interesting. The kicker remains that the article leads off by talking about two members of the Indians front office: Eric Wedge is a traditional, up-through-the-ranks-as-a-player baseball guy; Keith Woolner has two degrees from MIT, a master’s degree from Stanford, and a ton of experience working for software companies. The article treats these two men as the ying and yang of modern baseball, pointing out that both men have experience and knowledge that’s useful to their boss, Indians GM Mark Shapiro.

    The problem? The Indians stink this year.

    Nonetheless, there’s a great quote in the article from Wedge:

    “What I think people get in trouble with is when they go all feel or all numbers. You have to put it all together and look at everything, then make your best decision. You can’t have an ego about it.”

    The same holds true in business — if your strategy is simply “analyze the data,” you don’t really have a strategy. You’ve got to use your experience, your assessment of where your market is heading as the world changes, some real clarity about what you are and are not good at, an understanding of your competitors (who they are and where they’re stronger than you are), and then lay out your strategy. And stick to it. The data? It’s important! Use it while exploring different strategies to test a hypotheis here and there, and even to model different scenarios and how things will play out depending on different assumptions about the future. But, don’t sit back and wait for the data to set your strategy for you.

    Once you’ve set your strategy, you need to break that down into the tactics that you are going to employ. And the success/failure of those tactics need to be measured so that you are continuosly improving your operations. But don’t get caught up in thinking that the data is the start, the middle, and the end. If it was, we’d all just go out and buy SAS and let the numbers set our course.

    So, what about the goofy acronyms in the title of this post? Well:

    • VORP (Value Over Replacement Player) — a statistic that looks to compare how much more a player is worth than a base-level, attainable big leaguer playing the same position (Berkman had the highest VORP at the time of the article)
    • EqA (Equivalent Average) — think of this as Batting Average 2.0, but it takes into account different leagues and ballparks to try to make the measure as equitable as possible
    • FIP (Fielding Independent Pitching) — this is sort of ERA 2.0, but it tries to assess everything that a pitcher is solely responsible for, rather than simply earned runs

    The fact is, these are good metrics, even if they start to bend the “it has to be easy to understand” rule. In baseball, there have been a lot of people looking at a lot of data over a long period. My guess is that there were many fans and professionals who realized the shortcomings of batting average and ERA, and it was only a matter of time before someone started tuning these metrics and looking for new ones to fill in the gaps.

    At the end of the day, the Indians stink. And it’s a game. And there are countless variables at play that will never be fully captured and analyzable (the same holds true in business). Mark Shapiro will continue to have to make countless decisions based on his instincts, with data as merely one important input. Maybe they won’t stink next season.

    Analysis

    You Might Be Overanalyzing If…

    I was working with a client last week who was looking to update their lead scoring. This wasn’t any fancy-schmancy multidimensional lead scoring — it was plain ol’ pick-a-few-fields-and-assign-’em-some-values lead scoring. Which is a great place to start.

    In this case, the company was in the process of streamlining their registration form on one area of their web site. This was an experiment to see if we could improve their registration form conversion rates by reducing the number and complexity of the fields they required visitors to fill out. We took a good hard look at the fields and asked two things: 1) Do we really need to know this information up front? 2) Is the information “easy” to provide (an “Industry” list with 25 fields was deemed “hard,” because the visitor had to scan through the whole list and then make a judgment call as to which industry most fit his situation).

    The result was that we combined a couple of fields, removed a couple of fields, and reworded one question and the possible answers. So far, so good. The kicker was that these changes, while still giving us all of the same underlying information that the company was using to assess the quality of their leads, required changing the lead scoring formula. The formula was going from having three variables to two, because two of the scored variables had been merged into a single, much shorter, much clearer field.

    We interrupt this blog entry to provide an aside on cognitive dissonance

    The company’s existing three-variable lead score was fairly problematic. When qualitatively assessing a batch of leads, the Sales organization could always pick out a number of high-scoring leads whom they were not interested in calling, and they could pick out a number of low-scoring leads who they absolutely wanted to reach out to. “Our lead score is pretty awful,” was the general consensus.

    At the same time, the lead score was used at an aggregate level — by the same people — to assess the results from various lead generating activities. “We had 35 leads that scored over 1,000! This event was great!”

    We’ll go with the wiktionary defintion of cognitive dissonance: “a conflict or anxiety resulting from inconsistencies between one’s beliefs and one’s actions or other beliefs.” In this case, a strongly held belief that the lead scoring was fatally flawed, and an equally strongly held belief that the lead score was a great way to assess the results of lead gen efforts.

    Initially, we (I) actually let the latter belief prevail, and I struggled to come up with a new lead scoring formula and value weighting that would provide as similar as possible an assessment of each lead between the old scoring system and the new.

    And I kept hitting dead ends.

    In then occurred to me that, by going through the exercise to streamline the fields, we had actually gained some valuable insight into what the Sales organization did/did not see as important qualification criteria for the leads that were sent to them.

    So, I started over.

    The two scored fields that we were planning to continue to capture were “Job Role” and “Annual Revenue.” Job role was a hybrid of job title and department — a short list that really honed in on the types of people who were most likely to be influencers or decision-makers when it came to the company’s services. We’d discovered, while getting to those fields on the registration form, that if a company had greater than $25 million in revenue in any year, the Sales organization wanted to talk to them regardless of their role in the company. Likewise, there were a handful of job roles that, regardless of the (reported) annual revenue, Sales wanted to talk to them as well. So, we started by making sure that those “trump” values would put the lead over the qualification threshhold regardless of the other field’s value. We then worked backwards from there to the mid-tier fields — fields that, if the other field was promising, then Sales would want to talk to the lead. And so on from there. This was much more an exercise in logic than an exercise in analysis. But, it made more sense than the lead score it was replacing.

    As a check, we compared a sample of leads using both the old and new scoring methods. We highlighted a random set of leads that would have moved from below the qualification threshhold in the old scoring system to above it in the new, and vice versa. The majority of these shifts made sense. And, overall, we were looking like we would be qualifying a slightly higher percentage of leads under the new scoring system. We patted ourselves on the back, summarized the changes, the logic, and the before-vs-after results…and headed down to Sales to make sure they were looped in and could identify any gaping holes in our logic.

    Instead…they honed in on two things:

    • The slight increase in leads that would be qualified using the new system
    • One lead who had a very low level job title…at a >$1 billion company — she was not qualified under the old system but became qualified under the new

    Things then got a bit ugly, which was unfortunate. Cognitive dissonance again. The old system let plenty of not-good leads through to Sales and kept just as many good leads out. And it was not really fixable by simply tweaking the formula. It was broken.

    The new system took input directly from the Sales organization and, using the two attributes they cared about the most, applied a logical approach. But, lead scoring is not perfect. The only way to have a “perfect” lead score is to ask your leads 50 questions, check the veracity of all of their answers, and build up a very complex system for taking all of those variables into account. In a way, multidimensional lead scoring is a step in that direction…without putting an undue burden on the lead to answer so many questions, and without requiring a PhD and a Cray supercomputer to develop the right formula.

    But, lead scoring is really simply intended to identify the “best” leads, to disqualify the clearly bad leads, and to leave a pretty big gray area where the quality of the lead simply isn’t known. It’s then up to the individual situation to determine where in that gray area to put the qualification threshold. The higher the threshhold, the fewer false positives and more false negatives there will be. The lower the threshold, the fewer false negatives, but the more false positives.

    “Analysis paralysis” is a cliché, but it’s a well-warranted one. Looking for perfection when you shouldn’t expect it to exist can be crippling.

    Analysis, Presentation, Reporting

    The "Action Dashboard" — Avinash Mounts My Favorite Soapbox

    Avinash Kaushik has a great post today titled The “Action Dashboard” (An Alternative to Crappy Dashboards. As usual, Avinash is spot-on with his observations about how to make data truly useful. He provides a pretty interesting 4-quadrant dashboard framework (as a transitional step to an even more powerful dashboard). I’ve gotten red in the face more times than I care to count when it comes to trying to get some of the concepts he presents across. It’s a slow process that requires quite a bit of patience. For a more complete take on my thoughts check out my post over on the Bulldog Solutions blog.

    And, yes, I’m posting here and pointing to another post that I wrote on a completely different blog. We’ve recently re-launched the Bulldog Solutions blog — new platform, and, we hope, with a more focussed purpose and strategy. What I haven’t fully worked out yet is how to determine when to post here and when to post there…and when to post here AND there (like this post).

    It may be that we find out that we’re not quite as ready to be as transparent as we ought to be over on the corporate blog, in which case this blog may get some posts that are more “my fringe opinion” than will fly on the corporate blog. I don’t know. We’ll see. I know I’m not the first person to face the challenge of contributing to multiple blogs (I’ve also got my wife’s and my personal blog…but that one’s pretty easy to carve off).

    Analysis, Analytics Strategy

    Complex Processes and Analyses Therein

    Stéphane Hamel, it seems, is a bit peeved with Eric Peterson. These are two pretty big names in web analytics — Eric as one of the fathers of web analytics, and Stéphane as both a thought leader in the space as well as the creator of one of the most practical, useful web analytics supplemental tools out there — WASP: The Web Analytics Solution Profiler plugin for Firefox. With the plugin, you visit any site, and a sidebar will tell you what web analytics solutions it looks like it’s running. It’s pretty cool.

    I don’t know the full background of the current back-and-forth between these two guys, but I’m a huge fan of Stéphane, and my ears perked up when I read this observation in the post:

    Business Process Analysis implies understanding & improving a collection of interrelated tasks which solve a particular issue. Nothing new here… Most businesses face complex and “hard” processes, and the way to make them “easy” is by decomposing them into smaller sub-processes until they are manageable.

    For one thing, for a period of ~8 months, my job title was “Director of Business Process Analytics.” And, frankly, I was never sure what that meant. In hindsight, if I’d had these two sentences from Stéphane and if I’d replaced “Analytics” with “Analysis,” I would have seen a much clearer mapping from my label to what I was actually doing in the role.

    More important, though, is the concept of “decomposition.” I find myself preaching the Decomposition Doctrine regularly. And I believe in it strongly.

    As an example, when it comes to the Holy Grail of Marketing Analysis — calculating the ROI of your marketing spend — many, many B2B marketers start out looking for the correlation between leads generated and revenue. I have yet to see a case in B2B where this can be found with a sufficiently tight, sustained correlation to be meaningful. That actually makes sense. It’s like looking for a correlation between the state someone is born in and the achievement of a PhD. There’s a lot going on over time between Point A and Point Z.

    In the case of B2B marketing, decomposition makes sense. Decompose the process:

    • The lead-to-qualified lead sub-process
    • The qualified lead to sales accepted lead sub-process
    • The sales accepted lead to sales qualified lead sub-process
    • The sales qualified lead to close sub-process

    Each of these sub-processes have people who proceed to the next sub-process as well as people who do not — put simplistically: people who “fall out of the funnel.” But, you can further decompose — of the people who fell out, where did they fall out and why? And does that mean they are gone forever, or are there processes/subprocesses that can be used to reengage them in the future?

    The key here is that, from a theoretical perspective, if you link together all of the simpler sub-processes, then you’ve got an accurate representation of the more complex master process. The problem is that this is mostly true. There are probably other sub-processes that are unknown — those pesky “corner cases” that the real world insists on throwing at us. And, each sub-process likely experiences various anomalies over time. Add those together, and you’ve got a complex process that verges on the unanalyzable.

    On the other hand, if you focus on a sub-process, you can analyze what is going on, including accounting for the anomalies. “But, isn’t there a risk that you’ll be missing the forest for the trees?” you ask. Absolutely. That’s why it’s important to start with a high-level view of the whole process, with a clear picture of the components that go into it. If you simply pick a “simple sub-process” to focus on, without understanding how and where that fits into the big picture, you run the risk of rearranging deck chairs on the Titanic. On the other hand, if you simply try to “analyze the Titanic,” without some level of decomposition, you’re equally doomed.

    Analysis, Presentation

    Sometimes, the Data DOES Paint a Clear Picture

    I’ll admit right up front that this is the least value-add post on this blog to date. Part of me sincerely hopes that it holds that distinction indefinitely. But, I know me better than that, so no promises.

    We all have them. Those moments where someone says something — in person, in an e-mail, in an instant message — that triggers a completely random, but oddly inspired, response.

    What happened: One of my pet peeves is the cliche, “If you can’t measure it, don’t do it.” It sounds good, but I challenge any company to fully apply this overly simplistic maxim and survive. I’m all for having a bias towards measurement, but I get nervous when people speak in absolutes like this.

    Earlier this week, I fired off an internal e-mail proposing an initiative that was extremely low cost that seemed like a good idea to me. It really wasn’t an initiative where it made sense to try to quantify the benefits, though. I made a comment as such in the e-mail — that, despite it not being practical to measure the results, I still thought it was a good idea. (I was having one of the 15-20 snarky moments I have throughout any given day.) Two of the five people on the distribution list immediately responded with demands for an ROI estimate.

    FLASH!

    10 minutes later, and I’d fashioned the following chart in Excel and responded to the group with my analysis:

    The Bird

    Everyone had a good chuckle.

    Here’s the spreadsheet file itself. It’s as clean as clean can be, so feel free to snag it and put it to your own use. If you put it to use with entertaining results, I’d appreciate a quick comment with the tale. Or, if you make modifications to enhance the end result, I’d love to get a copy.

    Enjoy.

    Analysis, Analytics Strategy

    A Little Bit of Data Can Be a Time-Consuming Thing

    I had an experience over the past week that, in hindsight, I really should have been able to avoid. The situation was basically this: several different people had made comments in passing about how we were probably “overcommunicating” to our database. “Overcommunication” being the tactful way to say “spamming.” In this case, I can actually trace the perception back to at least two different highly anecdotal events, which then spawned comments that led to assumptions, and so on.

    Now, I am all for diligent database management, especially when it comes to how often and with what content we communicate with our contacts. My general sense was that we could be doing better, but we were far from reaching a crisis point (I lived through a situation at another company where we did reach that crisis point, and there were plenty of telltale signs leading up to that). “I can pull some quick data on that to at least get some basic facts circulated,” I innocently thought. And, that’s what I did.

    I knew going in that, while the data was one thing, the definition of “good” vs. “bad” was likely all over the map, so I wasn’t likely to change many people’s opinions as to the situation by simply sharing the data. So, I shot an e-mail out to a group of interested parties and told them I had the data, and I’d be happy to share it, if they shared with me their opinions as to what an acceptable maximum of communications per week and per month would be.

    As I suspected, I got a wide range of responses, and most of the responses had some form of qualifier — well-founded qualifiers regarding the type of communication, actually. So far, so good.

    I then shared the data that I had spent 15 minutes compiling in a way to make for easy analysis, still knowing that there was no clear good/bad definition, and there was no clear hypothesis being tested or action being planned that this analysis would influence. The data did show a few things unequivocably — really just highlighting that the concerns were somewhat well-founded and that discussions should continue amongst the people who already tacitly owned the situation. But, it also spawned requests for additional data that was more curiousity-driven than actionability-driven. Several people asked that the data be pulled with their particular qualifiers addressed. Most of these people were in no position to actually take any action based on the results. And, unfortunately, as reporting and analysis systems can sometimes be — applying the qualifiers would have turned the analysis into a highly manual, multiple man-hours exercise, whereas the high-level, basic pull was a 15-minute task.

    On the one hand, I could ding our data storage system. By golly, Tenet No. 1 of good BI/DW design is to design for flexibility, right? In this case, the system limitations are actually a boon — they give me an out for simply saying, “No,” rather than the much more involved discussion that begins, “Why?”

    It’s a punt, I realize. And not an out I would take if it was throwing anyone in IT under a bus.

    My point is that “interesting” can be a Siren Song that dwarfs the pragmatism of “actionability.”

    Analysis, Presentation, Reporting

    Depth vs. Breadth, Data Presentation vs. Absorption, Frank and Bernanke

    For anyone who knows me or follows this blog, it will be no surprise that I can get a bit…er…animated when it comes to data visualization. Partly, this may be from my background in Art and Design. I got out of that world as quickly as possible, when I realized that I lacked the underlying wiring to really do visual design well.

    As a professional data practitioner, I also see effective data visualization as being a way to manage the paradox of business data: the world of business is increasingly complex, yet the human brain is only able to comprehend a finite level of complexity. And, while I love to bury myself up to my elbows in complex systems and processes, I’m the first person to admit that my eyes glaze over when I’m presented with a detailed balance sheet (sorry, Andy). A picture is worth a thousand words. A chart is worth a thousand data points. That’s how we interpret data most effectively — by aggregating and summarizing it in a picture.

    So, it’s pretty important that the picture be “drawn” effectively. I had a boss for a year or two who flat-out was much closer to Stephen Hawking-ish than he was to Homer Simpson when it came to raw brainpower. He took over the management of a 50-person group, and promptly called the whole group together and presented slide after slide of data that “clearly showed”…something or other. The presentation has become semi-legendary for those of us who witnessed it. The fellow was facing a room of blank-confused-bored-bewildered gazes by the time he hit his third slide. Now, to his credit, he learned from the experience. He still looks at fairly raw data…but he’s careful as to how and where he shares it.

    All that is a lengthy preamble to a Presentation Zen post I read this evening about Depth vs. Breadth of presentations. It’s a simple concept (meaning I can understand it), with some pretty good, rich examples to back it up. The fundamental point is that none of us spend very much time thinking about what to cut from our presentations. I would extend that to say we don’t spend very much time thinking about what data not to share or show. It’s easy to see this as a case for “make the data support what you want it to,” which it is not. At all! Really, it’s more about focussing on showing the data — and only the data — that directly relates to the objectives you are measuring or the hypotheses that you are testing.

    Then, focus on presenting that data in a way that makes it clear as to what story it is telling. You do the hard work of interpreting the data. Then, highlight what is coming out of that intepretation. If there is ambiguity, highlight that, too. If there is a clear story, and your audience gets it, and you then introduce an anomaly, you’re much more likely to have a fruitful, engaging discussion about it. You will learn, and your audience will retain!

    In the end, this is a riff on a bit of a tangent, I realize. Robert Frank presents some fairly alarming evidence of college professors aiming for broad and deep…and not gaining any better retention than the slide-happy, chart-crazy PowerPoint users provide in the business setting. He goes on to talk about how, in his teaching, he makes a point, repeats it, comes at it from a different angle, makes the students think about it, and then repeats it again. He goes for deep. His students, I’m sure, leave his introductory economics class with a thoroughly embedded (and accurate) understanding of “opportunity cost” (having seen the term mis-applied more than once in my day…and still having to struggle to get to the correct answer…and barely…and barely in time…in his presentation…I applaud that!).

    I’m not arguing for simplicity for simplicity’s sake. I’m arguing for going deep, understanding the complexity, and then distilling it down to a narrative, cleanly presented, that leaves your audience with takeaways that are accurate and absorbed.

    And…on that note, have any of you read The Economic Naturalist? It sounds like it would be right up my alley. It’s just a bonus that, if I ever actually attended something that could be labeled a “cocktail party,” I could talk about how I’d “read some of Bernanke’s work!”

    Analysis, Analytics Strategy, Reporting, Social Media

    Bounce Rate is not Revenue

    Avinash Kaushik just published a post titled History Is Overrated (Atleast For Us, Atleast ForNow). The point of that post is that, in the world of web analytics, it can be tempting to try to keep years of historical data…usually “for trending purposes.” Unfortunately, this can get costly, as even a moderately trafficked site can generate a lot of web traffic data. And, even with a cost-per-MB for storage of a fraction of a penny, the infrastructure to retain this data in an accessible format can get expensive. Avinash makes a number of good points as to why this really isn’t necessary. I’m not going to reiterate those here.

    The post sparked a related thought in my head, which is the title of this post: bounce rate is not revenue. Obviously, bounce rate (the % of traffic to your site that exits the site before viewing a second page) is not revenue. And, bounce rate doesn’t necessarily correlate to revenue. It might correlate in a parallel universe where there is a natural law that no dependent variable can have more than 2 independent variables. But, here on planet Earth, there are simply too many moving parts between the bounce rate and revenue for this to actually happen.

    But.

    That’s not really my point.

    What jumped out at me from Avinash’s post, as well as some of the follow-up comments, was that, at the end of the day, most companies measure their success on some form of revenue and profitability. Realizing that there is incredible complexity in calculating both of these when it comes to GAAP and financial accounting, what these two measures are trying to get at, and what they mean, are fairly clear intuitively. And, it’s safe to say that these are going to be key measures for most companies 10, 20, or 50 years from now, just as they were key measures for most companies 50 years ago.

    Sales organizations are typically driven by revenue — broken down as sales quotas and results. Manufacturing departments are more focussed on profitability-related measures: COGS, inventory turns, first pass yields, etc.  Over the past 5-10 years, there has been a push to take measurement / data-driven decision-making into Marketing. And, understandably, Marketing departments have balked. Partly, this is a fear of “accountability” (although Marketing ROI is not the same as accountability, it certainly gets treated that way) Partly, this is a fear of figuring out something that can be very, very, very difficult.

    But, many companies are giving this a go. Cost Per Lead (CPL) is a typical “profitability” measure. Lead Conversion is a typical “revenue” measure. That is all well and good, but the internet is adding complexity at a rapid pace. Pockets of the organization are embracing and driving success with new web technologies, as well as new ways to analyze and improve content and processes through web analytics. No one was talking about “bounce rate” 5 years ago and, I’d be shocked if anyone is talking about bounce rate 5 years from now.

    Social media, new media, Web 2.0 — call it what you like. It’s changing. It’s changing fast. Marketing departments are scrambling to keep up. In the end, customers are going to win…and Marketing is going to be a lot more fun. But we’ve got a lonnnnnnnnng period of rapidly changing definitions of “the right metrics to look at” for Marketing.

    While it is easy to get into a mode of too constantly reevaluating what your Marketing KPIs are, it is equally foolish to think that this is a one-time exercise that will not need to revisited for several years.

    Oh, what exciting times we live in!

    Analysis

    Book Review Part 2 of 2: Super Crunchers

    It’s amazing what an airplane flight can do when it comes to finishing up books that have been lingering. Especially when you pack the “fun” book in your checked baggage so you can’t cheat. Sucks when you arrive at the hotel to find that your “fun” book is, apparently, still sitting in the bedroom at home! Argh!

    Part 2 will be a bit less harsh, as Ayres does eventually touch on some of my bigger issues. Sort of. I’m still not satisfied with his treatment, though.

    Ayres does, at least once, throw in a pretty critical caveat: “As long as you have a large enough dataset, almost any decision can be crunched.” That’s a HUGE caveat, even in our increasingly data-capturing world. As a matter of fact, I like to talk about the data explosion. But, when I discuss it, it’s as a warning that you can’t just see “getting the raw data” as the biggest challenge in making data-driven decisions. My claim is that the bigger challenge is developing discipline about how you approach that data. While Ayres does periodically speak to the fact that there is real skill and creativity required to develop the hypotheses that you want to test, he does not put nearly enough emphasis on this point. And, not only that there needs to be diligence in developing the hypothesis, but also considerable rigor in determining how that hypothesis will be tested.

    Ayres walks through a pretty fascinating example of super crunching gone awry in the case of John Lott, who did some super crunching that demonstrated that concealed handgun laws brought down the crime rate. According to Ayres, Lott made an error in the data prep for his analysis that, when corrected, did not show this at all. Ayres uses the example more to preach that even data-oriented people can still get caught up emotionally and refuse to face hard facts. While that is undoubtedly true, Ayres also misses the opportunity to speak in any real depth as to the amount of data prep work that needs to be done to normalize and cleanse data before actually running a regression. He does mention this…but it is very brief.

    More good stuff: Ayres devotes a good chunk of a chapter to explaining (and illustrating) just how bad humans are at gauging the quality of their own intuition. Many of the points he makes here echo Daniel Gilbert’s points in Stumbling on Happiness. Points very well taken. But — and I’m sure Ayres would say that my next statements prove his point and that I’m just one more person in denial — things get stretched pretty far at times here. For instance, Ayres claims that the safety procedures that flight attendants follow as a hard script almost all of the time make them more effective than a less structured approach. Seeing as I was sitting on a flight, where I had easily tuned out the flight attendant’s script, I had a “Gimme a break!” moment. Everyone who travels at all has occasionally stumbled across a Southwest (and it even happened to me on United once) flight where the flight attendant has a little more fun. Everyone listens! And, they are clearly sticking to the content of the script, if not the specific language.

    Oops. Language. I’ll nitpick just a little bit. Ayres uses “digitalize” a lot. Maybe he can’t help it — he’s a lawyer AND an academic, so why go with the shorter, common synomym — digitize — when a longer word will suffice. He also defines CDI as “consumer data integration” in the context of Acxiom Corporation’s services. While Google indicates this is one possible explanation of the acronym, even Acxiom seems to use the breakdown that I’m much more familiar with: customer data integration. Splitting hairs a little bit, I realize. But, it’s an indication that Ayres really isn’t all that familiar with CDI, which is a descriptor of a host of technologies that are trying to actually solve all of the complexity of normalizing data from disparate data source that Ayres barely acknowledges. Ayres also uses “commodification” a lot, and I was prepared to zing him there, too…but Google shows this as a more common word than it’s synomym, commodization. So, I learned something there!

    Two more specific beef examples and I’ll wrap up.

    Ayres devotes a good chunk of writing to various plans that use super crunching to predict the likelihood of recidivism for inmates who are being paroled. The statement that really got me was that, according to the data, an inmate who received a score of four or higher when his history was plugged into a certain model would have a 55% chance of committing another sex offense within 10 years. Ayres then jumps into example of one such inmate who received a score of four, was paroled anyway, and promptly disappeared. The 55% is what gets me, though. Ayres ignores that this is far from an overwhelming indication. It may very well be my liberal bias that I don’t think that it’s a slam dunk to keep everyone locked up knowing that 45% of these people would NOT commit another sex offense within ten years. Ayres completely glosses over this question, which seems to be an ethical one worth addressing. I actually made a note — Minority Report — when I read this. I was thinking of the Philip K. Dick-inspired Steven Spielberg movie that starred Tom Cruise. In the movie, citizens are arrested and based on “foreknowledge” — a vision by one of three “pre-cogs” (it’s a very small leap to turn these specially-endowed humans into a super crunching computer) that the person is going to commit a murder in the future. The movie is chilling in a 1984 kind of way. And, ultimately, condemns persecuting someone for something they have not yet done. Ayres sounds just a little too much like the films main antagonist — Director Lamar Burgess (played by Max von Sydow) — for my comfort. Interestingly, later in the book, Ayres refers to the same film…but not at all for the same reasons. Rather, he references the personalized ads that Tom Cruise is bombarded with at all times while walking down a street in the movie (which really isn’t that far of a stretch from today, given the proliferation of RFID technology).

    Second specific beef has to do with NASA history. Ayres points to Gus Grissom as one of the astronauts who balked at the idea that the Mercury splashdown capsules would be designed so that that would not be able to be opened from the inside. Ayres then points out that, because the astronauts wan out, Gus Grissom was able to panic upon splashdown and blow the hatch open prematurely. Which he did. And almost drowned. First off, Ayres gives no acknowledgement to the fact that there are a lot of reasons to believe (as NASA does) that the hatch blew open due to an equipment malfunction — not because Grissom did anything inappropriate. Second, Gus Grissom was killed in a fire inside a training capsule for the Apollo 1 mission. I honestly don’t know enough of my history here to know if the training capsule could be opened from the inside, or, if not, if it would have made a difference. My guess is that the fire was so fast that it wouldn’t have mattered. But, on both accounts, Grissom seems like a spectacularly lousy example of Ayres’s point.

    So, I’ll wrap up. The book has left a dirty taste in my mouth. Ayres delivered a book that he hopes will sell with the gee-whiz factor. My guess is that Bantam Books smelled “Jump on the Freakonomics Bandwagon” money, and, as bandwagon books, movies, and TV shows are prone to do, it underdelivers. The book dramatically downplays the challenges involved in super crunching — not just challenges in the “it’s hard” sense, but challenges in the “large, clean, relevant data sets” are not the same as “large amounts of raw data.” He actually includes examples of super crunching where, when the results came out, they were applied on a limited scale because the basic approach to the analyses were called into question. So, how about a chapter on “design of experiments?” He does have good points, and his assertion that companies could benefit from having a designated “devil’s advocate” to question — hard — any analysis is a great idea. There is definitely a shock of wheat here and there in the book. Unfortunately, Ayres spends most of his writing about the chaff.

    Analysis

    Book Review Part 1 (of ?): Super Crunchers

    I promised in an earlier post that I would read Ian Ayres new book, Super Crunchers, and it’s taken me longer to get to it than I’d planned. And, while it’s a fairly easy read, it’s not getting a nightly hit on my nightstand because it tends to get me worked up.

    I’m fully aware that I got the book with preconceived notions that it might set me off. I’m doing my darnedest, though, to be objective in my observations about it. And I haven’t finished it, yet, but I’m going to lose track of the different points I want to make if I don’t start recording them.

    My fundamental beef with the book is that it picks and pulls anecdotes that paint a picture of how easy it is and how common it is becoming to do super crunching. So far, it has done very little to articulate all the ways that super crunching does not or cannot work…either “yet” or “ever.” Much more on that to come. To me, it reads mostly like an effort to jump on the Freakonomics bandwagon. Freakonomics was original and interesting. Super Crunchers is a poor, poor follow-on.

    Ayres does reference some great books in his opening pages. James Surowiecki’s The Wisdom of Crowds is a must-read. And I thoroughly enjoyed Moneyball (although I’m an admitted college baseball junkie). I’m ambivalent about Freakonomics, but generally feel positive about it. It may be that it just seems to play too much to the mass market, and I like to think I’m higher brow than that. 😉

    Ayres views the world with a heavy, heavy academic bias. As such, he thinks data is cleaner than it is, people are easier to change than they are, and time machines exist. I’ll tackle the last one first. Ayres presents super crunching as an easy, two-party formula: 1) the ability to run regressions (just plug the date into the software and press “Go” is what is sounds like), and 2) A/B testing is a no-brainer (and just occurs in the world naturally if you look for it).

    Here’s where the time machine comes in. On page 69, he expounds on how politicians are using super crunching. And then he extends how they could use it even more as follows:

    “Want to know how negative radio ads will influence turnout of both your and your opponent’s voters? Run the ads at random in some cities and not in others.”

    Stop and think about that one for a minute. Clearly, we’re talking about a statewide or federal election. You’re on a 2-, 4-, or 6-year election cycle. You run negative ads in some cities and not others. The election happens. You lose. You assess voter turnout. Great. Now what? The election is OVER!

    Earlier in the same paragraph, Ayers started at a more reasonable level by discussing political scientists. Extend that to political consultants, and you’ve got some validity in super crunching. The RNC and DNC could probably learn a lot by exerting some influence and doing some randomized testing on this sort of thing with all of their candidates and then applying that knowledge in the next election (they will get pushback from their candidates, who are much more concerned about winning their current election than they are about providing data for the common good). But Ayers very explicitly dives down to a prescription for one candidate’s strategy in one campaign against one opponent. And that just doesn’t work.

    Why does this irk me so much? Because the casual reader might not realize what a joke that statement is. And, thus, might walk away from the book with a vague notion that super crunching can be applied to every situation with ease!

    A more general criticism is that Ayers picks mammoth, established companies that have massive data sets to work from for most of his examples. When I worked at a $600 million company, which was by no means mammoth, but still generated a lot of data, we frequently found that we didn’t have enough data to answer the most burning questions with any statistical validity. Why? Because: 1) the real world is very, very noisy, and it’s really hard to control for all of the likely independent variables (in Ayers election example, there was no mention of the candidates’ issues and how those may resonate differently in different cities regardless of what ads were run), 2) businesses operate in the fourth dimension (time) — if they have a six-month sales cycle and they want to run a test of how something early in the cycle influences events late in the sales cycle, they can be in a pickle when it comes to actually capturing enough data to run a legitimate regression, and 3) the data is NEVER as clean as folk in academia or who haven’t really sat down and understood the processes think it is.

    I’ll jump ahead to Ayres’s discussion of evidence-based medicine. He does let in a sliver of “it ain’t always a slam dunk” when he discusses (pp. 90 and 91) a study that linked excessive caffeine consumption to increased risk of heart disease…without controlling for the fact that there was a correlation between caffeine consumption and the likeliness that the subject smoked. That reminded me of a huge commercial super crunching gaff that used A/B testing and regression heavily…but that bit one of the top consumer brands of all time in the butt: Coca-Cola.

    Coke doesn’t show up in Ayres’s index, so I’m assuming I’m not going to stumble across the New Coke fiasco. Two different books have given reasons for that flop: in Stumbling on Happiness (I’m almost positive that was the book, but I read Blink at the same time and there was a lot of overlap in the concepts), Daniel Gilbert made the point that the test was flawed. Coke assumed that a single sip in a blind taste test would be a valid indicator. It turns out that your first sip of a soft drink gives you one experience, while the last 80% of the drink gives you another experience — after your taste buds / brain have adjusted to the “new” experience. In Waiting for Your Cat to Bark, the Eisenberg brothers attributed the failure more to the fact that the Coke flavor was sufficiently ingrained in the soul of Coke drinkers that, by golly, they didn’t want that flavor to change! Either way (and it was likely a combination of both) this was a super crunching disaster!

    Back to evidence-based medicine. There is some absolutely fascinating stuff in this chapter. And there is no refuting that the medical profession is wildly behind the times when it comes to adopting information technology. And Ayres outlines some pretty interesting initiatives on that front that I genuinely hope succeed. But, again, there were some alarming statements in the chapter that illustrated Ayres’s academic naivete. He plugs in a quote from Dr. Joseph Britto around a feature under development in his (very cool) Isabel system that will enable doctors to enter their notes in a patient’s medical record through a “structured flexible set of input fields to capture essential data.” Britto is quoted as saying, “If you have structured fields, you then force a physician to go through them and therefore the data that you get are much richer that had you left him on his own to write case notes, where they tend to be brief.”

    Whoa!!!!

    TWO big beefs with this line of thought.

    Beef No. 1: I HAVE gone through many situations where someone who wants to use the data (and, therefore, wants data that is comprehensive and complete) makes a massive logical leap that the answer is simply to make the system “force” the people who are inputting the data to provide more data and to provide it in a way that tees it up for valid, straightforward analysis. Life. Just. Don’t. Work. That. WAY! The only time I have seen this work is when immediate, direct value to the person inputting the data is realized by the change. “Do it this way, and you will feed a much larger aggregate data set that you will ultimately benefit from” simply does not work. “Required fields” don’t work. If you introduce a process that takes as little as 15 seconds longer than the old process, those required fields quickly become populated with a single character or the first value on the list (for structured inputs). Taking a top-down approach and auditing and enforcing better data input requires a Herculean effort, but it’s doable. What concerns me is that neither Britto nor Ayres acknowledge that it’s not as simple as a system deployment. The data analyst’s approach is going to be towards capturing everything about the patient, regardless of whether it seems relevant. That way, data mining may find things that no one thought were relevant but turned out to be. Unfortunately, 95% of that “comprehensive” data is not going to be relevant, and there’s no way to identify where that 95% is. Britto and Ayres want to put the burden of capturing all of that data on the doctors, which is simply not practical.

    My current company is actually experimenting with a tool on the CRM side that acknowledges this challenge. ShadeTree Technology is a plug-in for Salesforce.com that, at the end of the day, is trying to drive more comprehensive, more consistent, cleaner CRM data. BUT, they’re taking an approach that the only way that will work is if the tool provides immediate, direct benefit to the people who are using it. We are early on in our implementation of the tool, but the fact that ShadeTree understands this concept is a key reason that we are implementing it.

    Beef no. 2: I don’t have a medical background, but I have three kids. Our oldest wound up in the hospital for a week when he was six years old. He is a very bright and articulate kid. What landed him in the hospital was a pain that was somewhere in his hip/upper thigh area, deep under the skin. X-rays, bone scans, and CAT scans weren’t able to pinpoint the location of the issue. It took an MRI to find a small abscess. And it took a battery of tests to figure out that he’d gotten salmonella in his bloodstream that, ultimately caused the abscess. How is this relevant? Because we had a helluva time trying to understand exactly what his symptoms were. “Upper thigh” is different from “hip.” Even with a willing, articulate kid, identifying “pain in the bone” vs. “pain in the muscle” was damn near impossible. Humans are not numbers and data. “Pain” is subjective. “Small brown spots on the skin” has numerous subjective elements (another example from this chapter). Ayres wildly glosses over the challenges here, and that bothers me.

    To be clear, I’m not saying that none of these initiatives will work. And, for all I know, Britto actually is much more sensitive to the issues than Ayres makes him sound. What sticks in my craw is that Ayres continually presents the world as a place where clean, comprehensive, easily analyzed data already exists or where it’s going to be relatively easy to collect it. That is painting an unrealistic picture. It may sell books, but it’s not all that actionable and, I claim, will actually hinder serious readers’ ability to actually improve their own usage of data.

    More to come in Part 2…

    Analysis, Reporting

    More Data Is Better

    I had two discussions yesterday that centered around a similar topic. Both were with people who felt that more data is, by definition, better in the CRM space.

    One of the discussions centered on deliverables for a service offering. It’s somewhat a best practice in the services industry to manufacture some sort of hard deliverable so that the customer goes away with something that is tangible, even if the real value they received was a service that did not result in any tangible goods. That makes sense, and it’s why consultants almost always deliver some form of post-engagement report to their clients.

    But, this can get tricky if the “real” deliverable is data of some sort. It is tempting to make the tangible deliverable simply a binder of the all of that data sliced and graphed in enough ways to make a reasonably hefty book. The nice thing about data is that it doesn’t take very much to make a really complex-looking chart, and one small data set can be presented in countless ways.

    The problem is that this is exactly the worst way to go about actually getting value from data. Spewing out charts quasi-randomly is a terrible way to get from data to information, and from information to action.

    I agree that, for many customers of the service, this approach might work in the short term. In their minds, they believe “more data is better,” and it’s hard to argue that a dead tree, thinly sliced, bound, and covered with pretty pictures isn’t “more data.” Some of these customers may actually flip through the charts and ponder each one in succession. More, I would guess, will look at the first couple of pages and then set the whole book aside with the best intentions to sift through it later. In both cases, if asked if the data was useful, they are likely to respond, “Yes. Very.” But let’s not probe deeper and ask, “What actionable insights did you get from the report?” More often than not, this question will result in an awkward silence.

    Think about it, though. It’s human nature to not want to admit, when you were given what you thought you wanted, and maybe even what you expected, asked for, and eagerly awaited, that it’s really not all that useful.

    To build a long-term, lasting relationship with a services customer, isn’t it better to focus on what really will give them long-term value? Spend the time up front helping them articulate their objectives and goals for the service you are delivering. Establish success metrics up front that are meaningful. This may be harder than you think, but, it’s just like project management — the up front work will pay huge dividends. And, by working with the customer to make sure you have clearly articulated their objectives, and then stayed focused on their objectives throughout the process, and then delivered a report that reports on how well those objectives were met, there is a much stronger, longer term relationship in the making.

    As a matter of fact, reporting that some of the objectives were not met, along with analysis and speculation as to why, can be a powerful customer relationship tool. It actually builds trust and shows a high level of integrity: “We did not meet all of your objectives — we met some and exceeded others, but we also missed in one or two areas. We’re not happy about that, and we’ve tried to understand what happened. We’d like to work through that with you so we both can learn and improve going forward.”

    More data is not always better. The right data is best.

    As for the second person, I’ll save that for another blog entry.