Home Data-Driven Thinking Google’s Data-Driven Attribution Model Isn’t Perfect, But It Is Progress

Google’s Data-Driven Attribution Model Isn’t Perfect, But It Is Progress

SHARE:
Dmitri Kazanski, Head Of Product, North America, MGID

Data-Driven Thinking” is written by members of the media community and contains fresh ideas on the digital revolution in media.

Today’s column is written by Dmitri Kazanski, head of product for North America at MGID.

Last click is the most commonly used attribution. Why? Because it’s very simple – but it’s also clearly flawed.

A user’s path in the funnel is affected by multiple touch points, including the ad impressions that are seen or heard and not clicked. Assigning all the credit to the last click is as good as assigning all the credit for one’s fitness level to one’s last workout.

I bet you can’t remember the last time you clicked a Geico ad. But if you live in the US, you can easily fill in the blanks in the following sentence: “A 15-minute call could save you 15% or more on ___  ____________.”

The next time you need car insurance, more likely than not, you’ll type “Geico” into your browser, after which you might click the first link you see: an AdWords link. The insurance quote you’re given will be counted as a lead by Geico, but here’s an important question: Does the ad you clicked deserve full credit for the lead?

Google AdWords, which supports six attribution models, recently changed its default from last click to a complex model Google calls “data-driven attribution.” The name is rather unfortunate. All attribution models, including last click, are driven by data. Google might as well call it an “electricity-powered” attribution model.

In principle, the idea behind data-driven attribution sounds great. The example given by Google appears to indicate a model that correlates conversions to certain events, such as clicks on particular ads. The credit is then spread across the events that correlate the most with the conversions. 

Unfortunately, not much is known about how the model is built or how exactly it works. It’s a black box that might be powered by a regression or a neural net, among other things – who knows.

As someone who works with predictive modeling, I wonder if Google’s “data-driven attribution” model accounts for context and interactions.

In the example provided by Google, it’s possible that the ad for “Bike tour New York” might have a stronger correlation with conversions than “Bike tour Brooklyn waterfront” across all traffic. However, when the traffic comes from within the New York area, the more specific ads, such as “Bike tour Brooklyn waterfront,” might perform better. 

Secondly, the new default attribution model does not appear to explain how ad views that do not result in clicks count toward the attribution, if at all.

Google mentions “holdback experiments” as a way to calibrate the model and arrive at incrementality, which is encouraging. In my view, strictly controlled holdback experiments are the gold standard of attribution and incrementality measurement. This works as follows:

  • A certain percentage, say 10%, of the target audience is held back as a control. The users in the control group are not exposed to the ads.
  • After the campaign is complete, the advertiser shares its list of buyers with the provider.
  • Some of the participants in the control group will end up converting anyway. The difference in the percentage (and monetary value) of the conversions between the control group and the exposed group represents the true incrementality of the campaign.

In practice, this attribution study will be challenging to implement. Usually, it involves resolving the identities of both converted users and exposed users. Doing so presents obvious privacy-related challenges. Clearly, Google cannot do this type of study for every campaign, but at least such studies appear to be used for calibration.

The new default attribution solution should answer the question as to which of Google’s campaign components contributed to the most conversions. It won’t, however, answer the question of incrementality or the question of which components of advertisers’ overall spend produced the most conversions.

Still, it is a step in the right direction.

Follow MGID (@MGID) and AdExchanger (@adexchanger) on Twitter.

Must Read

Meta’s NewFronts Message To Advertisers: Embrace The Noise

Can a good sales presentation offset the impact of a very bad news week? That’s a question for Meta, which collected two guilty verdicts in court this week for failing to protect children and creating additive products.

AI Helps Manscaped Trim Social Chatter Down To The Bare Essentials

Meet Clamor, a new social listening product that pulls cultural insights from online conversations in real time. Clamor helped Manscaped freshen up its marketing, including for this year’s Super Bowl.

A man talking to a robot

How Red Roof Is Bringing In More Customers With Zeta’s Voice-Activated AI Agent

Hotel chain Red Roof is using Zeta’s new voice-activated AI agent to guide its campaign creation, deployment timing and audience development.

Privacy! Commerce! Connected TV! Read all about it. Subscribe to AdExchanger Newsletters
Jean-Paul Schmetz, Chief of Ads, Brave

Why Ad-Blocking Browser Brave Introduced Its Own Ads

Brave’s chief of ads Jean-Paul Schmetz on competition in the search and browser markets, the fallout from the Google Search antitrust ruling and whether AI search will help smaller upstarts compete with Big Tech.

Vizio Helps Walmart Cut A Bigger Slice Of The CTV Ad Pie

Walmart and Vizio announced at NewFronts that unified account logins are coming to smart TVs using Vizio’s operating system.

Comic: CTV Tracking

Carl’s Jr. And Hardee’s Marketing Goes Regional With Amazon Ads’ Streaming Media

The age-old question for streaming TV advertisers is, how to target the viewers they want while reaching the scale their businesses need. The quick-serve restaurant operator CKE, which owns Carl’s Jr. and Hardee’s, sought an answer in a case study with Attain and Amazon Ads.