Home On TV & Video To Make TV Attribution Better, We Need To All Get On The Same Page

To Make TV Attribution Better, We Need To All Get On The Same Page

SHARE:

On TV And Video” is a column exploring opportunities and challenges in advanced TV and video.

 Today’s column is written by Jane Clarke, managing director and CEO at the Coalition for Innovative Media Measurement (CIMM).

Attribution, or the measurement and assignation of an outcome following an ad exposure, has revolutionized media planning and buying with the promise of being able to not only assess the impact of spend but optimize it moving forward.

While there are many different approaches to attribution, with each vendor applying a unique solution, the underlying data definitions that serve as the basis for their analysis should be the same.

And currently they are not – a shortcoming that is inhibiting the overall growth and impact of attribution as a measurement discipline. It’s not that the attribution analyses aren’t valuable, it’s that buyers can’t feel confident that there isn’t hidden error confounding their results.

This reality was made clear in an analysis of TV attribution providers just completed for CIMM and the 4A’s Media Measurement Task Force by Sequent Partners and Janus Strategy and Insights.

The research found that ad occurrence and exposure data are highly inconsistent across providers. The reason was not only the differences between the actual data, but primarily the methodology being used to convert that data into final ad occurrence files and exposure data, including weighting, editing and other data processing rules.

The situation is analogous to everyone working from their own “set of facts,” defining reality in different ways. That is why, if you had a variety of providers examine the same campaign, you would in all probability end up with very different conclusions about ROAS.

Data on ad occurrence and frequency varies greatly from attribution provider to attribution provider because currently there are no standards defining the methodology that should govern processing of this data. Television attribution results will become more transparent, consistent and reliable when providers adopt more stringent media measurement standards.

The Media Rating Council has an effort underway to develop outcome-based measurement standards. Ideally, when it comes to how data processing is defined, these standards should address:

  • Weighting. Providers need to implement a robust panel weighting scheme that addresses variables common to TV viewing such as DMA, HH size/Presence of Children and Income/Education/Occupation.
  • Unification. A standard process for unifying the database for ROI measurement is needed, providing a common base of viewers where there is both an opportunity for exposure and an opportunity for a response, such as website visitation, retail traffic or purchase.
  • Exposure Qualification. There needs to be agreement as to what standard exposure criteria is. Should it be one second, three seconds, five seconds, 10 seconds or a one-minute schedule at 300 GRPs?
  • Occurrences. There has to be rigorous quality control in the re-creation of as-run schedules and in the evaluation of Reach reporting from exposure data across schedules.

For attribution to fulfill its promise and for buyers and sellers alike to fully embrace it for ROAS analysis, we need a common approach to defining the starting point of occurrence and exposure data.

Subscribe

AdExchanger Daily

Get our editors’ roundup delivered to your inbox every weekday.

In an industry where proprietary exclusivity is all, standards are sometimes bristled at.

Certainly, attribution providers can and should go to market with their own unique approaches. But all we need to start on the same page. And to accomplish that, we need standards to govern how occurrence and exposure is being defined and processed from data, to ensure accurate and consistent assessment of ROAS.

 

Must Read

John Gentry, CEO, OpenX

‘I Am A Lucky And Thankful Man’: Remembering OpenX CEO John ‘JG’ Gentry

To those who knew him, John “JG” Gentry wasn’t just a CEO. He was a colleague who showed up with genuine care and curiosity.

Prebid Takes Over AdCP’s Code For Creating Sell-Side AI Agents

The group that turned header bidding software into an open standard is bringing the same approach to publisher-side AI agents.

Meta logo seen on smartphone and AI letters on the background. Concept for Meta Facebook Artificial Intelligence. Stafford, UK, May 2, 2023

Meta Bets That Its Ad Machine Can Fund Its AI Dreams

Meta is channeling its booming ad revenue into a $135 billion AI drive to power its “personal superintelligence” future.

Privacy! Commerce! Connected TV! Read all about it. Subscribe to AdExchanger Newsletters
Comic: Header Bidding Rapper (Wrapper!)

Microsoft To Stop Caching Prebid Video Files, Leaving Publishers With A Major Ad Serving Problem

Most publishers have no idea that a major part of their video ad delivery will stop working on April 30, shortly after Microsoft shuts down the Xandr DSP.

AdExchanger's Big Story podcast with journalistic insights on advertising, marketing and ad tech

Guess Its AdsGPT Now?

Ads were going to be a “last resort” for ChatGPT, OpenAI CEO Sam Altman promised two years ago. Now, they’re finally here. Omnicom Digital CEO Jonathan Nelson joins the AdExchanger editorial team to talk through what comes next.

Comic: Marketer Resolutions

Hershey’s Undergoes A Brand Update As It Rethinks Paid, Earned And Owned Media

This Wednesday marks the beginning of Hershey’s first major brand marketing campaign since 2018