Home On TV & Video To Make TV Attribution Better, We Need To All Get On The Same Page

To Make TV Attribution Better, We Need To All Get On The Same Page

SHARE:

On TV And Video” is a column exploring opportunities and challenges in advanced TV and video.

 Today’s column is written by Jane Clarke, managing director and CEO at the Coalition for Innovative Media Measurement (CIMM).

Attribution, or the measurement and assignation of an outcome following an ad exposure, has revolutionized media planning and buying with the promise of being able to not only assess the impact of spend but optimize it moving forward.

While there are many different approaches to attribution, with each vendor applying a unique solution, the underlying data definitions that serve as the basis for their analysis should be the same.

And currently they are not – a shortcoming that is inhibiting the overall growth and impact of attribution as a measurement discipline. It’s not that the attribution analyses aren’t valuable, it’s that buyers can’t feel confident that there isn’t hidden error confounding their results.

This reality was made clear in an analysis of TV attribution providers just completed for CIMM and the 4A’s Media Measurement Task Force by Sequent Partners and Janus Strategy and Insights.

The research found that ad occurrence and exposure data are highly inconsistent across providers. The reason was not only the differences between the actual data, but primarily the methodology being used to convert that data into final ad occurrence files and exposure data, including weighting, editing and other data processing rules.

The situation is analogous to everyone working from their own “set of facts,” defining reality in different ways. That is why, if you had a variety of providers examine the same campaign, you would in all probability end up with very different conclusions about ROAS.

Data on ad occurrence and frequency varies greatly from attribution provider to attribution provider because currently there are no standards defining the methodology that should govern processing of this data. Television attribution results will become more transparent, consistent and reliable when providers adopt more stringent media measurement standards.

The Media Rating Council has an effort underway to develop outcome-based measurement standards. Ideally, when it comes to how data processing is defined, these standards should address:

  • Weighting. Providers need to implement a robust panel weighting scheme that addresses variables common to TV viewing such as DMA, HH size/Presence of Children and Income/Education/Occupation.
  • Unification. A standard process for unifying the database for ROI measurement is needed, providing a common base of viewers where there is both an opportunity for exposure and an opportunity for a response, such as website visitation, retail traffic or purchase.
  • Exposure Qualification. There needs to be agreement as to what standard exposure criteria is. Should it be one second, three seconds, five seconds, 10 seconds or a one-minute schedule at 300 GRPs?
  • Occurrences. There has to be rigorous quality control in the re-creation of as-run schedules and in the evaluation of Reach reporting from exposure data across schedules.

For attribution to fulfill its promise and for buyers and sellers alike to fully embrace it for ROAS analysis, we need a common approach to defining the starting point of occurrence and exposure data.

In an industry where proprietary exclusivity is all, standards are sometimes bristled at.

Certainly, attribution providers can and should go to market with their own unique approaches. But all we need to start on the same page. And to accomplish that, we need standards to govern how occurrence and exposure is being defined and processed from data, to ensure accurate and consistent assessment of ROAS.

 

Must Read

AdExchanger Senior Editors Anthony Vargas and Alyssa Boyle.

POSSIBLE 2026: AdExchanger's Hot Takes

AdExchanger Senior Editors Alyssa Boyle and Anthony Vargas share their takeaways from three days chatting about agentic AI at POSSIBLE.

Reddit Reports A 75% Boost In Q1 Ad Revenue As It Reaches For 100 Million Daily US Users

Generative AI search has pushed traffic off a cliff across most of the internet, but not on social platforms. Reddit included.

POSSIBLE 2026: Can AI Help Agencies Finally Break Down Those Silos?

Domenic Venuto, indie agency Horizon Media’s chief product and data officer, sat down with AdExchanger during POSSIBLE at the Fontainebleau in Miami to unpack the role of AI in today’s media and advertising landscape.

Privacy! Commerce! Connected TV! Read all about it. Subscribe to AdExchanger Newsletters

Google Touts Its AI Ad Tech Adoption And New AI Max Features

Google announced new features and ad types for AI Max, its AI-based bidding product for search and shopping or sponsored product ads. The company also touted “hundreds of thousands” of advertisers using AI Max.

Hand pressing blue AI button on keyboard. Digital collage of artificial intelligence interface.

Meta’s Ad Machine Is Purring, So Why Did Its Stock Drop?

Meta’s Q1 call sounded like an AI and hardware pitch, but under the hood it was still about one thing: investing in AI to squeeze more money out of its ads business.

Alphabet Exceeds $100 Billion In Q1 And Its Profits Almost Doubled

Alphabet earned $109.9 billion in Q1 this year, up from $90.2 billion a year ago. And that’s not even the truly gobsmacking number.