Understand Lift measurement statuses and metrics in Google Ads

In this article, we explain the meaning of statuses within your study and Brand Lift reporting. You can click the links below to go directly to your area of interest.

Statuses

“X% Lift”

An X% lift indicates that we've detected high enough lift based on the number of responses we received to generate a report. For example, a 5% increase in the Absolute brand lift column indicates that your ads influenced your audience's positive feelings towards your brand or product by +5%. Learn more about the different Brand Lift metrics

“Not enough data”

“Not enough data” means that based on the date range you’ve selected in your account, the number of Brand Lift survey responses received in that date range is below the minimum threshold required to surface results.

Fix “not enough data”

There could be multiple reasons for not getting enough data for your study or an individual slice. To fix it, make sure that:

  • You spend your budget in full.
  • The actual spend in your campaigns meets the minimums, and not just the budget.

If your campaigns are spending enough but are still not getting Brand Lift results, check for the following:

Is your CPV bid too low?

Low traffic can indicate that you're getting outbid. Raise your bid to win more impressions and generate traffic. However, keep in mind that, if you raise your bid, you’ll spend your budget faster (assuming those impressions lead to views). When you use your budget faster, you’ll have less unique viewers and less potential for more viewers to fill out a survey.

Note: If you have an unlimited budget, “bid too high” isn’t an issue.

Recommendation: If your traffic is low despite broad targeting, consider raising your bid. If raising your bid means you are hitting your budget cap, consider raising your budget to accommodate the higher bid.

Is your campaign configuration negatively affecting the survey control group?

If a Brand Lift study uses campaigns that target audiences that viewed the ad video before, it currently does not build a control group.

For example, let’s say you create a study with Video A. Next, you create a second study in which you target a YouTube list of “Viewers who watched Video A as an ad”. With this setup, you won’t be able to build a control group. You may have progress, but it will only be on the exposed side, so you can't expect results to post.

Another example would be Brand lift study that uses campaigns that targets audiences who saw the first ad of a video ad sequence (VAS) campaigns. With VAS campaign subtypes you can create sequences of ads you want users to view in a certain order (i.e. ‘Show users Ad A then Ad B then Ad C’). Let’s say you create a campaign to add to your brand lift study that targets an audience list ‘Viewers who watched Video B as an Ad’ and then use Ad C as your creative. Because all users that saw Ad C would have had to have seen Ad B first, this means your ‘control group’ will be primarily composed of users that have already seen your ad within the VAS campaign.

Such configurations mean the study can’t build a control group because targeted users who will view your ad have already seen it. If only viewers who are eligible to be entered into the study are going to be blocked, your control group won’t progress. In this case, you shouldn’t expect results to post.

Is the campaign targeting too narrow?

The following study and campaign setup configurations may sometimes reduce the number of survey responses that your study will be able to gather. The extent to which they slow down survey response collection varies depending on the degree to which they’re narrowing your targeting reach.

Audiences (particularly retargeting), placements, keywords, and topics

More restrictive targeting types, such as placements, keywords, and retargeting, reduce the number of eligible viewers and can lead to less impressions. Less impressions and less viewers in turn mean that there's less potential for viewers to fill out a survey.

Small geography

Too small of a geography might limit unique viewers, which reduces your odds of getting enough responses. Ideally, studies are run at the country level, but you can also target smaller geographies, as long as there’s a large enough population of viewers.

Recommendation: Monitor your traffic closely as the study is progressing. If you aren’t spending in full, broaden any overly restrictive targeting by expanding geography or removing overly restrictive targeting types like placements or keywords.

Are you issuing surveys that might have a low response rate by showing non-English surveys in all languages?

Your survey can only serve in one language. If you target multiple or “All languages”, you’re serving your survey to viewers who don’t speak that language. These viewers are likely to dismiss your survey. Thus, targeting multiple or “All languages'' isn't recommended, as this could lead to a negative experience for many viewers. If your survey is in English, depending on the country, you can target “All languages”, because English is a commonly spoken second language in many countries. Note that even in this case, it isn’t a recommended practice.

Recommendation: In your campaign targeting, have the language you target match the language of the survey. Avoid targeting multiple geographies that speak different languages unless you know there's a high number of bilingual users or if your survey is in English, which tends to be the most common second language of bilingual speakers.

Are there too many campaigns (or Video experiment arms) in the Lift Measurement Configuration (LMC)?

Too many campaigns (or Video experiment arms) in the LMC result in lower impressions per campaign/Video experiment arm. Use of Video experiments with many experiment arms may result in “Not Enough Data” at the campaign level if your campaign traffic isn’t large enough for each experiment arm.

Recommendation: If campaign level data is important to you, be conscientious of the number of experiment arms/campaigns within an arm that you add.

Additionally, inclusion of many campaigns in the same study (especially with overlapping targeting) may result in “Not Enough Data” at the campaign level. If you add more campaigns, this means that, at the campaign level, you need enough responses per campaign or reporting slice (for example, device, demo, or ad). If that level of reporting is a priority, this is something to keep in mind when you think about how many campaigns to add to the same study.

Avoid adding lots of campaigns to your study if campaign level reporting is a priority for you. If it is, consider running multiple studies with one campaign per study or use video experiments to ensure you don’t have cross contamination across studies.

For reach-focused campaigns, are you showing multiple ads to the same viewer?

If your goal is to reach as many viewers as possible, set a frequency cap on your campaign so that more viewers can view the ad and then respond to the survey. Without a frequency cap, you could be serving ads to the same user multiple times, but sending the survey only once, thus reducing the number of unique survey responses. It’s important to note that while adding a frequency cap helps you avoid getting ‘Not Enough Data’ it may hurt your odds of getting lift, as we've generally seen that higher frequency of impressions typically leads to more lift.

“No lift detected”

Sometimes a study that has ended with enough survey responses will still show “No lift detected”. This happens when there was no statistically significant difference between the survey responses from viewers who watched your video ad and those who didn’t. If you don't have lift at the study level, check if you have lift in specific segments (for example, age, gender, campaign or device). Consider focusing on those segments with positive lift.

As with any media channel, some metrics are more difficult to move than others. Some audiences are more difficult to reach than others. It’s normal for video campaigns to have no lift on certain metrics and audiences.

Below are a few things you can do to improve your campaign’s set up, creative or targeting to increase the chances of seeing lift.

Set up your study correctly

  • Select your competitor answer choices carefully
    • Mismatch in the competitor’s brand or product compared to your brand or product might lead them to be selected more than you. For example, if you’re a small beverage company and choose a globally recognized soda brand as your competitor in the brand lift survey answer choices, viewers might choose them more, resulting in no lift for your brand.
  • Ensure you entered your brand or product as the “Preferred Answer”
    • If you didn’t enter the advertised brand or product as the “Preferred answer”, the study ran with the wrong parameter. You can make edits and use Re-measurement to re-enable your study with the correct brand or product and competitors.
  • If the creative is focused on a product, choose the right product category
    • If your creative focused on a specific product, and you measured the impact on the brand, you’ll likely have “No lift detected”. Unfortunately, the study ran with too large a scope. You should wait for the next campaign to measure its effectiveness. You’ve learned that this creative is too product specific to move the overall brand.

Improve your creative

Quality of the creative plays a huge role in getting lift. Check if your ad is following the ABCDs of effective YouTube creative. Contact your account manager for detailed guidance on improving your creative.

For light-branded ads, if your brand or product name isn’t present, appears late in the ad, or is too subtle, the audience won’t attribute the creative back to the brand or product advertised. To correct this, consider adding branding, like an icon, watermark, or banner, earlier in the ad. You can also change the script to integrate the brand or product more clearly.

To lift lower funnel metrics, such as conversions, adding the branding early probably won’t be enough to cause a significant lift. The creative needs to be more persuasive. Consider moving the main argument to earlier in the creative, or include more arguments in the ad script.

Limit exposure to your creative outside the lift study

If a creative has been seen by viewers before the Brand Lift study has been launched, it’s possible that the control group (group that does not view the ad) has been contaminated, resulting in no lift. This will make them the control group respond similarly to how your exposed users respond, thus reducing ‘lift’. To minimize creative contamination:

  • Avoid running YouTube Video campaigns with non-Youtube channels like TV and other ad platforms.
  • Avoid multiple brand lift studies with the same or similar creative (unless using Video Experiments)
  • Avoid leaving out other video campaigns from your brand lift study with a similar creative

Target the right viewers for your campaign

Sometimes a campaign may not be tailored to the target audience. Watch the creative and ask yourself: is this message relevant to the target audience? If you have lift on certain segments, consider targeting them only with that creative. For the other segments, consider tailoring the creatives.

Brand Lift metrics

Brand Lift measurement data is available in most tables in Google Ads, including "Campaign", "Ad Group", "Demographics", and more. You can also view results at the "Product" or "Brand" level in the "Lift Measurement" table.

To check your Brand Lift measurement data:

  1. In your Google Ads account, click the Goals icon Goals Icon.
  2. Click the Measurements drop down in the section menu.
  3. Click Lift measurement.
  4. Click the columns icon A picture of the Google Ads columns icon.
  5. ClickModify columns.
  6. Select Brand lift, then click Apply.

You can also segment your measurement data by a specific metric (such as "Ad recall", "Awareness", "Consideration", "Favorability", and "Purchase Intent"):

  1. Click the segment icon Segment.
  2. Select Brand lift type to find the measurement data for your chosen metric.

Lifted users

This shows the estimated number of users in a sample survey whose perception of your brand changed as a result of your ads, extended to the overall reach of the campaign. It shows the difference in positive responses to your brand or product surveys between the group of users who saw your ad and the group who didn’t. For example, your ads could result in a lift in consideration (or awareness, or ad recall) with regard to your brand or product after seeing your ads.

The “lifted users” metric doesn’t necessarily measure unique users. A user may become lifted more than once during the course of your campaign.

Note: Using your data to re-engage lifted users is not supported.

Lifted users (co-viewed)

This metric is similar to the "lifted user" metric, but it also takes co-viewing into consideration. When multiple people watch YouTube on a connected TV (CTV) device together and view an ad at the same time, it could lead to more lifted users for your campaign. This metric includes lifted users from co-viewed impressions on CTV devices. Because there are no profiles available for co-viewers, they are treated as the same audience profile as the users who responded to surveys for the study.

Cost per lifted user

This shows the average cost for a lifted user who's now thinking about your brand after seeing your ads. Cost per lifted user is measured by dividing the total cost of your campaign by the number of lifted users. You can use this metric to understand the cost to change someone’s mind about your brand in terms of brand consideration, ad recall, or brand awareness.

Absolute Brand Lift

This metric shows the difference in positive responses to brand or product surveys between the group of people who saw your ads (the exposed group) and the group withheld from seeing your ads (the baseline group). This metric is calculated by subtracting the positive response rate of the baseline group from the exposed group. Absolute Brand Lift measures how much your ads influenced your audience’s positive feelings towards your brand or product. For example, an increase from 20% to 40% in the positive survey responses between the 2 surveyed groups represents an absolute lift of 20%.

Absolute Brand Lift and campaign performance

Absolute lift doesn’t necessarily reflect your overall brand lift performance. It is better to focus on a metric like cost per lifted user as the primary success metric of your campaign, because it factors in both reach and cost. View the following table:

Campaign Cost Cost per 1,000 impressions (CPM) Reach Absolute lift Lifted users Cost-per-lifted user
Campaign 1 $100 $15 6,666 10% 667 $0.15
Campaign 2 $100 $5 20,000 5% 1,000 $0.10
Difference n/a 66% 200% 50% 60% 33%

If you look at absolute lift only, Campaign 1 seems to perform better than Campaign 2. But at the same cost, Campaign 2 drove 50% more lifted users, at a 66% lower CPM, and with a 33% more efficient cost per lifted user.

Headroom Brand Lift

This measures the impact your ads had on increasing positive feelings towards your brand or product compared to the positive growth potential your brand or product could've gotten. This metric is calculated by dividing absolute lift by 1 minus the positive response rate of the baseline group. For example, an increase from 20% to 40% in the positive survey responses between the exposed group and the baseline groups represents a headroom lift of 25%.

Relative Brand Lift

This describes the difference in positive responses to brand or product surveys between users who saw your ads, versus users who were stopped from viewing your ads. This difference is then divided by the number of positive responses from the group of users who didn’t view your ads. The result measures how much your ads influenced your audience’s positive perception of your brand. For example, an increase from 20% to 40% in the positive survey responses between the two surveyed groups represents a relative lift of 100%.

Since survey responses can’t be collected for the entire exposed and baseline groups, this data is calculated from the responses that have been collected, which gives you an estimated number within a certain range. Usually, the confidence interval is 90%, so you can expect that in 90% of the cases, the true lift number will be within that range (if you were to have reached everyone).

Baseline positive response rate

This defines how often users who were stopped from seeing your ads responded positively to your brand. Use this metric to better understand how positive responses to your brand were influenced by general media exposure and other factors, not by seeing the ads in your campaigns.

Exposed survey responses

This metric shows the number of survey responses from people who saw your ads.

Note: If you view a low number in this column, that indicates that there aren’t enough survey responses yet. Continue running your campaigns and check back soon.

Baseline survey responses

This metric describes the number of survey responses from people who were withheld from seeing your ads.

Note: If you view a low number in this column, that indicates that there aren’t enough survey responses yet. Continue running your campaigns and check back soon.

Exposed positive response rate

This defines how often users who saw your ads responded positively to your brand.

Confidence interval

When talking about lift metrics like absolute lift, “point estimate” is usually referred to, which is the most likely lift generated by the ad. However, in Google Ads, you can also find a confidence interval for all brand lift metrics which is an estimated range in which your result could fall. This range is defined by an upper and lower bound, which are the highest and lowest values where your lift is likely to actually be. Lift results use 80% 2-sided confidence intervals, which means that there is an 80% chance that the true lift is between the lower bound and upper bound. This also means that you have a 90% chance that the lift is greater than the lower bound. For example, you may notice that your relative lift is 35%, which is the point estimate. However, you can also find that the confidence interval goes from 30% to 40%, which means that there is an 80% chance that the true lift is between 30%, the lower bound, and 40%, the upper bound. Another way to look at this is that there is a 90% chance that lift is greater than the 30%.

Certainty of lift

The certainty of lift, available only for brand lift, is an important metric to understand the reliability of your lift results. It represents the likelihood that the measured lift is generated by your campaigns, and not due to chance. The certainty of lift is calculated as 1 - p-value and can sometimes be referred to as the “statistical significance” or the “confidence” of lift results. The p-value tells you how likely your lift results would be if the ads were actually ineffective. Thus, a high certainty, corresponding to a low p-value, indicates that results are unlikely to have happened purely by chance and is a strong indication that your ads generated lift.

Note: The certainty of lift isn’t available to all accounts at this time. If you do not find it in your account, then you can only view results with the highest certainty, which is more than 90%.

Related links

Was this helpful?

How can we improve it?
Search
Clear search
Close search
Google apps
Main menu
11902721131397159120
true
Search Help Center
true
true
true
true
true
73067
false
false
false