Google releases August 2025 spam update

Google released its August 2025 spam update today, the company announced at 12:05 p.m. This is Google’s first announced algorithm update since the June 2025 core update. It is Google’s first spam update of 2025 and the first since December.

Timing. Google called this a “normal spam update” and it will take a “few weeks” to finish rolling out.

The announcement. Google announced:

  • “Today we released the August 2025 spam update. It may take a few weeks to complete. This is a normal spam update, and it will roll out for all languages and locations. We’ll post on the Google Search Status Dashboard when the rollout is done.”

Previous spam updates. Before today, Google’s last spam update was released Dec. 19 and finished rolling out Dec. 26; it was more volatile than the June 2024 spam update, which was released June 20, 2024 and completed rolling out June 27, 2024.

Why we care. This is the first Google algorithm update since the June 2025 core update. It’s unclear what type of spam this update is targeting, but if you see any ranking or traffic changes in the next few weeks, it could be due to this update.

Read more at Read More

ChatGPT’s answers came from Google Search after all: Report

ChatGPT Google unmasking

Multiple tests have suggested ChatGPT is using Google Search. Well, a new report seems to confirm ChatGPT is indeed using Google Search data.

  • OpenAI quietly used (and may still be using) a Google Search scraping service to power ChatGPT’s answers on real-time topics like news, sports, and finance, according to The Information.

The details. OpenAI used SerpApi, an 8-year-old scraping firm, to extract Google results.

  • Google has reportedly long tried to block SerpApi’s crawler, though it’s unclear how effective those efforts have been.
  • Other SerpApi customers reportedly include Meta, Apple, and Perplexity.

Zoom out. This revelation contrasts with OpenAI’s public stance that ChatGPT search relies on its own crawler, Microsoft Bing, and licensed publisher data.

Meanwhile. OpenAI CEO Sam Altman recently dismissed Google Search, saying:

  • “I don’t use Google anymore. I legitimately cannot tell you the last time I did a Google search.” 

Well, based on this news, it seems like he probably is using Google Search all the time within his own product.

Why we care. Google’s search index remains the foundation of online discovery – so much so that even its biggest AI search rival appears to be using it to partially power ChatGPT. This is yet another reminder that SEO isn’t going anywhere just yet. If Google’s results are valuable to OpenAI, they remain essential for driving visibility, traffic, and business outcomes.

The report. OpenAI Is Challenging Google—While Using Its Search Data (subscription required)

Read more at Read More

Historic recurrence in search: Why AI feels familiar and what’s next

Historic recurrence in search- Why AI feels familiar and what’s next

Historic recurrence is the idea that patterns repeat over time, even if the details differ.

In digital marketing, change is the only constant.

Over the last 30 years, we’ve seen nonstop shifts and transformations in platforms and tactics.

Search, social, and mobile have each gone through their own waves of evolution. 

But AI represents something bigger – not just another tactic, but a fundamental shift in how people research, evaluate, and buy products and services.

Estimates vary, but Gartner projects that AI-driven search could account for 25% of search volume by the end of 2026.

I suspect the true share will be much higher as Google weaves AI deeper into its results.

For digital marketers, it can feel like we need a crystal ball to predict what’s next. 

While we don’t have magical foresight, we do have the next best thing: lessons from the past.

This article looks back at the early days of search, how user behavior evolved alongside technology, and what those patterns can teach us as we navigate the AI era.

The early days: Wild and wonderful queries

If you remember the early web – AltaVista, Lycos, Yahoo, Hotbot – search was a free-for-all. 

People typed in long, rambling queries, sometimes entire sentences, other times just a few random words that “felt” right.

There were no search suggestions, no “people also ask,” and no autocorrect. 

It was a simpler time, often summed up as “10 blue links.”

Google Search - 10 blue links

Searchers had to experiment, refine, and iterate on their own, and the variance in query wording was huge.

For marketers, that meant opportunity. 

You could capture traffic in all sorts of unexpected ways simply by having relevant pages indexed.

Back then, SEO was, in large part, about one thing: existing in the index.

Dig deeper: A guide to Google: Origins, history and key moments in search

Google’s rise: From exploration to efficiency

Anyone working in digital marketing in the early 2000s will remember. 

From Day 1, Google felt different. The quality of its results was markedly better.

Then came Google Suggest in 2008, quietly changing the game. 

Suddenly, you didn’t have to finish typing your thought. Google would complete it for you, based on the most common searches.

Research from Moz and others at the time showed that autocomplete reduced query length and variance. 

People defaulted to Google’s suggestions because it was faster and easier.

This marked a significant shift in our behavior as searchers. We moved from sprawling, exploratory queries to shorter, more standardized ones.

It’s not surprising. When something can be achieved with less effort, human nature drives us toward the path of least resistance.

Once again, technology had changed how we search and find information.

Mobile, voice, and the second compression

The shift to mobile accelerated this compression.

Tiny keyboards and on-the-go contexts meant people typed as little as possible.

Autocomplete, voice input, and “search as you type” all encouraged brevity.

At the same time, Google kept rolling out features that answered questions directly, creating a blended, multi-contextual SERP.

The cumulative effect? Search behavior became more predictable and uniform.

For marketers running Google Ads or tracking performance in Google Analytics and Search Console, this shift came with another challenge: less data. 

Long-tail keywords shrank, while most traffic and budget concentrated on a smaller set of high-volume terms.

Once again, our search behavior – and the insights we could glean from it – had evolved.

Zero-click search and the walled garden

By the late 2010s, zero-click searches were on the rise. 

Google – and even social platforms – wanted to keep users inside their ecosystems.

More and more questions were answered directly in the search results. 

Search got smarter, and shorter queries could deliver more refined results thanks to personalization and past interactions.

Google started doing everything for us.

Search for a flight? You’d see Google Flights.

A restaurant? Google Maps. 

A product? Google Shopping. 

Information? YouTube

You get the picture.

For businesses built on organic traffic, this shift was disruptive. 

But for users, it felt seamless – arguably a better experience, even if it created new challenges for optimizers.

Get the newsletter search marketers rely on.


Quality vs. brevity

This shift worked – until it didn’t. 

One common complaint today is that search results feel worse

It’s a complicated issue to unpack. 

  • Have search results actually gotten worse? 
  • Or are the results as good as ever, but the underlying sites have declined in quality?

It’s tricky to call. 

What is certain is that as traffic declined, many sites got more aggressive – adding more ads, more pop-ups, and sneakier lead gen CTAs to squeeze more value from fewer clicks.

The search results themselves have also become a bewildering mix of ads, organic listings, and SERP features. 

To deliver better results from shorter queries, search engines have had to guess at intent while still sending enough clicks to advertisers and publishers to keep the ecosystem running.

And as traffic-starved publishers got more desperate, user experience took a nosedive. 

Anyone who has had to scroll through a food blogger’s life story – while dodging pop-ups and auto-playing ads – just to get to a recipe knows how painful this can be.

It’s this chaotic landscape that, in part, has driven the move to answer engines like ChatGPT and other large language models (LLMs). 

People are simply tired of panning for gold in the search results.

The AI era: From compression back to conversation

Up to this point, the pattern has been clear: the average query length kept getting shorter.

But AI is changing the game again, and the query-length pendulum is now swinging sharply in the opposite direction.

Tools like ChatGPT, Claude, Perplexity, and Google’s own AI Mode are making it normal to type or speak longer, more detailed questions again.

We can now:

  • Ask questions instead of searching for keywords. 
  • Refine queries conversationally. 
  • Ask follow-ups without starting over. 

And as users, we can finally skip the over-optimized lead gen traps that have made the web a worse place overall.

Here’s the key point: we’ve gone from mid-length, varied queries in the early days, to short, refined queries over the last 12 years or so, and now to full, detailed questions in the AI era.

The way we seek information has changed once more.

We’re no longer just searching for sources of information. We’re asking detailed questions to get clear, direct answers.

And as AI becomes more tightly integrated into Google over the coming months and years, this shift will continue to reshape how we search – or, more accurately, how we question – Google.

Dig deeper: SEO in an AI-powered world: What changed in just a year

AI and search: Google playing catch-up

Google was a little behind the AI curve.

ChatGPT launched in late 2022 to massive buzz and unprecedented adoption.

Google’s AI Overviews – frankly underwhelming by comparison – didn’t roll out until mid-2024. 

After launching in the U.S. in mid-June and the U.K. in late July 2025, Google’s full AI Mode is now available in 180 countries and territories around the world.

Now, we can ask more detailed, multi-part questions and get thorough answers – without battling through the lead gen traps that clutter so many websites.

The reality is simple: this is a better system.

This is progress.

Want to know the best way to boil an egg – and whether the process changes for eggs stored in the fridge versus at room temperature? Just ask.

Google will often decide if an AI Overview is helpful and generate it on the fly, considering both parts of your question.

  • What is the best way to boil an egg?
  • Does it differ if they are from the fridge?

The AI Overview answers the question directly. 

And if you want to keep going, you can click the bold “Dive deeper in AI Mode” button to continue the conversation.

Dive deeper in AI Mode

Inside AI Mode, you get streamlined, conversational answers to questions that traditional search could answer – just without the manual trawling or the painfully over-optimized, pop-up-heavy recipe sites.

From shorter queries to shorter journeys

Stepping back, we can see how behavior is shifting – and how it ties to human nature’s tendency to seek the path of least resistance.

The “easy” option used to be entering short queries and wading through an increasingly complex mix of results to find what you needed.

Now, the path of least resistance is to put in a bit more effort upfront – asking a longer, more refined question – and let the AI do the heavy lifting.

A search for the best steak restaurant nearby once meant seven separate queries and reviewing over 100 sites. That’s a lot of donkey work you can now skip.

It’s a subtle shift: slightly more work up front, but a far smoother journey in return.

This change also aligns with a classic computing principle: GIGO – garbage in, garbage out. 

A more refined, context-rich question gives the system better input, which produces a more useful, accurate output.

Historic recurrence: The pattern revealed

Looking back, it’s clear there’s a repeating cycle in how technology shapes search behavior.

The early web (1990s)

  • Behavior: Long, experimental, often clumsy queries.
  • Why: No guidance, poor relevance, and lots of trial-and-error.
  • Marketing lesson: Simply having relevant content was often enough to capture traffic.

Google + Autocomplete (2000s)

  • Behavior: Queries got shorter and more standardized.
  • Why: Google Suggest and smarter algorithms nudged users toward the most common phrases.
  • Marketing lesson: Keyword targeting became more focused, with heavier competition around fewer, high-volume terms.

Mobile and voice era (2010s–early 2020s)

  • Behavior: Even shorter, highly predictable queries.
  • Why: Tiny keyboards, voice assistants, and SERP features that answered questions directly.
  • Marketing lesson: The long tail collapsed into clusters. Zero-click searches rose. Winning visibility meant optimizing for snippets and structured data.

AI conversation era (2023–present)

  • Behavior: Longer, natural-language queries return – now in back-and-forth conversations.
  • Why: Generative AI tools like ChatGPT, Gemini, and Perplexity encourage refinement, context, and multi-step questions.
  • Marketing lesson: It’s no longer about just showing up. It’s about being the best answer – authoritative, helpful, and easy for AI to surface.

Technology drives change

The key takeaway is that technology drives changes in how people ask questions.

And tactically, we’ve come full circle – closer to the early days of search than we’ve been in years.

Despite all the doom and gloom around SEO, there’s real opportunity in the AI era for those who adapt.

What this means for SEO, AEO, LLMO, GEO – and beyond

The environment is changing.

Technology is reshaping how we seek information – and how we expect answers to be delivered.

Traditional search engine results are still important. Don’t abandon conventional SEO.

But now, we also need to optimize for answer engines like ChatGPT, Perplexity, and Google’s AI Mode.

That means developing deeper insight into your customer segments and fully understanding the journey from awareness to interest to conversion. 

  • Talk to your customers. 
  • Run surveys. 
  • Reach out to those who didn’t convert and ask why. 

Then weave those insights into genuinely helpful content that can be found, indexed, and surfaced by the large language models powering these new platforms.

It’s a brave new world – but an incredibly exciting one to be part of.

Read more at Read More

How to tell if Google Ads automation helps or hurts your campaigns

How to tell if Google Ads automation helps or hurts your campaigns

Smart BiddingPerformance Max, and responsive search ads (RSAs) can all deliver efficiency, but only if they’re optimizing for the right signals.

The issue isn’t that automation makes mistakes. It’s that those mistakes compound over time.

Left unchecked, that drift can quietly inflate your CPAs, waste spend, or flood your pipeline with junk leads.

Automation isn’t the enemy, though. The real challenge is knowing when it’s helping and when it’s hurting your campaigns.

Here’s how to tell.

When automation is actually failing

These are cases where automation isn’t just constrained by your inputs. It’s actively pushing performance in the wrong direction.

Performance Max cannibalization

The issue

PMax often prioritizes cheap, easy traffic – especially branded queries or high-intent searches you intended to capture with Search campaigns. 

Even with brand exclusions, Google still serves impressions against brand queries, inflating reported performance and giving the illusion of efficiency. 

On top of that, when PMax and Search campaigns overlap, Google’s auction rules give PMax priority, meaning carefully built Search campaigns can lose impressions they should own.

A clear sign this is happening: if you see Search Lost IS (rank) rising in your Search campaigns while PMax spend increases, it’s likely PMax is siphoning traffic.

Recommendation

Use brand exclusions and negatives in PMax to block queries you want Search to own. 

Segment brand and non-brand campaigns so you can track each cleanly. And to monitor branded traffic specifically, tools like the PMax Brand Traffic Analyzer (by Smarter Ecommerce) can help.

Dig deeper: Performance Max vs. Search campaigns: New data reveals substantial search term overlap

Auto-applied recommendations (AAR) rewriting structure

The issue

AARs can quietly restructure your campaigns without you even noticing. This includes:

  • Adding broad match keywords. 
  • “Upgrading” existing keywords to broader match types.
  • Adding new keywords that are sometimes irrelevant to your targeting.

Google has framed these “optimizations” as efficiency improvements, but the issue is that they can destabilize performance. 

Broad keywords open the door to irrelevant queries, which then can spike CPA and waste budget.

Recommendation

First, opt out of AARs and manually review all recommendations moving forward. 

Second, audit the changes that have already been made by going to Campaigns > Recommendations > Auto Apply > History. 

From there, you can see what change happened on what date, which allows you to go back to your campaign data and see if there are any performance correlations. 

Dig deeper: Top Google Ads recommendations you should always ignore, use, or evaluate

Modeled conversions inflating numbers

The issue

Modeled conversions can climb while real sales or MQLs stay flat. 

For example, you may see a surge in reported leads or purchases in your ads account, but when you look at your CRM, the numbers don’t match up. 

This happens because Google uses modeling to estimate conversions where direct measurement isn’t possible. 

If Google doesn’t have full tracking, it fills gaps by estimating conversions it can’t directly track, based on patterns in observable data. 

When left unchecked, the automation will double down on these patterns (because it assumes they’re correct), wasting budget on traffic that looks good but won’t convert.

Recommendation

Tell the automation what matters most to your business. 

Import offline or qualified conversions (via Enhanced Conversions, manual uploads, or CRM integration). 

This will ensure that Google optimizes for real revenue and not modeled noise.

When automation is boxed in: Reading the signals

Not every warning in Google means automation is failing. 

Sometimes the system is limited by the goals, budget, or inputs you’ve set – and it’s simply flagging that.

These diagnostic signals help you understand when to adjust your setup instead of blaming the algorithm.

Limited statuses (red vs. yellow)

The issue

A Limited status doesn’t always mean your campaign is broken. 

  • If you see a red Limited label, this means your settings are too strict. That could mean that your CPA or ROAS targets are unrealistic, your budget is too low, etc. 
  • Seeing a yellow Limited label is more of a caution sign. It’s usually tied to low volume, limited data, or the campaign is still learning.

Recommendation

If the status is red, loosen constraints gradually: raise your budget and ease up CPA/ROAS targets by 10–15%. 

If the status is yellow, don’t panic. This is Google’s version of telling you that they could use more money, if possible, but it’s not vital to your campaign’s success.

Responsive search ads (RSAs) inputs

The issue

RSAs are built in real-time from the headlines and descriptions you have already provided Google. 

At a minimum, advertisers are required to write 3 headlines with a maximum of 15 (and up to 4 descriptions). The fewer the assets you give the system, the less flexibility it will have. 

On the other hand, if you’re running a small budget and give the RSAs all 15 headlines and 4 descriptions, there is no way Google will be able to collect enough data to figure out which combinations actually work.

The automation isn’t failing with either. You’ve either given it too little information or too much with too little spending. 

Recommendation

Match asset volume to the budget allocated to the campaign. 

  • If you’re unsure, aim to write between 8-10 headlines and 2-4 descriptions.
  • If each headline/description isn’t distinct, don’t use it. 

Conversion reporting lag and attribution issues

The issue

Sometimes, Google Ads reports fewer conversions than your business actually sees. 

This isn’t necessarily an automation failure. It’s often just a matter of when the conversion is counted. 

By default, Google reports conversions on the day of the click, not the day the actual conversion happened. 

That means if you check performance mid-week, you might see fewer conversions than your campaign has actually generated because Google attributes them back to the click date. 

The data usually “catches up” as lagging conversions are processed.

Recommendation

Use the Conversions (by conversion time) column alongside the standard conversion column.

Conversions (by conversion time) column

This helps you separate true performance drops from simple reporting delays. 

If discrepancies persist beyond a few days, investigate the tracking setup or import accuracy. Just don’t assume automation is broken just because of timing gaps.

Get the newsletter search marketers rely on.


Where to look in the Google Ads UI

Automation leaves a clear trail within Google Ads if you know where to look. 

Here are some reports and columns to help spot when automation is drifting.

Bid Strategy report: Top signals 

The issue

The bid strategy report shows some of the signals Smart Bidding relies on when there is enough data. 

The “top signals” can sometimes make sense, and at other times, they can be a bit misleading. 

If the algorithm relies on weak signals (e.g., broad search themes and a lack of first-party data), its optimizations will be weak, too.

Bid Strategy report: Top signals 

Recommendation

Make checking your Top Signals a regular activity. 

If they don’t align with your business, fix the inputs. 

  • Improve conversion tracking.
  • Import offline conversions.
  • Reevaluate search themes.
  • Add customer/remarketing lists.
  • Expand your negative keyword list(s). 

Impression share metrics

The issue

When a campaign underdelivers, it’s tempting to assume automation is failing, but looking at Impression Share (IS) metrics tends to reveal the real bottleneck. 

By looking at Search Lost IS (budget), Search Lost IS (rank), and Absolute Top IS together, you can separate automation problems from structural or competitive ones.

How to use IS metrics as a diagnostic tool.

  • Budget problem
    • High Lost IS (budget) + low Lost IS (rank): Your campaign isn’t struggling. It just doesn’t have enough budget to run properly.
    • Recommendation: Raise the budget or accept capped volume.
  • Targets too aggressive
    • High Lost IS (rank) + low Absolute Top IS: If your Lost IS (rank) is high and your budget is adequate, your CPA/ROAS targets are likely too aggressive, causing Smart Bidding to underbid in auctions.
    • Recommendation: Loosen targets gradually (10-15%).

Scripts to keep automation honest

Scripts give you early warnings so you can step in before wasted spend piles up.

Anomaly detection

  • The issue: Automation can suddenly overspend or underspend when conditions in the marketplace change, but you often won’t notice until reporting lags.
  • Recommendation: Use an anomaly detection script to flag unusual swings in spend, clicks, or conversions so you can investigate quickly.

Query quality (N-gram analysis)

  • The issue: Broad match and PMax can drift into irrelevant themes (“free,” “jobs,” “definition”), wasting budget on low-quality queries.
  • Recommendation: Run an N-gram script to surface recurring poor-quality terms and add them as negatives before automation optimizes toward them.

Budget pacing

  • The issue: Google won’t exceed your monthly cap, but daily spend will be uneven. Pacing scripts help you spot front-loading.
  • Recommendation: A pacing script shows you how spend is distributed so you can adjust daily budgets mid-month or hold back funds when performance is weak.

Turning automation into an asset

Automation rarely fails in dramatic ways – it drifts. 

Your job isn’t to fight it, but to supervise it: 

  • Supply the right signals.
  • Track when it goes off course.
  • Step in before wasted spend compounds.

The diagnostics we covered – impression share, attribution checks, PMax insights, and scripts – help you separate real failures from cases where automation is simply following your inputs.

The key takeaway: automation is powerful, but not self-policing. 

With the right guardrails and oversight, it becomes an asset instead of a liability.

Read more at Read More

Global expansion and hyperlocal focus redefine the next chapter of retail media networks by DoorDash

Retail media networks are projected to be worth $179.5 billion by 2025, but capturing share and achieving long-term success won’t hinge solely on growing their customer base. With over 200 retail media networks now competing for advertiser attention, the landscape has become increasingly complex and crowded. The RMNs that stand out will be those taking a differentiated approach to meeting the evolving needs of advertisers.

The industry’s concentration creates interesting dynamics. While some platforms have achieved significant scale, nearly 70% of RMN buyers cite “complexity in the buying process” as their biggest obstacle. That tension, between explosive growth and operational complexity, is forcing the industry to evolve beyond traditional approaches.

As the landscape matures, which strategies will define the next wave of growth: global expansion, hyperlocal targeting, or both?

The evolution of retail media platforms

To understand where the industry is heading, it’s worth examining how successful platforms are addressing advertisers’ core challenges. Lack of measurement standards across platforms continues to frustrate advertisers who want to compare performance across networks. Manual processes dominate smaller networks, making campaign management inefficient and time-consuming.

At the same time, most retailers lack the digital footprint necessary for standalone success. This has created opportunities for platforms that can solve multiple problems simultaneously: standardization, automation, and scale.

DoorDash represents an interesting case study in this evolution. The platform has built its advertising capabilities around reaching consumers at their moment of local need across multiple categories. With more than 42 million monthly active consumers as of December 2024, DoorDash provides scale and access to high-intent shoppers across various categories spanning restaurants, groceries and retail.

The company’s approach demonstrates how platforms can address advertiser pain points through technology. DoorDash’s recent platform announcement showcases this evolution: the company now serves advertisers with new AI-powered tools and expanded capabilities. Through its acquisition of ad tech platform Symbiosys, a next-generation retail media platform, brands can expand their reach into digital channels, such as search, social, and display, and retailers can extend the breadth of their retail media networks.

Global expansion meets local precision

International expansion presents both opportunities and challenges for retail media networks. Europe’s retail media industry is projected to surpass €31 billion by 2028,. This creates opportunities for networks that can solve the technology puzzle of operating across multiple geographies.

The challenge lies in building platforms that work seamlessly across countries while maintaining local relevance. International expansion requires handling different currencies, regulations, and cultural contexts—capabilities that many networks struggle to develop.

DoorDash’s acquisition of Wolt illustrates how platforms can achieve global scale while maintaining local connections. The integration enables brands to manage campaigns across Europe and the U.S. through a single interface—exactly the kind of operational efficiency that overwhelmed advertisers seek.

The combined entity now operates across more than 30 countries, with DoorDash and Wolt Ads crossing an annualized advertising revenue run rate of more than $1 billion in 2024. What makes this expansion compelling isn’t just the scale—it’s how the integration maintains neighborhood-level precision across diverse geographies.

Wolt has transformed from a food delivery platform into what it describes as a multi-category “shopping mall in people’s pockets.”

The hyperlocal advantage: context beats demographics

Here’s what’s really changing the game: the shift from demographic targeting to contextual precision. Privacy regulations favor contextual targeting over behavioral tracking, but that’s not the only reason smart networks are going hyperlocal.

Location-based intent signals provide dramatically higher conversion probability than traditional demographics. Real-time contextual data—weather patterns, local events, proximity to fulfillment—influences purchase decisions in immediate, actionable ways that broad demographic targeting simply can’t match.

DoorDash built its entire advertising model around this insight, reaching consumers at the exact moment of local need across multiple categories. The platform provides scale and access to high-intent shoppers with contextual precision. A recent innovation that exemplifies this approach is Dayparting for CPG brands, which enables advertisers to target users in their local time zones—a level of time-based precision that distinguishes hyperlocal platforms from broader retail media networks.

In one example, Unilever applied Dayparting to focus on late-night and weekend windows for its ice cream campaigns, aligning ad delivery with peak demand periods. Over a two-week period, 77% of attributed sales were new-to-brand, demonstrating the power of contextual timing in driving incremental reach.

Major brands, including Unilever, Coca-Cola, and Heineken, utilize both DoorDash and Wolt platforms for hyperlocal targeting, proving the model is effective for both endemic and non-endemic advertisers seeking neighborhood-level precision.

Technology evolution: measurement and automation

The technical requirements for next-generation retail media networks extend far beyond basic advertising capabilities. Self-serve functionality has become standard for international geographies—not because it’s trendy, but because manual campaign management doesn’t scale across dozens of countries with different currencies, regulations, and cultural contexts.

Cross-country campaign management requires unified dashboards that manage complexity while maintaining simplicity for advertisers. Automation isn’t optional anymore; it’s necessary to compete with established players who’ve built machine learning into their core operations.

But here’s what’s really transforming measurement: new attribution methodologies that go beyond traditional ROAS. When platforms can integrate fulfillment data with advertising exposure, they enable real-time performance tracking that connects ad spend to actual business outcomes rather than just clicks and impressions.

Progress on standardization continues through IAB guidelines addressing measurement consistency, alongside industry pushes for technical integration standards. The challenge lies in balancing standardization with differentiation—networks need to offer easy integration and consistent measurement while maintaining unique value propositions.

In a move toward addressing advertisers’ need for measurement consistency, DoorDash recognized that restaurant brands valued both click and impression-based attribution for their sponsored listing ads, and recently introduced impression-based attribution and reporting in Ads Manager. This has enabled restaurant brands to gain a deeper understanding of performance and results driven on DoorDash.

Global technology challenges add another layer of complexity: multi-currency transactions, local payment methods, regulatory compliance across countries, and cultural adaptation while maintaining platform consistency. These aren’t afterthoughts for international platforms, they’re core competencies that determine success or failure.

Industry outlook: consolidation and opportunity

Retail media is heading toward consolidation, but not in the way most people expect. Hyperlocal networks are positioned to capture share from undifferentiated RMNs that compete solely on inventory volume. Geographic specialization is becoming a viable alternative to traditional scale-focused approaches.

Simultaneously, community impact measurement is gaining importance for brand strategy. Marketers are discovering that advertising dollars spent on local commerce platforms create multiplier effects—supporting neighborhood businesses and strengthening local economies in ways that traditional e-commerce advertising doesn’t achieve.

The networks that understand this dynamic, that can offer global platform capabilities with genuine local industry expertise, are the ones positioned to define retail media’s next chapter. Success requires technology integration that enables contextual and location-based targeting, plus measurement solutions that prove incrementality beyond traditional metrics.

The path forward

As retail media networks mature, success lies not in choosing between global scale and local relevance, but in achieving both simultaneously. The DoorDash-Wolt combination provides a compelling blueprint, demonstrating how technology platforms can enable international expansion while deepening neighborhood-level connections.

For marketers navigating this evolution, the fundamental question shifts from “where should we advertise?” to “how can we reach consumers at their moment of need?” Networks that answer this effectively—through global reach, hyperlocal precision, or ideally both, will write retail media’s next chapter.Interested to learn more about DoorDash Ads? Get started today.

Read more at Read More

Google Ads expands PMax Channel Reporting to account level

Your guide to Google Ads Smart Bidding

Performance Max (PMax) advertisers just got a major visibility upgrade: Channel Reporting is now available at the account level, not just within individual campaigns.

How it works:

  • View and compare all PMax campaigns in a single reporting overview.
  • Segment by conversion metrics to understand what’s driving results.
  • Identify performance patterns across channels without jumping campaign to campaign.

Why we care. Until now, channel performance data was siloed within each PMax campaign. The new account-level reporting makes it easier to spot trends, compare results, and optimize across campaigns.

The big picture. Google notes that channel data is available for PMax campaigns “at this time” — a phrasing that suggests the feature could expand to other campaign types down the road.

Bottom line. More visibility, less friction. This change gives advertisers a faster, more complete view of PMax performance — and hints at broader reporting upgrades ahead.

First seen. This update was first picked up by Jun von Matt IMPACT’s Head of Google Ads, Thomas Eccel.

Read more at Read More

Why community is the antidote to AI overload in search marketing

Why community is the antidote to AI overload in search marketing

In 2025, people aren’t just searching for answers anymore.

They’re looking for genuine responses from the people they trust most: 

  • Creators.
  • Communities.
  • Fellow brand supporters. 

In many ways, community has become an algorithm of its own.

AI-powered tools like Google Gemini, ChatGPT, and Perplexity have made knowledge more accessible than ever. 

But in doing so, they’ve also flattened it. 

Answers feel repetitive, citations pull from the same limited sources, and brand voices risk becoming interchangeable.

That’s where community comes in. 

While generative AI commoditizes information, community restores individuality. 

It offers what no model can compress into tokens: 

  • Authentic connection.
  • Lived experience.
  • Trust.

When democratized information becomes homogenized

I can still remember when Google – and later YouTube – made information feel democratized, putting knowledge at our fingertips like never before.

But with the rise of AI, that same accessibility now comes at a cost: everything starts to sound the same.

Every brand competing for similar keywords risks becoming interchangeable, sounding the same in AI-generated summaries that deliver information without distinction. 

Meanwhile, authority is concentrated into a small set of repeatedly cited sources, so users encounter little variation in what LLMs surface.

In some ways, this is similar to traditional SEO

But there’s an important difference: websites once gave us the chance to “get our brand over” and show what made our solution unique. 

That’s what feels lost in an AI-driven search experience.

Still, within this sameness lies opportunity. 

While brands fight for visibility inside an AI Overview, those with the strongest communities can not just stand apart – but truly stand alone.

Dig deeper: SEO for user activation, retention and community

Community as the differentiator

AI responses are built around compression – getting audiences to an answer as quickly and concisely as possible. 

Community, on the other hand, expands.

AI platforms tend to generalize first, then personalize only when prompted. 

Community works the opposite way: it personalizes from the start.

In my view, that’s the kind of user experience audiences will ultimately prefer – and it’s how brands will become the choice within their niche.

Think about:

  • A Reddit thread that discusses your product specifically. That’s not just another citation. It’s a living testimonial, open to being challenged or reinforced in real time.
  • A Discord server filled with engaged users doesn’t just provide customer support – it showcases the culture and identity your brand is building.
  • Social comment threads around a creator’s content show personality, emotion, and authenticity that no LLM can replicate.

Ultimately, your community gives your brand the one thing AI can’t compress or flatten into tokens: individuality. 

In a world of sameness, community is what gives your brand its voice back.

UGC? Hello, UGT! 

User-generated content (UGC) has long been viewed as central to search marketing – a key driver of discoverability.

That’s still true, but the conversation has matured. 

It’s no longer just about “content.” What truly matters now is user-generated trust (UGT).

This may sound like a subtle mindset shift, but it changes everything. 

Search marketing teams should focus on the real, ongoing conversations within communities that validate products and learn how to leverage those conversations wherever possible.

That’s where genuine user advocacy emerges. And it’s advocacy that increasingly shows up in SERPs and AI responses.

Whether it’s a YouTube video featured in search results, a Reddit thread highlighted in an AI answer, or a TikTok creator’s series, UGT creates organic momentum. 

It sends signals to both people and algorithms that your brand is credible, trustworthy, and the preferred choice.

Backlinks can be gamed, and citations scraped or manufactured.

UGC is about output. And UGT? It’s about advocacy and credibility – and that’s exactly what search marketing teams need to drive lasting results.

Dig deeper: Advanced tactics to maximize the SEO value of user-generated content

Get the newsletter search marketers rely on.


How owned and earned communities build trust

When thinking about your brand community, there are two key considerations.

Owned communities

These are spaces where conversations and culture consistently reinforce or evolve your positioning and shape how you’re perceived.

They include:

  • Brand Discord channels. 
  • Slack groups.
  • Reddit forums.

Earned communities

These are the building blocks your brand participates in – where authenticity can either strengthen or undermine trust.

Think of:

  • Reddit threads not owned by your brand.
  • Facebook groups.
  • Quora discussions.
  • Comment threads.
  • Other spaces where people gather independently. 

Both owned and earned communities play a critical role in the smartest strategies. 

By seeding and nurturing conversations where your community and broader audience already gather.

By cultivating a “home” that is uniquely yours, you protect your brand against the homogenizing effect of AI-driven search.

Dig deeper: The rise of forums: Why Google prefers them and how to adapt

Why community is the secret sauce of search everywhere

Here’s a sobering thought: AI is only going to get better. 

They’ll become more skilled at surfacing consensus and amplifying shared rhetoric.

But consensus doesn’t drive differentiation. Community does.

A backlink can be replicated. A feature in a listicle can be matched. That’s just search marketing ping-pong.

A community, on the other hand, can’t be scraped, cloned, or copied.

Brands that invest in their communities today aren’t just building engagement.

They’re building something much more powerful: a moat of differentiation and individuality.

Community resists the sameness of AI-driven search. It’s what ensures your brand’s voice doesn’t just show up, but truly stands out.

LEGO Ideas: Community in action

One brand that proves the power of community as a competitive advantage – especially in an AI-driven world – is LEGO.

Through its LEGO Ideas platform, the company has turned its community into a creative engine for product ideation and a discovery layer that informs both content and product development.

Fans submit their ideas and vote on their favorites. The best and most popular are turned into real products. 

Everything from pop culture tie-ins (recently, a Wallace and Gromit set was greenlit) to architectural replicas has emerged through this process.

So why is this powerful from a search perspective?

There are two key reasons.

Authenticity at scale, organically

Every submission, vote, and comment is UGT in action. 

The community validates which ideas deserve attention, creating a visible signal of credibility long before a set hits the shelves.

Fan conversations fuel visibility

The conversations, forums, and social amplification around these fan-led projects fuel organic visibility. 

A single fan concept can spark thousands of blog posts, Reddit threads, YouTube videos, and TikToks – surfacing LEGO in contexts no AI citation list could ever replicate.

While competitors battle for presence in AI summaries or listicle roundups – vying to be labeled “Best Construction Toy” – LEGO has built differentiation through something much harder to copy: a living, breathing community that fuels product innovation, search visibility, and brand preference.

No amount of AI summarization can flatten a brand’s individuality when its users are constantly creating new stories about it.

Dig deeper: How to use social and forum data to inform next-level SEO strategies

From presence to preference: Why community wins

The future of search can’t be about simply being listed. 

Visibility alone is no longer enough.

Brands need to feel alive – human – and that happens through community.

AI can summarize anything, but it can’t replicate belonging. 

That’s why the brands investing in their communities today will be the ones that win tomorrow.

They won’t just be seen; they’ll be chosen. They’ll become the preference.

So, start by asking: where is my community already thriving? 

Listen, nurture, and amplify. 

That’s how you turn presence into preference in a world where every brand shows up.

Read more at Read More

The future of remarketing? Microsoft bets on impressions, not clicks

The future of remarketing? Microsoft bets on impressions, not clicks

There’s a shift happening in digital advertising. 

For years, remarketing hinged on clicks: someone had to visit your site, trigger a pixel, and leave behind a trail you could follow with ads. 

But what if you could build your remarketing audience before they ever click?

That is the core promise of impression-based remarketing – a Microsoft Advertising-exclusive capability that lets advertisers build audiences (or exclusions) simply from users seeing their ads. 

No click. No form fill. Just an impression.

In a world of privacy shifts, AI-driven search, and fractured attention spans, this approach may not just be a nice-to-have – it could be the future.

(Disclosure: I work as Microsoft’s product liaison, and the perspectives shared here reflect my role inside Microsoft Advertising.)

What is impression-based remarketing? 

Impression-based remarketing is Microsoft Advertising’s super-powered audience targeting method. 

Instead of waiting for a user to take an action such as visiting your site, it lets you track and segment audiences based solely on ad visibility. 

Here is how it works in plain terms: 

If your ad is displayed on Bing search results, native placements, Copilot, or other Microsoft inventory, the person who saw it can be added to a remarketing list. 

That list can then be used for targeting, exclusions, or bid adjustments across eligible campaigns. 

Key operational details: 

  • You can define up to 20 sources (campaigns or ad groups that feed your remarketing lists). 
  • The audience membership window can be 1-30 days (seven days is often the sweet spot for balancing awareness and consumer sentiment). 
  • Any campaign type can be a source, but not all can be a target. For example, Premium streaming can feed lists but cannot be targeted directly. 
  • Emerging surfaces like Copilot impressions are eligible as both sources and targets, though granular reporting is not yet fully available. To clarify, only Showroom ads (currently in closed beta) can specifically target Copilot placements. 
  • If you use autobidding, Microsoft’s system will factor in your bid adjustments, meaning a +20% bid really will raise CPC or CPM. 

In short, it is the ability to remarket to people who have only seen your ad, which opens up a broader, top-of-funnel opportunity while respecting the growing limitations on tracking. 

Dig deeper: Microsoft Advertising expands remarketing list sources to 20 campaigns

How to use it – functionally and strategically 

Think of impression-based remarketing in two phases: 

  • Functional setup: The technical nuts and bolts. 
  • Strategic execution: Deciding which campaigns feed the lists, which campaigns target them, and what creative to use. 
Microsoft Ads impression-based remarketing - How to use it – functionally and strategically

Functional setup 

  • Build your audience lists
    • Identify the campaigns or ad groups that will act as sources. 
    • These are the ads whose impressions will populate your lists. 
  • Create associations
    • Associate your sources with the target campaigns where you will use the audiences for targeting, exclusions, or bid adjustments. 
    • At least one audience ad must be in your associations to make all campaign types eligible to target. 
  • Decide on membership duration
    • Seven days is often ideal to balance recency with volume, but your industry’s buying cycle may warrant shorter or longer windows. 
  • Layer on bid strategies
    • Keep in mind that bid adjustments impact CPC or CPM directly under auto-bidding. 
Microsoft Ads impression-based remarketing - Functional setup

Strategic execution 

This is where impression-based remarketing can go from “neat” to “needle-moving.” 

Empathize with the customer journey 

A first-time viewer is not ready for the same message as a warm lead. 

The most common mistake in Impression-based remarketing is running the same creative to people regardless of where they are in the funnel. 

For example: 

  • Cold audience (first exposure): Focus on brand awareness and curiosity hooks. 
  • Warm audience (saw an ad, maybe interacted with other brand assets): Lean into unique value propositions and proof points. 
  • Hot audience (familiar, showing intent signals): Shift toward urgency, offers, or clear conversion CTAs. 

Tailor messaging to decision-makers vs. influencers 

Not all buyers are the same. In B2B, especially, the person seeing your ad may not be the one signing the check. 

  • Decision-maker personas respond to concrete ROI, cost, terms, and support benefits. 
  • Influencer personas, those who need to convince the buyer, often respond better to emotional appeals, user stories, or tips on how to get leadership buy-in. 

Use micro-steps in the buyer’s journey 

Since the trigger is just an impression, do not assume you can skip stages. 

Instead of expecting someone to leap from “saw ad” to “buy,” map out micro-conversions: 

  • Move from awareness to engagement (click, video view). 
  • From engagement to consideration (content download, add to cart). 
  • From consideration to decision (purchase, sign-up). 

Sometimes this means setting ad groups, not just campaigns, as your sources and targets to allow for precise audience control.

Budget with conversion thresholds in mind 

If your targeting is too narrow, you might never gather enough impressions to reach performance significance.

Budgets should align with the audience sizes needed to meet your conversion goals.

Microsoft Ads impression-based remarketing - Strategic execution

Get the newsletter search marketers rely on.


Why impressions are the future 

The shift to impression-based remarketing is not just about Microsoft offering a new targeting lever. 

It’s about survival in a rapidly changing ecosystem. 

1. Privacy is rewriting the rules 

With cookie deprecation, consent restrictions, and stricter data privacy laws, the reliable, click-based remarketing audiences of the past are disappearing. 

An impression, recorded server-side, does not rely on a user’s browser for tracking, making it a more resilient signal. 

2. AI-powered search changes user behavior 

As conversational AI like Microsoft Copilot, ChatGPT, and other assistants take center stage, the traditional search journey (“type query → click site → take action”) is being replaced. 

In many cases, users will get answers without ever clicking a link. 

This means advertisers must reach and influence people before they click, or even without them clicking at all. 

Dig deeper: How Microsoft Ads compares to Google Ads and when to use it

3. Sentiment and recall become the new metrics 

The old metrics, such as CTR, do not tell the whole story when much of the journey happens off-site. 

The future winners will be brands that: 

  • Create memorable touchpoints. 
  • Build positive sentiment before a user enters the buying stage. 
  • Stay top-of-mind when the moment of need arises. 

Impression-based remarketing allows you to intentionally re-engage based on visibility alone, which aligns perfectly with these goals. 

4. Redeeming undervalued placements 

Historically, advertisers have excluded certain placements, such as mobile games or sites with high ad density, because they seemed “low quality” in a click-through world. 

Those same environments can be very effective for brand imprinting. 

The user might not click in the moment, but repeated impressions in familiar contexts can drive recall later. 

Impression-based remarketing allows you to capitalize on these “slow burn” touches without overvaluing accidental clicks.  

Takeaways for advertisers 

If you are planning campaigns for the holiday season or for the AI-driven world we are already stepping into, here is the checklist to make impression-based remarketing work for you: 

  • Set it up now
    • Build your sources and associations. 
    • Keep the target list broad, but be selective with your sources. 
  • Map the journey
    • Identify what someone needs to see first, second, and third. 
    • Create dedicated creative for each stage. 
  • Respect personas
    • Decision-makers and influencers need different messaging. 
    • Avoid “one size fits all” creative blasts. 
  • Budget for volume and thresholds
    • Without enough impressions, your targeting power fades. 
    • Ensure campaigns have enough spend to feed the machine. 
  • Think beyond clicks
    • Use impression-based lists to drive brand familiarity, not just immediate conversions. 
    • Measure impact with recall and sentiment studies where possible. 

Impression-based remarketing: From feature to future

Impression-based remarketing is not just another targeting option. 

It is a structural shift in how advertisers can build relationships with their audiences. 

In a clickless, AI-mediated future, it lets you control the who and when of your targeting, even if the how of user interaction changes completely. 

Microsoft might have positioned it as a feature, but for savvy advertisers, it is a competitive moat. 

Dig deeper: How to maximize your Google Ads remarketing campaigns

Read more at Read More

Google traffic to news publishers is steady, but it isn’t traditional Search

Google traffic news sites

Google has remained a stable source of traffic to news publishers over the past year. Although many websites have seen their traffic significantly impacted by Google’s AI Overviews, Chartbeat data shows that for 565 U.S. and UK news publishers:

  • Search referrals made up 19% of traffic in July, little changed since early 2019.
  • Google dominates search traffic: 96% of publisher referrals.
How publisher traffic referral types are stacking up.

Yes, but. “Search” here includes Google Discover, which is not traditional search. Discover is now the primary driver of Google referrals.

Why we care. Search traffic hasn’t collapsed. However, the stability is somewhat masked by a shift from traditional Google Search to Google Discover.

Dig deeper. Google says AI is boosting Search. Yes, but…

Direct traffic is shaky. Efforts to build a loyal, “type-in” audience have largely stalled, leaving publishers more dependent on Google and aggregators. Direct traffic to homepages and landing pages has fallen to 11.5% from a pandemic-era high of 16.3%.

Social keeps sinking. Social’s decline means fewer diversified referral sources:

  • Facebook referrals are down 50% since 2019, despite a recent bump.
  • X traffic is down 75% vs. 2019.
  • Only Reddit is surging – up 220% since 2019, boosted by Google visibility and an AI training deal (but it still sends less referrals than Facebook and X).

The report. Publisher traffic sources: Google steady but social and direct referrals are down, as reported by PressGazette

Read more at Read More

Google finally gives visibility into Search Partner Network placements

Why campaign-specific goals matter in Google Ads

Advertisers can now see exactly where their Search, Shopping, and App campaign ads are running across the Search Partner Network (SPN), with full site-level impression data.

How it works:

  • Reports list all SPN sites where your ads appeared.
  • Impression data is broken down at the site level.
  • Works like existing placement reports in Performance Max.

Why we care. Transparency has long been a sticking point with SPN. This update gives advertisers the visibility they’ve been asking for – and the ability to make smarter, brand-safe decisions.

The big picture. This change empowers advertisers to:

  • Audit brand suitability more effectively.
  • Optimize spend by analyzing which sites drive value.
  • Gain tighter control over campaign performance.

First seen. This update was first noted by Anthony Higman, founder and CEO of ADSQUIRE. He is still skeptical of Search Partner Networks despite it being an answer to a request advertisers have made for years:

  • “Still Most Likely Wont Be Participating In The Search Partner Network But This Is Unprecedented And What ALL Advertisers Have Been Requesting For Decades Now!!!”

Bottom line. Advertisers finally have the transparency and control needed to run on SPN confidently and optimize placements for better results.

Read more at Read More