Posts

Why LLM-only pages aren’t the answer to AI search

Why LLM-only pages aren’t the answer to AI search

With new updates in the search world stacking up in 2026, content teams are trying a new strategy to rank: LLM pages.

They’re building pages that no human will ever see: markdown files, stripped-down JSON feeds, and entire /ai/ versions of their articles.

The logic seems sound: if you make content easier for AI to parse, you’ll get more citations in ChatGPT, Perplexity, and Google’s AI Overviews.

Strip out the ads. Remove the navigation. Serve bots pure, clean text.

Industry experts such as Malte Landwehr have documented sites creating .md copies of every article or adding llms.txt files to guide AI crawlers.

Teams are even building entire shadow versions of their content libraries.

Google’s John Mueller isn’t buying it.

  • “LLMs have trained on – read and parsed – normal web pages since the beginning,” he said in a recent discussion on Bluesky. “Why would they want to see a page that no user sees?”
JohnMu, Lily Ray on BlueSky

His comparison was blunt: LLM-only pages are like the old keywords meta tag. Available for anyone to use, but ignored by the systems they’re meant to influence.

So is this trend actually working, or is it just the latest SEO myth?

The rise of ‘LLM-only’ web pages

The trend is real. Sites across tech, SaaS, and documentation are implementing LLM-specific content formats.

The question isn’t whether adoption is happening, it’s whether these implementations are driving the AI citations teams hoped for.

Here’s what content and SEO teams are actually building.

llms.txt files

A markdown file at your domain root listing key pages for AI systems.

The format was introduced in 2024 by AI researcher Simon Willison to help AI systems discover and prioritize important content. 

Plain text lives at yourdomain.com/llms.txt with an H1 project name, brief description, and organized sections linking to important pages.

Stripe’s implementation at docs.stripe.com/llms.txt shows the approach in action:

markdown# Stripe Documentation

> Build payment integrations with Stripe APIs

## Testing

- [Test mode](https://docs.stripe.com/testing): Simulate payments

## API Reference

- [API docs](https://docs.stripe.com/api): Complete API reference

The payment processor’s bet is simple: if ChatGPT can parse their documentation cleanly, developers will get better answers when they ask, “how do I implement Stripe.”

They’re not alone. Current adopters include Cloudflare, Anthropic, Zapier, Perplexity, Coinbase, Supabase, and Vercel.

Markdown (.md) page copies

Sites are creating stripped-down markdown versions of their regular pages.

The implementation is straightforward: just add .md to any URL. Stripe’s docs.stripe.com/testing becomes docs.stripe.com/testing.md.

Everything gets stripped out except the actual content. No styling. No menus. No footers. No interactive elements. Just pure text and basic formatting.

The thinking: if AI systems don’t have to wade through CSS and JavaScript to find the information they need, they’re more likely to cite your page accurately.

/ai and similar paths

Some sites are building entirely separate versions of their content under /ai/, /llm/, or similar directories.

You might find /ai/about living alongside the regular /about page, or /llm/products as a bot-friendly alternative to the main product catalog. 

Sometimes these pages have more detail than the originals. Sometimes they’re just reformatted.

The idea: give AI systems their own dedicated content that’s built for machine consumption, not human eyes. 

If a person accidentally lands on one of these pages, they’ll find something that looks like a website from 2005.

JSON metadata files

Dell took this approach with their product specs.

Instead of creating separate pages, they built structured data feeds that live alongside their regular ecommerce site.

The files contain clean JSON – specs, pricing, and availability.

Everything an AI needs to answer “what’s the best Dell laptop under $1000” without having to parse through product descriptions written for humans.

You’ll typically find these files as /llm-metadata.json or /ai-feed.json in the site’s directory.

# Dell Technologies

> Dell Technologies is a leading technology provider, specializing in PCs, servers, and IT solutions for businesses and consumers.

## Product and Catalog Data

- [Product Feed - US Store](https://www.dell.com/data/us/catalog/products.json): Key product attributes and availability.

- [Dell Return Policy](https://www.dell.com/return-policy.md): Standard return and warranty information.

## Support and Documentation

- [Knowledge Base](https://www.dell.com/support/knowledge-base.md): Troubleshooting guides and FAQs.

This approach makes the most sense for ecommerce and SaaS companies that already keep their product data in databases. 

They’re just exposing what they already have in a format AI systems can easily digest.

Dig deeper: LLM optimization in 2026: Tracking, visibility, and what’s next for AI discovery

Real-world citation data: What actually gets referenced

The theory sounds good. The adoption numbers look impressive. 

But do these LLM-optimized pages actually get cited?

The individual analysis

Landwehr, CPO and CMO at Peec AI, ran targeted tests on five websites using these tactics. He crafted prompts specifically designed to surface their LLM-friendly content.

Some queries even contained explicit 20+ word quotes designed to trigger specific sources.

Landwehr - LLM experiment 1

Across nearly 18,000 citations, here’s what he found.

llms.txt: 0.03% of citations

Out of 18,000 citations, only six pointed to llms.txt files. 

The six that did work had something in common: they contained genuinely useful information about how to use an API and where to find additional documentation. 

The kind of content that actually helps AI systems answer technical questions. The “search-optimized” llms.txt files, the ones stuffed with content and keywords, received zero citations.

Markdown (.md) pages: 0% of citations

Sites using .md copies of their content got cited 3,500+ times. None of those citations pointed to the markdown versions. 

The one exception: GitHub, where .md files are the standard URLs. 

They’re linked internally, and there’s no HTML alternative. But these are just regular pages that happen to be in markdown format.

/ai pages: 0.5% to 16% of citations

Results varied wildly depending on implementation. 

One site saw 0.5% of its citations point to its/ai pages. Another hit 16%. 

The difference? 

The higher-performing site put significantly more information in their /ai pages than existed anywhere else on their site. 

Keep in mind, these prompts were specifically asking for information contained in these files. 

Even with prompts designed to surface this content, most queries ignored the /ai versions.

JSON metadata: 5% of citations

One brand saw 85 out of 1,800 citations (5%) come from their metadata JSON file. 

The critical detail here is that the file contained information that didn’t exist anywhere else on the website. 

Once again, the query specifically asked for those pieces of information.

Landwehr - LLM experiment 1

The large-scale analysis

SE Ranking took a different approach

Instead of testing individual sites, they analyzed 300,000 domains to see if llms.txt adoption correlated with citation frequency at scale.

Only 10.13% of domains, or 1 in 10, had implemented llms.txt. 

For context, that’s nowhere near the universal adoption of standards like robots.txt or XML sitemaps.

During the study, an interesting relationship between adoption rates and traffic levels emerged.

Sites with 0-100 monthly visits adopted llms.txt at 9.88%. 

Sites with 100,001+ visits? Just 8.27%. 

The biggest, most established sites were actually slightly less likely to use the file than mid-tier ones.

But the real test was whether llms.txt impacted citations. 

SE Ranking built a machine learning model using XGBoost to predict citation frequency based on various factors, including the presence of llms.txt.

The result: removing llms.txt from the model actually improved its accuracy. 

The file wasn’t helping predict citation behavior, it was adding noise.

The pattern

Both analyses point to the same conclusion: LLM-optimized pages get cited when they contain unique, useful information that doesn’t exist elsewhere on your site.

The format doesn’t matter. 

Landwehr’s conclusion was blunt: “You could create a 12345.txt file and it would be cited if it contains useful and unique information.”

A well-structured about page achieves the same result as an /ai/about page. API documentation gets cited whether it’s in llms.txt or buried in your regular docs.

The files themselves get no special treatment from AI systems. 

The content inside them might, but only if it’s actually better than what already exists on your regular pages.

SE Ranking’s data backs this up at scale. There’s no correlation between having llms.txt and getting more citations. 

The presence of the file made no measurable difference in how AI systems referenced domains.

Dig deeper: 7 hard truths about measuring AI visibility and GEO performance

What Google and AI platforms actually say

No major AI company has confirmed using llms.txt files in their crawling or citation processes.

Google’s Mueller made the sharpest critique in April 2025, comparing llms.txt to the obsolete keywords meta tag: 

  • “[As far as I know], none of the AI services have said they’re using LLMs.TXT (and you can tell when you look at your server logs that they don’t even check for it).”

Google’s Gary Illyes reinforced this at the July 2025 Search Central Deep Dive in Bangkok, explicitly stating Google “doesn’t support LLMs.txt and isn’t planning to.”

Google Search Central’s documentation is equally clear: 

  • “The best practices for SEO remain relevant for AI features in Google Search. There are no additional requirements to appear in AI Overviews or AI Mode, nor other special optimizations necessary.”

OpenAI, Anthropic, and Perplexity all maintain their own llms.txt files for their API documentation to make it easy for developers to load into AI assistants. 

But none have announced their crawlers actually read these files from other websites.

The consistent message from every major platform: standard web publishing practices drive visibility in AI search. 

No special files, no new markup, and no separate versions needed.

What this means for SEO teams

The evidence points to a single conclusion: stop building content that only machines will see.

Mueller’s question cuts to the core issue: 

  • “Why would they want to see a page that no user sees?” 

If AI companies needed special formats to generate better responses, they would tell you. As he noted:

  • “AI companies aren’t really known for being shy.” 

The data proves him right. 

Across Landwehr’s nearly 18,000 citations, LLM-optimized formats showed no advantage unless they contained unique information that didn’t exist anywhere else on the site. 

SE Ranking’s analysis of 300,000 domains found that llms.txt actually added confusion to their citation prediction model rather than improving it.

Instead of creating shadow versions of your content, focus on what actually works.

Build clean HTML that both humans and AI can parse easily. 

Reduce JavaScript dependencies for critical content, which Mueller identified as the real technical barrier: 

  • “Excluding JS, which still seems hard for many of these systems.” 

Heavy client-side rendering creates actual problems for AI parsing.

Use structured data when platforms have published official specifications, such as OpenAI’s ecommerce product feeds

Improve your information architecture so key content is discoverable and well-organized.

The best page for AI citation is the same page that works for users: well-structured, clearly written, and technically sound. 

Until AI companies publish formal requirements stating otherwise, that’s where your optimization energy belongs.

Dig deeper: GEO myths: This article may contain lies

Read more at Read More

SEO in 2026: What will stay the same

SEO in 2026 what will stay the same

Around the turn of the year, search industry media fills up with reviews and predictions. Bold, disruptive ideas steal the spotlight and trigger a sense of FOMO (fear of missing out).

However, sustainable online sales growth doesn’t come from chasing the next big trend. In SEO, what truly matters stays the same.

FOMO is bad for you 

We regularly get excited about the next big thing. Each new idea is framed as a disruptive force that will level the playing field.

Real shifts do happen, but they are rare. More often, the promised upheaval fades into a storm in a teacup.

Over the years, search has introduced many innovations that now barely raise an eyebrow. Just a few examples:

  • Voice search.
  • Universal Search.
  • Google Instant.
  • The Knowledge Graph.
  • HTTPS as a ranking signal.
  • RankBrain.
  • Mobile-first indexing.
  • AMP.
  • Featured snippets and zero-click searches.
  • E-A-T and E-E-A-T.
  • Core Web Vitals.
  • Passage indexing.
  • AI Overviews.

Some claimed these developments would revolutionize SEO or wipe it out entirely. That never happened.

The latest addition to the SEO hype cycle, LLMs and AI, fits neatly into this list. After the initial upheaval, the excitement has already started to fade.

The benefits of LLMs are clear in some areas, especially coding and software development. AI tools boost efficiency and significantly shorten production cycles.

In organic search, however, their impact remains limited, despite warnings from attention-seeking doomsayers. No AI-driven challenger has captured meaningful search market share.

Beyond ethical concerns about carbon footprint and extreme energy use, accuracy remains the biggest hurdle. Because they rely on unverified inputs, LLM-generated answers often leave users more confused than informed.

AI-driven platforms still depend on crawling the web and using core SEO signals to train models and answer queries. Like any bot, they need servers and content to be accessible and crawlable.

Without strong quality controls, low-quality inputs produce inconsistent and unreliable outputs. This is just one reason why Google’s organic search market share remains close to 90%.

It also explains why Google is likely to remain the dominant force in ecommerce search for the foreseeable future. For now, a critical mass of users will continue to rely on Google as their search engine of choice.

It’s all about data 

Fundamentally, it makes little difference whether a business focuses on Google, LLM-based alternatives, or both. All search systems depend on crawled data, and that won’t change.

Fast, reliable, and trustworthy indexing signals sit at the core of every ranking system. Instead of chasing hype, brands and businesses are better served by focusing on two core areas: their customers’ needs and the crawlability of their web platforms.

Customer needs always come first.

Most users do not care whether a provider uses the latest innovation. They care about whether expectations are met and promises are kept. That will not change.

Meeting user expectations will remain a core objective of SEO.

Crawlability is just as critical. A platform that cannot be properly crawled or indexed has no chance in competitive sectors such as retail, travel, marketplaces, news, or affiliate marketing.

Making sure bots can crawl a site, and algorithms can clearly understand the unique value of its content, will remain a key success factor in both SEO and GEO for the foreseeable future.

Won’t change: Uncrawled content won’t rank

Other factors are unlikely to change as well, including brand recognition, user trust, ease of use, and fast site performance.

These factors have always mattered and will continue to do so. They only support SEO and GEO if a platform can be properly crawled and understood. That is why regular reviews of technical signals are a critical part of a successful online operation.

Won’t change: server errors prevent indexing by any bot

At the start of a new year, you should resist the fear of missing out on the latest novelty. Following the herd rarely helps anyone stand out.

A better approach is to focus on what is certain to remain consistent in 2026 and beyond.

What to do next

Publishers can breathe a sigh of relief. There is no need to rush into a new tool just because everyone else is. Adopt it if it makes sense, but no tool alone will make a business thrive.

Focus on what you do best and make it even better. Your customers will notice and appreciate it.

At the same time, make sure your web platform is fast and reliable, that your most relevant content is regularly re-crawled, and that bots clearly understand its purpose. These are the SEO and GEO factors that will endure.

Holistic SEO is both an art and a science. While it is far more complex in 2026, it is the unchanging foundational signals that matter most.

Read more at Read More

Web Design and Development San Diego

Some Google AI Overviews now use Gemini 3 Pro

Google now uses Gemini 3 Pro to generate some AI Overviews in Google Search. Google said for more complex queries Gemini 3 Pro is used for AI Overview.

Your customers search everywhere. Make sure your brand shows up.

The SEO toolkit you know, plus the AI visibility data you need.

Start Free Trial
Get started with

Semrush One Logo

This was previously announced for AI Mode results back in November and then in December Google began using Gemini 3 Flash for AI Mode. Now, Google is taking Gemini 3 Pro to AI Overviews for complex queries.

Gemini 3 Pro is used to generate AI Overviews for complex queries in English, globally for Google AI Pro & Ultra subscribers.

What Google said. Robby Stein, VP of Product at Google Search wrote:

  • “Update: AI Overviews now tap into Gemini 3 Pro for complex topics.”
  • “Behind the scenes, Search will intelligently route your toughest Qs to our frontier model (just like we do in AI Mode) while continuing to use faster models for simpler tasks.”
  • “Live in English globally for Google AI Pro & Ultra subs.”

Why we care. The AI Overviews may be very different today than it was a week or so ago. Google will continue to improve its Gemini models and work those upgraded models into Google Search, including AI Overviews and AI Mode.

Read more at Read More

Perplexity AI User and Revenue Statistics

Founded in 2022, Perplexity offers an AI-powered search engine.

AI tools offer a new way to search for factual information, where Perplexity stands out as an AI-native search engine that combines large language models with real-time web search.

With a valuation of $20 billion and a growing user base of 30 million monthly active users, Perplexity is one of the fastest-growing tech startups challenging Google’s dominance with its AI-native search engine.

From the number of Perplexity active users to company revenue, we’ll cover the latest stats about the popular AI search engine on this page.

Key Perplexity Stats

  • Perplexity has 30 million monthly active users.
  • Perplexity processes around 600 million search queries a month.
  • Lifetime downloads of Perplexity mobile apps reached 80.5 million to date.
  • Perplexity’s annualized recurring revenue reportedly reached nearly $200 million.

Perplexity Monthly Active Users

According to the latest data, Perplexity AI has around 30 million monthly active users worldwide as of April 2025.

As of April 2025, Perplexity AI has 30 million monthly active users worldwide

That’s up from 10 million monthly active users reported in January 2024.

Here’s a table with the Perplexity AI’s monthly active user base since March 2023:

Date Perplexity AI Monthly Active Users
March 2023 2 million
January 2024 10 million
April 2025 30 million

Sources: The Verge, Perplexity AI, Perplexity AI

Perplexity Search Volume

According to Perplexity AI CEO, the search engine processes around 600 million queries per month as of April 2025. That’s an increase from 400 million reported in October 2024.

Search engine processes around 600 million queries per month as of April 2025

Here’s an overview of Perplexity AI monthly search volume over time since X:

Date Perplexity AI Monthly Search Queries
July 2024 250 million
October 2024 400 million
April 2025 600 million

Sources: The Verge, TechCrunch

Perplexity Website Traffic

According to the latest estimates, the Perplexity AI website received 239.7 million visits worldwide in November 2025, showing a 13.21% decrease compared to October 2025.

Perplexity AI website received 239.7 million visits worldwide in November 2025

Here’s a website traffic breakdown of the Perplexity AI website since September 2025:

Date Perplexity AI Website Traffic
September 2025 194.37 million
October 2025 276.5 million
November 2025 239.97 million

Source: Semrush

Perplexity App Downloads

According to recent estimates, Perplexity AI app downloads across Google Play and App Store reached an estimated lifetime downloads of 80.5 million to date, including 5.1 million in November 2025 alone.

Perplexity App Downloads

Perplexity AI had the highest number of app downloads in October 2025, with 15.5 million monthly installs worldwide.

Here’s a table with Perplexity AI app downloads over time since January 2024:

Date Perplexity AI App Downloads
January 2024 0.98 million
February 2024 0.84 million
March 2024 0.75 million
April 2024 0.63 million
May 2024 0.75 million
June 2024 0.79 million
July 2024 0.72 million
August 2024 0.8 million
September 2024 1 million
October 2024 1.27 million
November 2024 1.73 million
December 2024 1.6 million
January 2025 1.82 million
February 2025 2.88 million
March 2025 4 million
April 2025 2.94 million
May 2025 2.56 million
June 2025 2.62 million
July 2025 5.52 million
August 2025 8.84 million
September 2025 11.98 million
October 2025 15.45 million
November 2025 5.1 million

Source: Appfigures

Perplexity Revenue

Perplexity’s annual recurring revenue reportedly reached nearly $200 million as of September 2025, up from $100 million in March 2025.

Perplexity's annual recurring revenue reportedly reached nearly $200 million as of September 2025

Sources: TechCrunch, Perplexity

Perplexity Funding

Perplexity raised a total of $1.22 billion across 7 publicly disclosed funding rounds to date.

Perplexity Funding

Here’s a table with information on Perplexity AI’s latest funding rounds to date:

Date Funding Round, Amount
March 28, 2023 Series A, $28.8 million
January 4, 2024 Series B, $73.6 million
April 23, 2024 Series C, $63 million
August 9, 2024 Series C, $250 million
December 18, 2024 Series D, $500 million
July 18, 2025 Series E, $100 million
September 10, 2025 Series E, $200 million

Source: Tracxn

The post Perplexity AI User and Revenue Statistics appeared first on Backlinko.

Read more at Read More

Web Design and Development San Diego

Inside SearchGuard: How Google detects bots and what the SerpAPI lawsuit reveals

Google SearchGuard

We fully decrypted Google’s SearchGuard anti-bot system, the technology at the center of its recent lawsuit against SerpAPI.

After fully deobfuscating the JavaScript code, we now have an unprecedented look at how Google distinguishes human visitors from automated scrapers in real time.

What happened. Google filed a lawsuit on Dec. 19 against Texas-based SerpAPI LLC, alleging the company circumvented SearchGuard to scrape copyrighted content from Google Search results at a scale of “hundreds of millions” of queries daily. Rather than targeting terms-of-service violations, Google built its case on DMCA Section 1201 – the anti-circumvention provision of copyright law.

Your customers search everywhere. Make sure your brand shows up.

The SEO toolkit you know, plus the AI visibility data you need.

Start Free Trial
Get started with

Semrush One Logo

The complaint describes SearchGuard as “the product of tens of thousands of person hours and millions of dollars of investment.”

Why we care. The lawsuit reveals exactly what Google considers worth protecting – and how far it will go to defend it. For SEOs and marketers, understanding SearchGuard matters because any large-scale automated interaction with Google Search now triggers this system. If you’re using tools that scrape SERPs, this is the wall they’re hitting.

The OpenAI connection

Here’s where it gets interesting: SerpAPI isn’t just any scraping company.

OpenAI has been partially using Google search results scraped by SerpAPI to power ChatGPT’s real-time answers. SerpAPI listed OpenAI as a customer on its website as recently as May 2024, before the reference was quietly removed.

Google declined OpenAI’s direct request to access its search index in 2024. Yet ChatGPT still needed fresh search data to compete.

The solution? A third-party scraper that pillages Google’s SERPs and resells the data.

Google isn’t attacking OpenAI directly. It’s targeting a key link in the supply chain that feeds its main AI competitor.

The timing is telling. Google is striking at the infrastructure that powers rival search products — without naming them in the complaint.

What we found inside SearchGuard

We fully decrypted version 41 of the BotGuard script – the technology underlying SearchGuard. The script opens with an unexpectedly friendly message:

Anti-spam. Want to say hello? Contact botguard-contact@google.com */

Behind that greeting sits one of the most sophisticated bot detection systems ever deployed.

BotGuard vs. SearchGuard. BotGuard is Google’s proprietary anti-bot system, internally called “Web Application Attestation” (WAA). Introduced around 2013, it now protects virtually all Google services: YouTube, reCAPTCHA v3, Google Maps, and more.

In its complaint against SerpAPI, Google revealed that the system protecting Search specifically is called “SearchGuard” – presumably the internal name for BotGuard when applied to Google Search. This is the component that was deployed in January 2025, breaking nearly every SERP scraper overnight.

Unlike traditional CAPTCHAs that require clicking images of traffic lights, BotGuard operates completely invisibly. It continuously collects behavioral signals and analyzes them using statistical algorithms to distinguish humans from bots – all without the user knowing.

The code runs inside a bytecode virtual machine with 512 registers, specifically designed to resist reverse engineering.

How Google knows you’re human

The system tracks four categories of behavior in real time. Here’s what it measures:

Mouse movements

Humans don’t move cursors in straight lines. We follow natural curves with acceleration and deceleration – tiny imperfections that reveal our humanity.

Google tracks:

  • Trajectory (path shape)
  • Velocity (speed)
  • Acceleration (speed changes)
  • Jitter (micro-tremors)

A “perfect” mouse movement – linear, constant speed – is immediately suspicious. Bots typically move in precise vectors or teleport between points. Humans are messier.

Detection threshold: Mouse velocity variance below 10 flags as bot behavior. Normal human variance falls between 50-500.

Keyboard rhythm

Everyone has a unique typing signature. Google measures:

  • Inter-key intervals (time between keystrokes)
  • Key press duration (how long each key is held)
  • Error patterns
  • Pauses after punctuation

A human typically shows 80-150ms variance between keystrokes. A bot? Often less than 10ms with robotic consistency.

Detection threshold: Key press duration variance under 5ms indicates automation. Normal human typing shows 20-50ms variance.

Scroll behavior

Natural scrolling has variable velocity, direction changes, and momentum-based deceleration. Programmatic scrolling is often too smooth, too fast, or perfectly uniform.

Google measures:

  • Amplitude (how far)
  • Direction changes
  • Timing between scrolls
  • Smoothness patterns

Scrolling in fixed increments – 100px, 100px, 100px – is a red flag.

Detection threshold: Scroll delta variance under 5px suggests bot activity. Humans typically show 20-100px variance.

Timing jitter

This is the killer signal. Humans are inconsistent, and that’s exactly what makes us human.

Google uses Welford’s algorithm to calculate variance in real-time with constant memory usage – meaning it can analyze patterns without storing massive amounts of data, regardless of how many events occur. As each event arrives, the algorithm updates its running statistics.

If your action intervals have near-zero variance, you’re flagged.

The math: If timing follows a Gaussian distribution with natural variance, you’re human. If it’s uniform or deterministic, you’re a bot.

Detection threshold: Event counts exceeding 200 per second indicate automation. Normal human interaction generates 10-50 events per second.

The 100+ DOM elements Google monitors

Beyond behavior, SearchGuard fingerprints your browser environment by monitoring over 100 HTML elements. The complete list extracted from the source code includes:

  • High-priority elements (forms): BUTTON, INPUT – these receive special attention because bots often target interactive elements.
  • Structure: ARTICLE, SECTION, NAV, ASIDE, HEADER, FOOTER, MAIN, DIV
  • Text: P, PRE, BLOCKQUOTE, EM, STRONG, CODE, SPAN, and 25 others
  • Tables: TABLE, CAPTION, TBODY, THEAD, TR, TD, TH
  • Media: FIGURE, CANVAS, PICTURE
  • Interactive: DETAILS, SUMMARY, MENU, DIALOG

Environmental fingerprinting

SearchGuard also collects extensive browser and device data:

Navigator properties:

  • userAgent
  • language / languages
  • platform
  • hardwareConcurrency (CPU cores)
  • deviceMemory
  • maxTouchPoints

Screen properties:

  • width / height
  • colorDepth / pixelDepth
  • devicePixelRatio

Performance:

  • performance.now() precision
  • performance.timeOrigin
  • Timer jitter (fluctuations in timing APIs)

Visibility:

  • document.hidden
  • visibilityState
  • hasFocus()

WebDriver detection: The script specifically checks for signatures that betray automation tools:

  • navigator.webdriver (true if automated)
  • window.chrome.runtime (absent in headless mode)
  • ChromeDriver signatures ($cdc_ prefixes)
  • Puppeteer markers ($chrome_asyncScriptInfo)
  • Selenium indicators (__selenium_unwrapped)
  • PhantomJS artifacts (_phantom)

Why bypasses become obsolete in minutes

Here’s the critical discovery: SearchGuard uses a cryptographic system that can invalidate any bypass within minutes.

The script generates encrypted tokens using an ARX cipher (Addition-Rotation-XOR) – similar to Speck, a family of lightweight block ciphers released by the NSA in 2013 and optimized for software implementations on devices with limited processing power.

But there’s a twist.

The magic constant rotates. The cryptographic constant embedded in the cipher isn’t fixed. It changes with every script rotation.

Observed values from our analysis:

  • Timestamp 16:04:21: Constant = 1426
  • Timestamp 16:24:06: Constant = 3328

The script itself is served from URLs with integrity hashes: //www.google.com/js/bg/{HASH}.js. When the hash changes, the cache invalidates, and every client downloads a fresh version with new cryptographic parameters.

Even if you fully reverse-engineer the system, your implementation becomes invalid with the next update.

It’s cat and mouse by design.

The statistical algorithms

Two algorithms power SearchGuard’s behavioral analysis:

  • Welford’s algorithm calculates variance in real time with constant memory usage – meaning it processes each event as it arrives and updates a running statistical summary, without storing every past interaction. Whether the system has seen 100 or 100 million events, memory consumption stays the same.
  • Reservoir sampling maintains a random sample of 50 events per metric to estimate median behavior. This provides a representative sample without storing every interaction.

Combined, these algorithms build a statistical profile of your behavior and compare it against what humans actually do.

SerpAPI’s response

SerpAPI’s founder and CEO, Julien Khaleghy, shared this statement with Search Engine Land:

“SerpApi has not been served with Google’s complaint, and prior to filing, Google did not contact us to raise any concerns or explore a constructive resolution. For more than eight years, SerpApi has provided developers, researchers, and businesses with access to public search data. The information we provide is the same information any person can see in their browser without signing in. We believe this lawsuit is an effort to stifle competition from the innovators who rely on our services to build next-generation AI, security, browsers, productivity, and many other applications.”

The defense may face challenges. The DMCA doesn’t require content to be non-public – it prohibits circumventing technical protection measures, period. If Google proves SerpAPI deliberately bypassed SearchGuard protections, the “public data” argument may not hold.

What this means for SEO – and the bigger picture

If you’re building SEO tools that programmatically access Google Search, 2025 was brutal.

In January, Google deployed SearchGuard. Nearly every SERP scraper suddenly stopped returning results. SerpAPI had to scramble to develop workarounds – which Google now calls illegal circumvention.

Then in September, Google removed the num=100 parameter – a long-standing URL trick that allowed tools to retrieve 100 results in a single request instead of 10. Officially, Google said it was “not a formally supported feature.” But the timing was telling: forcing scrapers to make 10x more requests dramatically increased their operational costs. Some analysts suggested the move specifically targeted AI platforms like ChatGPT and Perplexity that relied on mass scraping for real-time data.

See the complete picture of your search visibility.

Track, optimize, and win in Google and AI search from one platform.

Start Free Trial
Get started with

Semrush One Logo

The combined effect: traditional scraping approaches are increasingly difficult and expensive to maintain.

For the industry: This lawsuit could reshape how courts view anti-scraping measures. If SearchGuard qualifies as a valid “technological protection measure” under DMCA, every platform could deploy similar systems with legal teeth.

Under DMCA Section 1201, statutory damages range from $200 to $2,500 per circumvention act. With hundreds of millions of alleged violations daily, the theoretical liability is astronomical – though Google’s complaint acknowledges that “SerpApi will be unable to pay.”

The message isn’t about money. It’s about setting precedent.

Meanwhile, the antitrust case rolls on. Judge Mehta ordered Google to share its index and user data with “Qualified Competitors” at marginal cost. One hand is being forced open while the other throws punches.

Google’s position: “You want our data? Go through the antitrust process and the technical committee. Not through scraping.”

Here’s the uncomfortable truth: Google technically offers publishers controls, but they’re limited. Google-Extended allows publishers to opt out of AI training for Gemini models and Vertex AI – but it doesn’t apply to Search AI features including AI Overviews.

Google’s documentation states:

“AI is built into Search and integral to how Search functions, which is why robots.txt directives for Googlebot is the control for site owners to manage access to how their sites are crawled for Search.”

Court testimony from DeepMind VP Eli Collins during the antitrust trial confirmed this separation: content opted out via Google-Extended could still be used by the Search organization for AI Overviews, because Google-Extended isn’t the control mechanism for Search.

The only way to fully opt out of AI Overviews? Block Googlebot entirely – and lose all search traffic.

Publishers face an impossible choice: accept that your content feeds Google’s AI search products, or disappear from search results altogether.

Your move, courts.

Dig deeper

This analysis is based on version 41 of the BotGuard script, extracted and deobfuscated from challenge data in January 2026. The information is provided for informational purposes only.

Read more at Read More

Web Design and Development San Diego

GEO myths: This article may contain lies

GEO myths- This article may contain lies

Less than 200 years ago, scientists were ridiculed for suggesting that hand washing might save lives.

In the 1840s, it was shown that hygiene reduced death rates, but the underlying explanation was missing.

Without a clear mechanism, adoption stalled for decades, leading to countless preventable deaths.

The joke of the past becomes the truth of today. The inverse is also true when you follow misleading guidance.

Bad GEO advice (I don’t like this acronym, but will use it because it seems to be the most popular) will not literally kill you. 

That said, it can definitely cost money, cause unemployment, and lead to economic death.

Not long ago, I wrote about a similar topic and explained why unscientific SEO research is dangerous and acts as a marketing instrument rather than real scientific discovery. 

This article is a continuation of that work and provides a framework to make sense of the myths surrounding AI search optimization.

I will highlight three concrete GEO myths, examine whether they are true, and explain what I would do if I were you.

If you’re pressed for time, here’s a TL;DR:

  • We fall for bad GEO and SEO advice because of ignorance, stupidity, cognitive biases, and black-and-white thinking.
  • To evaluate any advice, you can use the ladder of misinference – statement vs. fact vs. data vs. evidence vs. proof.
  • You become more knowledgeable if you seek dissenting viewpoints, consume with the intent to understand, pause before you believe, and rely less on AI.
  • You currently:
    • Don’t need an llms.txt.
    • Should leverage schema markup even if AI chatbots don’t use it today.
    • Have to keep your content fresh, especially if it matters for your queries.

Before we dive in, I will recap why we fall for bad advice.

Recap: Why we fall for bad GEO and SEO advice

The reasons are:

  • Ignorance, stupidity, and amathia (voluntary stupidity).
  • Cognitive biases, such as confirmation bias.
  • Black-and-white thinking.

We are ignorant because we don’t know better yet. We are stupid if we can’t know better. Both are neutral. 

We suffer from amathia when we refuse to know better, which is why it’s the worst of the three.

We all suffer from biases. When it comes to articles and research, confirmation bias is probably the most prevalent. 

We refuse to see flaws in how we see things and instead seek out flaws, often with great effort, in rival theories or remain blind to them.

Lastly, we struggle with black-and-white thinking. Everything is this or that, never something in between. A few examples:

  • Backlinks are always good.
  • Reddit is always important for AI search.
  • Blocking AI bots is always stupid.

The truth is, the world consists of many shades of gray. This idea is captured well in the book “May Contain Lies” by Alex Edmans

He says something can be moderate, granular, or marbled:

  • Backlinks are not always good or important, as they lose their impact after a certain point (moderate).
  • Reddit isn’t always important for AI search if it’s not cited at all for the relevant prompt set (granular).
  • Blocking some AI bots isn’t always stupid because, for some business models and companies, it makes perfect sense (marbled).

The first step to get better is always awareness. And we all are sometimes ignorant, (voluntarily or involuntarily) stupid, suffer from biases or think black and white.

Let’s get more practical now that we know why we fall for bad advice.

Dig deeper: Most SEO research doesn’t lie – but doesn’t tell the truth either

How I evaluate GEO (and SEO) advice and protect myself from being stupid

One way to save yourself is the ladder of misinference, once again borrowing from Edmans’ book. It looks like this:

The ladder of misinference

To accept something as proof, it needs to climb the rungs of the ladder. 

On closer inspection, many claims fail at the last rung when it comes to evidence versus proof. 

To give you an example:

  • Statement: “User signals are an important factor for better organic performance.”
  • Fact: Better CTR performance can lead to better rankings.
  • Data: You can directly measure this on your own site, and several experiments showed the impact of user signals long before it became common knowledge.
  • Evidence: There are experiments demonstrating causal effects, and a well-known portion of the 2024 Google leak focuses on evaluating user signals.
  • Proof: Court documents in Google’s DOJ monopoly trial confirmed the data and evidence, making this universally true.

Fun fact: Rand Fishkin and Marcus Tandler both said that user signals matter many years ago and were laughed at, much like scientists in the 1800s. 

At the time, the evidence wasn’t strong enough. Today, their “joke” is now the truth.

If I were you, here’s what I would do:

  • Seek dissenting viewpoints: You only truly understand something when you can argue in its favor. The best defense is steelmanning your argument. To do that, you need to fully understand the other side.
  • Consume with the intent to understand: Too often, we listen to reply, which means we don’t listen at all and instead converse with ourselves in our own heads. We focus on our own arguments and what we will say next. To understand, you need to listen actively.
  • Pause before you share and believe: False information is highly contagious, so sharing half-truths or lies is dangerous. You also shouldn’t believe something simply because a well-known person said it or because it’s repeated over and over again.
  • Don’t use AI to summarize (perhaps controversial): AI has significant flaws when it comes to summarization. For example, prompts that ask for brief summaries increase hallucinations, and source material can put a veil of credibility and trust over the response.

We will see why the last point is a big problem in a second.

The prime example: Blinding AI workslop

I decided against finger-pointing, so there is no link or mention of who this is about. With a bit of research, you might find the example yourself.

This “research” was promoted in the following way:

  • “How AI search really works.”
  • Requiring a time investment of weeks.
  • 19 studies and six case studies analyzed.
  • Validated, reviewed, and stress-tested.

To quote Edmans:

  • “It’s not for the authors to call their findings groundbreaking. That’s for the reader to judge. You need to shout about the conclusiveness of your proof or the novelty of your results. Maybe they’re not strong enough to speak for themselves. … It doesn’t matter what fancy name you give your techniques or how much data you gather. Quantity is no substitute for quality.”

Just because something took a long time does not mean the results are good. 

Just because the author or authors say so does not mean the findings are groundbreaking.

According to the HBR, AI workslop is:

  • “AI-generated work content that masquerades as good work, but lacks the substance to meaningfully advance a given task.”

I don’t have proof this work was AI-generated. It’s simply how it felt when I read it myself, with no skimming or AI summaries. 

Here are a few things that caught my attention:

  • It doesn’t deliver what it claims. It purports to explain how AI search works, but instead lists false correlations between studies that analyzed something different from what the analysis claims.
  • Reported sample sizes are inaccurate.
  • Studies and articles are mishmashed.
  • One source is a “someone said something that someone said something that someone said.”
  • Cited research didn’t analyze or conclude what is claimed in the meta-analysis.
  • The “correlation coefficient” isn’t a correlation coefficient, but a weighted score.
  • To be specific, it misdates the GEO study as 2024 instead of 2023 and claims the research “confirms” that schema markup, lists, and FAQ blocks significantly improve inclusion in AI responses. A review of the study shows that it makes no such claims.

This analysis looks convincing on the surface and masquerades as good work, but on closer inspection, it crumbles under scrutiny.

Disclaimer: I specifically wanted to highlight one example because it reflects everything I wrote about in my last article and serves as a perfect continuation. 

This “research” was shared in newsletters, news sites, and roundups. It got a lot of eyeballs.

Let’s now take a look at the three, in my opinion, most pervasive recommendations for influencing the rate of your AI citations.

Dig deeper: Forget the Great Decoupling – SEO’s Great Normalization has begun

Get the newsletter search marketers rely on.


The most common GEO myths: Claims vs. reality

‘Build an llms.txt’

The claims for why this should help:

  • AI chatbots have a centralized source of important information to use for citations.
  • It’s a lightweight file that makes it easier for AI crawlers to evaluate your domain.

When viewed through the ladder of misinference, the llms.txt claim is a statement. 

Some parts are factual – for example, Google and others crawl these files, and Google even indexes and ranks them for keywords – and there is data to support that. 

However, there is no data or evidence showing that llms.txt files boost AI inclusion. There is certainly no proof.

The reality is that llms.txt is a proposal from 2024 that gained traction largely because it was amplified by influencers. 

It was repeated often enough to become one of the more tiring talking points in black-and-white debates.

One side dismisses it entirely, while the other promotes it as a secret holy grail that will solve all AI visibility problems.

The original proposal also stated:

  • “We furthermore propose that pages on websites that have information that might be useful for LLMs to read provide a clean markdown version of those pages at the same URL as the original page, but with .md appended.”

This approach would lead to internal competition, duplicate content, and an unnecessary increase in total crawl volume. 

The only scenario where llms.txt makes sense is if you operate a complex API that AI agents can meaningfully benefit from.

(There’s a small experiment showing that neither llms.txt nor .md files have an impact on AI citations.)

So, if I were you, here’s what I would do:

  • On a quarterly basis:
    • Check whether companies such as OpenAI, Anthropic, and Google have openly announced support.
    • Review log files to see how crawl volume to llms.txt changes over time. You can do this without providing an llms.txt file.
  • If it is officially supported, create one according to published documentation guidelines.

At the moment, no one has evidence – or proof – that an llms.txt meaningfully influences your AI presence.

‘Use schema markup’

The claims for why this should help:

  • Machines love structured data.
  • Generally, the advice “make it as easy as possible” holds true.
  • Microsoft said so.”

The last point is egregious. No one has a direct quote from Fabrice Canel or the exact context in which he supposedly said this.

For this recommendation, there is no solid data or evidence.

The reality is this:

  • For training
    • Text is extracted and HTML elements are stripped.
    • Tokenization after pretraining destroys coherent code if markup makes it through to this step.
    • The existence of LLMs is based on structuring unstructured content.
    • They can handle schema and write it because they are trained to do so.
    • That doesn’t mean your individual markup plays a role in the knowledge of the foundation model.
  • For grounding
    • There is no evidence that AI chatbots access schema markup.
    • Correlation studies show that websites with schema markup have better AI visibility, but there are many rival theories that could explain this.
    • Recent experiments (including this and this) showed the opposite. The tools AI chatbots can access don’t use the HTML.
    • I recently tested this in Perplexity Comet. Even with an open DOM, it hallucinated schema markup on the page that didn’t match what was actually there.

Also, when someone says they use structured data, that can – but does not have to – mean schema. 

All schema is structured data, but not all structured data is schema. In most cases, they mean proper HTML elements such as tables and lists. 

So, if I were you, here’s what I would do:

  • Use schema markup for supported rich results.
  • Use all relevant properties in your schema markup.

You might ask why I recommend this. To me, solid schema markup is a hygiene factor of good SEO. 

Just because AI chatbots and agents don’t use schema today doesn’t mean they won’t in the future.

“One could say the same for llms.txt.” That’s true. However, llms.txt has no SEO benefits.

Schema markup doesn’t help us improve how AI systems process our content directly.

Instead, it helps improve signals they frequently look at, such as search rankings, both in the top 10 and beyond for fan-out queries.

‘Provide fresh content’

The claims for why this should help:

  • AI chatbots prefer fresh content.
  • Fresh content is important for some queries and prompts.
  • Newer or recently updated content should be more accurate.

Compared with llms.txt and schema markup, this recommendation stands on a much more solid foundation in terms of evidence and data.

The reality is that foundation models contain content up to the end of 2022. 

After digesting that information, they need fresh content, which means cited sources, on average, have to be more recent.

If freshness is relevant to a query – OpenAI, Anthropic, and Perplexity use freshness as a signal to determine whether to use web search – then finding fresh sources matters.

There is research supporting this hypothesis from Ahrefs, Generative Pulse, and Seer Interactive

More recently, a scientific paper also supported these claims.

A few words of caution about that paper:

  • The researchers used API results, not the user interface. Results differ because of chatbot system prompts and API settings. Surfer recently published a study showing how large those differences can be.
  • Asking a model to rerank is not how the model or chatbot actually reranks results in the background.
  • The way dates were injected was highly artificial, with a perfect inverse correlation that may exaggerate the results.

That said, this recommendation appears to have the strongest case for meaningfully influencing AI visibility and increasing citations.

So, if I were you, here’s what I would do:

  • Add a relevant date indicating when your content was last updated.
  • Keep update dates consistent:
    • On-page.
    • Schema markup.
    • Sitemap lastmod.
  • Update content regularly, especially for queries where freshness matters. Fan-out queries from AI chatbots often signal freshness when a date is included.
  • Never artificially update content by changing only the date. Google stores up to 20 past versions of a web page and can detect manipulation.

In other words, this one appears to be legitimate.

Dig deeper: The rise of ‘like hat’ SEO: When attention replaces outcomes

Escaping the vortex of AI search misinformation

We have to avoid shoveling AI search misinformation into the walls of our industry. 

Otherwise, it will become the asbestos we eventually have to dig out.

An attention-grabbing headline should always raise red flags. 

I understand the allure of believing what appears to be the consensus or using AI to summarize. It’s easier. We’re all busy.

The issue is that there was already too much content to consume before AI. Now there’s even more because of it. 

We can’t consume and analyze everything, so we rely on the same tools not only to generate content, but also to consume it.

It’s a snake-biting-its-own-tail problem. 

Our compression culture risks creating a vortex of AI search misinformation that feeds back into the training data of the AI chatbots we both love and hate. 

We’re already there. AI chatbots sometimes answer GEO questions from model knowledge.

Take the time to think for yourself and get your hands dirty. 

Try to understand why something should or shouldn’t work. 

And never take anything at face value, no matter who said it. Authority isn’t accuracy.

P.S. This article may contain lies.

Read more at Read More

AI Search Watcher Looker Studio Connector – Integration Guide

Google Looker Studio (formerly Data Studio) is a free, browser-based reporting tool that helps you turn raw data into clear dashboards and shareable reports. For marketers and agencies, it’s…

The post AI Search Watcher Looker Studio Connector – Integration Guide appeared first on Mangools.

Read more at Read More

Choosing the right WordPress SEO plugin for your business – Yoast vs Rank Math 

Selecting an SEO plugin for your WordPress site is one of the most important decisions you’ll make for your online presence. It’s not just about installing software; it’s about choosing a long-term partner that will grow with your business, adapt to changing search algorithms, and support you in the age of AI. While the market offers several options, understanding what truly matters is key. Two of the most popular plugins in the market today are Yoast and Rank Math. Therefore, factors such as reliability, innovation, ecosystem, and trust help you make a choice that will serve your business for years to come. 

This guide provides an in-depth comparison of the key differentiating factors between Yoast and Rank Math. We will understand why millions of websites worldwide have made Yoast their trusted comrade in the search business. 

Key takeaways

  • Choosing an SEO plugin like Yoast SEO impacts your online presence and future growth.
  • Yoast offers reliability with over 15 years of experience and millions of active installations, unlike newer competitors.
  • Innovations such as AI integration and a unified schema graph set Yoast apart from other plugins.
  • Yoast provides comprehensive support, education, and a multi-platform ecosystem tailored for long-term success.
  • Trust industry leaders like Microsoft and Spotify who use Yoast SEO to enhance their online visibility.

What really matters when choosing an SEO plugin

When evaluating WordPress SEO plugins, it’s easy to get distracted by feature lists and flashy interfaces. But experienced marketers, agencies, and business owners know that the best tools are defined by much more than what they promise on paper. 

The questions that matter most: 

  • Can you trust this plugin to work reliably as your business scales? 
  • Will the company behind it still be innovating five years from now? 
  • What happens when you need help before a critical deadline? 
  • Does the plugin anticipate future SEO trends, or just react to them? 
  • Is this a tool you install, or an ecosystem that supports your growth and development? 

These aren’t trivial questions. Your SEO plugin touches essential pages on your site, influences the content you publish, and directly impacts your ability to be found by potential customers.  
Choosing poorly can lead to migration headaches, compatibility issues, and lost rankings. Choosing wisely means peace of mind, ongoing innovation, and a solid foundation to build upon. 

Why legacy and proven trust matter in SEO plugins

Trust isn’t given. It’s earned. Yoast has defined the WordPress SEO landscape for over 15 years, with more than 13 million active installations and over 850 million downloads. This extensive legacy reflects a consistent track record of innovation, stability, and trust. Brands such as The Guardian, Microsoft, Spotify, and others rely on Yoast SEO as a foundation for their SEO strategies. This depth of experience is invaluable as SEO requires ongoing adaptation to algorithm changes and new technologies. 

While Rank Math is an ambitious and feature-rich plugin with a growing user base, its presence in the market is relatively recent. For businesses seeking a proven solution with a long-standing heritage, Yoast’s established positioning offers confidence that the plugin will continue to evolve and provide reliable support for years to come. 

Innovation that shapes the industry

Yoast has always been at the forefront of defining what modern SEO looks like. This isn’t a reactive development; it’s proactive innovation that anticipates where search is heading. Both plugins invest in innovation, but Yoast’s leadership in integrating AI and collaboration with Google sets it apart. 

AI and Automation 

We have introduced an industry-first AI-powered optimization toolset, including: 

  • AI Generate: Creates multiple optimized title and meta description variations instantly, giving you professionally crafted options in seconds instead of struggling for the perfect phrasing.
  • AI Optimize: Scans your content and provides precise, actionable suggestions to improve keyphrase placement, sentence structure, and readability, teaching you SEO best practices while you write. 
  • AI Summarize: Instantly generates bullet-point summaries of your content, making it more scannable and engaging for readers who skim before diving deep. 
  • AI Brand Insights: This is where Yoast truly separates from the pack. As AI platforms like ChatGPT reshape how people find information, AI Brand Insights, included in the Yoast SEO AI+ package, tracks how your brand appears in AI-generated responses. You can monitor your AI visibility, compare it against competitors, and ensure AI platforms accurately represent your business. 

While Rank Math includes helpful automation features such as AI keyword suggestions, Yoast’s AI integration is more comprehensive and positioned as a core pillar of modern SEO strategy. 

Schema markup that search engines can understand

While many plugins output disconnected structured data, Yoast SEO automatically generates a unified semantic graph on every page, linking your organization, content, authors, and products through a single JSON-LD structure that search engines and AI platforms can interpret consistently. 

What makes this different 

Automatic and invisible: 
Yoast outputs rich structured data representing your content, business, and relationships without requiring technical configuration. You focus on creating content; Yoast handles the complexity of structured data behind the scenes. 

Single unified graph format: 
Instead of fragmented schema markup, Yoast creates one cohesive graph structure per page, connecting all entities with unique IDs. When plugins output conflicting schema, search engines can’t reliably interpret your site. Yoast’s unified graph ensures consistent interpretation at scale, whether Google, ChatGPT, or any API is reading your content. 

Minimal configuration: 
Choose whether your site represents a person or organization; Yoast handles the rest automatically. Specialized blocks like FAQ and How-To map directly to correct schema types and link into the graph without additional setup. 

Why this matters for AI-driven search 

As AI platforms increasingly rely on structured data to understand websites, Yoast’s approach of creating a full semantic model of your site positions you for how search and discovery are evolving. The framework scales reliably from 100 to 100,000 pages while maintaining valid entity relationships. For developers, Yoast’s Schema API provides clean filters to extend or customize the graph without breaking its integrity. 

Rank Math and other plugins support Schema markup, but Yoast’s unified graph framework represents a fundamentally different approach: automatic generation, consistent entity relationships, and architecture built for scale. 

Continuous algorithm adaptation

Search engines make thousands of updates every year. Google alone rolls out over 5,000 algorithm changes annually. Now, as search engines evolve to incorporate AI tooling and platforms like ChatGPT reshape the way people discover information, the SEO landscape is changing faster than ever.  

Most website owners can’t possibly track these shifts across traditional search AND emerging AI platforms, let alone understand their implications. Yoast’s dedicated SEO team monitors every significant update, from Google algorithm changes to how AI platforms index and reference content, and proactively adjusts the plugin to ensure your site stays optimized for both traditional and AI-driven discovery.  

When you use Yoast, you’re not just getting software. You’re getting a team of experts working behind the scenes to keep your SEO strategy current across the entire discovery ecosystem. 

An ecosystem built to support your SEO workflow

Yoast offers an ecosystem beyond the plugin. While Yoast SEO itself is a plugin, Yoast provides a comprehensive ecosystem to support your growth: 

  • 24/7 real human expert support available for Yoast SEO Premium users. It ensures that you get fast, knowledgeable help when you need it. 
  • Yoast SEO Academy offers comprehensive SEO education, covering a range of topics from basics to advanced, with accompanying certifications. 
  • A massive knowledge base and community for continuous learning and troubleshooting. 
     

Multi-Platform Support 

Your business doesn’t exist on WordPress alone. That’s why Yoast extends beyond a single platform: 

  • Yoast SEO for Shopify: Brings Yoast’s trusted optimization to Shopify stores, helping ecommerce businesses improve product visibility and drive more sales. 
  • Yoast WooCommerce SEO: Specifically designed for WooCommerce stores with automated product schema, smart breadcrumbs, and ecommerce-focused content analysis. 

This ecosystem approach means Yoast grows with your business, supporting you across platforms as your needs evolve. Rank Math primarily focuses on the WordPress environment with a strong feature set, but lacks the same breadth of educational resources and multi-platform reach. 

Stability and reliability at enterprise-grade scale

Flashy features attract attention. Rock-solid reliability keeps businesses running. Yoast rigorously tests every update for compatibility and performance across different WordPress versions and server configurations. This commitment ensures: 

  • Backward compatibility: Updates maintain existing functionality without requiring extensive reconfiguration 
  • WordPress core integration: Seamless compatibility with new WordPress releases 
  • Performance at any scale: Optimized for sites ranging from personal blogs to high-traffic enterprise installations 

With over 15 years in the market and more than 13 million active installations, Yoast has proven its reliability across millions of sites, hosting environments, and various use cases. 

Rigorous testing and quality assurance 

Yoast maintains strict development standards that prioritize stability above rapid feature deployment. Every update undergoes extensive testing across the latest WordPress versions, most PHP configurations, and common plugin combinations before release. 

This disciplined approach means Yoast users rarely experience plugin conflicts, broken updates, or compatibility issues that plague WordPress sites using less mature plugins. 

Backward compatibility 

Major updates usually shake the functionality of plugins and software. However, Yoast maintains backward compatibility, ensuring that updating your plugin doesn’t suddenly break critical SEO features or require extensive reconfiguration. 

WordPress core compatibility 

As a plugin deeply integrated with WordPress development, Yoast maintains close relationships with the WordPress core team. This ensures seamless compatibility with new WordPress releases, often supporting new versions on launch day while other plugins scramble to catch up. 

Performance optimized for scale 

Whether you run a small blog or an enterprise site with millions of pages, Yoast performs efficiently without slowing down your site. The plugin is engineered for performance, using best practices for database queries, resource loading, and caching integration. 

Enterprises trust Yoast precisely because it scales reliably. Small teams appreciate that the same plugin powering major corporations works flawlessly on their modest sites, too. 

Ready to make a difference with Yoast SEO Premium?

Explore Yoast SEO Premium and the Yoast SEO AI+ package to discover advanced tools built for serious marketers.

Get Yoast SEO Premium Only $118.80 / year (ex VAT)

Where Yoast takes the lead

While comprehensive feature-by-feature comparisons can be overwhelming, certain capabilities distinguish truly professional SEO plugins from the rest. Here’s where Yoast’s innovation and depth shine through. 

AI-powered optimization 

Yoast leads the industry in AI integration for SEO optimization: 

  • AI-generated titles and meta descriptions 
  • Real-time content optimization suggestions 
  • An instant content summarization plugin 
  • AI Brand Insights for tracking your presence in AI search platforms 

No competing plugin offers this comprehensive AI integration designed specifically for modern SEO workflows. 

Schema Graph 

Yoast’s Schema implementation creates a complete structured data graph connecting your organization, content, authors, and brand identity. This goes far beyond basic Schema markup, providing search engines with rich context that improves your chances of appearing in knowledge panels, rich results, and AI-generated answers. 

Smart internal linking 

Yoast SEO Premium includes intelligent internal linking suggestions that analyze your content and recommend relevant pages to link to. This isn’t just a list of posts; it’s context-aware suggestions that strengthen your site architecture and improve crawlability. 

Advanced redirect manager 

Managing redirects is critical when restructuring sites, changing URLs, or handling broken links. Yoast’s redirect manager offers: 

  • Automatic redirects when you change a post URL 
  • Bulk CSV import/export for large-scale migrations 
  • REGEX support for complex redirect patterns 
  • Full redirect history and management 

WooCommerce-specific optimization 

If you run an online store, Yoast WooCommerce SEO provides: 

  • Automated product schema markup (price, availability, reviews) 
  • Smart breadcrumbs for product categories 
  • Ecommerce-focused content analysis 
  • Duplicate content prevention for product variations 

Comprehensive crawl settings 

Advanced users appreciate Yoast’s granular control over crawl optimization, robots.txt management, and indexation settings, giving technical SEO professionals the precision they need without overwhelming casual users. 

Bot blocker for LLM training control 

As AI companies scrape the web to train large language models, Yoast gives you control over whether your content is used for AI training via Bot Blocker. This cutting-edge feature addresses a concern most plugins haven’t even acknowledged yet. 

Recognized and trusted by industry leaders 

The company you keep says a lot about who you are. When the world’s most recognized brands trust Yoast to power their WordPress SEO, it’s a powerful testament to the quality, reliability, and effectiveness of our solutions. 

Global brands* using Yoast include: 

  • The Guardian 
  • Microsoft 
  • Spotify 
  • Rolling Stones 
  • Taylor Swift 
  • Facebook 
  • eBay 

These organizations have teams of developers, SEO experts, and decision-makers who have evaluated every available option. They chose Yoast, not because it was the newest, but because it was the best. 

*Disclaimer: Based on third party data sources.

Industry Recognition: 

  • Global Search Awards Finalist: Recognized among the world’s leading SEO solutions 
  • Women’s Choice Awards Winner: Acknowledged for excellence and customer satisfaction 

Yoast isn’t just popular, it’s the default choice for WordPress SEO professionals worldwide. 

Understanding what you really need

Before making your final decision, consider what matters most for your specific situation: 

If you value reliability and stability: Choose a plugin with a proven track record of consistent updates, compatibility, and performance. Longevity matters because it signals the company will be around to support you for years to come. 

If innovation matters to your strategy: Look for a plugin that anticipates SEO trends rather than reacting to them. AI integration, Schema excellence, and algorithm adaptation separate forward-thinking tools from those playing catch-up. 

If support is critical: Consider whether you need community forums or access to real SEO experts who can troubleshoot complex issues quickly. When your business relies on organic traffic, response time is crucial. 

If education is important: Some plugins provide features; others teach you how to use them effectively. Comprehensive training resources and certifications demonstrate a commitment to your success. 

If you’re building for the long term: Think about whether this plugin will grow with your business. Multi-platform support, scalability, and an ecosystem approach ensure that your investment pays dividends for years to come. 

Make the choice that drives real growth

Choosing an SEO plugin isn’t about finding the tool with the longest feature list; it’s about finding the one that best suits your needs. It’s about partnering with a company that shares your commitment to long-term growth, innovation, and excellence. 

Over 13 million websites trust Yoast SEO because it delivers on these promises: 

  • Reliability: 15+ years of consistent innovation and stability 
  • Trust: Used by global brands and industry leaders 
  • Innovation: Leading the industry in AI integration and Schema excellence 
  • Support: 24/7 access to real SEO professionals 
  • Education: Comprehensive training through Yoast Academy 
  • Ecosystem: Multi-platform support and continuous learning resources 
  • Stability: Enterprise-grade performance at any scale 

When you choose Yoast, you’re not just installing a plugin; you’re joining millions of websites that have made the strategic decision to partner with the most trusted name in WordPress SEO. 

A smarter analysis in Yoast SEO Premium

Yoast SEO Premium has a smart content analysis that helps you take your content to the next level!

Get Yoast SEO Premium Only $118.80 / year (ex VAT)

The post Choosing the right WordPress SEO plugin for your business – Yoast vs Rank Math  appeared first on Yoast.

Read more at Read More

How brands can respond to misleading Google AI Overviews

Misleading -Google AI Overview

Google’s AI Overviews feature has become the face of our search engine results.

Type almost any question into your Google search bar, and the first answer you receive will be AI generated.

Many are thrilled about this. Others are wary.

Marketers and those in the online reputation management (ORM) field are among those urging caution.

Why? Because Google AI Overviews are often littered with information stemming from online forums like Reddit and Quora. 

And oftentimes, this user-generated content can be inaccurate — or entirely false. 

Why Google AI Overviews heavily rely on content from Reddit and Quora

But how and why have Google AI Overviews come to rely on user-generated content forums?

The answer is quite simple. Google AI Overviews sources much of its information from “high-authority” domains. These happen to be platforms like Reddit and Quora.

Google also prioritizes “conversational content” and “real user experiences.” They want searchers to receive answers firsthand from other online humans.

Furthermore, Google places the same amount of weight on these firsthand anecdotes as it does on factual reporting. 

How negative threads end up on AI summaries

Obviously, the emphasis placed on Reddit and Quora threads can lead to issues. Especially for professionals and those leading product- or service-driven organizations.

Many of the Reddit threads that rise to the surface are those that are complaint-driven. Think of threads where users are asking, “Does Brand X actually suck?” or “Is Brand Z actually a scam?”

The main problem is that these threads become extremely popular. AI Overviews gather the consensus of many comments and combine them into a single resounding answer. 

In essence, minority opinions end up being represented as fact.

Additionally, Google AI Overviews often resurface old threads that lack timestamps. This can lead to the resurfacing of outdated, often inaccurate information. 

Patterns that SEO, ORM, and brands are noticing

Those in the ORM field have been noticing troubling patterns in Google AI Overviews for a while now. For instance, we’ve identified the following trends:

  • Overwhelming Reddit criticism: Criticism on Reddit rises to the top at alarming rates. Google AI Overviews even seem to ignore official responses from brands at times, instead opting for the opinions of users on forum platforms.
  • Pros vs. cons summaries: These sorts of lists are supposed to implore balance. (Isn’t that the entire point of identifying both the pros and cons of a brand?) However, sites like Reddit and Quora tend to accentuate the negative aspects of brands, at times ignoring the pros altogether. 
  • Outdated content resurfacing: As mentioned in the previous section, outdated content can hold far too much value. Aa troubling amount of “resolved issues” gain prevalence in the Google AI Overviews feature.

The amplification effect: AI can turn opinion into fact

We live in an era defined by instantaneous knowledge.

Gen Z takes in information at startling rates. What’s seen on TikTok is absorbed as immediate fact. Instagram is where many turn to get both breaking news and updates on the latest brands

This has led to an amplification effect, where algorithms quickly turn opinion into fact. We’re seeing it widely across social media, and now on Google AI Overviews, too.

On top of what we listed in the previous section, those in the ORM realm are noticing the following take effect:

  • Nuance-less summarization: Because AI Overviews take such overwhelming negative criticism from Reddit, we’re getting less nuanced responses. The focus in AI Overviews is often one-sided and seemingly biased, featuring emotional, extreme language. 
  • Feedback loops: As others in the ORM field have pointed out, many citations in Overview come from deep pages. It’s also common to see feedback loops wherein one negative Reddit thread can hold multiple citations, leading to quick AI validation.
  • Enhanced trust in AI Overviews: Perhaps most troubling of all has been society’s immediate jump to accept AI Overviews and all the answers it has to offer. Many users now turn to Google’s feature as their ultimate encyopledia — without even caring to view the citations AI Overviews has listed. 

Misinformation and bias create risk

All in all, the rise of information from Reddit and Quora on AI Overviews has led to enhanced risk for businesses and entrepreneurs alike.

False statements and defamatory claims posted online can be accepted as fact. And incomplete narratives or opinion-based criticism floating around on forums are filtered through the lens of AI Overviews.

Making matters worse is that Google does not automatically remove or filter AI summaries that are linked to harmful content. 

This can be damaging to a company’s reputation, as users absorb what they see on AI Overviews at face value. They take it as fact, even though it might be fiction.

Building a reputation strategy for false AI-driven searches

As a business owner, it’s critical to have response strategies in place for Google AI Overviews. 

Working with an ORM team is a critical first step. They might suggest the following measures:

  • Monitoring online forums: Yes, our modern world dictates that you stay on top of online forums like Reddit and Quora. Monitor the name of your business and the top players on your team. If you’re aware of the dialogue, you’re already one step ahead.
  • Creating “AI-readable” content: It’s also important to always be creating content designed to land on AI Overviews. This content should boost your platform on search engines, be citation-worthy, and push down less favorable results.
  • Addressing known criticism: Ever notice criticism directed at your brand? Seek to address it with proper business practices. Respond to online reviews kindly, suppress or remove negative content with your ORM team, and establish your business as a caring practice online.
  • Coordinating various teams: It’s imperative to establish the right teams around your business. We already mentioned ORM, but what about your legal, SEO, and PR teams? Have the right experts in place to deal with any controversies before they arise.

Also, remember to keep an eye on the future. Online reputation management is constantly evolving, and if your intention is to manage and elevate your brand, you must evolve with the times.

That means staying up-to-date with AI literacy and adapting to new KPIs, including sentiment framing, source attribution, and AI visibility. 

Staying on top of Google AI Overviews

We live in a new age. One where AI Overviews dictates much of what searches think and react to.

And the honest truth is that much of the knowledge AI Overviews gleans comes from user-dominated forums like Reddit and Quora.

As a brand manager, you can no longer be idle. You have to act. You have to manage the sources that Google AI Overviews summarizes, constantly staying one step ahead.

If you don’t, then you’re not properly managing your search reputation. 

Read more at Read More

What 107,000 pages reveal about Core Web Vitals and AI search

Core Web Vitals AI visibility

As AI-led search becomes a real driver of discovery, an old assumption is back with new urgency. If AI systems infer quality from user experience, and Core Web Vitals (CWV) are Google’s most visible proxy for experience, then strong CWV performance should correlate with strong AI visibility.

The logic makes sense.

Faster page load times result in smoother page load times, increased user engagement, improved signals, and AI systems that reward the outcome (supposedly)

But logic is not evidence.

To test this properly, I analysed 107,352 webpages that appear prominently in Google AI Overviews and AI Mode, examining the distribution of Core Web Vitals at the page level and comparing them against patterns of performance in AI-driven search and answer systems. 

The aim was not to confirm whether performance “matters”, but to understand how it matters, where it matters, and whether it meaningfully differentiates in an AI context.

What emerged was not a simple yes or no, but a more nuanced conclusion that challenges prevailing assumptions about how many teams currently prioritise technical optimisation in the AI era.

Why distributions matter more than scores

Most Core Web Vitals reporting is built around thresholds and averages. Pages pass or fail. Sites are summarized with mean scores. Dashboards reduce thousands of URLs into a single number.

The first step in this analysis was to step away from that framing entirely.

When Largest Contentful Paint was visualized as a distribution, the pattern was immediately clear. The dataset exhibited a heavy right skew. 

Median LCP values clustered in a broadly acceptable range, while a long tail of extreme outliers extended far beyond it. A relatively small proportion of pages were horrendously slow, but they exerted a disproportionate influence on the average.

Cumulative Layout Shift showed a similar issue. The majority of pages recorded near-zero CLS, while a small minority exhibited severe instability. 

Again, the mean suggested a site-wide problem that did not reflect the lived reality of most pages.

This matters because AI systems do not reason over averages, if they reason on user engagement metrics at all. 

They evaluate individual documents, templates, and passages of content. A site-wide CWV score is an abstraction created for reporting convenience, not a signal consumed by an AI model.

Before correlation can even be discussed, one thing becomes clear. Core Web Vitals are not a single signal, they are a distribution of behaviors across a mixed population of pages.

Correlations

Because the data was uneven and not normally distributed, a standard Pearson correlation was not suitable. Instead, I used a Spearman rank correlation, which assesses whether higher-ranking pages on one measure also tend to rank higher or lower on another, without assuming a linear relationship.

This matters because, if Core Web Vitals were closely linked to AI performance, pages that perform better on CWV would also tend to perform better in AI visibility, even if the link was weak.

I found a small negative relationship. It was present, but limited. For Largest Contentful Paint, the correlation ranged from -0.12 to -0.18, depending on how AI visibility was measured. For Cumulative Layout Shift, it was weaker again, typically between -0.05 and -0.09.

These relationships are visible when you look at large volumes of data, but they are not strong in practical terms. Crucially, they do not suggest that faster or more stable pages are consistently more visible in AI systems. Instead, they point to a more subtle pattern.

The absence of upside, and the presence of downside

The data do not support the claim that improving Core Web Vitals beyond basic thresholds improves AI performance. Pages with good CWV scores did not reliably outperform their peers in AI inclusion, citation, or retrieval.

However, the negative correlation is instructive.

Pages sitting in the extreme tail of CWV performance, particularly for LCP, were far less likely to perform well in AI contexts. 

These pages tended to exhibit lower engagement, higher abandonment, and weaker behavioral reinforcement signals. Those second-order effects are precisely the kinds of signals AI systems rely on, directly or indirectly, when learning what to trust.

This reveals the true shape of the relationship.

Core Web Vitals do not act as a growth lever for AI visibility. They act as a constraint.

Good performance does not create an advantage. Severe failure creates disadvantage.

This distinction is easy to miss if you examine only pass rates or averages. It becomes apparent when examining distributions and rank-based relationships.

Why ‘passing CWV’ is not a differentiator

One reason the positive correlation many expect does not appear is simple. Passing Core Web Vitals is no longer rare.

In this dataset, the majority of pages already met recommended thresholds, especially for CLS. When most of the population clears a bar, clearing it does not distinguish you. It merely keeps you in contention.

AI systems are not selecting between pages because one loads in 1.8 seconds and another in 2.3 seconds. They are selecting between pages because one explains a concept clearly, aligns with established sources, and satisfies the user’s intent, whereas the other does not.

Core Web Vitals ensure that the experience does not actively undermine those qualities. They do not substitute for them.

Reframing the role of Core Web Vitals in AI strategy

The implication is not that Core Web Vitals are unimportant. It is that their role has been misunderstood.

In an AI-led search environment, Core Web Vitals function as a risk-management tool, not acompetitive strategy. They prevent pages from falling out of contention due to poor experience signals.

This reframing has practical consequences for developing an AI visibility strategy.

Chasing incremental CWV gains across already acceptable pages is unlikely to deliver returns in AI visibility. It consumes engineering effort without changing the underlying selection logic AI systems apply.

Targeting the extreme tail, however, does matter. Pages with really bad performance generate negative behavioral signals that can suppress trust, reduce reuse, and weaken downstream learning signals.

The objective is not to make everything perfect. It is to ensure that the content you want AI systems to rely on is not compromised by avoidable technical failure.

Why this matters

As AI systems increasingly mediate discovery, brands are seeking controllable levers. Core Web Vitals feel attractive because they are measurable, familiar, and actionable.

The risk is mistaking measurability for impact.

This analysis suggests a more disciplined approach. Treat Core Web Vitals as table stakes. Eliminate extreme failures. 

Protect your most important content from technical debt. Then shift focus back to the factors AI systems actually use to infer value, such as clarity, consistency, intent alignment, and behavioral validation.

Core Web Vitals: A gatekeeper, not a differentiator

Based on an analysis of 107,352 AI visible webpages, the relationship between Core Web Vitals and AI performance is real, but limited.

There is no strong positive correlation. Improving CWV beyond baseline thresholds does not reliably improve AI visibility.

However, a measurable negative relationship exists at the extremes. Severe performance failures are associated with poorer AI outcomes, mediated through user behavior and engagement.

Core Web Vitals are therefore best understood as a gate, not a signal of excellence.

In an AI-led search landscape, this clarity matters.

Read more at Read More