All posts
GTM

The Promptability Score: How to Grade Whether AI Agents Can Actually Recommend Your Product

By Beatriz8 min read

Abstract neural network visualization

PMM Mindset · April 2026


We already score products for usability, conversion, and retention.

Now we need another score: promptability.

What Is Promptability?

Promptability is how reliably an AI agent can evaluate and recommend your product for a specific use case when asked by a buyer.

Not how good your marketing sounds. Not your conversion rate. Not your NPS.

Your recommendability score in an agent-mediated buying flow.

If agents are becoming part of the shortlisting layer — and they are — then weak promptability creates silent pipeline loss. Buyers never reach you because their agent filtered you out, not because you weren't good enough.

The same way SEO blind spots cost you organic traffic, promptability blind spots will cost you AI-referred traffic.

The Promptability Score (0-100)

Score each category from 0 to 20. Higher = better for agent evaluation.

| Dimension | What It Measures | Low Score Signal | |-----------|------------------|------------------| | Clarity | Can an agent identify what you are, who you're for, and where you're weak? | Vague messaging, generic claims, unclear ICP language | | Verifiability | Can key claims be validated from credible sources? | Unsupported claims, outdated docs, no proof assets | | Comparability | Can an agent compare you against alternatives fairly? | Defensive comparison content, no category framing | | Accessibility | Can agent systems parse your content efficiently? | Fragmented pages, inconsistent taxonomy, buried details | | Freshness | Is your information current and trustworthy? | Stale pricing, outdated docs, mismatched narrative |


Dimension 1: Clarity (0-20)

Clarity measures whether an agent can quickly extract the four essential pieces of product identity:

  1. -->What the product is — One-sentence description that a human and a model can both parse
  2. -->Who it is for — ICP language that maps to job title, use case, and company stage
  3. -->What problems it solves — Specific outcomes, not aspirational claims
  4. -->Where it is weak — Honest constraints that prevent a bad fit

Why This Matters for Agents

LLMs are pattern matchers. When a buyer asks their agent "find me a code review tool for a 10-person startup on a budget," the agent needs to match your product to that query.

If your homepage says "We help developers ship better code faster" without specifics, the agent has nothing to match against. Generic language gets filtered out.

How to Score Your Clarity

| Score | Description | |-------|-------------| | 15-20 | ICP, use cases, and constraints are explicit on key pages. An agent can answer "is this for X?" accurately. | | 8-14 | General sense of who you serve, but ICP language is implied, not explicit. Some ambiguity in agent matching. | | 0-7 | Generic messaging, no specific ICP language, aspirational claims without constraints. High risk of misclassification. |

How to Improve

  • -->Audit your homepage, pricing page, and comparison pages for explicit ICP statements
  • -->Add "best for" and "not for" language to product descriptions
  • -->Replace "we help teams ship faster" with "for teams that need automated code review before merging"

Dimension 2: Verifiability (0-20)

Verifiability measures whether an agent can validate your claims from credible sources.

Key questions:

  • -->Can the agent find concrete metrics on your pricing page?
  • -->Are case studies linked from high-intent pages, or buried in a separate section?
  • -->Do you have changelog or release notes that prove you're actively maintained?
  • -->Are there third-party validations (security certifications, analyst coverage, user reviews)?

The Trust Problem

Agents are trained to weight credible sources. A claim from your homepage carries less weight than a metric from a case study, which carries less weight than a third-party review.

If your proof assets are disconnected from your claims, agents can't connect the dots.

How to Score Your Verifiability

| Score | Description | |-------|-------------| | 15-20 | Every major claim links to a verifiable source. Case studies include specific outcomes. Pricing is transparent with plan boundaries. | | 8-14 | Some proof exists but not consistently linked to claims. Verification requires agent to hunt across pages. | | 0-7 | Marketing claims with no backing. No case studies, no metrics, no third-party validation. |

How to Improve

  • -->Audit claims on your top 5 pages: can each claim be validated by clicking a link?
  • -->Consolidate proof assets into one canonical page, linked from everywhere
  • -->Add specific metrics to case studies: "reduced deployment time by 40%" not "improved efficiency"

Dimension 3: Comparability (0-20)

Comparability measures whether an agent can compare you against alternatives fairly.

This is uncomfortable for most marketing teams. Comparability implies acknowledging that alternatives exist — and being honest about where you win and where you lose.

But agents are built to evaluate multiple options. If you don't provide comparison data, the agent will generate comparisons from whatever it can scrape — often your competitors' marketing materials.

What Comparability Looks Like

  • -->Explicit differentiators stated in your own words
  • -->Use-case-based comparison pages (not feature matrices)
  • -->"When to choose us vs. alternatives" guidance
  • -->Migration scenarios and constraints documented

How to Score Your Comparability

| Score | Description | |-------|-------------| | 15-20 | You provide the comparison framework. Agent can answer "when should someone choose X over Y?" accurately. | | 8-14 | Some comparison content exists but it's defensive or incomplete. Agent has to infer the rest. | | 0-7 | No comparison content. Agent has to generate comparisons from external sources — often inaccurate. |

How to Improve

  • -->Publish honest "when to choose us" pages by use case
  • -->Add "not the right fit if..." language to your positioning
  • -->Create comparison content that acknowledges alternatives' strengths — it builds trust and gives agents accurate data

Dimension 4: Accessibility (0-20)

Accessibility measures whether agent systems can parse your content efficiently.

This is the technical layer of promptability.

Agents don't read the way humans do. They process text. And their ability to extract accurate information depends on:

  • -->Structured headings and information hierarchy
  • -->Machine-readable product details (schema, clean HTML)
  • -->Clean information architecture without orphaned pages
  • -->Consistent taxonomy across pages

The Fragmentation Problem

Many companies have good content — it's just buried across dozens of pages with inconsistent naming. An agent searching for "what's included in the enterprise plan" might pull from outdated pricing pages or feature lists that contradict each other.

How to Score Your Accessibility

| Score | Description | |-------|-------------| | 15-20 | Information architecture is clean. Agent can find any key detail without hunting. Taxonomy is consistent. | | 8-14 | Most information exists but architecture is complex. Agent may need multiple queries to assemble a full picture. | | 0-7 | Fragmented content, inconsistent naming, orphaned pages. High risk of agent pulling contradictory or outdated information. |

How to Improve

  • -->Audit your information architecture: can an agent find every key detail in 3 clicks or fewer?
  • -->Standardize taxonomy: pick one name for each feature and use it everywhere
  • -->Audit for schema markup on key pages
  • -->Run a "what's the agent's picture of us?" test by querying your own product through an AI search

Dimension 5: Freshness (0-20)

Freshness measures whether your information is current and trustworthy.

This is the most overlooked dimension because most teams treat content as "done" once published.

But in an agent-mediated flow, stale information is worse than no information. A buyer who acts on outdated pricing or features gets a bad experience — and blames you, not the agent.

What Freshness Includes

  • -->Regular content updates (not "last updated" dates from 2023)
  • -->Release notes and changelogs linked from key pages
  • -->Timestamped proof content (case studies with dates, not generic testimonials)
  • -->Consistent narrative across all channels — not a product page that says X and a LinkedIn post that implies Y

How to Score Your Freshness

| Score | Description | |-------|-------------| | 15-20 | All key pages show recent updates. Changelog is linked from product pages. Case studies are dated and specific. | | 8-14 | Core pages are current but secondary content is stale. Some channels have outdated messaging. | | 0-7 | No evidence of updates. Pricing, features, or messaging contradictions across channels. |

How to Improve

  • -->Set a quarterly "promptability audit" cadence — not just content audit
  • -->Link release notes from your homepage and pricing page
  • -->Date your case studies and proof content
  • -->Audit for cross-channel consistency at least quarterly

Scoring Bands

| Score | Band | Implication | |-------|------|-------------| | 80-100 | High promptability | Strong recommendation readiness. Your product can be evaluated accurately by agents. | | 60-79 | Moderate promptability | Visible but inconsistent. Vulnerable in head-to-head comparisons. | | Below 60 | Low promptability | Likely underrepresented or misrepresented in agent outputs. Silent pipeline risk. |

Use this as a decision tool, not a vanity score. A 60 that identifies fixable gaps is more valuable than an 85 with blind spots.


How to Run a Promptability Audit

Cadence

  • -->Baseline: Once per quarter
  • -->Spot-check: Monthly for priority use cases or new product launches
  • -->Trigger: After any major product, pricing, or positioning change

Participants

Don't run this as a solo audit. Promptability is cross-functional by nature:

  • -->PMM (positioning and messaging)
  • -->Content lead (clarity, freshness)
  • -->Docs lead (accessibility)
  • -->Product representative (accuracy of claims)

Output

  1. -->Scorecard: Scores for each dimension with supporting evidence
  2. -->Top 5 gaps: Prioritized list of what to fix first
  3. -->Owners and due dates: Assigned accountability

Example Remediation Plan

If Comparability is low:

  • -->Publish honest comparison pages by use case
  • -->Add "when we are not the right fit" guidance
  • -->Include migration scenarios and constraints

If Verifiability is low:

  • -->Refresh case studies with concrete metrics
  • -->Consolidate proof assets into one canonical page
  • -->Link proof from high-intent pages

If Accessibility is low:

  • -->Audit and flatten information architecture
  • -->Standardize feature naming across all pages
  • -->Add schema markup to key product pages

The Bottom Line

Promptability is not a copywriting project.

It is an operational score for how well your product can be evaluated in agent-mediated buying journeys.

Teams that measure it early will fix representation gaps before they become pipeline problems. Teams that ignore it will discover — too late — that their product is being filtered out by agents they didn't know were evaluating it.

The question isn't whether agents will be part of the buying process.

It's whether you'll be ready when they are.


Next: The Skillability Score: Writing Prompts That AI Can Actually Use