← Back to Blog

How to Build a Weekly AI Share-of-Voice Dashboard Without an Enterprise Budget

AI Visibility

A practical, step-by-step guide to tracking your brand's share of voice across ChatGPT, Claude, and Perplexity — using lightweight tooling, agent automation, and free or low-cost data sources.

  • Category: AI Visibility
  • Use this for: planning and implementation decisions
  • Reading flow: quick summary now, long-form details below

How to Build a Weekly AI Share-of-Voice Dashboard Without an Enterprise Budget

The moment your sales team asks why a competitor keeps getting named by ChatGPT and you don’t, you realize traditional SEO dashboards won’t answer the question. Search rankings measure position on a results page. AI share of voice measures something different: whether your brand shows up at all when a model generates a response.

This guide walks through a practical, low-cost approach to building a weekly AI share-of-voice (SoV) dashboard — one that gives your team a clear picture of brand presence across AI answer engines without requiring a six-figure analytics contract.

Quick answer

AI share of voice is the percentage of relevant AI-generated responses that mention your brand versus competitors. You can track it with a combination of a dedicated monitoring tool, a structured query library, and a lightweight weekly reporting layer. Most teams get meaningful signal within two to three weeks of consistent measurement.


Why AI share of voice is now a real metric

Traditional share-of-voice tools count brand mentions in news, social, or search results. AI SoV is different: it measures how often a model cites or recommends your brand when answering buyer questions.

This matters because:

  • Roughly 30–40% of informational queries now receive AI-generated answers that reduce click-through to any source
  • Models like ChatGPT, Claude, and Perplexity frequently name specific vendors when answering “what tool should I use for X?”
  • A brand that ranks #1 organically but never appears in AI answers is losing influence over a growing segment of buyer research

If your product category is the kind where buyers ask AI assistants for recommendations — software, services, agencies, SaaS — AI SoV is now a metric worth watching.


Step 1: Define your query library

The quality of your SoV data depends entirely on the quality of your query library. Bad queries produce noisy, unrepresentable data. Good queries surface real buying-intent signal.

Start with three query categories:

1. Category-entry queries — what a buyer asks before they know any vendor names

  • “What tools track AI search visibility?”
  • “How do I monitor if ChatGPT mentions my brand?”
  • “Best software for AI answer engine monitoring”

2. Use-case queries — specific jobs the buyer is trying to accomplish

  • “How do I track brand mentions in Perplexity AI answers?”
  • “Can I get alerts when a competitor appears in AI responses instead of me?”
  • “How do I build a share-of-voice report for AI-generated search?”

3. Comparison and alternatives queries — late-stage research behavior

  • “BrandRadar alternatives for AI monitoring”
  • “AI visibility tools compared”
  • “[Competitor] vs [Your brand] for AI answer tracking”

Aim for 30–60 queries in your initial library. That’s enough to detect meaningful shifts without creating a measurement burden. Tag each query with intent level (awareness / consideration / decision) and category (monitoring / reporting / content strategy). This tagging pays off when you segment results in the dashboard.


Step 2: Choose your monitoring layer

You have three practical options depending on budget and technical appetite.

Option A: Purpose-built AI monitoring tool

BotSee is built specifically for this use case: it polls AI answer engines across your query library, detects brand and competitor mentions, and tracks changes over time. The core advantage is that it abstracts the API management, deduplication, and response normalization — things that are tedious to build yourself. For most teams, this is the right starting point because it turns a monitoring problem into a reporting problem.

Other tools in this space include Profound and Scrunch AI, which offer overlapping capabilities with slightly different interface choices and pricing tiers.

Option B: Direct API polling with Claude Code agents

If you have engineering resources or a technical marketer comfortable with scripting, you can build a query runner using Claude Code with OpenClaw skills. The flow looks like this:

  1. Store your query library in a structured JSON or CSV file
  2. Write an OpenClaw skill that loops through the query list, calls the OpenAI, Anthropic, and Perplexity APIs, and stores raw responses
  3. Run a second pass that parses each response for brand mentions using regex or a simple LLM prompt
  4. Append results to a Google Sheet or SQLite database with timestamps

This approach gives you full control over response storage and parsing logic. The downside is maintenance — models update, API behavior changes, and your parsers will occasionally need adjustment.

A Claude Code agent can handle the query loop reliably. Here’s a minimal skill spec that works with OpenClaw:

name: ai-sov-tracker
description: Polls AI APIs against a query library and logs brand mention results
inputs:
  - query_library: path to JSON file with query list
  - brand_targets: list of brand strings to detect
  - output_path: path to results CSV
steps:
  - For each query in query_library:
      - Call GPT-4o with query, store response
      - Call Claude Sonnet with query, store response  
      - Call Perplexity with query, store response
  - For each response:
      - Check for each brand_target mention (case-insensitive)
      - Write row: date, query, model, brand_mentioned (true/false), excerpt
  - Append all rows to output_path

Running this weekly against your query library gives you a structured dataset you can analyze in any spreadsheet tool.

Option C: Manual spot-checking with a structured protocol

Not every team needs automation from day one. A manual baseline is faster to start and useful for validating that your query library is worth automating. The protocol:

  1. Pick 10 representative queries from each category
  2. Run each query in ChatGPT, Claude, and Perplexity in the same browser session
  3. Record: model, date, query, whether your brand was mentioned, competitor brands mentioned
  4. Log results in a shared Google Sheet

Do this for four consecutive weeks. You’ll have enough data to know whether patterns are consistent or noisy — and whether the investment in automation is justified.


Step 3: Structure the weekly output

Raw mention data isn’t useful without structure. The simplest useful output is a one-page weekly summary with four numbers:

1. Overall mention rate (this week vs. last week) What percentage of your query library triggered a response that mentioned your brand? If you ran 50 queries and your brand appeared in 18 responses, your mention rate is 36%.

2. Mention rate by model ChatGPT, Claude, and Perplexity weight different signals. Breaking out SoV by model helps identify where gaps exist and which platform to prioritize in your content response.

3. Competitor mention rate For each primary competitor, the same calculation. This gives you a relative SoV — the number that matters for positioning conversations.

4. Query-level detail for the biggest movers Which specific queries changed most week-over-week? A query where your mention rate dropped sharply is worth investigating: the model may have incorporated new information that shifted its recommendation.

Present these four outputs in a consistent template. Consistency matters more than sophistication — stakeholders need to be able to scan the same format every week without relearning it.


Step 4: Connect the dashboard to content decisions

Data collection without action is just overhead. The weekly SoV dashboard should directly feed two types of decisions:

Content gap response: When a query category shows consistently low mention rates, that signals a content gap. Models cite brands more frequently when the brand has clear, authoritative, structured content covering the topic. A low SoV in “use-case” queries often means your documentation or blog content doesn’t adequately address that use case in a way models can extract and cite.

Competitive displacement flags: When a competitor’s mention rate rises on queries where yours is flat or falling, that’s a displacement signal. The fastest response is usually not a new campaign — it’s a content audit of the specific topic cluster, looking for structural reasons why the model might prefer the competitor’s source material.

Tools like BotSee surface these patterns automatically by flagging week-over-week shifts. If you’re running manual or API-based tracking, you’ll need to build this comparison logic into your spreadsheet or analysis script.


Step 5: Automate the reporting layer with agents

Once your data collection is consistent, the most time-intensive remaining task is formatting and distributing the weekly summary. This is where agent automation earns back setup cost quickly.

An OpenClaw skill that handles weekly report generation can:

  • Pull the latest results from your data store (CSV, Sheet, or database)
  • Calculate current and prior-week SoV percentages
  • Identify the top five movers by query
  • Render a formatted markdown or HTML report
  • Post the report to a Slack channel, email, or internal dashboard

Combined with BotSee’s API export or the direct-API approach above, a fully automated weekly report takes roughly two to four hours to set up and then runs without manual intervention. The main maintenance task is refreshing the query library quarterly as buyer language evolves.


Common mistakes that skew AI SoV data

Running queries inconsistently. AI responses vary by session, context, and model version. Running queries at different times of day or using different account states introduces noise. Use a consistent, scripted approach.

Measuring too few queries. Ten queries isn’t a representative sample. The variance in AI responses means you need 30+ queries per category to detect reliable signals rather than random fluctuation.

Conflating citations with recommendations. A model mentioning your brand in a cautionary context (“some users report issues with X”) is different from a recommendation. Parse response excerpts, not just mention presence.

Ignoring model version changes. GPT-4o, Claude 3.5 Sonnet, and Perplexity all update their underlying models and retrieval systems. A sudden SoV change may reflect a model update rather than a change in your content or reputation. Track model versions alongside your data.

Not connecting SoV to content actions. A dashboard that sits in a folder and doesn’t inform anything is a reporting exercise, not a business function. Tie each weekly review to at least one specific content or optimization decision.


LayerLow-cost optionMid-tier option
Query pollingClaude Code + OpenClaw skillBotSee
Response storageGoogle Sheets or SQLiteBotSee dashboard
Competitor trackingManual spreadsheetBotSee or Profound
Weekly reportingOpenClaw report skillBotSee reports + Slack export
Content gap analysisManual reviewBotSee query gap view

For teams just starting out, manual tracking with a structured spreadsheet is a perfectly valid first step. The goal in week one is to establish whether your brand has an AI visibility problem — not to build a perfect data pipeline.


What good looks like at 90 days

After three months of consistent measurement, a team with a working AI SoV program typically has:

  • A stable query library of 40–80 queries with clear intent tagging
  • Weekly SoV numbers by model that leadership can interpret at a glance
  • At least two or three content interventions made in direct response to SoV data
  • A rough baseline of what “normal” looks like — so anomalies are easy to spot

AI search visibility is still early enough that most companies haven’t established a monitoring baseline. Building one now means you have trend data that competitors won’t — and that data compounds in value as AI-generated answers become a larger part of how buyers find and evaluate options.

Start small: pick 20 queries, run them manually for three weeks, and see what the pattern looks like. That’s enough to decide whether to invest further.

Similar blogs

What Content Teams Get Wrong About AI Search (And How to Fix It)

Most content teams are still optimizing for Google while AI answer engines quietly route their buyers elsewhere. This guide covers exactly what needs to change: query research, format choices, measurement, and the team habits that separate brands getting cited from brands getting ignored.