The analysis on this page is derived from AI platform responses, providing a detailed view of brand visibility and positioning

What’s the best way to analyze which prompts include my brand?

Prompt

Last updated 15th September 2025

Domain count
writesonic.com
18
medium.com
16
authoritas.com
15
Citations count
30
0
20
10
Brands Discovered
29
15
14
0

Original answer

Short answer: use a hybrid pipeline — start with a curated list of brand variants + deterministic checks (exact/regex, domain/handle matching), add fuzzy matching for misspellings, and add a semantic layer (embeddings or a classifier) to catch paraphrases and implied mentions. Monitor performance with a labeled sample and iterate with human-in-the-loop.

Detailed, practical plan

  1. Build your brand vocabulary
  • Canonical names, legal forms, abbreviations, common misspellings, punctuation/diacritics variants.
  • Domains, social handles, hashtags, common product names, acronyms.
  • Example for "Acme Co.": ["acme", "acme co", "acme.co", "acme-corp", "@acme", "#acme"].
  1. Preprocess text
  • Normalize unicode, lowercase, remove zero-width chars, optionally strip punctuation except in handles/URLs.
  • Extract URLs, @handles, and hashtags separately (they’re high-precision signals).
  1. Fast deterministic checks (high precision)
  • Exact token match with word-boundary regex (catches canonical forms).
  • URL/handle/domain match (domain acme.co or @acme is definitive).
  • Use word-boundary regex to avoid partial matches (e.g., "\bacme\b" so "macme" doesn’t match).

Example (Python): import re pattern = re.compile(r'\b(acme|acme\s+co|acme.co|acmecorp)\b', re.I) bool(pattern.search(text))

  1. Fuzzy matching (catch typos)
  • Use RapidFuzz or Levenshtein-based scoring for short strings.
  • Compare tokens/ngrams from the prompt against aliases. Tune threshold (typical start: 80–90 for short names).
  • Be conservative (higher threshold) for short names — low thresholds produce false positives.
  1. Semantic matching / paraphrase detection (catch implied mentions)
  • Use sentence embeddings (e.g., sentence-transformers) and cosine similarity between prompt embeddings and one or more "brand description" embeddings (a short canonical description, product names, slogans).
  • For scale, index embeddings with FAISS, Milvus, Pinecone.
  • Typical cosine thresholds: 0.6–0.8 (must be tuned on your labeled data).

Example (sentence-transformers): from sentence_transformers import SentenceTransformer, util model = SentenceTransformer('all-MiniLM-L6-v2') brand_vec = model.encode("Acme Company - maker of widgets", convert_to_tensor=True) prompt_vec = model.encode(prompt, convert_to_tensor=True) sim = util.cos_sim(prompt_vec, brand_vec) match = sim >= 0.65 # tune threshold

  1. Supervised classifier (if you have labeled examples)
  • Train a binary text classifier (transformer fine-tune or classical TF-IDF + logistic regression) to label prompts as "mentions brand" / "doesn't".
  • Good when you need to capture tone, context, or disambiguate homonyms (e.g., “apple” the fruit vs Apple Inc.).
  • Evaluate with precision/recall/F1 and use calibration/thresholding to meet business goals (e.g., favor precision if false positives are costly).
  1. Disambiguation & context
  • Use NER and dependency parsing (spaCy) when a token could be multiple things.
  • If brand name matches a common word, require co-occurring signals (product name, domain, context words) to confirm.
  1. Human-in-the-loop + active learning
  • Sample false positives/negatives regularly and add them to training/alias lists.
  • Use an active learning loop: surface low-confidence detections for human review to improve classifier and alias list.
  1. Scale & infrastructure
  • Small volumes: regex + RapidFuzz on incoming text.
  • Medium/Large: precompute embeddings, index with FAISS/Milvus/Pinecone for fast nearest-neighbor queries, run real-time inference on new prompts.
  • Add streaming ingestion, metrics, and autoscaling as needed.
  1. Quality metrics and thresholds
  • Create a labeled test set and measure precision, recall, F1. Track these over time.
  • Choose threshold based on whether you care more about false positives (precision) or false negatives (recall).
  • Monitor rate of “unknown / low-confidence” detections as a signal for updating aliases or retraining.
  1. Privacy, logging, and compliance
  • Avoid storing PII unnecessarily. Anonymize or hash content if required.
  • Follow retention and consent rules for user-submitted prompts.

Quick recommended starter stack

  • Extraction & rules: Python re, URL parsing
  • Fuzzy matching: RapidFuzz
  • NER/Parsing: spaCy
  • Semantic detection: sentence-transformers (all-MiniLM-*), embeddings + FAISS or Pinecone for scale
  • Classifier: scikit-learn for TF-IDF models; Hugging Face transformers for fine-tuning
  • Monitoring: Kibana/Looker/Grafana, small labeled-review app for human validation

Practical rollout (minimal effort → full solution)

  1. Implement exact/regex + handle/domain matching + alias list (fast, high precision).
  2. Add RapidFuzz for common misspellings.
  3. If you still miss cases or need semantic coverage, add embeddings + similarity search.
  4. If you have labels or need contextual accuracy, train a classifier and run human-in-the-loop for low-confidence items.

If you want, I can:

  • Draft the initial alias/regex set given your brand name.
  • Provide ready-to-run Python snippets tailored to your tech stack.
  • Outline an evaluation dataset sampling plan and thresholds to start with.

Which would you like next?

Join BrandRadar to track your LLM score

Stay ahead of the curve. Get insights, benchmarks, and personalized alerts on your brand's LLM performance.

Rank
Brand
Topic
LLM
Sentiment
1
🥇

Acme Co.

30%
90%
0%
0%
Neutral
2
🥈

Apple

28%
85%
0%
0%
Neutral
3
🥉

Acme Corp

28%
0%
0%
85%
Neutral
4

Google Alerts

27%
0%
0%
80%
Neutral
5

Python

25%
75%
0%
0%
Neutral
6

SEMrush

25%
0%
0%
75%
Neutral
7

rapidfuzz

23%
70%
0%
0%
Neutral
8

Meltwater

23%
0%
0%
70%
Neutral
9

spaCy

22%
65%
0%
0%
Neutral
10

Mention

22%
0%
0%
65%
Neutral
11

SentenceTransformers

20%
60%
0%
0%
Neutral
12

Brandwatch

20%
0%
0%
60%
Neutral
13

all-MiniLM-L6-v2

18%
55%
0%
0%
Neutral
14

Mentionlytics

18%
0%
0%
55%
Neutral
15

FAISS

17%
50%
0%
0%
Neutral
16

BuzzSumo

17%
0%
0%
50%
Neutral
17

Milvus

15%
45%
0%
0%
Neutral
18

Ahrefs

15%
0%
0%
45%
Neutral
19

Pinecone

13%
40%
0%
0%
Neutral
20

Evertune

13%
0%
0%
40%
Neutral
21

Hugging Face

12%
35%
0%
0%
Neutral
22

scikit-learn

12%
35%
0%
0%
Neutral
23

Kibana

12%
35%
0%
0%
Neutral
24

Looker

12%
35%
0%
0%
Neutral
25

Grafana

12%
35%
0%
0%
Neutral
26

BrandLight

12%
0%
0%
35%
Neutral
27

ChatGPT

12%
0%
0%
35%
Neutral
28

Gemini

12%
0%
0%
35%
Neutral
29

Perplexity

12%
0%
0%
35%
Neutral
Domain
Title
LLM
URL
ibm.com
Gemini
insight7.io
Gemini
rellify.com
Gemini
analyticsvidhya.com
Gemini
quora.com
Gemini
harmonize.ai
Gemini
johnsnowlabs.com
Gemini
codefixer.com
Gemini
medium.com
Gemini
shopify.com
Gemini
mdpi.com
Gemini
askdaman.com
Gemini
mention.com
Gemini
mentionlytics.com
Gemini
bloggerjet.com
Gemini
explodingtopics.com
Gemini
meltwater.com
Gemini
authoritas.com
Gemini
aivisibility.io
Gemini
seoboost.com
Gemini
dejan.ai
Perplexity
mediashower.com
Perplexity
prowly.com
Perplexity
writesonic.com
Perplexity
promptmonitor.io
Perplexity
promptdrive.ai
Perplexity
irenechan.co
Perplexity
everworker.ai
Perplexity
mentionlytics.com
Perplexity
triplewhale.com
Perplexity
Logo© 2025 BrandRadar. All Rights Reserved.
BrandRadar.org - What’s the best way to analyze which prompts include my brand?