general self-reported-attribution

self-reported-attribution

This skill should be used when the user asks to "set up self-reported attribution", "add a how-did-you-hear field", "design a self-reported attribution form", "track where leads come from with surveys", "build a how did you find us question", "implement self-reported attribution in HubSpot", "set up HDYHAU", "design an attribution survey", "add attribution questions to forms", or any variation of implementing self-reported attribution for B2B SaaS.
Download .md

Self-Reported Attribution

Self-reported attribution asks the prospect directly: "How did you hear about us?" It captures dark funnel sources that software attribution misses. Podcast mentions, word of mouth, community conversations, social posts seen but not clicked. Software tracks clicks. Self-reported attribution tracks awareness.

The principle: software attribution and self-reported attribution answer different questions. Software tells you which click path led to the form fill. Self-reported tells you what actually made them aware of you. Use both. Neither is complete alone.

Why Self-Reported Attribution Matters

What software misses

Source Software tracks it? Self-reported catches it?
Google paid click → form fill Yes (UTM parameters) Sometimes (prospect may not remember the ad)
Podcast mention → direct visit → form fill No (shows as "direct" traffic) Yes ("I heard you on [podcast]")
Colleague recommendation → form fill No (shows as "direct") Yes ("A colleague recommended you")
LinkedIn post → later search → form fill No (shows as "organic search") Yes ("I saw your post on LinkedIn")
Community mention (Slack, Discord) → form fill No (shows as "direct") Yes ("Someone mentioned you in [community]")
Conference conversation → form fill No (shows as "direct") Yes ("I met your team at [event]")

Attribution gap

Software attribution typically captures 40-60% of true sources. The remaining 40-60% shows as "direct" or "organic." Self-reported attribution fills that gap, especially for dark funnel sources that drive 50%+ of B2B pipeline.


The Question Design

Primary question

"How did you first hear about us?"

Not "how did you find us" (implies the form fill path, not awareness). Not "where did you learn about us" (implies education, not discovery). "How did you first hear about us" asks about initial awareness.

Answer options

Option What it captures Notes
Google search Organic discovery Don't split paid vs organic here. Software handles that distinction
LinkedIn (post or ad) LinkedIn content or advertising Combine post and ad. The prospect rarely distinguishes
Podcast Podcast guest appearances or ads Follow up with "Which podcast?" as a text field
Colleague or friend Word of mouth The highest-value attribution source. Track volume carefully
Community (Slack, Discord, forum) Community mentions Follow up with "Which community?"
Conference or event In-person Follow up with "Which event?"
Blog or content Content marketing Software usually captures this too. Cross-reference
Review site (G2, Capterra) Review-driven Software captures G2 clicks. Self-reported catches G2 browsing
Other (please specify) Everything else Always include. Free-text catches sources you didn't anticipate

Question design rules

  • Single-select, not multi-select. You want the FIRST source, not every source. Multi-select produces data that's impossible to analyze cleanly
  • 8-12 options maximum. More than 12 and prospects skip the question. Fewer than 8 and you miss important sources
  • Include "Other" with a text field. Every quarter, review "Other" responses. If a new source appears frequently, add it as its own option
  • "Colleague or friend" is always an option. Word of mouth is typically 20-40% of B2B pipeline sources. If you don't include it, you can't measure it
  • No "N/A" or "Prefer not to say." These are escape hatches that reduce data quality. If the field is optional, a prospect who doesn't want to answer will skip it. If required, they'll pick something

Where to Place the Question

Placement options

Placement Pros Cons Recommended?
Demo request form Highest-intent moment. Data tied to pipeline Adds friction to the highest-value form Yes (if kept to 1 question)
Post-form thank-you page No form friction. Asked after conversion Lower completion rate (30-50% vs 80%+ on-form) Yes (as fallback)
Sign-up flow (PLG) Captures product-led sources Friction in sign-up reduces conversion Yes (only if sign-up is low-friction)
Email survey post-meeting No form friction. Can ask detailed follow-ups Low response rate (10-20%). Memory bias No (too late, too low response)
SDR asks on discovery call Can probe for detail ("which podcast?") Inconsistent. SDRs forget. Hard to standardize Supplementary only

Placement rules

  • Put it on the demo request form. This is the primary placement. One additional dropdown field adds minimal friction. The data is tied directly to the highest-intent action
  • Make it required on the demo form. Optional fields get 40-60% fill rates. Required gets 95%+. The data is too valuable to leave optional. One dropdown takes 3 seconds
  • If form friction is a concern, use the thank-you page. Show the question immediately after form submission. "One quick question before your confirmation: how did you first hear about us?" Completion rate drops to 30-50% but zero form friction
  • Never rely solely on SDR-collected attribution. SDRs forget to ask. They interpret answers differently. They skip it when the call is going well. Use form-based collection as the primary source. SDR-collected is supplementary detail

Implementation

HubSpot implementation

1. Create a contact property:
   - Name: "How did you hear about us"
   - Field type: Dropdown select
   - Options: [your 8-12 options]
   - Group: Contact information

2. Add to demo request form:
   - Position: last field before submit button
   - Required: Yes
   - Label: "How did you first hear about us?"

3. Create a second property for detail:
   - Name: "Attribution detail"
   - Field type: Single-line text
   - Conditional: shows only when "Podcast", "Community",
     "Conference", or "Other" is selected
   - Label: "Which one?" or "Please specify"

4. Create a report:
   - X-axis: "How did you hear about us"
   - Y-axis: Count of contacts
   - Filter: created in last 90 days
   - Drill-down: by lifecycle stage (MQL, SQL, Won)

Salesforce implementation

1. Create a picklist field on Lead and Contact:
   - API name: How_Did_You_Hear__c
   - Values: [your 8-12 options]

2. Create a text field for detail:
   - API name: Attribution_Detail__c
   - Length: 255

3. Add both fields to the web-to-lead form

4. Map lead field to contact field on lead conversion

5. Create a report:
   - Report type: Leads with converted info
   - Group by: How_Did_You_Hear__c
   - Show: conversion rate, pipeline value per source

Analyzing Self-Reported Data

Primary analysis

Analysis How to calculate What it tells you
Source distribution % of form fills per source Where awareness is coming from
Source-to-pipeline Pipeline $ per source Which sources produce the most pipeline
Source-to-closed-won Revenue per source Which sources produce the most revenue
Source quality Win rate per source Which sources produce the best prospects
Source trend Month-over-month change per source Which sources are growing or declining

Cross-referencing with software attribution

Self-reported says Software says What's happening
Podcast Direct Prospect heard podcast, typed URL directly. Software missed the source. Self-reported is correct
Google search Google organic Both agree. High confidence in this attribution
Colleague LinkedIn ad Prospect saw the ad AND got a recommendation. Self-reported captures the higher-intent source
LinkedIn post Direct Prospect saw the post, later visited directly. Self-reported captures the awareness source
Other: "Twitter" Organic social Both agree on social, but self-reported is more specific about the platform

Analysis rules

  • Report self-reported and software attribution separately. They answer different questions. Don't merge them into one field. Present both to leadership
  • Self-reported is more accurate for dark funnel. Podcasts, word of mouth, communities. Trust self-reported over software for these sources
  • Software is more accurate for click-path. Paid search, paid social, email campaigns. Trust software for these. Prospects don't always remember clicking an ad
  • Review "Other" responses quarterly. New sources emerge. If "TikTok" or "YouTube" starts appearing in Other, add it as a dedicated option
  • Track at the pipeline level, not just the lead level. "30% of leads say podcast" is interesting. "Podcast leads convert to pipeline at 2x the rate of Google leads" is actionable

Common Pitfalls

Data quality issues

Problem Cause Fix
60% select "Google search" Options are too generic. Prospects default to the most familiar option Add more specific options. Split "Google search" into "Searched for a solution" and "Searched for us by name"
"Other" is 25% of responses Missing common sources from the dropdown Review "Other" text responses. Add any source that appears 5%+ of the time
Responses don't match software data at all Prospect confuses "how did you find us today" with "how did you first hear about us" Clarify the question: "How did you FIRST hear about us?" Emphasis on first awareness
SDR-collected data conflicts with form data SDR probed deeper and got a different answer Both can be valid. Store separately. The form captures initial awareness. The SDR captures the nuanced story

Measurement

Metric Definition Target Frequency
Fill rate % of form submissions that include the attribution field > 90% (if required) Weekly
Source coverage % of responses that use a named option (not "Other") > 85% Monthly
Dark funnel capture % of responses that are dark funnel sources (podcast, WOM, community) Track trend, no fixed target Monthly
Source-to-pipeline correlation Does each source produce proportional pipeline? Track by source Quarterly
Cross-reference alignment % of responses where self-reported and software agree 40-60% (disagreement is expected for dark funnel) Quarterly

Pre-Implementation Checklist

  • [ ] Question wording is "How did you first hear about us?" (not "find us" or "learn about us")
  • [ ] 8-12 answer options covering all major channels including dark funnel
  • [ ] "Colleague or friend" (word of mouth) is included
  • [ ] "Other (please specify)" with text field is included
  • [ ] Single-select, not multi-select
  • [ ] Placed on the demo request form (or highest-intent conversion point)
  • [ ] Field is required
  • [ ] Follow-up text field for podcast name, community name, event name
  • [ ] CRM property created and mapped
  • [ ] Report built showing source distribution and source-to-pipeline
  • [ ] Plan to review "Other" responses quarterly and add new options

Anti-Pattern Check

  • Making the field optional. Fill rate drops to 40-60%. You lose half your attribution data. Make it required. One dropdown takes 3 seconds. The friction is negligible
  • Using multi-select instead of single-select. Prospect selects "Google search" and "LinkedIn" and "Colleague." Now you have attribution soup. Single-select forces them to identify the FIRST source
  • Only tracking at the lead level. "30% of leads say podcast." So what? Track through to pipeline and closed-won. "Podcast leads generate $2M pipeline vs $800K for Google leads" is actionable
  • No "Other" option. You have 8 sources in the dropdown. A prospect heard about you on a niche industry podcast that doesn't fit any category. They pick "Google search" as a default. You lose the real attribution. Always include "Other" with a text field
  • Never updating options. Your dropdown has been the same for 18 months. Meanwhile, you started a YouTube channel, launched a community, and appeared on 12 podcasts. Update options quarterly based on "Other" responses and new channel investments
  • Replacing software attribution with self-reported. They answer different questions. Software tracks click paths. Self-reported tracks awareness. Use both. Report them separately. Let leadership see both views
  • Asking on a post-meeting email survey. 10-15% response rate. The prospect has already forgotten where they first heard about you. Memory bias is worst at this stage. Ask at the moment of highest intent: the form fill
Want agents that use skill files like this?
We customize skill files for your brand voice and methodology, then run content agents against them.
Book a call