Self-Reported Attribution
Self-reported attribution asks the prospect directly: "How did you hear about us?" It captures dark funnel sources that software attribution misses. Podcast mentions, word of mouth, community conversations, social posts seen but not clicked. Software tracks clicks. Self-reported attribution tracks awareness.
The principle: software attribution and self-reported attribution answer different questions. Software tells you which click path led to the form fill. Self-reported tells you what actually made them aware of you. Use both. Neither is complete alone.
Why Self-Reported Attribution Matters
What software misses
| Source |
Software tracks it? |
Self-reported catches it? |
| Google paid click → form fill |
Yes (UTM parameters) |
Sometimes (prospect may not remember the ad) |
| Podcast mention → direct visit → form fill |
No (shows as "direct" traffic) |
Yes ("I heard you on [podcast]") |
| Colleague recommendation → form fill |
No (shows as "direct") |
Yes ("A colleague recommended you") |
| LinkedIn post → later search → form fill |
No (shows as "organic search") |
Yes ("I saw your post on LinkedIn") |
| Community mention (Slack, Discord) → form fill |
No (shows as "direct") |
Yes ("Someone mentioned you in [community]") |
| Conference conversation → form fill |
No (shows as "direct") |
Yes ("I met your team at [event]") |
Attribution gap
Software attribution typically captures 40-60% of true sources. The remaining 40-60% shows as "direct" or "organic." Self-reported attribution fills that gap, especially for dark funnel sources that drive 50%+ of B2B pipeline.
The Question Design
Primary question
"How did you first hear about us?"
Not "how did you find us" (implies the form fill path, not awareness). Not "where did you learn about us" (implies education, not discovery). "How did you first hear about us" asks about initial awareness.
Answer options
| Option |
What it captures |
Notes |
| Google search |
Organic discovery |
Don't split paid vs organic here. Software handles that distinction |
| LinkedIn (post or ad) |
LinkedIn content or advertising |
Combine post and ad. The prospect rarely distinguishes |
| Podcast |
Podcast guest appearances or ads |
Follow up with "Which podcast?" as a text field |
| Colleague or friend |
Word of mouth |
The highest-value attribution source. Track volume carefully |
| Community (Slack, Discord, forum) |
Community mentions |
Follow up with "Which community?" |
| Conference or event |
In-person |
Follow up with "Which event?" |
| Blog or content |
Content marketing |
Software usually captures this too. Cross-reference |
| Review site (G2, Capterra) |
Review-driven |
Software captures G2 clicks. Self-reported catches G2 browsing |
| Other (please specify) |
Everything else |
Always include. Free-text catches sources you didn't anticipate |
Question design rules
- Single-select, not multi-select. You want the FIRST source, not every source. Multi-select produces data that's impossible to analyze cleanly
- 8-12 options maximum. More than 12 and prospects skip the question. Fewer than 8 and you miss important sources
- Include "Other" with a text field. Every quarter, review "Other" responses. If a new source appears frequently, add it as its own option
- "Colleague or friend" is always an option. Word of mouth is typically 20-40% of B2B pipeline sources. If you don't include it, you can't measure it
- No "N/A" or "Prefer not to say." These are escape hatches that reduce data quality. If the field is optional, a prospect who doesn't want to answer will skip it. If required, they'll pick something
Where to Place the Question
Placement options
| Placement |
Pros |
Cons |
Recommended? |
| Demo request form |
Highest-intent moment. Data tied to pipeline |
Adds friction to the highest-value form |
Yes (if kept to 1 question) |
| Post-form thank-you page |
No form friction. Asked after conversion |
Lower completion rate (30-50% vs 80%+ on-form) |
Yes (as fallback) |
| Sign-up flow (PLG) |
Captures product-led sources |
Friction in sign-up reduces conversion |
Yes (only if sign-up is low-friction) |
| Email survey post-meeting |
No form friction. Can ask detailed follow-ups |
Low response rate (10-20%). Memory bias |
No (too late, too low response) |
| SDR asks on discovery call |
Can probe for detail ("which podcast?") |
Inconsistent. SDRs forget. Hard to standardize |
Supplementary only |
Placement rules
- Put it on the demo request form. This is the primary placement. One additional dropdown field adds minimal friction. The data is tied directly to the highest-intent action
- Make it required on the demo form. Optional fields get 40-60% fill rates. Required gets 95%+. The data is too valuable to leave optional. One dropdown takes 3 seconds
- If form friction is a concern, use the thank-you page. Show the question immediately after form submission. "One quick question before your confirmation: how did you first hear about us?" Completion rate drops to 30-50% but zero form friction
- Never rely solely on SDR-collected attribution. SDRs forget to ask. They interpret answers differently. They skip it when the call is going well. Use form-based collection as the primary source. SDR-collected is supplementary detail
Implementation
HubSpot implementation
1. Create a contact property:
- Name: "How did you hear about us"
- Field type: Dropdown select
- Options: [your 8-12 options]
- Group: Contact information
2. Add to demo request form:
- Position: last field before submit button
- Required: Yes
- Label: "How did you first hear about us?"
3. Create a second property for detail:
- Name: "Attribution detail"
- Field type: Single-line text
- Conditional: shows only when "Podcast", "Community",
"Conference", or "Other" is selected
- Label: "Which one?" or "Please specify"
4. Create a report:
- X-axis: "How did you hear about us"
- Y-axis: Count of contacts
- Filter: created in last 90 days
- Drill-down: by lifecycle stage (MQL, SQL, Won)
Salesforce implementation
1. Create a picklist field on Lead and Contact:
- API name: How_Did_You_Hear__c
- Values: [your 8-12 options]
2. Create a text field for detail:
- API name: Attribution_Detail__c
- Length: 255
3. Add both fields to the web-to-lead form
4. Map lead field to contact field on lead conversion
5. Create a report:
- Report type: Leads with converted info
- Group by: How_Did_You_Hear__c
- Show: conversion rate, pipeline value per source
Analyzing Self-Reported Data
Primary analysis
| Analysis |
How to calculate |
What it tells you |
| Source distribution |
% of form fills per source |
Where awareness is coming from |
| Source-to-pipeline |
Pipeline $ per source |
Which sources produce the most pipeline |
| Source-to-closed-won |
Revenue per source |
Which sources produce the most revenue |
| Source quality |
Win rate per source |
Which sources produce the best prospects |
| Source trend |
Month-over-month change per source |
Which sources are growing or declining |
Cross-referencing with software attribution
| Self-reported says |
Software says |
What's happening |
| Podcast |
Direct |
Prospect heard podcast, typed URL directly. Software missed the source. Self-reported is correct |
| Google search |
Google organic |
Both agree. High confidence in this attribution |
| Colleague |
LinkedIn ad |
Prospect saw the ad AND got a recommendation. Self-reported captures the higher-intent source |
| LinkedIn post |
Direct |
Prospect saw the post, later visited directly. Self-reported captures the awareness source |
| Other: "Twitter" |
Organic social |
Both agree on social, but self-reported is more specific about the platform |
Analysis rules
- Report self-reported and software attribution separately. They answer different questions. Don't merge them into one field. Present both to leadership
- Self-reported is more accurate for dark funnel. Podcasts, word of mouth, communities. Trust self-reported over software for these sources
- Software is more accurate for click-path. Paid search, paid social, email campaigns. Trust software for these. Prospects don't always remember clicking an ad
- Review "Other" responses quarterly. New sources emerge. If "TikTok" or "YouTube" starts appearing in Other, add it as a dedicated option
- Track at the pipeline level, not just the lead level. "30% of leads say podcast" is interesting. "Podcast leads convert to pipeline at 2x the rate of Google leads" is actionable
Common Pitfalls
Data quality issues
| Problem |
Cause |
Fix |
| 60% select "Google search" |
Options are too generic. Prospects default to the most familiar option |
Add more specific options. Split "Google search" into "Searched for a solution" and "Searched for us by name" |
| "Other" is 25% of responses |
Missing common sources from the dropdown |
Review "Other" text responses. Add any source that appears 5%+ of the time |
| Responses don't match software data at all |
Prospect confuses "how did you find us today" with "how did you first hear about us" |
Clarify the question: "How did you FIRST hear about us?" Emphasis on first awareness |
| SDR-collected data conflicts with form data |
SDR probed deeper and got a different answer |
Both can be valid. Store separately. The form captures initial awareness. The SDR captures the nuanced story |
Measurement
| Metric |
Definition |
Target |
Frequency |
| Fill rate |
% of form submissions that include the attribution field |
> 90% (if required) |
Weekly |
| Source coverage |
% of responses that use a named option (not "Other") |
> 85% |
Monthly |
| Dark funnel capture |
% of responses that are dark funnel sources (podcast, WOM, community) |
Track trend, no fixed target |
Monthly |
| Source-to-pipeline correlation |
Does each source produce proportional pipeline? |
Track by source |
Quarterly |
| Cross-reference alignment |
% of responses where self-reported and software agree |
40-60% (disagreement is expected for dark funnel) |
Quarterly |
Pre-Implementation Checklist
- [ ] Question wording is "How did you first hear about us?" (not "find us" or "learn about us")
- [ ] 8-12 answer options covering all major channels including dark funnel
- [ ] "Colleague or friend" (word of mouth) is included
- [ ] "Other (please specify)" with text field is included
- [ ] Single-select, not multi-select
- [ ] Placed on the demo request form (or highest-intent conversion point)
- [ ] Field is required
- [ ] Follow-up text field for podcast name, community name, event name
- [ ] CRM property created and mapped
- [ ] Report built showing source distribution and source-to-pipeline
- [ ] Plan to review "Other" responses quarterly and add new options
Anti-Pattern Check
- Making the field optional. Fill rate drops to 40-60%. You lose half your attribution data. Make it required. One dropdown takes 3 seconds. The friction is negligible
- Using multi-select instead of single-select. Prospect selects "Google search" and "LinkedIn" and "Colleague." Now you have attribution soup. Single-select forces them to identify the FIRST source
- Only tracking at the lead level. "30% of leads say podcast." So what? Track through to pipeline and closed-won. "Podcast leads generate $2M pipeline vs $800K for Google leads" is actionable
- No "Other" option. You have 8 sources in the dropdown. A prospect heard about you on a niche industry podcast that doesn't fit any category. They pick "Google search" as a default. You lose the real attribution. Always include "Other" with a text field
- Never updating options. Your dropdown has been the same for 18 months. Meanwhile, you started a YouTube channel, launched a community, and appeared on 12 podcasts. Update options quarterly based on "Other" responses and new channel investments
- Replacing software attribution with self-reported. They answer different questions. Software tracks click paths. Self-reported tracks awareness. Use both. Report them separately. Let leadership see both views
- Asking on a post-meeting email survey. 10-15% response rate. The prospect has already forgotten where they first heard about you. Memory bias is worst at this stage. Ask at the moment of highest intent: the form fill