Research

Agent 5

Funnel Specs

Synthesizes all research from Agents 1-4 into 5 meaningfully different funnel approach specifications.

Feeds into: Prototyper

Executive Summary

Funnel Variation Specs — Executive Summary

Research Foundation

Synthesis of research from 34+ competitor funnels (Agents 1-3), Twofold product analysis and user sentiment (Agent 4), team feedback with 35 directives (Agent 10), PostHog funnel data, and NUX experiment results shaped these 5 variations.

Key insights that drove the specs:

  1. Quiz funnels convert 30-50% vs 2-5% for landing pages. Every variation uses a quiz-first approach. No landing pages.
  2. One question per screen is universal best practice. BetterHelp, Noom, Calm, Guardio all use it. All 5 variations adopt this pattern.
  3. The /new page is the critical bottleneck. 50% desktop / 35% mobile users press "start recording." The NUX experiment showed +24.84% activation with guided first action. Every variation redesigns the /new page with strong guidance and "not ready to record" alternatives.
  4. Mobile-first is non-negotiable. Facebook ads → mobile phones → Facebook in-app browser. All designs fit in a single mobile viewport. No scrolling to answer questions.
  5. No demo notes, no payment-first, no pre-quiz intros, multiple choice only. Hard constraints from team feedback honored in all variations.
  6. Email signup preferred for FB funnel. Google SSO won't work in the FB browser (38.8% Google vs 87.3% email conversion). All variations use email-based signup.
  7. Skip onboarding for cold traffic. Quiz captures specialty and workflow data that the current 4-step onboarding collects. All variations remove the onboarding wizard entirely.

Current Flow Assessment

All 5 variations converge on the same structural decisions:

Current StepDecisionRationale
Landing pageReplace with quizQuiz-first converts 6-10× better for cold traffic
Signup pageReplace — embed in resultsReduces pages. Results page creates conversion momentum.
Onboarding (all 4 steps)RemoveQuiz already captures specialty, practice data, and preferences. Welcome screen is redundant.
NUX DialogReplace with guided first actionCurrent NUX lets 47% skip. Guided action improves activation +24.84%. All variations remove the skip option and provide 2-3 engagement paths.
/new pageHeavily ModifyPersonalized based on quiz answers. Mobile-first. Dictation/sample alternatives for users not ready to record. Matches app design.

Shared pattern: Every variation follows the flow: FB Ad → Quiz (in FB browser) → Results + Signup → FB Browser Ejection → Personalized /new Page. Both session-based and account-based ejection approaches are specced for team evaluation.

The 5 Variations

1. Speed Start — Rapid Personalization Quiz

5-question quiz, 90 seconds to signup. Bets that busy mobile clinicians need the fastest path from curiosity to configured product. Simplest to build, lowest content requirements. Risk: too shallow to create commitment. /new page: 3 equal-weight first-action options (record/dictate/sample).

2. Deep Commitment — Extended Investment Quiz

13-question quiz with 3 value interstitials, 4-5 minutes to signup. Bets that deeper investment creates higher signup AND activation rates through sunk cost. Users who spend 4 minutes answering personal questions don't want to waste that effort. Draws from Noom (96 screens) and BetterHelp (35+ questions). Risk: high drop-off from quiz length increases CAC. /new page: honors the user's stated preferred note-creation method (Q10).

3. Pain Path — Segmented Entry Points

4 separate quiz entry points (time/compliance/burnout/balance), 7 questions each, 2-3 minutes. Bets that message match from ad to /new page drives higher conversion at every step. Each segment gets its own emotional arc, social proof, results, and /new page treatment. Draws from Hims (concern-specific funnels) and Calm (topic-specific quizzes). Risk: 4× content creation effort; requires per-segment ad creative and analytics. /new page: continues the emotional arc of the specific segment.

4. ROI Reveal — Quantified Outcome Calculator

9-question quiz focused on practice data (patients, time/note, hours, hourly rate), 3 minutes. Calculates a personalized ROI dashboard: hours saved/week, dollars saved/month, extra patients possible. Bets that concrete numbers persuade rational professionals better than emotional appeals. Draws from Guardio (vulnerability score) and Lemonade (value-first quote). Risk: hourly rate question feels intrusive; calculation skepticism; cold for emotionally-motivated buyers. /new page: savings tracker with progress bar gamifying documentation savings.

5. Peer Proof — Specialty-Led Social Validation

7-question quiz where every screen reveals specialty-specific peer data, 2-3 minutes. Bets that clinicians adopt what their peers adopt. Every question triggers a peer stat: "73% of therapists say the same thing." Results show what [specialty] providers experience. /new page shows "How [specialty] providers create their first note" with usage percentages. Draws from SimplePractice's community approach and the universal principle that peer recommendation is the #1 adoption driver for clinical tools. Risk: specialty-specific data may be thin for niche specialties; peer stat on every screen may fatigue.

Test 1: Variation 1 (Speed Start) vs Variation 2 (Deep Commitment)

Why first: This tests the fundamental question — does quiz length help or hurt? The answer to this shapes all subsequent optimization. Speed Start is the easiest to build (baseline). Deep Commitment tests the sunk-cost hypothesis. Run as an A/B test splitting Facebook traffic 50/50.

What we learn: Optimal quiz length for clinician cold traffic. Whether higher investment improves activation enough to offset lower signup volume.

Test 2: Winner of Test 1 vs Variation 5 (Peer Proof)

Why second: Peer Proof tests a fundamentally different conversion mechanism (social proof vs personalization). If peers outperform personalization, it reshapes the entire marketing strategy. Peer Proof has moderate implementation complexity and tests a novel approach no competitor is using.

What we learn: Whether peer validation outperforms individual personalization for clinicians.

Test 3: Winner of Tests 1-2 vs Variation 3 (Pain Path) or Variation 4 (ROI Reveal)

Why third: Pain Path requires the most content (4× paths) and should only be built if segmented messaging shows promise. ROI Reveal tests the data-driven approach. Choose between them based on learnings from Tests 1-2:

  • If Test 2 winner was more emotional (Peer Proof or Deep Commitment): test ROI Reveal as the analytical counterpoint.
  • If Test 2 winner was more analytical (Speed Start): test Pain Path for emotional segmentation.

What we learn: Whether a different persuasion style (emotional segmentation or rational data) can beat the current winner.

Shared Elements

All 5 variations share these elements regardless of approach:

  1. Quiz-first, no landing page. First screen is always Q1.
  2. No pre-quiz intro page. Quiz starts with the first question immediately.
  3. One question per screen. Multiple choice only. No text input.
  4. Signup embedded in results page. No separate signup page.
  5. Facebook browser ejection. Both session-based and account-based approaches specced.
  6. No onboarding wizard. Quiz replaces onboarding entirely.
  7. Guided /new page. No skip option. 2-3 first-action paths including "not ready to record" alternatives (dictation, sample audio).
  8. Mobile-first. No scrolling to answer. Single CTA per screen. Large tap targets.
  9. App-matching design. Blue/indigo primary. White/light gray background. Clean typography matching Twofold app, not website.
  10. HIPAA badge + "No credit card" on signup. Trust signals at the decision point.
  11. Email-based signup (not Google SSO) for the Facebook browser context.
  12. Free trial: 7 days, all features, no credit card.
  13. Hide trial countdown/pricing on first /new visit. Show only after first note generated — pricing pressure on first visit increases drop-off.
  14. Mobile: emphasize Dictate/Sample over Record on /new page. Mobile users from Facebook ads are unlikely to have a patient present. Dictation/sample should be the highlighted default on mobile.

Key Assumptions to Validate

  1. Does quiz-first outperform the current landing page for Facebook traffic? No variation preserves the current landing page. If quiz-first underperforms, we need to reconsider.
  2. Can we remove onboarding entirely? All variations skip the 4-step wizard. If activation drops without the specialty/preferences setup flow, we may need to reintroduce a lightweight version.
  3. Does Facebook browser ejection lose users? The extra step of switching browsers is untested. If ejection drop-off is severe, we need to optimize the in-browser experience.
  4. Do "not ready to record" alternatives actually drive activation? Dictation and sample audio are offered as alternatives, but we don't know if users who choose them convert to real recordings later.
  5. Does removing the skip option on /new page help or hurt? The NUX experiment showed guided action improves activation, but forced guidance could feel coercive.
  6. What's the minimum viable quiz length? Variation 1 (5Q) vs Variation 2 (13Q) directly tests this.
  7. Is specialty-specific peer data available at sufficient scale? Variation 5 requires per-specialty user counts and testimonials. For rare specialties, this may not be feasible.
  8. Does app-matching design help or hurt cold traffic conversion? All variations currently default to the Twofold app aesthetic (blue/indigo, clean, minimal). However, the highest-converting competitor funnels (Noom, BetterHelp, Calm, Hims) all use conversion-optimized designs that differ from their products — higher-contrast CTAs, bolder social proof, direct-response layouts. At least one variation should A/B test a conversion-optimized design against the app-matching default to validate whether brand consistency helps or hurts for cold Facebook traffic.
  9. Does the guided activation experience transfer to the native app? The NUX experiment showed +24.84% activation on desktop web but did NOT improve activation on the native app. All 5 variations design guided /new pages for the web experience. If users convert through the FB funnel (mobile web) but later switch to the native app, the guided activation and quiz-derived personalization may not transfer. Need to validate: (a) what percentage of FB funnel users eventually download the native app, (b) whether native app activation rates lag behind web for funnel-acquired users, and (c) whether carrying quiz personalization into the native app improves native activation.

Output Files (7)