Toprank ads-landing
Score and diagnose Google Ads landing pages. Use when asked to audit a landing page, check landing page quality, diagnose high-CTR but low-conversion-rate ad groups, improve Quality Score's Landing Page Experience component, or compare an ad group's messaging against its landing page. Trigger on "landing page audit", "landing page score", "landing page quality", "why is my conversion rate low", "LPX", "landing page experience", "ad to page match", or when `/ads-audit` surfaces a high-CTR / low-CVR ad group.
git clone https://github.com/nowork-studio/toprank
T=$(mktemp -d) && git clone --depth=1 https://github.com/nowork-studio/toprank "$T" && mkdir -p ~/.claude/skills && cp -r "$T/google-ads/ads-landing" ~/.claude/skills/nowork-studio-toprank-ads-landing && rm -rf "$T"
google-ads/ads-landing/SKILL.mdSetup
Read and follow
../shared/preamble.md — it handles MCP detection, token, and account selection. If config is already cached, this is instant.
Landing Page Scoring + Diagnostic
Google Ads campaigns fail on the landing page more often than in the auction. A great RSA that sends traffic to a slow, unfocused, or mismatched page burns budget twice — once on the click, once on the lost conversion. This skill scores landing pages on 5 weighted dimensions and emits concrete fixes.
When to run this
| Trigger | Source |
|---|---|
| User explicitly asks for a landing page audit | Direct invocation |
Layer 4 finds ad groups with CTR > account avg AND CVR < 50% of account avg | Auto-handoff |
| Quality Score diagnosis flags "Landing Page Experience: Below Average" on high-spend keywords | routes here |
is about to write copy for a page the user hasn't validated | Preflight check |
Only score pages that actually run ad traffic. Don't score random marketing pages.
Reference Documents
— The 5-dimension weighted rubric, thresholds, and evidence fields (read this before running any score)references/scoring-rubric.md
— Uses../ads-audit/references/business-context.md
for brand voice and differentiators to check message match{data_dir}/business-context.json
— Read only if the user's goal is QS improvement specifically../ads/references/quality-score-framework.md
Phase 1: Resolve the target pages
Figure out which URLs to score. In priority order:
- User supplied a URL — score that one page, skip discovery.
- User supplied an ad group or campaign name — pull
for that ad group and extract unique final URLs. Normalize (strip tracking params, preserve path + query that affects routing).listAds - Auto-handoff from
— the handoff passes the specific ad groups flagged in Layer 4. Pull their final URLs./ads-audit - No arguments — pull
for the whole account, rank final URLs by spend (last 30 days), and propose the top 3 to score. Ask the user to confirm.listAds
De-duplicate aggressively. Many ads point to the same final URL — score each unique URL once, then map back to all the ad groups that use it.
Phase 2: Gather signal (parallel)
Run all of these in a single tool-use turn:
- WebFetch the landing page — capture visible headline, subheadline, primary CTA text, form fields, trust signals, body copy tone. Also capture the full HTML so we can spot script bloat and above-the-fold content.
- PageSpeed Insights API call — use
via WebFetch. No API key needed for single-URL queries. Extract LCP, CLS, INP, TTI, performance score, and the top 3 opportunities (https://www.googleapis.com/pagespeedonline/v5/runPagespeed?url={url}&strategy=mobile&category=performance&category=accessibility&category=best-practices&category=seo
).lighthouseResult.audits - Pull the referring ad copy — from the ad group(s) that send traffic here, get headlines and descriptions via
. This is the message-match baseline.listAds - Read
— for brand voice, differentiators, offers, target audience. If it's missing, point the user to{data_dir}/business-context.json
first. Don't guess the business./ads-audit - Pull conversion data for the ad group(s) —
+ search term report for the ad groups pointing here. Used to calculate CVR and to ground dollar-impact estimates.listAds
If any single call fails, continue — note the gap in the report rather than blocking. PageSpeed Insights can rate-limit; if it does, fall back to manual timing annotation ("PSI unavailable — could not score Page Speed") and deflate the final grade's confidence rather than skipping the dimension.
Phase 3: Score the page
Read
references/scoring-rubric.md and score each dimension 0-100 with evidence. The dimension scores are real measurements (PageSpeed Insights numbers, word-for-word copy comparison, form field counts, etc.) — they're not artificial ratings, they're observations.
Compute the weighted composite only as an internal reference number for the dollar-lift formula below. Do not surface it as a letter grade. The user sees the dimension-level measurements and the estimated dollar lift — the composite is plumbing.
internal_composite = 0.25 * Message Match + 0.25 * Page Speed + 0.20 * Mobile Experience + 0.15 * Trust Signals + 0.15 * Form & CTA
Dollar lift is the headline. If
business-context.json.unit_economics has aov_usd + profit_margin, compute the estimated monthly lift from raising the composite by 15 points (see ../shared/ppc-math.md):
Target lift = min(+15, 90 - internal_composite) # cap at 90 internal Assumed CVR lift = target_lift / 100 * 0.5 # cap at 50% relative lift Current conversions = ad group conversions from last 30d Additional conversions = current_conversions * assumed_CVR_lift Additional revenue = additional_conversions * AOV Additional profit = additional_conversions * AOV * profit_margin
Present the lift as
fixing this page is worth ~$X/mo in profit — never as a guarantee. The 50% cap on CVR lift and the 15-point cap on score improvement keep estimates out of fantasy territory. If unit_economics isn't available, skip the dollar line entirely rather than making up a number — the dimension measurements still stand on their own.
Phase 4: Deliver the report
Max 60 lines. Lead with the dollar lift (when available) and the single biggest fix. No letter grade.
# Landing Page — [URL] Ads sending traffic here: [N ad groups] · [X clicks/mo] · [$Y spent/mo] · CVR [Z%] [If unit_economics available] **Estimated lift from top 3 fixes: ~$X/mo in profit** [If unit_economics is missing] _(Dollar lift unavailable — no verified AOV/margin. Confirm unit economics in business-context.json for sharper estimates.)_ **Biggest leak:** [one sentence naming the dimension and the specific observation, e.g. "LCP is 5.8s on mobile — 2.8s slower than the 3s threshold that kills conversion rate."] ## Measurements | Dimension | Measurement | Top Finding | |-----------|-------------|-------------| | Message Match | [word-for-word verdict: Match / Drift / Broken] | [one line citing ad H1 vs page H1] | | Page Speed | LCP Xs · INP Xms · CLS X · PSI perf score X | [top blocking audit from Lighthouse] | | Mobile Experience | PSI accessibility X · [mobile-specific issue count] | [one line: e.g. "No click-to-call, form below fold"] | | Trust Signals | [review count, years in business, cert count] | [one line: e.g. "Zero named testimonials, copyright 2023"] | | Form & CTA | [field count] fields · CTA text: "[button]" · [above/below fold] | [one line: e.g. "11 fields for a free quote"] | ## Fix First (top 3, ranked by estimated $ lift) 1. **[Action]** — est. +$X/mo · `<time_to_fix>` Evidence: [the actual text/number from the page or PSI audit] 2. **[Action]** — est. +$X/mo · `<time_to_fix>` Evidence: [...] 3. **[Action]** — est. +$X/mo · `<time_to_fix>` Evidence: [...] ## Message Match Detail Ad headline: "[actual headline from top-spending ad]" Page H1: "[actual H1 from landing page]" Observation: [Match / Drift / Broken] — [one-line rationale citing the specific words that match or don't] ## Handoff [Pick one:] - Page speed dominates the problem → "Share these fixes with your developer: [list]" - Message mismatch dominates → "Run /ads-copy to rewrite ads to match the page, or update the page to match the ads" - Form friction dominates → "Reduce form to [specific fields]. Every removed field is ~10% more conversions"
Writing back to history
Append the score to
{data_dir}/landing-page-history.json so re-audits can show deltas:
{ "pages": { "https://example.com/services/roofing": { "history": [ { "date": "2026-04-14", "internal_composite": 67, "dimensions": { "message_match": 72, "page_speed": 45, "mobile": 80, "trust": 70, "form_cta": 65 }, "psi_mobile_lcp_s": 4.2, "psi_mobile_cls": 0.15, "psi_mobile_inp_ms": 320, "estimated_lift_usd_per_month": 380, "ad_groups": ["Tukwila Search - Roofing"], "monthly_spend": 1240.50, "monthly_cvr": 2.1, "biggest_leak": "Page Speed — LCP 4.2s on mobile" } ] } } }
internal_composite is stored for trend tracking only — it's the internal reference number used by the dollar-lift formula, never shown to the user as a letter grade. On subsequent runs against the same URL, diff the raw dimension measurements and the dollar lift: LCP 4.2s → 2.1s · Page Speed 45 → 78 · estimated lift $380/mo → $120/mo remaining. Three measurements moved, no artificial grade flip.
Rules
- Never score a page without WebFetch'ing it. The rubric demands evidence. No WebFetch = no score. Ask the user to help if the page is gated or requires auth.
- Never report a PSI number you didn't measure. If PSI failed, say "PSI unavailable" — don't estimate.
- One page at a time unless the user asks for multiple. Scoring three pages in one turn creates unreadable reports. Batch only when explicitly requested.
- Don't rewrite copy here. This skill diagnoses the page. Handoff to
for new headlines or/ads-copy
for bid/negative/budget moves./ads - Margin-aware dollar impact requires verified unit economics. If
, appendunit_economics.source == "inferred_from_template"
to the lift line._(using industry defaults — confirm your AOV/margin for sharper estimates)_ - Always persist. Every scored page goes into
, even if the user doesn't ask — future audits depend on the baseline.landing-page-history.json