You're Using AI Wrong for Landing Pages
I know exactly how most people use AI to "optimize" their landing page. They open ChatGPT, type "write me a landing page for my SaaS product that does X," and paste whatever comes back into their Framer template. Or they're slightly more sophisticated: they paste in their existing copy and say "make this better."
Both approaches produce the same thing: vaguely professional text that sounds like every other landing page on the internet. "Streamline your workflow." "Unlock the power of." "The all-in-one platform for." You've read a thousand of these pages. You've probably bounced from a thousand of these pages.
The problem isn't AI. The problem is using AI without knowing what's actually broken on your page. It's the equivalent of walking into a doctor's office and saying "make me healthier" without running any tests. You might get generic advice — drink more water, sleep better — but you won't get the intervention that actually moves the needle.
Over the past year, I've been refining a workflow that uses AI at every step but never lets AI drive blindly. It's five steps: Audit, Prioritize, Generate, Edit, Verify. Each step uses AI differently, and the order matters more than you think. Skip the audit and your AI-generated copy will be confidently wrong. Skip the verification and you won't know if you made things better or worse.
Here's the whole process, step by step, with the exact tools and prompts I use at each stage.
The 5-Step Workflow at a Glance
The AI Landing Page Optimization Loop
This is a loop, not a line. Step 5 feeds back into Step 1. You run it until the scores plateau or you hit diminishing returns.
The key insight behind this workflow: AI is excellent at generating options and terrible at diagnosing problems. If you ask ChatGPT "what's wrong with my landing page?" you'll get a laundry list of generic observations. But if you feed it specific audit data — "My hero section scored 4/10 because the headline is vague and there's no supporting proof" — it generates dramatically better output. The audit step creates the context that makes the generation step useful.
Let me walk through each step in detail.
Step 1: Audit What You Have
This is where 90% of people skip straight to the AI prompt. Don't.
Before you generate a single line of copy, you need a baseline. What's actually working on your page? What's broken? And — critically — what's broken relative to everything else? Because a page with a weak hero and strong trust signals needs a completely different intervention than a page with a strong hero and no social proof.
Start by running your page through the landing page analyzer. You'll get scores across multiple dimensions: First Impression, Copy & Messaging, CTA, Trust & Social Proof, Visual Design, and more. These dimension scores are the diagnostic that makes everything else work.
An overall score of 52 tells you "your page needs work." Dimension scores tell you where. A 52 with hero at 7/10 and trust at 2/10 is a very different page than a 52 with hero at 3/10 and trust at 6/10. The interventions are completely different. The prompts you'll use are completely different. Without dimension-level data, you're guessing.
Here's a real example from a page I optimized last month. B2B analytics SaaS, overall score of 48. Decent-looking page. Clean design. The founder was convinced the problem was the headline because he'd rewritten it nine times already. Here's what the audit showed:
| Dimension | Score | Key Issue |
|---|---|---|
| First Impression / Hero | 6/10 | Headline was actually fine. Subheadline was vague. |
| Copy & Messaging | 5/10 | Feature-heavy. No customer language. |
| CTA | 3/10 | One generic "Get Started" button. No friction reducers. |
| Trust & Social Proof | 2/10 | Zero testimonials. No logos. No metrics. |
| Visual Design | 7/10 | Clean layout. Good hierarchy. |
The headline wasn't the problem. Trust was the problem. The page had zero social proof. No testimonials, no logos, no customer count, no case study references, no third-party badges. It was a well-designed page asking visitors to trust a completely unverified stranger. No amount of headline rewriting was going to fix that.
Without the audit, this founder would have spent another month A/B testing headlines. With the audit, he spent a weekend collecting three customer testimonials and adding a logo bar. The page score jumped to 67. Conversion rate went from 2.1% to 3.8%.
If you want to go deeper than the automated tool, use the landing page audit prompt to run a manual review with ChatGPT or Claude. Feed it your page's HTML or a full-page screenshot and ask it to score each dimension. The combination of automated scoring plus AI-assisted review gives you an extremely clear picture of where to focus.
Step 2: Prioritize by Impact
Now you know what's broken. The temptation is to fix everything at once. Resist it.
Not all dimensions carry equal weight for conversion. Here's the weighting we use at roast.page, based on analyzing the correlation between dimension scores and actual conversion outcomes across thousands of pages:
Dimension Weights by Conversion Impact
The math for prioritization is simple: Impact = Weight x Gap. The gap is the difference between your current score and, say, 8 out of 10. A dimension weighted at 20% with a score of 3/10 (gap of 5) has an impact score of 100. A dimension weighted at 10% with a score of 6/10 (gap of 2) has an impact score of 20. Fix the first one first.
Going back to our B2B analytics example:
| Dimension | Score | Weight | Gap (to 8) | Impact | Priority |
|---|---|---|---|---|---|
| Trust & Social Proof | 2/10 | 15% | 6 | 90 | #1 |
| CTA Effectiveness | 3/10 | 15% | 5 | 75 | #2 |
| Copy & Messaging | 5/10 | 20% | 3 | 60 | #3 |
| First Impression / Hero | 6/10 | 20% | 2 | 40 | #4 |
| Visual Design | 7/10 | 10% | 1 | 10 | #5 |
Clear as day: Trust first, CTA second, then Copy. Visual design is last — it's already good and it carries the lowest weight. This is exactly the kind of prioritization that people get wrong when they go by gut feel. The founder's instinct was to rewrite the headline (priority #4). The data said to add testimonials (priority #1).
Here's the mistake I see constantly: people fix what's easy instead of what matters. Tweaking button colors is easy. Collecting customer testimonials requires sending awkward emails. Guess which one teams do first? The button colors. And then they wonder why nothing changed.
Step 3: Generate With the Right Prompts
Now — now — you can open ChatGPT. Or Claude. Or whatever AI tool you prefer. But not with a generic "make my page better" prompt. You know exactly which dimension to fix, and you should use a prompt designed for that specific dimension.
This is the step where most AI-for-landing-pages advice starts. Which is exactly why it fails. Without Steps 1 and 2, you're generating solutions to problems you haven't identified. With Steps 1 and 2, you're generating targeted fixes for specific, measured weaknesses. The difference in output quality is enormous.
Here's my prompt mapping. Match your lowest-scoring dimension to the right prompt set:
Weak Hero (below 5/10)
Weak Copy (below 5/10)
Weak CTA (below 5/10)
Weak Trust (below 5/10)
Let me walk through a concrete example. Say your hero score is 4/10. The audit says: "Headline is vague — 'The Future of Productivity' communicates nothing specific. No subheadline supporting the claim. No product visual above the fold."
Here's what I'd do. First, I'd pull up the hero section headlines prompt and feed it three things:
- The audit findings — Paste in exactly what the analyzer said about your hero. The specific criticisms. The score breakdown.
- Your product context — What you do. Who it's for. What outcome your best customers care about. (Not features. Outcomes.)
- Your current headline — So the AI knows what to improve upon, not what to imitate.
The prompt output will give you 8-12 headline variations. Some will be terrible. Some will be generic. But 2-3 will contain a kernel of something real — a phrasing, an angle, a structure — that you can build on. That's the value. AI doesn't write your final headline. It generates a range of options that you'd never come up with staring at a blank page because the AI doesn't have your cognitive biases about your own product.
The single biggest quality lever in AI-generated landing page copy is the specificity of the input. Vague input = vague output. When you feed the AI your audit scores, your customer interviews, your competitor URLs, and your specific product metrics — the output goes from "Boost your productivity with our cutting-edge solution" to "Cut your weekly reporting time from 4 hours to 15 minutes." Same AI, same model, 10x better output. The audit data from Step 1 is the context injection that makes everything work.
Then run the outputs through the Headline Analyzer to check clarity, specificity, and emotional pull. Kill the ones that score low. Refine the survivors.
For CTA weakness, the process is similar but you use the CTA optimization prompt. Feed it your current CTA, the audit critique, and your conversion goal. It'll generate button copy, friction reducers ("No credit card required," "Set up in 2 minutes"), and urgency elements. Our CTA analysis found that the combination of specific button text + friction reducer below the button outperforms generic CTAs by 30%+. The AI can generate dozens of these combinations in seconds.
Step 4: Edit and Ship
This is where I lose people. They're excited about the AI outputs from Step 3, and they want to copy-paste them straight into production. Do not do this.
AI output is draft one. It's a starting point. It's the raw material, not the finished product. And there's a specific reason why: AI optimizes for plausibility, not accuracy. It generates text that sounds correct and reads well. But "sounds correct" and "is specifically true about your product" are different things. AI doesn't know that your onboarding takes 3 minutes, not 5. It doesn't know that your customers are primarily mid-market, not enterprise. It doesn't know that the word "simple" is how your competitor positions, not you.
Every piece of AI-generated copy needs to pass through a three-question filter before it goes live:
The 3-Question Filter for AI Copy
-
Is this specific to my product?
Could you swap in a competitor's name and have the sentence still work? If yes, it's too generic. "Our platform helps teams collaborate more effectively" is true of Slack, Notion, Asana, Monday, Basecamp, and approximately 4,000 other products. Rewrite it until it could only describe your product. -
Would a competitor's page say the same thing?
Related but different from #1. This catches positioning problems. If your AI-generated copy positions you the same way your competitor positions themselves, you've just surrendered your differentiation. The copy needs to highlight what makes you different, not what makes you similar. -
Does this add information or just words?
AI loves filler. "Seamlessly integrate with your existing workflow" — what information does this convey? Almost none. Compare: "Works with Salesforce, HubSpot, and Pipedrive. Connect in under 2 minutes." Same idea, but with actual information. Every sentence on your landing page should earn its place by telling the visitor something they didn't know before they read it. If it's just taking up space, cut it. Your copy quality depends on ruthless editing.
Here's my actual editing process. I print (yes, on paper) the AI-generated options alongside my existing copy. I read each one out loud. The ones that make me cringe when spoken aloud get cut immediately — if it feels unnatural in your mouth, it'll feel unnatural on the page. The survivors get fact-checked against my actual product. Then I Frankenstein the best pieces together: a structure from option A, a specific phrase from option C, a CTA framing from option F.
The result is copy that's better than what I'd write from scratch (because the AI explored angles I wouldn't have considered) and better than what the AI generated (because I added specificity, brand voice, and factual accuracy). The whole point of AI in this workflow is augmentation, not replacement. If your final copy is 100% AI-written, you've done it wrong. If it's 0% AI-influenced, you're leaving quality on the table.
Preserving brand voice
One thing that trips people up: AI-generated copy often has a recognizable "AI voice." Slightly too polished. Too many transition words. A tendency toward parallel structure that no human would naturally use. The telltale "Whether you're... or..." construction.
The fix: before you start Step 3, feed the AI 3-5 paragraphs of your existing copy that you genuinely like. Tell it to match that voice. This won't make the output perfect, but it'll get it 70% of the way there. The remaining 30% is your editing job — and honestly, it's the most important 30%. Your voice is your brand's fingerprint. Don't let an AI sand it off.
Step 5: Verify the Improvement
You've audited. You've prioritized. You've generated and edited. The new copy is live. Now comes the step that separates professionals from amateurs: measure whether it actually worked.
Run the updated page through the analyzer again. Compare the new dimension scores against your baseline from Step 1. The questions to answer:
- Did the target dimension improve? If you focused on Trust (was 2/10), did it actually go up? By how much?
- Did any other dimensions drop? This happens more often than you'd think. Adding three testimonials above the fold might boost trust but push your CTA below the fold, hurting your CTA score. Adding a longer headline might improve specificity but break the visual balance the designer worked hard on.
- Is the overall score higher? Dimension-level improvements should roll up into overall score improvement. If a dimension went up but the overall score stayed flat, something else regressed. Find it.
This is normal. Landing page optimization is always about tradeoffs. A longer, more specific headline improves copy scores but might hurt visual balance. More trust signals above the fold improve trust but push the CTA down. The solution is not to avoid the fix — it's to fix both. Add the testimonials AND restructure the layout so the CTA stays visible. Write the specific headline AND adjust the font size to maintain visual hierarchy. Optimization is iterative. That's why this is a loop.
I recommend running 2-3 full loops of this workflow before shifting to live A/B testing. Each loop takes 2-4 hours depending on the changes involved. By the third loop, you've typically addressed the top 3-4 dimension weaknesses and your score has moved meaningfully — I typically see 15-25 point improvements across three iterations. At that point, the audit-level improvements have diminishing returns and it's time to let real visitor behavior data guide the next phase.
The before/after data also gives you something invaluable: a paper trail for your decisions. When your CEO asks "why did you change the headline?" you don't say "I thought it sounded better." You say "The audit showed our hero scored 4/10 with a specificity gap. The new headline tested at 7/10. Overall score went from 48 to 71." That's a different conversation entirely.
The 6 Mistakes That Kill This Workflow
I've watched dozens of teams attempt AI-assisted landing page optimization. The workflow above works when followed. Here's how people break it:
This is the big one. Asking AI to "improve my landing page" without audit data is like asking a doctor to prescribe medication without running tests. The AI will generate plausible-sounding improvements that may have nothing to do with your actual weaknesses. If your CTA is the problem but the AI rewrites your headline, you've spent time and possibly made things worse.
I cannot stress this enough. Raw AI output has a sameness to it. Visitors might not consciously recognize AI-generated copy, but they recognize generic copy. And generic copy doesn't convert. Every piece of AI output needs human editing for specificity, accuracy, and voice. These copy mistakes are exactly what unedited AI generates.
People default to fixing what they know how to fix. Designers fix visual design. Copywriters fix copy. Engineers fix page speed. But the data might say trust is the bottleneck. The prioritization math in Step 2 exists to override your instincts with evidence. Use it.
The right mental model: AI gives you a range of options to react to, not a finished product to accept. When you get 10 headline variations, the value isn't "pick the best one." It's "notice which angle resonates, then write an 11th version that combines the best angle with your specific knowledge." That 11th version is almost always better than any of the 10.
You made changes. Cool. Did they work? If you don't re-run the audit, you don't know. Plenty of "improvements" make things worse. I once watched a team add so many trust signals that the page became cluttered, dropping the Visual Design score by more than the Trust score improved. Net negative. Without the verification step, they would have celebrated a page that performed worse than the original.
One pass through this workflow is better than nothing. Three passes is where the compounding kicks in. Each loop addresses the next-highest-impact weakness. After three loops, you've fixed the top 3-4 issues and the page is in a fundamentally different place. Don't stop at one. The best landing pages aren't built in a single pass — they're iterated into existence.
The Workflow Works Because of Context
Let me be direct about why this works and why "just use AI" doesn't.
Large language models are pattern-matching engines. They generate text that fits the patterns in their training data. When you give them zero context about your specific page — "write me a landing page" — they match against the average of all landing pages they've seen. You get average output. Mediocre. The pattern of mediocrity, perfectly replicated.
When you give them rich, specific context — audit scores, dimension breakdowns, your product specifics, your customer language, your competitive positioning — they match against much more specific patterns. The output is dramatically more useful. Not because the AI got smarter, but because you gave it better input.
That's the whole point of this workflow. Steps 1 and 2 create the context. Step 3 uses the context. Step 4 adds your unique knowledge. Step 5 validates the result. Each step makes the next step better. Remove any step and the chain breaks.
If you want to try this right now, start with Step 1. Run your page through roast.page and look at your dimension scores. Find the lowest one. Then go to the audit prompt to dig deeper into what's specifically wrong. By the time you reach Step 3, you'll have the context that makes AI actually useful — and you'll wonder why you ever tried to use it without a diagnostic first.
The full loop takes a few hours. The pages that come out the other side are measurably, provably better. Not because AI is magic, but because you gave it the right job at the right step.