Why Next.js 15 Is the Future of Web Development
Exploring the features that make Next.js 15 the go-to framework for building modern, performant web applications.
Search is changing. More people are asking ChatGPT "who's a good web developer in Utah?" instead of typing it into Google. If your business isn't structured to appear in AI-generated answers, you're invisible to a growing slice of your market.
This is the gap between traditional SEO and what I call **Generative Engine Optimization (GEO)** — optimizing your content to be surfaced and recommended by AI models, not just ranked in search results.
AI language models answer questions by pattern-matching against what they've seen during training — and what they can retrieve via live web search (in tools like Perplexity, Claude.ai, and ChatGPT with browsing). To get recommended, your site needs to:
1. Be crawlable by AI training and retrieval systems
2. Contain clear, factual, machine-readable information about who you are and what you offer
3. Be cited as a source across the web (reviews, directories, linked mentions)
4. Use structured data that explicitly describes your services, location, and expertise
One of the most actionable things you can do right now is add an `llms.txt` file to your site root. Proposed in 2024 and gaining rapid adoption, this is essentially a `robots.txt` for AI — a plain-text Markdown file that gives language models a concise, structured summary of your site.
> One-sentence description of what you do and for whom.
- [Service Name](/services/service): What it does and who it's for.
- Email: you@yourdomain.com
- Location: City, State
Place this at `yourdomain.com/llms.txt`. AI crawlers from Anthropic, OpenAI, and others are already looking for this file.
Schema.org markup helps both traditional search engines and AI parsing understand your content without inference. For a service business, the most valuable schemas are:
- **`ProfessionalService` or `LocalBusiness`** — describes your business entity
- **`Person`** — describes the founder/owner with credentials and expertise
- **`FAQPage`** — surfaces answers directly in search and AI results
- **`Article`** — marks up blog posts with authorship, date, and topic
A complete `Person` schema communicates to AI models exactly what you're an expert in:
{
"@type": "Person",
"name": "Your Name",
"jobTitle": "Full-Stack Software Engineer",
"knowsAbout": ["React", "Next.js", "Software Architecture", "SEO"],
"url": "https://yourdomain.com"
}
By default, AI training crawlers follow `User-agent: *` rules. But explicitly listing them signals that you welcome their indexing:
User-agent: GPTBot
Allow: /
User-agent: ClaudeBot
Allow: /
User-agent: PerplexityBot
Allow: /
User-agent: Google-Extended
Allow: /
This has no downside — if you want to appear in AI answers, you want these crawlers on your site.
AI models summarize. They pull out the clearest, most specific facts from a page. Vague marketing copy ("we deliver excellence") gets ignored. Specific, factual claims get extracted and cited:
- ❌ "We build great websites for businesses."
- ✅ "Starter websites from $1,500, built with Next.js and React, delivered in 2–4 weeks."
Write your About, Services, and Pricing pages as if you're answering specific user questions. Use clear headings. Include numbers, locations, technologies, and timelines.
AI models trained on web data weight mentions in third-party sources heavily. If five different sites describe you as "a Next.js developer based in Utah," that pattern gets reinforced. Practical steps:
- Claim and complete your Google Business Profile
- Get listed on Clutch, UpWork, and relevant industry directories
- Ask satisfied clients to leave Google reviews that mention your specialty
- Guest post or contribute quotes to industry publications
GEO is harder to measure than SEO, but you can track:
- **Direct mentions in AI tools**: Periodically ask ChatGPT, Claude, and Perplexity "who is a good [your specialty] in [your location]" and note if you appear
- **Traffic from AI referrals**: Perplexity and some AI tools send referral traffic you can track in GA4
- **Citation growth**: Monitor brand mentions using tools like Google Alerts
GEO isn't magic. It's the same fundamentals as SEO — clear content, structured data, authoritative sources — applied to how AI models retrieve and synthesize information rather than how search algorithms rank pages. The businesses that win in AI-powered search will be the ones that make it easiest for machines to understand exactly what they offer, for whom, and at what price.
Start with `llms.txt`. Add your schemas. Be specific. That's most of it.