Author: Gary Crossey

  • Episode 1.2: Question-Based Content: The Secret Sauce of AEO

    Episode 2: Question-Based Content

    Page is being updated.

  • Episode 1.1: From SEO to AEO: Why Your Content Needs to Speak AI

    Episode 1: From SEO to AEO

    Pages are being updated.

  • Episode 2.4: RAG-Aware Content Patterns

    Episode 2.4: RAG-Aware Content Patterns

    Welcome back to AEO Decoded – I’m Gary Crossey, and if you’re joining us for Episode 2.4 of Season 2, you’re in for a treat!

    Today we’re tackling Episode 2.4 of Season 2: RAG-Aware Content Patterns – and I promise this one’s going to be pure dead brilliant! Over the 10 episodes of Season 2, we’re diving into advanced AEO strategies that separate good optimization from world-class optimization, and today’s topic is absolutely critical for anyone serious about winning in the age of AI answers.

    Last episode, we explored conversation patterns and follow-up funnels – how to map those natural question trees and keep AI systems coming back to your content. This week, we’re going deeper into the mechanics: how do LLMs actually ingest, chunk, and retrieve your content when they’re generating answers?

    If you caught Season 1, you learned the fundamentals of question-based content in Episode 2: Question-Based Content: The Secret Sauce of AEO. Today, we’re building on that foundation to understand exactly how your content gets processed at a technical level – and more importantly, how to structure it so your passages win retrieval every single time.

    If you’ve been following since Season 1, you know we’ve built something special here – a community that’s genuinely excited about the cutting edge of AEO. We’re still a tight-knit group, but that’s exactly what makes this so powerful. We’re ahead of the curve on advanced RAG optimization, and I’m grateful to have such an engaged audience joining me on this journey. Keep the questions and feedback coming – you lot are brilliant!

    Today we’re diving deep into RAG-aware content patterns – stick with me for the next 15 minutes and you’ll walk away with strategies you can implement right away to make your content citation-worthy.

    Let me tell you a story about something that happened a few months back working with a client. They had brilliant content – detailed, accurate, authoritative – but ChatGPT kept citing their competitors instead of them. Frustrated doesn’t even begin to cover it, so it doesn’t.

    We dug into the problem and discovered something fascinating. Their content was structured in these massive, flowing paragraphs – beautiful prose, really, like reading a proper novel. But here’s the thing: when an LLM processes that content for retrieval, it has to break it into chunks. And their lovely flowing prose? It was getting chopped up in all the wrong places.

    Imagine taking a perfectly good Ulster fry and running it through a blender. Sure, all the ingredients are still there – the sausage, the bacon, the potato bread – but you’ve lost what made it special. That’s what was happening to their content.

    We restructured everything using RAG-aware patterns – clear semantic boundaries, explicit passage markers, citation-friendly formatting. Within weeks, their citation rate tripled. Same information, same expertise, but now structured in a way that LLMs could actually work with.

    That’s the power of understanding RAG systems. It’s not enough to have great content anymore – you need content that survives the journey from your page through the chunking process, into the embedding space, and out the other side as a citation. And that’s exactly what we’re covering today.

    3. Overview (1.5-2 minutes / 195-300 words)

    So what exactly is RAG, and why should you care? RAG stands for Retrieval-Augmented Generation – it’s the technology that lets LLMs like ChatGPT, Claude, and Perplexity pull in fresh information from the web to generate accurate, up-to-date answers.

    Here’s how it works in simple terms: When someone asks a question, the system doesn’t just rely on what it learned during training. Instead, it searches for relevant content, retrieves specific passages, and uses those passages to construct an answer. Think of it like how you’d prepare for a pub quiz – you don’t need to memorize everything, you just need to know where to look things up quickly.

    But here’s the critical bit: these systems don’t read your content the way humans do. They break it into chunks (usually 500-1000 tokens), convert those chunks into mathematical representations called embeddings, and then search through millions of these embeddings to find the most relevant passages for any given query.

    This process – chunking, embedding, retrieving – is where most content fails. Your brilliant 2,000-word article gets chopped into pieces, and if those pieces don’t make sense on their own, they won’t get retrieved. If they don’t get retrieved, they don’t get cited. Simple as that.

    In Season 1, we covered question-based content and FAQ patterns. Those fundamentals are still critical – but now we’re adding another layer. We’re not just structuring content for AI understanding; we’re structuring it to survive the retrieval process. That’s the advanced game, and that’s what separates content that gets cited from content that gets ignored.

    Alright folks, it’s time for ‘The Breakdown’ – where we take those fancy-pants AI concepts and break them down into bite-sized morsels that won’t give you digital indigestion!

    Let’s start with the chunking process, because this is where everything begins. When an LLM encounters your content, it doesn’t process the whole thing at once. Instead, it breaks it into smaller pieces – typically 500-1000 tokens, which is roughly 350-700 words. Think of it like cutting a cake: the system needs reasonably sized pieces it can work with.

    But here’s the problem: most chunking algorithms are dead simple. They look for paragraph breaks, heading tags, or just count tokens and cut when they hit the limit. If your content doesn’t have clear semantic boundaries, you end up with chunks that start mid-thought and end mid-sentence. That’s like serving someone half a sandwich – technically edible, but not exactly appetizing.

    The first RAG-aware pattern: Create explicit semantic boundaries.

    Every major idea in your content should be self-contained within natural chunk-sized sections. Use headings liberally – not just H1s and H2s, but H3s for sub-concepts. Each section under a heading should be able to stand alone and make sense without requiring the reader to have seen what came before.

    Here’s a practical example: Instead of writing “As mentioned earlier, this technique…” write “This technique (introduced in the section above on entity graphs)…” Give context within each passage. It’s a wee bit redundant for human readers, but it’s absolutely critical for chunked retrieval.

    The second pattern: Front-load your key information.

    In journalism, they call it the inverted pyramid – most important information first, supporting details after. In RAG optimization, it’s even more critical. The first sentence of each section should contain the core claim or answer, because that sentence is what determines whether the entire chunk gets retrieved.

    Remember back in Season 1 when we talked about question-based content? This is where that foundation pays off. If your section starts with “What is entity disambiguation?” followed immediately by a clear definition, that chunk has a much higher chance of being retrieved for related queries than if you buried the definition three paragraphs down after historical context.

    The third pattern: Use passage markers and anchor points.

    This is where we get a bit technical, but stay with me – it’s pure class when you see it in action. HTML anchor tags (those <a id="section-name"> bits in your code) aren’t just for creating jump links. They’re semantic markers that help chunking algorithms identify logical boundaries.

    Similarly, structured elements like lists, tables, and callout boxes create natural chunk boundaries. An LLM processing your content sees these as discrete units of information. A well-formatted comparison table, for instance, will often be chunked as a single unit – which means it gets retrieved as a complete, citation-worthy piece of information.

    The fourth pattern: Optimize for passage-level relevance.

    Here’s where embeddings come in. When your content gets chunked, each chunk is converted into a mathematical representation – a vector in high-dimensional space, if you want to get technical about it. But all you need to know is this: chunks with clear topic focus and relevant terminology get better embeddings.

    What does that mean practically? Each section should focus on ONE concept and use the terminology someone would actually use when asking about that concept. Don’t get creative with synonyms just to avoid repetition. If you’re writing about “schema markup,” use that exact phrase multiple times in that section. Consistent terminology leads to stronger semantic signals.

    The fifth pattern: Build citation-worthy passages.

    Not all retrieved passages get cited. LLMs have citation criteria – they prefer passages that include attributable claims, specific data, clear expertise signals, and proper context. Think about what makes a passage citation-worthy:

    • Does it include specific, verifiable information (not just generalities)?
    • Does it demonstrate clear expertise (author credentials, institutional backing, methodology)?
    • Does it provide proper context (definitions, scope, limitations)?
    • Is it structured clearly (with logical flow and explicit conclusions)?

    Here’s an example of the difference: “RAG systems are important for AI” versus “According to research from Stanford’s AI Lab (2024), RAG systems improved answer accuracy by 40% compared to non-retrieval methods, particularly for queries requiring current information or domain-specific expertise.”

    Which one would you cite? Exactly.

    The final bit about RAG awareness: understand that retrieval is competitive. When an LLM searches for relevant passages, it’s ranking them. Your passage isn’t just competing to be good enough – it’s competing to be better than thousands of other passages on the same topic. That’s why these patterns matter so much. They’re not about gaming the system; they’re about making your genuinely valuable content accessible in the format these systems need.

    Now let’s get practical about how you actually implement this, so it is.

    Step 1: Audit your existing content for chunk-ability. Take your most important pages and mentally divide them into 350-700 word sections. Do those sections make sense on their own? If not, you need restructuring. This isn’t a quick job – plan for 2-3 hours per major piece of content.

    Step 2: Add semantic structure. Go through and add H3 headings for every distinct concept. Each heading should be a clear topic label – not clever or creative, just descriptive. “How RAG Systems Work” beats “The Magic Behind the Curtain” every single time for retrieval purposes.

    Step 3: Rewrite opening sentences. Look at the first sentence of each section. Does it contain the key information? Can someone understand the main point from that sentence alone? If not, rewrite it. Front-load those key claims.

    Step 4: Add passage markers. If you have access to your site’s HTML, add anchor IDs to major sections. Format: <h3 id="topic-name">Your Heading</h3>. This helps chunking algorithms and also enables deep linking.

    Step 5: Enhance citation-worthiness. Add specific data points, dates, sources, and expertise signals. Include phrases like “According to…”, “Research shows…”, “Analysis of X reveals…”. These signal authoritative, citable information.

    Pro tip from the Method Q playbook: Create a “RAG optimization checklist” and run every major piece of content through it before publishing. Check for: clear headings, front-loaded information, explicit context in each section, specific data points, and proper semantic boundaries. Takes 10 minutes and dramatically improves your citation rate.

    Common pitfall to avoid: Don’t over-optimize to the point where your content becomes robotic. Yes, you want clear structure and explicit information, but it still needs to be readable for humans. The sweet spot is content that works for both audiences – properly structured for machines, still engaging for people.

    Timeline for results: Unlike some AEO strategies that take months, RAG optimization can show results quickly. We’ve seen citation rate improvements within 2-3 weeks of restructuring content, because LLMs are constantly re-crawling and re-indexing. The faster these systems update, the faster you see results from optimization.

    ⚡ Q&A Lightning Round — Your Burning Questions Answered!

    Now, let’s tackle some common questions about RAG-aware content patterns:

    Q: How do I know what chunk size to optimize for?

    A: The standard is 500-1000 tokens, but aim for the lower end (500-700) to be safe. Different systems use different chunk sizes, so optimizing for smaller chunks ensures your content works across platforms. As a rule of thumb, keep major sections under 500 words with clear breaks between concepts.

    Q: Should I create separate pages for each topic or keep everything in comprehensive guides?

    A: Both approaches work, but comprehensive guides with clear section structure often perform better. The key is making each section independently valuable. Think of it like building a page that’s actually 10 mini-pages stitched together – each section should be chunk-sized and self-contained.

    Q: How much does this impact my existing SEO?

    A: The brilliant news is that RAG-aware patterns actually improve traditional SEO too! Clear headings, front-loaded information, and well-structured content are exactly what Google has been recommending for years. You’re not choosing between SEO and RAG optimization – you’re doing both.

    Q: What about technical content with complex explanations?

    A: Complex topics need even more structure. Break them into smaller conceptual chunks, use analogies in your opening sentences, and create progressive disclosure – start with the simple explanation, then add layers of detail in subsequent sections. Each section can stand at a different complexity level.

    Q: How do I measure if my RAG optimization is working?

    A: Great question! Monitor citation rates using tools that track AI-generated content (we’ll cover this in detail in Episode 2.8). But even without specialized tools, you can manually check: search for your content topics in ChatGPT, Claude, and Perplexity. Are you being cited? That’s your primary success metric.

    Q: Is this worth doing for older content or just new content?

    A: Start with your highest-traffic, most important content first. Older content that’s still relevant absolutely deserves RAG optimization – in fact, it might benefit even more because it’s already established authority. Prioritize based on traffic and business value, not publication date.

    Remember, implementing these patterns isn’t about perfection – it’s about progress. Start with one piece of content, apply these principles, and see how it performs. Then iterate and scale up. You’ll be sorted rightly before you know it!

    7. Actionable Takeaway (1 minute / 130-150 words)

    Let’s wrap it up with the takeaway section. This section will give you that one actionable item you can work on.

    Here’s your one key action item from today: Take your single most important piece of content – your flagship article, your core service page, whatever drives your business – and apply the “chunk test.” Read through it and mentally break it into 500-word sections. For each section, ask: “If someone only saw this chunk, would they understand the key point?” If the answer is no, restructure that section.

    Add a clear H3 heading, rewrite the opening sentence to front-load the key information, and ensure the section includes proper context. Do this for your entire piece, section by section. It’ll take 2-3 hours, but this single exercise will dramatically improve your content’s retrieval and citation rates.

    Connect this to your broader Season 2 learning by thinking about how RAG-aware patterns integrate with the entity graphs, schema stacks, and conversation patterns we’ve covered. It all works together to create content that wins in AI systems.

    Before we go, let’s leave you with this:

    RAG isn’t just another acronym to memorize. It’s the difference between content that hopes to be found and content that expects to be cited. When you chunk it right, front-load your answers, and build citation-worthy passages, you’re not just playing the SEO game anymore — you’re playing on the same field as the models themselves.

    Next week in Episode 2.5, we’re sliding straight into “Multimodal Evidence Design for LLMs” — how to make your images, charts, audio, and video sing the same song as your text so AI can pull proof from every corner of your content. It’s going to be pure class, and a wee bit wild.

    If this episode hit home, go back to Season 1, Episode 2: Question-Based Content: The Secret Sauce of AEO and the FAQ patterns we kept coming back to all season. That’s the rhythm section. RAG-aware content is the solo on top.

    Head over to AEODecoded.ai to join the newsletter. You’ll get:

    • The downloadable RAG optimization checklist
    • Behind-the-scenes breakdowns
    • And a few extra riffs I only share with subscribers

    And if you’ve got a question you want me to tackle on air — whether you’re a listener like Maya or a SaaS team trying to make sense of your analytics — send it to admin@irishguy.us with “AEO Decoded” in the subject line. The best ones make it into the Lightning Round.

    Alright, that’s the strategy talk done.

    I’m Gary Crossey, helping you make your content speak AI fluently — so your pages don’t just chase clicks, they earn the answer.

    Now, since we’ve been talking RAG all episode…

    it’s only right we close out with a little “RAG-time blues” of our own.

    Roll the tune — let’s chunk it right one last time.

  • Episode 2.3: Conversation Patterns and Follow-Up Funnels

    Hello my lovely listeners, welcome back to AEO Decoded. I’m your host, Gary Crossey, and I’m absolutely chuffed you’ve joined me for Episode 2.3 of Season 2!

    Today we’re diving into “Conversation Patterns and Follow-Up Funnels” – and if that sounds a wee bit technical, don’t worry. By the end of these 15 minutes, you’ll understand exactly how to map the natural flow of questions your audience asks and structure your content so AI systems can guide users deeper into your expertise.

    Over the 10 episodes of Season 2, we’re diving into advanced AEO strategies that separate good optimization from world-class optimization. And today’s topic? It’s absolutely critical because conversational AI systems don’t just answer one question and stop – they’re designed to keep the dialogue going, to anticipate follow-up questions, and to escalate depth when users want more detail.

    If you caught Season 1, you’ll remember Episode 6 where we explored Conversation Design – creating content for dialogue, not just display. We also covered Question-Based Content in Episode 2, where we learned to structure content around the specific questions your audience asks. Today, we’re taking those fundamentals and turning them into a sophisticated system that keeps AI assistants coming back to your content, question after question after question.

    Last week in Episode 2.2, we tackled Advanced Schema Stacks and Harmonization – making sure all those structured data layers work together without contradicting each other. Today, we’re focusing on the human side of AI interaction: understanding how conversations naturally flow and building content architecture that mirrors those patterns.

    This is my personal outlet because, truth be told, not many people are talking about advanced AEO yet – but they will be! So if you’re interested, please reach out. Your questions and experiences help shape this podcast into something truly valuable for our growing community.

    Today we’re diving deep into conversation patterns and follow-up funnels – stick with me for the next 15 minutes and you’ll walk away with strategies you can implement right away. Let’s get started!

    Hook/Story

    Right, let me tell you a wee story that perfectly illustrates why conversation patterns matter so much in AEO.

    A few months back, I was helping a client who runs a brilliant healthcare website – they’ve got fantastic content about various medical conditions, treatments, and wellness advice. Their traffic from traditional search was grand, but they noticed something peculiar: when people found their content through AI assistants, they’d get one answer and then… nothing. The conversation would end there, so it would.

    Meanwhile, their competitor – who honestly had less comprehensive content – was getting cited multiple times in the same conversation. Users would ask a follow-up question, and the AI would pull from that same competitor’s site again. And again. It was like the AI had developed a wee crush on their competitor’s content!

    So we did some digging, and here’s what we discovered: The competitor had mapped out natural question progressions. When someone asked “What causes migraines?”, they didn’t just answer that question in isolation. They anticipated the next natural questions: “How long do migraines typically last?”, “What’s the difference between a migraine and a regular headache?”, “What treatments are available?”, and “When should I see a doctor?”

    But here’s the brilliant bit – they didn’t just create separate articles for each question. They built a conversation flow with internal links that explicitly said things like “If you’re wondering about treatment options next, here’s what you need to know” or “Many people also ask about prevention strategies – let’s explore that.”

    The AI systems could follow these breadcrumbs, so to speak. They could escalate the conversation naturally, keeping users engaged while continuing to cite the same authoritative source. Pure dead brilliant, so it is!

    That’s what we’re building today – a system that doesn’t just answer isolated questions but anticipates and facilitates the entire conversation journey your audience wants to have.

    Overview

    So what exactly are conversation patterns and follow-up funnels, and why should you care about them in your AEO strategy?

    Think about how people actually search for information today. Nobody asks just one question and walks away satisfied – especially with complex topics. They ask a starter question, get an answer, then immediately think of two or three follow-up questions. It’s like peeling an onion, so it is – each layer reveals new questions underneath.

    Traditional SEO taught us to target individual keywords and rank for specific queries. That worked grand when people typed “best running shoes” into Google and clicked through a list of results. But conversational AI systems work completely differently. They’re designed to maintain context across multiple exchanges, to understand that “What about waterproof options?” relates back to the earlier question about running shoes, and to provide progressively deeper information as the conversation continues.

    This is where conversation patterns come in. A conversation pattern is essentially a map of the natural question progression your audience follows when learning about a topic. It’s the “next natural question” tree that branches out from any given starting point.

    And follow-up funnels? Those are the structured pathways you create in your content that guide users (and AI systems) through these question progressions, using micro-answers and strategic internal links to keep the conversation flowing and the citations pointing back to you.

    What makes this advanced rather than basic AEO is the systematic approach. In Season 1, we learned to answer individual questions clearly. Now we’re learning to architect entire conversation ecosystems – networks of interconnected content that serve as definitive resources for complete topic exploration.

    This fits into the bigger AEO picture because modern AI assistants are increasingly context-aware and conversation-focused. They’re not just retrieving isolated facts; they’re building narratives and guiding users through learning journeys. If your content can facilitate those journeys better than your competitors, you become the trusted source that AI systems return to again and again.

    The Breakdown

    Alright folks, it’s time for ‘The Breakdown’ – where we take those fancy-pants AI concepts and break them down into bite-sized morsels that won’t give you digital indigestion!

    Let’s start with the foundation: understanding how conversations actually flow in your subject area.

    1. Mapping Natural Question Progressions

    The first step in building effective conversation patterns is understanding the psychology of curiosity in your niche. Every topic has natural question sequences that people follow – and these aren’t random. They follow predictable patterns based on how humans learn and make decisions.

    There are typically four types of follow-up questions that emerge from any starting question:

    Clarification questions: These dig deeper into the original answer. If you answer “What is entity optimization?”, the natural clarification questions might be “How does entity optimization differ from keyword optimization?” or “What specific elements define an entity?”

    Application questions: These focus on implementation. After understanding what something is, people want to know how to do it. “How do I start with entity optimization?” or “What tools can help with entity optimization?”

    Context questions: These explore the broader landscape. “How does entity optimization fit into my overall AEO strategy?” or “What’s more important – entity optimization or structured data?”

    Consequence questions: These examine results and implications. “How long before I see results from entity optimization?” or “What happens if I don’t optimize for entities?”

    Your job is to map these question types for your core topics. Start with your most important content pieces and literally write out the follow-up questions in each category. I promise you, this exercise alone will transform how you think about content architecture.

    2. Creating Micro-Answers with Strategic Links

    Now here’s where the magic happens. Once you’ve mapped those question progressions, you don’t need to write a massive 5,000-word article that covers everything. Instead, you create what I call “micro-answers” – concise, direct responses that satisfy the immediate question while explicitly acknowledging the natural next questions.

    A micro-answer has three components. First, it provides a direct, clear response to the specific question – usually 100-200 words. Second, it includes context that helps AI systems understand how this answer relates to the broader topic. Third – and this is crucial – it explicitly signals what questions typically come next and provides deep links to those answers.

    For example, instead of just answering “What causes migraines?” and stopping there, you’d structure it like this:

    “Migraines are caused by a combination of genetic and environmental factors that affect blood flow and nerve signaling in the brain. [Direct answer with essential detail]

    Understanding the causes helps explain why certain treatments work better than others. [Context]

    Most people wondering about causes next want to know about duration and severity, treatment options, or prevention strategies. [Explicit signaling]”

    Then you provide clear internal links to each of those follow-up topics. The key is being explicit – don’t just rely on standard “related articles” links. Actually acknowledge the conversation flow in your content.

    3. Building Conversation Trees, Not Silos

    Traditional content strategy often creates silos – individual articles optimized for individual keywords with weak connections between them. Conversation pattern optimization requires thinking in trees rather than silos.

    A conversation tree has a trunk (your main topic), primary branches (major question categories), and smaller branches (specific follow-up questions). Every piece of content should know its place in the tree and provide pathways both up (to broader context) and out (to related branches).

    Here’s a practical example from the healthcare client I mentioned. Their migraine content tree looked like this:

    Trunk: “Understanding Migraines” (overview page) Branch 1: “Migraine Causes and Triggers” Sub-branches: Genetic factors, Environmental triggers, Hormonal influences Branch 2: “Migraine Symptoms and Types” Sub-branches: Migraine with aura, Chronic migraine, Hemiplegic migraine Branch 3: “Migraine Treatment Options” Sub-branches: Acute treatments, Preventive medications, Alternative therapies Branch 4: “Living with Migraines” Sub-branches: Lifestyle modifications, When to see a doctor, Emergency warning signs

    Every page in this tree includes explicit navigation that acknowledges conversation flow. The causes page doesn’t just link to treatments – it says “Once you understand what triggers your migraines, the next step is exploring treatment options that address your specific triggers.”

    4. Optimizing for Depth Escalation

    AI systems are getting better at understanding when users want surface-level information versus deep expertise. Your conversation patterns should accommodate both.

    Think of it like this: Some people want the Cliffs Notes version, while others want the full academic textbook. Your content should provide clear pathways for both, and AI systems should be able to identify which level is appropriate based on conversation context.

    This means creating content at multiple depth levels for the same topic and being explicit about those levels. Label beginner-friendly overviews, intermediate deep-dives, and advanced technical discussions. Use schema markup to signal content depth. Structure your internal linking so AI systems can escalate or de-escalate complexity based on user signals.

    5. Maintaining Attribution Through Conversation Flows

    Here’s the business reason this all matters: When AI systems can follow clear conversation pathways through your content, they’re more likely to continue citing you as the conversation progresses. Each additional citation reinforces your authority and increases the likelihood you’ll be referenced in future conversations on the same topic.

    Think of it like building trust in a real conversation. If someone gives you a good answer to your first question, you’re likely to ask them your second question too. AI systems work similarly – if your content successfully addresses the first query and explicitly facilitates the natural follow-up, you become the go-to source for the entire conversation thread.

    This is why isolated, comprehensive articles sometimes perform worse in conversational AI than networks of focused, interconnected pieces. The network structure mirrors how conversations actually unfold, making it easier for AI systems to traverse and cite multiple times.

    Practical Implementation

    Now let’s get practical about how you actually implement conversation pattern optimization in your content strategy.

    Step 1: Audit Your Top Content for Conversation Gaps

    Start by identifying your 10-15 most important content pieces. For each one, ask yourself: What’s the most common next question someone would have after reading this? Then check – do you have content that answers that question? Is it clearly linked from the original piece? If not, you’ve found a conversation gap.

    Step 2: Create Question Flow Maps

    For your core topics, literally draw out the question flow on paper or in a tool like Miro or Lucidchart. Put your primary question in the center, then branch out with natural follow-ups. Keep going until you’ve mapped 2–3 levels of depth. This visual map becomes your content architecture blueprint.

    Step 3: Write or Revise Content with Explicit Signaling

    As you create new content or update existing pieces, be explicit about conversation flow. Use phrases like “Most people next want to know about…” or “The natural follow-up question is…” or “If you’re wondering about X, here’s what you need to understand…”

    Don’t be subtle about this! AI systems benefit from explicit signals about information relationships.

    Step 4: Implement Strategic Internal Linking

    Your internal links should tell a story about how concepts connect. Instead of generic “learn more” links, use descriptive anchor text that acknowledges the conversation flow: “Explore treatment options for migraine prevention” or “Understand the difference between migraine types.”

    Step 5: Test with AI Assistants

    This is crucial – actually test your conversation flows with ChatGPT, Claude, or Perplexity. Ask the initial question, see what gets cited, then ask natural follow-ups. Does the AI continue citing your content? If not, where does the conversation thread break? Those break points show you where to strengthen your content connections.

    From working with Method Q clients, I can tell you the timeline for seeing results varies. Usually within 4-6 weeks of implementing strong conversation patterns, you’ll notice AI systems citing your content more frequently and across multiple turns in the same conversation. The key is consistency – don’t just optimize one piece; build the entire conversation ecosystem.

    Common pitfalls to avoid: Don’t create circular links that send users in loops. Don’t over-optimize with too many internal links that become distracting. And don’t force unnatural connections just to build links – the conversation flow should always feel organic and helpful.

    Q&A Lightning Round

    Now, let’s tackle some common questions about conversation patterns and follow-up funnels:

    Q: How many follow-up questions should I anticipate for each piece of content?

    A: Focus on the 3–5 most natural next questions. Going deeper than that can create overwhelming complexity. Remember, you’re mapping natural curiosity patterns, not trying to anticipate every possible question in the universe. Quality over quantity, so it is.

    Q: Should I create separate pages for each follow-up question or include everything on one page?

    A: It depends on complexity. For simple topics where follow-ups are quick, one comprehensive page works grand. For complex topics where follow-ups require substantial explanation, separate interconnected pages perform better. The key is making the structure match the natural learning progression.

    Q: How do I avoid duplicate content when addressing related questions on multiple pages?

    A: Focus each page on a specific aspect while providing unique value. Use the “hub and spoke” model – a central comprehensive piece with shorter, focused pieces that go deeper on specific angles. Each should have a distinct purpose and primary question it addresses.

    Q: What if my topic doesn’t have obvious follow-up questions?

    A: Every topic has follow-ups! If you’re stuck, look at “People Also Ask” boxes in Google, check question forums like Quora or Reddit in your niche, or analyze your own site search queries and customer service questions. The follow-ups are there – you just need to discover them.

    Q: How does this work with voice search and smart speakers?

    A: Brilliantly! Voice assistants are inherently conversational, so content optimized for conversation patterns performs especially well. Voice users are likely to ask multiple related questions in sequence, making your conversation architecture even more valuable.

    Q: Can I apply this to product pages and e-commerce content?

    A: Absolutely! Product conversations follow patterns like: What is it? → How does it work? → What makes it better than alternatives? → What do I need to use it? → How much does it cost? → What do other customers say? Map these for your products and watch your AI visibility improve.

    The encouraging news is this: Once you’ve built strong conversation patterns for your core topics, maintaining them becomes much easier. You’re creating a self-reinforcing system where each new piece naturally fits into the existing conversation architecture.

    Actionable Takeaway

    Let’s wrap it up with the takeaway section. This section will give you that one actionable item you can work on.

    Here’s your action item for the next week: Choose your single most important piece of content and create a conversation flow map for it. Identify the top 5 natural follow-up questions, then check if you have quality content addressing each one. If not, add those to your content calendar. If yes, update your original piece to explicitly acknowledge those follow-ups and include strategic internal links with conversation-aware anchor text.

    This exercise should take 1-2 hours but will give you immediate insight into where your conversation architecture is strong and where it needs strengthening. More importantly, you’ll start seeing your content through the lens of AI conversation patterns rather than isolated keyword targets – and that perspective shift is absolutely transformative.

    Closing & Promotion

    Next week in Episode 2.4, we’re diving into “RAG-Aware Content Patterns” – exploring how LLMs ingest, chunk, embed, and cite passages, and how to structure your content so it survives chunking and wins retrieval. It’s going to be class altogether!

    Enjoyed this episode? For foundations on this topic, revisit Season 1: Episode 6 on Conversation Design and Episode 2 on Question-Based Content. Those episodes lay the groundwork that today’s advanced strategies build upon.

    Don’t forget to visit aeodecoded.ai and sign up for our newsletter for exclusive resources and bonus content. We’ve got templates, checklists, and deeper dives that complement these podcast episodes.

    Subscribe for weekly drops, and submit your questions via the Q&A form at aeodecoded.ai. I’ll feature select questions in the Q&A lightning round. Your questions make this podcast better for everyone!

    Thanks for spending these 15 minutes with me. Until next time, I’m Gary Crossey, helping you make your content speak AI fluently. May your content always earn answers, not just clicks!

  • Episode 2.2: Advanced Schema Stacks and Harmonization

    Hello my lovely listeners, welcome back to AEO Decoded. I’m your host, Gary Crossey.

    Today we’re tackling something that sounds dead technical but is absolutely crucial for your AEO success: Advanced Schema Stacks and Harmonization. Now, I know what you’re thinking – “Gary, that sounds like something out of a computer science textbook!” But stick with me, because this is where the magic happens, so it is.

    Over the 10 episodes of Season 2, we’re diving into advanced AEO strategies that separate good optimization from world-class optimization. Last week in Episode 2.1, we explored entity graphs and how to build machine-readable authority. This week, we’re building on that foundation to look at how you layer multiple schema types together without creating a confused mess that AI parsers can’t understand.

    If you remember back to Season 1, Episode 3 on Structured Data, we covered the basics of making your content AI-friendly with schema markup. Today, we’re taking that to the next level – we’re talking about stacking Article with FAQ with HowTo with Breadcrumb, adding Organization and Person schemas, and even creating custom JSON-LD that all works together in perfect harmony.

    I’m happy to be bringing you Season 2 of AEO Decoded. Since launching in July 2025, we’ve built an incredible global community – listeners in over 30 countries, and folks consuming episodes within the first hour of release. This has grown from my personal outlet into something much bigger, and I’m grateful for every one of you who’s joined this journey. In my work with Method Q in Atlanta, I’m seeing firsthand how these strategies translate into real results for clients, and I’m excited to share those advanced insights with you throughout this season. So if you’re interested in taking your AEO practice to the next level, you’re in exactly the right place!

    Today we’re diving deep into schema harmonization – stick with me for the next 15 minutes and you’ll walk away with strategies you can implement right away to make your structured data sing like a proper choir instead of a confused jumble of notes.

    Let me tell you about something that happened a few months back. I was working with a client – brilliant content team, really knew their stuff – and they’d gone all-in on schema markup. And I mean all-in. They had Article schema, FAQ schema, HowTo schema, BreadcrumbList, Organization, Person – the works, so they did.

    On paper, it looked class. They were ticking every box. But here’s the thing: when we ran it through Google’s Rich Results Test and some AI parser tools, it was throwing more errors than a drunk fella trying to parallel park outside a Belfast pub on a Saturday night.

    The problem? They had contradictory data all over the shop. The Article schema said the author was “Marketing Team,” but the Person schema had a completely different name. The FAQ items referenced entities that didn’t exist in their Organization schema. The HowTo steps linked to URLs that the Breadcrumb schema said were in a completely different site structure. It was pure chaos.

    And here’s the kicker – to human eyes, everything looked grand. The page displayed perfectly. But to an AI trying to understand and cite this content? It was like trying to follow directions from three different people all shouting different things at once.

    That’s when it clicked for them – and hopefully for you too. Having schema markup isn’t enough anymore. You need harmonized schema markup. You need your different schema types working together like a well-rehearsed band, not like a bunch of musicians all playing different songs at the same time.

    This isn’t just about avoiding errors, mind you. When LLMs and answer engines are deciding which content to trust and cite, they’re looking at the coherence of your structured data. Contradictions and orphaned nodes? They’re red flags that make AI parsers less likely to use your content as a source.

    Overview

    So what exactly do we mean by “schema stacks and harmonization”? Let’s break down this fancy terminology into something you can actually work with.

    A schema stack is when you layer multiple types of structured data on the same page. Think of it like building a lasagna – you’ve got your Article schema as your base layer, then you add FAQ schema for common questions, maybe HowTo schema for step-by-step guides, Breadcrumb schema for navigation context, and Organization or Person schema to establish authority. Each layer adds different information that helps AI understand your content better.

    Harmonization is making sure all those layers actually work together without contradicting each other or leaving gaps. It’s ensuring that when your Article schema mentions “John Smith” as the author, your Person schema has matching information about John Smith. It’s making sure your FAQ items reference entities that actually exist in your Organization schema. It’s confirming that your internal links match what your Breadcrumb schema says about your site structure.

    This matters more now than ever because we’re not just optimizing for Google’s traditional search anymore. We’re optimizing for ChatGPT, Claude, Perplexity, Google’s AI Overviews, and dozens of other AI systems that are trying to understand and synthesize information from across the web. These systems are sophisticated enough to spot contradictions and coherence issues – and they’ll pass over your content if it doesn’t add up.

    In Season 1, we learned the basics of adding structured data to your pages. Today, we’re learning how to orchestrate multiple types of structured data into a coherent, machine-readable narrative about your content. We’re learning how to avoid the pitfalls that confuse parsers, and how to create custom JSON-LD that extends standard schemas when you need something more specific.

    By the end of this episode, you’ll understand how to audit your existing schema stacks, identify contradictions and orphaned nodes, and implement a harmonization strategy that makes your content a dream for AI parsers to understand and cite.

    The Breakdown

    Alright folks, it’s time for ‘The Breakdown’ – where we take those fancy-pants AI concepts and break them down into bite-sized morsels that won’t give you digital indigestion! Now, I’ve broken this down into a five-step system that’ll make schema harmonization much more manageable. Think of it as your roadmap from messy to harmonized – each step builds on the last, and by the end, you’ll have schema stacks that work together like a well-oiled machine. Let’s dive into each one.

    1. Understanding Common Schema Stack Patterns

    Let’s start with the most common schema combinations you’ll want to use, and why they work together.

    The Article + FAQ + Organization stack is your bread and butter for informational content. Your Article schema describes the main content piece – title, author, date published, date modified, description. Your FAQ schema sits inside it, marking up your Q&A sections so AI can extract specific question-answer pairs. And your Organization schema establishes who’s publishing this content and why they’re authoritative.

    Here’s the crucial bit: these need to reference each other consistently. If your Article says it’s published by “Acme Corporation,” your Organization schema better have “Acme Corporation” as the name, not “Acme Corp” or “ACME” or some variation. Consistency matters, so it does.

    The HowTo + Article + Breadcrumb stack is brilliant for instructional content. Your HowTo schema marks up your steps in a way that AI can understand the sequence and dependencies. Your Article schema provides the broader context. And your Breadcrumb schema shows where this content lives in your site hierarchy, helping AI understand topical relationships.

    The key here is making sure your HowTo steps don’t reference tools, materials, or concepts that aren’t defined elsewhere in your schema or content. If step 3 mentions “the mixing bowl from step 1,” your step 1 better explicitly mention that mixing bowl. Don’t make the AI guess.

    The Product + Review + Organization stack is essential for e-commerce or product-focused content. But here’s where folks often mess up – they have Review schema that references a product with a different name than what’s in their Product schema. Or they have aggregate ratings that don’t match the individual reviews. These contradictions are death for AI trust.

    2. The Principle of Single Source of Truth

    This is where most schema implementations fall apart, and it’s a concept we need to drill into your head: every entity should have one canonical definition, and everything else should reference that definition.

    Let’s use a real example. Say you’re writing about “Dublin City Centre.” In one schema block, you call it “Dublin City Centre.” In another, you call it “Dublin City Center” (American spelling). In a third, you call it “Dublin Downtown.” To a human, these are obviously the same place. To an AI parser? These look like three different entities, and now your content seems confused about its own subject matter.

    The solution is to pick one canonical name – let’s say “Dublin City Centre” – and use it consistently across all your schema markup. If you need to mention alternative names, use the “alternateName” property within your schema to explicitly mark them as alternatives to the same entity.

    This applies to people, places, organizations, products, concepts – anything that might be mentioned in multiple places. Define it once with full detail, then reference that definition consistently everywhere else.

    Here’s a pro tip from working with Method Q clients: create an internal entity registry. It’s literally just a spreadsheet or document where you list every entity you regularly reference in your content, with its canonical name, identifier (if it has one, like a Wikipedia URL), and any alternative names. Then your content creators can reference this registry when they’re implementing schema markup. Dead simple, massively effective.

    3. Avoiding Contradictory Data

    Contradictions are the silent killer of schema effectiveness. They’re easy to introduce and hard to spot unless you’re specifically looking for them.

    Common contradictions include:

    • Date mismatches: Your Article schema says published date is January 1, but your webpage’s visible date says January 15. Pick one and use it consistently.
    • Author attribution differences: Schema says “John Smith,” byline says “J. Smith,” About section says “John P. Smith.” These need to match exactly, or you need to explicitly mark them as variants of the same person.
    • URL inconsistencies: Your Breadcrumb schema shows a URL structure that doesn’t match your actual URLs, or your FAQ answers link to URLs that aren’t properly represented in your site schema. AI parsers notice this and get confused.
    • Numerical conflicts: Your Product schema says price is €50, but your structured data for the shopping cart says €55. Your Review schema shows 47 reviews, but your aggregate rating says “based on 52 reviews.” These conflicts tank trust.

    The way to catch these is through systematic validation. Don’t just validate that your schema is technically correct – validate that it’s internally consistent. That means actually reading through all your schema markup for a given page and checking that the same entities are described the same way every time they appear.

    4. Handling Orphaned Nodes

    An orphaned node is a piece of schema that references something that doesn’t exist anywhere else in your structured data or isn’t clearly defined in your content. It’s like saying “as we mentioned before” when you never actually mentioned it before – it confuses the AI and makes your content seem less authoritative.

    Common orphaned nodes:

    • References to undefined entities: Your FAQ answer mentions “our flagship product” but there’s no Product schema defining what that product is.
    • Links without context: Your HowTo step links to a tool or resource, but there’s no schema explaining what that resource is or why it’s relevant.
    • Dangling relationships: Your Person schema says someone “worksFor” an organization, but there’s no Organization schema for that organization on your site.

    The fix is to ensure every entity you reference is either fully defined in schema somewhere on your site, or is clearly identified with an external reference (like a Wikipedia URL or Wikidata ID) that AI parsers can look up.

    For entities that appear frequently across your site, consider creating dedicated pages for them with full schema markup, then reference those pages consistently. For example, if you frequently mention your company’s CEO, create a proper About page for them with complete Person schema, then reference that same Person identifier in every Article or content piece where they’re mentioned as author or subject matter expert.

    5. Custom JSON-LD for Advanced Cases

    Sometimes the standard schema types don’t quite cover what you need. That’s when you extend into custom JSON-LD – but you need to do it carefully so you don’t break the harmony we’ve been building.

    The trick with custom JSON-LD is to extend, don’t replace. You start with standard schema types as your foundation, then add custom properties that provide additional context specific to your domain.

    For example, say you’re in healthcare and you need to mark up clinical trial data. Schema.org has MedicalTrial, but maybe you need more granular data about trial phases, participant demographics, or specific endpoints. You can extend the base MedicalTrial schema with custom properties, but you keep the standard properties intact so general AI parsers can still understand the basics.

    The key is using standard vocabulary wherever possible, and only going custom when you absolutely need to. And when you do go custom, document it clearly and use consistent property names across your entire site.

    Practical Implementation

    Now let’s get practical about how you actually implement harmonized schema stacks on your site.

    Step 1: Audit your current schema implementation. Use Google’s Rich Results Test, Schema.org validator, and general JSON-LD validators to check every major template on your site. But don’t stop at technical validation – actually read through the schema and check for the consistency issues we’ve discussed.

    Step 2: Create your entity registry. List out every entity you regularly reference: your organization, key people, main products or services, locations, concepts. For each one, document the canonical name, any identifiers (URLs, IDs), and alternative names. This becomes your reference guide.

    Step 3: Define your standard schema stacks per content type. Article pages get this stack, product pages get that stack, how-to guides get this other stack. Document these standards so everyone on your team knows what to implement.

    Step 4: Implement with consistency checks. As you add or update schema markup, cross-reference your entity registry to ensure you’re using canonical names and identifiers. Set up a review process where someone checks new schema against existing schema for consistency before publishing.

    Step 5: Test not just validity, but coherence. After implementation, don’t just validate that your schema is technically correct. Load your page into AI tools like ChatGPT or Claude and ask them questions about your content. If the AI seems confused or gives wrong information, that’s often a sign of schema inconsistencies or contradictions.

    Tools you’ll want: Google’s Rich Results Test (free), Schema.org validator (free), a JSON-LD validator (free), and ideally a schema visualization tool that can show you how your different schema types relate to each other – there are several browser extensions and online tools for this.

    Common pitfalls to avoid: Don’t just copy-paste schema from different sources without checking for consistency. Don’t use automated schema generators without reviewing the output. Don’t set up schema markup once and never review it – as your content and site evolve, your schema needs to evolve with it while maintaining consistency.

    Timeline expectations: If you’re starting from scratch, plan 2-3 weeks to audit, set up your entity registry, and define your standards. Then it’s ongoing maintenance – maybe 1-2 hours per month to review new content and ensure consistency. But the payoff is enormous: coherent, harmonized schema can dramatically increase your chances of being cited by AI systems.

    Q&A Lightning Round

    Now, let’s tackle some common questions about schema stacks and harmonization:

    Q: How many schema types can I stack on one page before it becomes too much?

    A: There’s no hard limit, but I typically recommend 3-5 schema types per page as a sweet spot. More than that and you risk introducing inconsistencies, plus you’re probably over-engineering it. Focus on the schema types that add the most value for your content type. An article page might have Article, FAQ, Organization, Person, and Breadcrumb – that’s five and it’s grand. But don’t add Product schema just because you can if you’re not actually describing a product.

    Q: What if different sections of my organization use different names or data formats?

    A: This is common in larger organizations, and it’s exactly why you need that entity registry I mentioned. Work with stakeholders across departments to agree on canonical entity definitions, then enforce those consistently in schema markup even if different departments use different terminology internally. The external-facing schema needs to be consistent, so it does.

    Q: Should I use JSON-LD, Microdata, or RDFa for my schema markup?

    A: JSON-LD, hands down. It’s easier to manage, easier to keep consistent, and it’s Google’s recommended format. It also separates your schema from your HTML, which makes it much easier to audit and update without breaking your page layout. All the major AI systems support JSON-LD well.

    Q: How do I handle schema for dynamic content that changes frequently?

    A: You need schema that updates dynamically too. If you’re using a CMS, set up templates that generate schema automatically from your content database, pulling from those canonical entity definitions we talked about. Don’t manually write schema for frequently-changing content – that’s a recipe for inconsistencies.

    Q: What’s the best way to test if my schema harmonization is actually working?

    A: Three-pronged approach: First, use technical validators to catch structural issues. Second, use AI chat tools to ask questions about your content and see if they get confused. Third, monitor your appearance in AI-generated answers and rich results – if you’re showing up consistently and accurately, your schema harmony is working. If AI systems misquote you or seem confused about basic facts, you likely have consistency issues to fix.

    Q: Is it worth creating custom JSON-LD properties, or should I stick to standard schema.org vocabulary?

    A: Stick to standard vocabulary unless you have a specific, compelling reason to go custom. Standard schema.org properties are understood by all major AI systems. Custom properties might be ignored or misunderstood. Only extend into custom territory if you’re marking up domain-specific data that has no standard equivalent, and even then, use standard properties for the core information.

    Remember, the goal isn’t perfection – it’s coherence. A simpler schema stack that’s perfectly harmonized will outperform a complex stack with contradictions every single time.

    Actionable Takeaway

    Let’s wrap it up with the takeaway section. This section will give you that one actionable item you can work on.

    Here’s your action item for the next week: Audit one major content template on your site for schema contradictions.

    Pick your most important template – probably your blog article or main product page template. Pull up the schema markup for it. Now go through systematically and check:

    • Do all entity names match exactly everywhere they appear?
    • Do dates and numbers align with what’s visible on the page?
    • Does every entity referenced have a clear definition somewhere?
    • Do your different schema types tell the same story about the content?

    Document any inconsistencies you find. That’s your harmonization fix list. Fix those contradictions, and you’ll immediately have more coherent, trustworthy structured data that AI systems can parse with confidence.

    Just one template this week – but do it thoroughly. That’s how you build the foundation for proper schema harmonization across your entire site.

    Closing Comments

    Next week in Episode 2.3, we’re diving into Conversation Patterns and Follow-Up Funnels – how to map those “next natural question” trees and structure your content so conversational AI systems can keep escalating depth while keeping you as the cited source. It’s going to be class altogether.

    Enjoyed this episode? For foundations on this topic, revisit Season 1, specifically Episode 3: Structured Data: Making Your Content AI-Friendly, where we covered the basics that everything we talked about today builds upon.

    Don’t forget to visit AEODecoded.ai and sign up for our newsletter for exclusive resources and bonus content, including schema templates and harmonization checklists.

    Subscribe for weekly drops, and send questions to admin@irishguy.us. I’ll feature select questions in the Q&A lightning round.

    Thanks for spending these 15 minutes with me. Until next time, I’m Gary Crossey, helping you make your content speak AI fluently. May your content always earn answers, not just clicks!