<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Stories by Alden Do Rosario on Medium]]></title>
        <description><![CDATA[Stories by Alden Do Rosario on Medium]]></description>
        <link>https://medium.com/@aldendorosario?source=rss-841a3b5d13d5------2</link>
        <image>
            <url>https://cdn-images-1.medium.com/fit/c/150/150/1*d-NreRTsd_u3enyKgyJ_4A.png</url>
            <title>Stories by Alden Do Rosario on Medium</title>
            <link>https://medium.com/@aldendorosario?source=rss-841a3b5d13d5------2</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Fri, 15 May 2026 08:40:58 GMT</lastBuildDate>
        <atom:link href="https://medium.com/@aldendorosario/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[No, PageIndex Will Not “Kill” RAG, But It Is Indeed Excellent In Some Cases]]></title>
            <link>https://medium.com/@aldendorosario/no-pageindex-will-not-kill-rag-but-it-is-indeed-excellent-in-some-cases-11bc67473145?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/11bc67473145</guid>
            <category><![CDATA[vector-database]]></category>
            <category><![CDATA[rags]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Sat, 31 Jan 2026 13:27:29 GMT</pubDate>
            <atom:updated>2026-01-31T13:27:29.824Z</atom:updated>
            <content:encoded><![CDATA[<p><em>An independent benchmark revealing when tree-based RAG outperforms vector RAG — and when it can’t even be used</em></p><p>A <a href="https://x.com/_avichawla/status/1882365488498393262">viral tweet</a> recently claimed that <a href="https://github.com/VectifyAI/PageIndex">PageIndex</a>, a new open-source “reasoning-based RAG” system, achieved 98.7% accuracy on a financial benchmark without vector databases, chunking, or similarity search.</p><p>The AI community took notice. Some called it a “RAG killer.”</p><p>I spent the past week trying to benchmark PageIndex against leading RAG providers. The results tell a more nuanced story — and reveal a fundamental limitation that no one is talking about.</p><h3>What Is PageIndex?</h3><p><a href="https://github.com/VectifyAI/PageIndex">PageIndex</a> by VectifyAI takes a fundamentally different approach to document retrieval. Instead of the standard chunk-embed-retrieve pipeline, it:</p><ol><li>Builds a <strong>hierarchical tree index</strong> (like a semantic table of contents)</li><li>Uses <strong>LLM reasoning</strong> to navigate the tree and find relevant sections</li><li>Extracts content from identified sections for answer generation</li></ol><p>The idea is compelling: similarity search finds <em>similar</em> content, but reasoning finds <em>relevant</em> content.</p><p>When a question asks for a certification <em>date</em>, similarity search might return a certifications <em>table</em> — related but useless. Tree-based reasoning can navigate to the timeline section instead.</p><p>VectifyAI’s Mafin 2.5, powered by PageIndex, <a href="https://github.com/VectifyAI/Mafin2.5-FinanceBench">achieved 98.7% accuracy on FinanceBench</a>. But FinanceBench tests single-document question answering — each question targets a specific financial report.</p><blockquote>The question is: <strong>what happens when you have 1000 documents?</strong></blockquote><h3>The Scalability Problem</h3><p>Here’s what I confirmed through testing: <strong>PageIndex’s tree-based approach cannot practically scale to multi-document scenarios.</strong></p><p>It’s great for single-document use cases (for example: Financial documents), but falls short on large multi-document knowledgebases.</p><p>In our testing using Google’s simpleqa-verified dataset (a 1000-question benchmark dataset that has about 2795 documents ), building the index ran into major scalability problems.</p><p>Due to which: We had to fall back to standard vector search — the same approach it claims to replace.</p><p>PageIndex’s team has been transparent about this. In a <a href="https://x.com/PageIndexAI/status/2016913668114698341?s=20">public exchange</a> on X (Twitter), their official account noted that PageIndex is currently designed for single long document question answering, and that for multiple documents (more than 5), they support via other customized techniques.</p><p>They also acknowledged that the open-source version uses a sequential indexing process intended more as a proof of concept than an enterprise-ready system.</p><h3>The Benchmark: 100 Questions, 1000 Documents</h3><p>To evaluate PageIndex in a multi-document scenario, I tested what actually happens at scale: FAISS vector retrieval (the fallback when tree indices aren’t available) followed by GPT-5.1 answer generation.</p><p>I compared this against three commercial RAG providers, all answering the same 100 questions from <a href="https://huggingface.co/datasets/google/simpleqa-verified">SimpleQA-Verified</a> across 2,795 source documents:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*lJuwFsy132_g3Ra8xrqzLw.png" /></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/810/1*GNgXHyeMWKjf0_gnSUzIXg.png" /></figure><h3>Scoring</h3><blockquote>Quality = (correct - 4 x incorrect) / total</blockquote><p>The 4x penalty for incorrect answers reflects a design choice that favors precision over recall: a confident wrong answer costs four times as much as a correct answer earns.</p><p>This benefits conservative systems that abstain when uncertain. With a different penalty ratio, rankings would change.</p><p><strong>Note:</strong> These results are based on a 100-question sample from SimpleQA-Verified, not the full 1,000-question benchmark. With this sample size, the rankings should be treated as directional indicators rather than statistically definitive. The difference between adjacent providers (e.g., CustomGPT at 0.78 vs PageIndex at 0.69) may not be statistically significant at n=100.</p><p>When the pipeline does answer, it achieves <strong>96.4% accuracy</strong> (81 out of 84 attempted).</p><h3>The Core Trade-off</h3><p>These results reveal a fundamental trade-off in PageIndex’s design:</p><h3>Single-Document: Designed to Excel</h3><p>PageIndex is built for single-document deep analysis. When it can use tree-based reasoning on a known document, the structural navigation genuinely finds information that similarity search misses. PageIndex’s own FinanceBench results demonstrate this capability.</p><h3>Multi-Document: Falls Back to Standard RAG</h3><p>When PageIndex faces hundreds or thousands of documents, it can’t build tree indices fast enough. It falls back to FAISS vector search — and performs like any other vector RAG system, without the structural reasoning that makes it special.</p><p>This is the core insight: <strong>PageIndex’s strength (tree reasoning) is exactly the thing that can’t scale.</strong></p><h3>Where PageIndex Genuinely Excels</h3><p>Despite the multi-document limitations, PageIndex’s approach has real value:</p><p><strong>Single-document deep analysis.</strong> For financial reports, legal filings, technical manuals — any scenario where you know which document to search — tree-based reasoning navigates complex structure better than chunk-based similarity search.</p><p><strong>Structured documents.</strong> Documents with natural hierarchy (sections, subsections, numbered items) play to PageIndex’s strengths. The tree index mirrors the document’s own structure.</p><p><strong>Auditability.</strong> Every retrieval decision is traceable — which tree nodes were considered, which were selected, and why. This matters for compliance-heavy domains.</p><p><strong>Principled abstention.</strong> PageIndex says “I don’t know” rather than guessing wrong — a valuable property for high-stakes applications.</p><h3>The Honest Takeaway</h3><p>PageIndex will not “kill” RAG. Its core technology (tree reasoning) can’t scale to the multi-document retrieval scenarios where most RAG systems operate.</p><p>But PageIndex <em>is</em> excellent in some cases. For high-stakes, single-document analysis — legal review, financial due diligence, regulatory compliance — the combination of structural reasoning and principled abstention is genuinely valuable.</p><p>The real future likely involves <strong>hybrid approaches</strong>: vector retrieval for document discovery, tree-based reasoning for precise extraction within top candidates. PageIndex has demonstrated that LLM reasoning over document structure can outperform similarity search for within-document retrieval. That’s a meaningful contribution.</p><p>Not a RAG killer. But a valuable tool for specific, high-stakes use cases.</p><h3>Methodology &amp; Reproducibility</h3><p>Full benchmark code, data, and results are published at: <a href="https://github.com/adorosario/pageindex-rag-benchmark"><strong>github.com/adorosario/pageindex-rag-benchmark</strong></a></p><h3>Technical Details</h3><ul><li><strong>Questions</strong>: 100 from <a href="https://huggingface.co/datasets/google/simpleqa-verified">SimpleQA-Verified</a> (factual, single-answer)</li><li><strong>Documents</strong>: 2,795 indexed in FAISS (text-embedding-3-small, 81,868 chunks)</li><li><strong>Answer model</strong>: GPT-5.1 (temperature=0) for PageIndex fallback pipeline; each commercial provider uses its native model</li><li><strong>Judge</strong>: GPT-4.1-mini using the <a href="https://github.com/openai/simple-evals">simple-evals grader template</a></li><li><strong>Scoring</strong>: Quality = (correct — 4 x incorrect) / total (penalty_ratio=4.0)</li></ul><p><strong>Note</strong>: The 4x penalty ratio is a design choice that favors precision-oriented systems. Rankings change under different penalty ratios:</p><p>At 1x penalty (no extra punishment for wrong answers), OpenAI RAG (0.81) would rank 3rd ahead of PageIndex (0.78).</p><h3>Limitations</h3><ul><li><strong>Sample size</strong>: This benchmark uses 100 questions from the 1,000-question SimpleQA-Verified dataset. Results are directional indicators, not statistically definitive — differences between adjacent providers may not be significant at this sample size</li><li><strong>Each provider uses its own answer model</strong>, making this an end-to-end provider comparison rather than a retrieval-only comparison</li><li><strong>I designed the benchmark methodology</strong> and selected the providers, scoring formula, and penalty ratio using techniques from OpenAI’s recent “Why LLMs Hallucinate” paper <a href="https://arxiv.org/abs/2509.04664">https://arxiv.org/abs/2509.04664</a> — PS: Read the <a href="https://openai.com/index/why-language-models-hallucinate/">blog post</a>, its a must-read.</li></ul><h3>Disclosure</h3><p>I am the Founder of <a href="https://customgpt.ai">CustomGPT.ai</a>, one of the evaluated providers. I selected the providers, designed the benchmark methodology, and ran all evaluations. Full audit data is published for transparency so readers can verify the results independently.</p><p>While there might be perception of bias, this is also based on 3-years experience (and 10,000+ paid customers) dealing with RAG.</p><p><em>Alden Do Rosario is Founder of </em><a href="https://customgpt.ai"><em>CustomGPT.ai</em></a><em>. The full benchmark code and results are available at </em><a href="https://github.com/adorosario/pageindex-rag-benchmark"><em>github.com/adorosario/pageindex-rag-benchmark</em></a><em>.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=11bc67473145" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Why Your AI Agent Fails at Document Analysis (And What We Built Instead)]]></title>
            <link>https://ai.plainenglish.io/why-your-ai-agent-fails-at-document-analysis-and-what-we-built-instead-73addfc59147?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/73addfc59147</guid>
            <category><![CDATA[ai-agent]]></category>
            <category><![CDATA[retrieval-augmented-gen]]></category>
            <category><![CDATA[enterprise-software]]></category>
            <category><![CDATA[document-management]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Thu, 08 Jan 2026 19:03:52 GMT</pubDate>
            <atom:updated>2026-01-16T18:17:22.525Z</atom:updated>
            <content:encoded><![CDATA[<p>I’ve watched hundreds of enterprise teams try to use ChatGPT and Claude for document analysis.</p><p>They upload a contract. They get a summary. They’re impressed.</p><p>Then they ask a real question: “Does this match our standard terms?”</p><p>And the AI has nothing useful to say.</p><p>This isn’t a failure of the AI. It’s a failure of approach. Generic AI gives generic answers because it doesn’t have your context.</p><p>After building AI solutions for enterprises over the past three years at CustomGPT.ai, we finally closed this gap.</p><p>We call it <a href="https://customgpt.ai/document-analyst/?utm_source=medium&amp;utm_medium=social&amp;utm_campaign=document-analyst-launch-jan26&amp;utm_content=founder-post"><strong>Document Analyst</strong></a>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*xcP0VlkJVG2S6RxqjHndgQ.png" /><figcaption>Analyzing photos against building code regulations</figcaption></figure><h3><strong>The Summarization Trap</strong></h3><p>Here’s what generic AI does well: summarize, extract key points, categorize information. If you upload a contract, it’ll tell you the parties involved, the key dates, the main obligations.</p><p>That’s useful. It’s also table stakes.</p><p>Here’s what generic AI can’t do:</p><p>- Check if a vendor agreement matches your standard terms</p><p>- Flag compliance issues against YOUR regulatory guidelines</p><p>- Compare a resume against YOUR hiring criteria</p><p>Ask ChatGPT any of these questions and you get… silence. Or worse, a generic answer that sounds helpful but isn’t. Maybe even a hallucination.</p><p>Why? Because ChatGPT doesn’t know your standard terms. It doesn’t know your regulatory guidelines. It doesn’t know your hiring criteria.</p><p>The AI is doing exactly what it’s designed to do: analyze the document in front of it. The problem is that real document analysis requires context that exists outside that document.</p><p>A compliance team uploads marketing materials. ChatGPT flags generic issues — “Consider checking industry regulations.” But it doesn’t know your specific regulatory framework. It doesn’t know that the FTC changed rules last quarter. It doesn’t know your internal approval checklist.</p><p>So after the AI “helps,” the team still has to do the real work. Pull up the guidelines. Cross-reference manually. Ask three colleagues for context. Spend half a day on what should take minutes.</p><p>That’s the summarization trap. The AI looks helpful. The work doesn’t actually get done faster.</p><h3><strong>Retrieval vs. Analysis</strong></h3><p>If you’re building enterprise AI, you’ve probably implemented RAG - Retrieval Augmented Generation. Your AI can search your knowledge base and pull relevant information.</p><p>That’s necessary. It’s not sufficient.</p><p>Retrieval answers: “<em>What do we know about X?</em>”</p><p>Analysis answers: “<em>How does this new document compare to what we know?</em>”</p><p>The difference is subtle but critical. Retrieval finds information. Analysis reasons across information.</p><p>When your legal team uploads a vendor contract, they don’t want to search your knowledge base. They want the AI to automatically compare that contract against your templates, your standards, your past agreements — and tell them what’s different, what’s missing, what’s concerning.</p><p>That requires the AI to hold two things in context simultaneously: the uploaded document AND your institutional knowledge. Then reason across both.</p><p>The breakthrough we made wasn’t better summarization. It was recognizing that the uploaded document isn’t the only input. Your institutional knowledge is the lens through which that document should be analyzed.</p><p>Same contract uploaded to ChatGPT and to your Document Analyst-enabled agent produces completely different outputs. ChatGPT gives you a summary anyone could get. Your agent gives you insights specific to your business.</p><p>Generic AI gives generic answers. <strong>Your agent gives YOUR answers.</strong></p><h3><strong>How Document Analyst Works</strong></h3><p>We designed Document Analyst to be simple. In your agent’s settings, toggle on the Document Analyst action. That’s it.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zryif1siVq71XokUlbuh3w.png" /><figcaption>1-click enable the document analyst.</figcaption></figure><p>Your users now see an attachment icon in chat. They can upload files — contracts, transcripts, resumes, schematics, photos, spreadsheets. Any text or image file.</p><p>Here’s what happens next:</p><p>The agent treats the uploaded file as temporary context. Then it analyzes that file against your permanent knowledge base — the documents, data, and guidelines you’ve already uploaded to train your agent.</p><p>The result: insights that combine the new information with everything your organization knows.</p><p>Need to compare multiple documents? Upload several. The agent can analyze them side-by-side, spot differences, find patterns.</p><p>And critically: every answer comes with citations. You’re never blindly trusting the AI. You can trace every claim back to the source — whether that’s the uploaded document or your knowledge base. This is non-negotiable for enterprise use.</p><p>Under the hood, Document Analyst reasons across both contexts, finding connections, spotting gaps, and surfacing insights that matter specifically to your business. But you don’t need to understand the architecture. You just need to know it works.</p><h3><strong>What Teams Are Actually Doing With It</strong></h3><p>We launched Document Analyst in beta to a group of power users. Here’s what they’re doing:</p><h4><strong>Compliance Pre-Check</strong></h4><p>Compliance teams drop in marketing materials before they go live. The AI flags issues against their regulatory guidelines instantly. Human review still happens, but the first pass that used to take hours now takes minutes.</p><h4><strong>Contract Review</strong></h4><p>Legal uploads vendor agreements and checks them against standard templates. The AI spots what’s missing, what’s different, what’s concerning — before anyone signs. “Clause 4.2 conflicts with your standard NDA” is more useful than “This is a vendor agreement.”</p><h4><strong>Candidate Screening</strong></h4><p>HR compares batches of resumes against job requirements using their own hiring criteria. Hours of screening work, automated.</p><h4><strong>Technical Support</strong></h4><p>Engineers drop in spec sheets or incident reports and cross-reference against SOPs and troubleshooting guides. No more digging through folders to find the relevant documentation.</p><p>The pattern is consistent: tasks that used to require assembling context from multiple sources, multiple colleagues, multiple systems — now happen in a single conversation with your AI agent.</p><p>Hours of work. Minutes to complete.</p><h3><strong>What This Means for Enterprise AI</strong></h3><p>We’re at an inflection point in how enterprises use AI.</p><p>The first wave was “AI that knows everything” .. ChatGPT, Claude, tools that have broad knowledge but <strong>no specific context</strong>.</p><p>The second wave is “AI that knows <strong>YOUR</strong> everything” , agents that combine general intelligence with <strong>institutional knowledge</strong>.</p><p>Document Analyst is one capability in this second wave. The ability to analyze new information through the lens of what your organization already knows.</p><p>This isn’t about replacing human judgment. The compliance team still reviews the flagged issues. Legal still makes the call on contract terms. HR still interviews candidates.</p><p>What changes is the <strong>prep work</strong>. The context-gathering. The “let me check with three colleagues before I can start” phase. That’s what gets compressed.</p><p>Your team spent years building institutional knowledge. Policies, templates, guidelines, best practices, tribal knowledge captured in documents. That knowledge has always been valuable. Now it’s also accessible, for every document someone uploads, for every analysis someone needs.</p><p>We’re building more capabilities like this. Document Analyst is the first. There’s more coming. Lots more.</p><h3><strong>Try It</strong></h3><p>Document Analyst is available now on Premium and Enterprise plans. It works across all deployment types, including API, so you can build document analysis into your own workflows.</p><p>To enable it: go to your agent’s Actions page and switch the toggle. Your users will see an attachment icon in chat. Upload documents, ask questions, get answers grounded in your business data.</p><p>I’d genuinely love to hear what you analyze first. What’s the document workflow that’s been stuck on “manual review”? What’s the analysis that requires three colleagues to start?</p><p>Enable <a href="https://customgpt.ai/document-analyst/?utm_source=medium&amp;utm_medium=social&amp;utm_campaign=document-analyst-launch-jan26&amp;utm_content=founder-post"><strong>Document Analyst</strong></a> and let me know.</p><p>— -</p><p><em>Alden Do Rosario is the Founder &amp; CEO of CustomGPT.ai, named one of the “Top 7 Emerging Leaders in Generative AI” by GAI Insights. He previously served as CTO at Chitika.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=73addfc59147" width="1" height="1" alt=""><hr><p><a href="https://ai.plainenglish.io/why-your-ai-agent-fails-at-document-analysis-and-what-we-built-instead-73addfc59147">Why Your AI Agent Fails at Document Analysis (And What We Built Instead)</a> was originally published in <a href="https://ai.plainenglish.io">Artificial Intelligence in Plain English</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[“$500-an-Hour” Is an Endangered Species — How To Use AI To Modernize Your Consulting Practice]]></title>
            <link>https://medium.com/@aldendorosario/500-an-hour-is-an-endangered-species-how-to-use-ai-to-modernize-your-consulting-practice-b8aa6438325e?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/b8aa6438325e</guid>
            <category><![CDATA[saas]]></category>
            <category><![CDATA[consulting]]></category>
            <category><![CDATA[chatgpt]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[customgpt]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Mon, 04 Aug 2025 18:17:55 GMT</pubDate>
            <atom:updated>2025-08-04T18:17:55.437Z</atom:updated>
            <content:encoded><![CDATA[<h3>“$500-an-Hour” Is an Endangered Species — How To Use AI To Modernize Your Consulting Practice</h3><p><em>“If ChatGPT can serve fast, honest advice whenever someone wants it, why pay me for a 2‑hour billing block?”</em></p><p>My neighbor literally came to my door yesterday asking me this blunt question. He is a respected consultant with thousands of C‑suite clients and decades of credibility.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/940/1*LqVTUKNvx-5GJpN9K2vfNA.png" /></figure><p>Billing hourly feels safe, until the AI runs 24/7. And right now, an AI-powered advisory industry is scaling faster than ever.</p><p>Executives at <em>over 70%</em> of consulting firms now report using AI in client work. And global revenue in AI consulting is expected to grow from about <strong>$11 billion in 2025</strong> to nearly <strong>$90 billion by 2035</strong>.</p><p>If someone you hired spent the weekend patching their face-to-face into an AI coach without reducing value — or even increasing insights — would they notice?</p><p>Here’s a five‑step blueprint <em>for non‑tech founders, coaches, and consultants</em> who want to turn their knowledge into an “always‑on” AI advisor, and escape the $500‑per‑hour treadmill.</p><h3>Step 1: Extract What You Know — Transform It into Evergreen Products</h3><p>Your unique know‑how is hiding in your PPTs, PDFs, email templates, recorded calls, and workshop worksheets. Serialize it once, and you can reuse it 1,000 times without losing clarity.</p><p>Ask yourself:</p><ul><li>Which deliverables do you hand out at every engagement?</li><li>Which worksheets or frameworks get used over and over?</li></ul><p>These become the fuel for your AI product.</p><p>You don’t have to code. Today’s platforms let you package your proprietary content into a <strong>custom GPT</strong>, no development team needed. Just upload, define its service-level tone, and train it smartly:</p><ul><li>Clear, structured prompts to walk users through complex ideas? <strong>On</strong>.</li><li>In‑line citations so your assistant says “According to slide 10 of my IP” instead of hallucinating? <strong>Built in</strong>.</li><li>Clear brand and tone of voice to mimic your unique style? <strong>Easy</strong>.</li></ul><p>Once you do this, your value isn’t measured in hours, it lives in that GPT you’ve built.</p><h3>Step 2: Create a “Thinking Agent” — Choose a Reasoning-Capable Model</h3><p>What does “reasoning-capable” mean? Your assistant does more than spit out chunks of your PDF, it can guide clients using your logic.</p><p>Most consultants want models that:</p><ul><li>Know <em>why</em> they teach something (not just <em>what</em>)</li><li>Tell the user <em>how to think</em>, not only <em>what to do</em></li><li>Cite your material so the user knows the source</li></ul><p>ChatGPT now offers GPT‑powered agents tuned for reasoning tasks, not just writing. You feed in your assessments, frameworks, PDFs, or slide decks. The model builds vector embeddings, connects similar meanings, and gives you consistent guidance. You can even provide it custom instructions, so that it can mimic your particular style, tone-of-voice and methodology.</p><p>Think of it like a digital co‑pilot: instead of you walking the client through a 60-minute call, the AI prompts them with follow‑up questions, suggests next steps, and spots blindspots, all in your voice.</p><h3>Step 3: Run a Client Beta — Ask One Crucial Question</h3><p>Let your top‑tier client group (2–3 people) test the agent. Then ask them:</p><p>“Did the AI coach make you <em>faster</em>, <em>clearer</em>, or <em>richer</em> in just one session?”</p><p>If yes, you have <strong><em>product‑market fit</em></strong>.</p><p>Don’t over‑survey or over-think. You’re looking for <em>yes‑or‑no conviction</em>, not polite feedback forms.</p><h3>Step 4: Flip on a Sustainable Revenue Model — Monthly Subscriptions, $/client, $49–$499/month</h3><p>Once feedback is positive, you’re ready to monetize. The subscription model is not speculation, it’s structurally more stable than hourly or one‑off packages because:</p><ul><li>Clients pay for <strong>access</strong>, not calendar blocks</li><li>Churn <em>drops</em> the longer someone stays</li><li>You can bundle premium human time as an upsell, AI as your funnel</li></ul><blockquote><strong>Ninja tip</strong>: Have a general-purpose custom GPT embedded on your website with free advice. This is your lead magnet.</blockquote><p><strong><em>Real‑world example</em></strong><em>:</em> An accountant launched an AI advisor with access at $89/month and offered one-on-one VIP hours on top. Within a few months, their recurring SaaS business has hit <strong>7‑figures valuation (based on 6-figure ARR)</strong></p><p>Still skeptical? Research shows clients in subscription relationships report stronger progress because they <em>feel accountable</em>, and your content is getting consumed weekly, not occasionally.</p><h3>Step 5: Weaponize the “Data Flywheel” — Your Agent Learns Every Week</h3><p>This is the real magic: every conversation powers better performance.</p><p>When users ask questions:</p><ol><li>Their prompts tell you which concepts get used the most.</li><li>Their feedback reveals where the bot stumbles (“I meant X, not Y”).</li><li>Their usage signals you which framework needs rewrites or expansion.</li></ol><p>You log and categorize these interactions — then iterate:</p><ul><li>Improve phrasing</li><li>Add clarification articles or examples</li><li>Update your slide deck, spreadsheet, or workbook</li><li>Retrain or refresh your assistant’s memory</li></ul><p>This is the “data flywheel”: each iteration of learning makes the agent more accurate, so clients trust it, which leads to more use, and the cycle repeats.</p><p>It’s what turned Amazon’s recommendation system into a self‑improving loop, and you can build a micro‑version of that with your own content.</p><p>See my previous article below on how to do this.</p><p><a href="https://medium.com/operations-research-bit/ai-chat-logs-the-hidden-goldmine-your-company-hasnt-discovered-yet-2749b14b7e91">AI Chat Logs — The Hidden Goldmine Your Company Hasn’t Discovered Yet.</a></p><h3>Risk &amp; Trust: Don’t DIY — Purchase Platforms That Handle Security and Updates</h3><p><strong>Don’t neglect compliance</strong>, AI chat logs can be sensitive:</p><ul><li>Use services with built‑in logging that meet SOC‑2 Type 2 or ISO‑standards</li><li>Add system prompts to prevent injection or misuse</li><li>Version your knowledge base so if you add data, you can always revert or audit content</li></ul><p>Use platforms that abstract away the compliance layer so you can focus on content.</p><h3>Why This Strategy Works</h3><ol><li><strong>Speed to market</strong> ‑ You package your existing IP and ship in days — not months of coding.</li><li><strong>Equity, not hours</strong> ‑ You build content once; it pays you forever.</li><li><strong>Higher price elasticity</strong> ‑ Clients happily pay $199/month plus your VIP coaching upsell.</li><li><strong>Scalable defensibility</strong> ‑ Your content becomes the moat; AI is the multiplier.</li><li><strong>Sustainable value</strong> ‑ The more clients use it, the smarter it becomes, unlike linear hourly labor.</li></ol><h3>So — What Should You Do Tomorrow?</h3><ol><li>Pick one asset — say, “How I run my quarterly strategy workshop”</li><li>Transform it into a simple chat flow: checklist, decision branch, reflection prompt</li><li>Implement in <a href="https://openai.com/index/introducing-gpts/">ChatGPT</a> or <a href="https://customgpt.ai/">CustomGPT.ai</a> and have 2–3 clients try it as a helper to your live call</li><li>Ask them: <em>Did it help you make progress in half the time, or save $1K in human hours?</em></li><li>If they say yes, re-build the custom GPT in a business-grade, compliant platform and embed it behind a paywall. (PS: You might need a Fiverr developer to do this part)</li><li>Observe the chat logs and update the agent weekly based on patterns</li></ol><p><strong>For the non‑technical leader</strong>: Think of this as copy‑and‑paste from your brain into a box that runs itself. Soon, every thought you’d teach once becomes a question you answer once, and a tool your clients use daily.</p><p>If you’re still billing by the hour in 2025 — when rivals are running agent-powered advice on demand — you’re not just behind. You’re forfeiting future value.</p><p>So here’s a handshake: start today, beta test this week, and choose your first monthly‑access product by next Monday. Then, watch it grow above your highest billing threshold, in perpetuity.</p><p>Giving this a thought? Drop a comment below.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=b8aa6438325e" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Dear Support Departments, Please Stop Measuring Success by Ticket Volume — It’s a Vanity Metric]]></title>
            <link>https://medium.com/@aldendorosario/dear-support-departments-please-stop-measuring-success-by-ticket-volume-its-a-vanity-metric-9638fd3e133e?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/9638fd3e133e</guid>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[customer-success]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Mon, 30 Jun 2025 16:42:31 GMT</pubDate>
            <atom:updated>2025-06-30T16:42:31.507Z</atom:updated>
            <content:encoded><![CDATA[<h3>Dear Support Departments, Please Stop Measuring Success by Ticket Volume — It’s a Vanity Metric</h3><p>Every day, support leaders celebrate soaring ticket counts and tidy backlog burndowns as if those charts proved they were winning.</p><p>They’re not.</p><p>Ticket-volume dashboards tell you how many problems customers had to report — not how many you actually solved, nor how easy (or painful) the experience felt.</p><p>In fact, chasing volume targets rewards bad product design, punishes real customer empathy, and silently drains revenue you could be using to scale.</p><p>Below is a hard-look audit of the “ticket-volume cult,” the collateral damage it causes, and the metrics that truly predict loyalty (and profit).</p><p>And most importantly, what to do to drive those metrics.</p><h3>The Ticket-Volume Trap</h3><p>Ticket-count graphs look authoritative because they move every hour and spit out big round numbers.</p><p>But call any KPI a “workload thermometer” long enough and you’ll forget it was never meant to be a thermometer for <em>success</em>.</p><p>Even Geckoboard, which popularized the metric, lists “staff planning”, <strong>NOT customer happiness, </strong>as the primary use-case for volume charts.</p><h3>Five Things Ticket Counts Never Tell You</h3><ol><li><strong>First-Contact Resolution (FCR).</strong> Customers only “feel” a fix when the <em>first</em> interaction works; FCR drives CSAT upward while volume remains silent on the topic.</li><li><strong>Customer Effort.</strong> Gartner’s Customer Effort Score (CES) shows high-effort experiences create 96 %<em>more</em> churn, even if they produce identical ticket totals.</li><li><strong>Sentiment &amp; Word-of-Mouth.</strong> Net Promoter Score (NPS) can fall while volume stays flat, proving that ticket counts miss the emotional side of service.</li><li><strong>Product Quality Signals.</strong> Spikes in password-reset tickets inflate “good” volume stats while masking a UX flaw your engineers should fix</li><li><strong>Silent Suffering.</strong> Up to 81 % of customers try to self-solve before logging a ticket; their frustration disappears from your chart entirely.</li></ol><h3>Metrics That Actually Matter</h3><ul><li><strong>First-Contact Resolution (FCR)</strong> : Cuts repeat contacts and is directly correlated with CSAT.</li><li><strong>Customer Effort Score (CES)</strong> : A stronger predictor of future spend than NPS or CSAT.</li><li><strong>Deflection Rate</strong>: Modern AI assistants can deflect 20–90 % of incoming tickets without hurting satisfaction. (Source: <a href="https://customgpt.ai/ticket-deflection/">customgpt.ai</a> )</li><li><strong>Agent Quality Index (AQI)</strong> : Gen-AI quality-assurance tools now score every interaction for accuracy and empathy, not just 1% of random calls. (Source: <a href="https://www.mckinsey.com/capabilities/operations/our-insights/operations-blog/ai-mastery-in-customer-care-raising-the-bar-for-quality-assurance">mckinsey.com</a>)</li><li><strong>Customer-Performance Indicators (CPIs)</strong> : HBR’s term for metrics customers themselves value (<a href="https://medium.com/towards-artificial-intelligence/during-off-hours-the-ai-takes-over-customer-service-21d649416d1b">speed</a>, clarity, empowerment).</li></ul><h3>Enter AI for L0 Support : The Fastest Way Off the Treadmill</h3><p>Forward-thinking teams let AI handle Level 0 inquiries 24/7 , resetting passwords, answering questions from knowledgebases, explaining features, so humans can focus on exceptions.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*29hnb9RB9508k6-21GgcXw.png" /></figure><p>Most customers already expect AI to resolve routine issues, and companies <a href="https://medium.com/ai-in-plain-english/how-i-reduced-support-ticket-volume-by-90-for-my-saas-405a0d0cbc3c">deploying AI</a> are now reporting huge cost-to-serve reductions.</p><p><a href="https://blog.cubed.run/dear-support-departments-please-let-ai-do-90-of-your-job-037086f46427">Dear Support Departments, Please Let AI Do 90% Of Your Job</a></p><h3>Breaking the Ticket-Volume Addiction</h3><ol><li><strong>Stop Worshipping the Graph.</strong> Hide volume from exec dashboards for 30 days; watch how conversations change.</li><li><strong>Swap Inputs for Outcomes.</strong> Replace “tickets closed” with FCR, CES, and revenue retained.</li><li><strong>Deploy an AI Assistant As L0 Support.</strong> Deflect the low-value flood so your new metrics don’t get drowned in noise.</li><li><strong>Run a 90-Day Pilot.</strong> Pick a chronic issue queue, roll out AI deflection, and benchmark FCR + CES before/after.</li><li><strong>Reward Empathy, Not Speed.</strong> Tie bonuses to CES and quality-assurance scores, not handle-time.</li></ol><p><a href="https://medium.com/the-generator/ai-in-customer-service-is-freaking-people-out-heres-how-to-sway-your-team-c892431f4c59">AI in Customer Service Is Freaking People Out — Here’s How to Sway Your Team …</a></p><p>Measuring support by ticket volume is like judging a hospital by the number of incoming ambulances: the chart may look impressive, but it’s the opposite of success.</p><p>The forward-thinking support department belongs to teams that minimise <em>customer effort</em>, maximise <em>first-contact resolution</em>, and weaponise AI as a tireless L0 shield.</p><p>Retire the vanity metrics, and let your dashboards reflect the metrics your customers care about most.</p><p><a href="https://ai.plainenglish.io/the-absolute-lowest-hanging-fruit-in-generative-ai-is-ff0226da302b">The Absolute Lowest Hanging Fruit In Generative AI Is …</a></p><p><em>Your Looker report may never forgive you, but your customers will.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=9638fd3e133e" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Improve your efficiency in meetings with this simple Manus prompt.]]></title>
            <link>https://medium.com/prompt-prompts/improve-your-efficiency-in-meetings-with-this-simple-manus-prompt-e25382270a3b?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/e25382270a3b</guid>
            <category><![CDATA[chatgpt]]></category>
            <category><![CDATA[prompt]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[prompt-engineering]]></category>
            <category><![CDATA[manus]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Tue, 27 May 2025 16:57:01 GMT</pubDate>
            <atom:updated>2025-05-27T21:52:07.523Z</atom:updated>
            <content:encoded><![CDATA[<p>There used to be a time when people would walk into meetings (over Zoom or in-person) with zero preparation.</p><p>And then the meeting goes down this low-fidelity path of silly questions like “How is your dog?” or “What’s the weather like where you are”.</p><p>And even worse: Since everyone is so blatantly unprepared, nothing real gets done in the meeting, and the real questions go un-asked.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*SivcyVr2FARkBJncmmNmYw.png" /><figcaption>Image Credit : <a href="https://medium.com/u/841a3b5d13d5">Alden Do Rosario</a></figcaption></figure><p><strong>Manus AI to the rescue.</strong></p><h3>But wait — what is Manus AI?</h3><p>Manus AI is a groundbreaking general-purpose AI agent. Unlike traditional AI chatbots that merely respond to prompts, Manus is designed to autonomously execute complex tasks across various domains, effectively acting as a digital assistant that can think, plan, and act independently.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*4MqCkm00EJ_f6RZl2IG1PQ.png" /><figcaption>Image Credit : <a href="https://medium.com/u/841a3b5d13d5">Alden Do Rosario</a></figcaption></figure><p>Manus AI has garnered attention for its ability to handle tasks ranging from content creation and data analysis to workflow optimization and travel planning.</p><p>It operates by breaking down high-level instructions into manageable subtasks, executing them without the need for continuous human oversight.</p><p>In the context of meetings, Manus can prepare agendas, anticipate questions, and even generate comprehensive reports, ensuring that sessions are productive and focused.</p><h3>How to use Manus to prepare for the meeting?</h3><p>To prepare for your meeting, just give Manus some context about the attendee and what the meeting is about, and let it get to work. ‘</p><p>For example, in the prompt below, a candidate is coming in for an interview. So it makes sense to run some deep research over the resume, combined with the job description.</p><pre>[NAME] is coming in for an interview for this job [ROLE] that I have attached.<br><br>I need you to do a full fledged HUMINT exercise on him. <br><br>Scour the web and whatever sources you have access to and give me a full HUMINT report on him. <br><br>Cross check the resume, dig deep, do what it takes.<br><br>Don&#39;t ask questions. Do whatever is needed. use your judgement.</pre><p>Here is the type of report you can expect:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*kjfL8b5Lj9Sv-8O_WUJ4Jg.png" /><figcaption>Image Credit : <a href="https://medium.com/u/841a3b5d13d5">Alden Do Rosario</a></figcaption></figure><p><strong>Why this works</strong>: The fidelity of the meeting will be increased 2–5X because you will be much more educated about the person. Instead of vague small talk like “What did you do in college?”, you can be like “Hey, I saw you scored 21 points against X — lets talk about that game” or “Dude, tell me about why you got suspended in junior year”</p><h3>Can this be used for sales meetings too?</h3><p>Absolutely — doing this for sales meetings is bound to improve your close rate. Here is a sample prompt for sales meetings (please modify it to your own situation)</p><pre>I have a meeting with the attendees shown below - its a 30-min zoom call.<br><br>here are the attendees:<br>```<br>Copy-paste attendees from your calendar invite. <br>```<br><br>Here is the agenda:<br>```<br>Include agenda of the meeting (typically copy-pasted from calendar invite or email) <br>```<br><br>Here is my goal for this meeting: <br>```<br>Clearly state the goal of the meeting (like close the deal, setup demo, etc) <br>```<br><br>I want you to deeply research the participants and then prepare me with the anticipated questions they will have based on their roles. Give me a full report of each participant, what he/she does and the questions I should anticipate from them.</pre><p>I’ve been using this approach to prepare for sales meetings recently and the attendees truly appreciate the preparation and deep research.</p><p>And best of all: Someone else (the AI!) did all the hard work, something that would previously have taken me ages to do (which means I would skip it!)</p><blockquote>One key thing to remember: If the meeting is critical, do spend a few minutes to cross-check the facts (it’s AI after all and can hallucinate sometimes!)</blockquote><h3>What is PROMPT PROMPTS?</h3><p>It’s a new publication established as a library for useful AI prompts. You’ll get fast, no-nonsense guides and tips from a range of prompt engineers.</p><p>Keep up with our newest AI resources — <strong>follow the publication now!</strong></p><p><a href="https://medium.com/prompt-prompts">Prompt Prompts</a></p><h4>Write for PROMPT PROMPTS</h4><p>If you have the AI skills and want to contribute to <strong>Prompt Prompts</strong>, you can email to be added to our writers: <strong>PromptPrompts@jimtheaiwhisperer.com</strong></p><p>You’ll be sent a template to follow. <strong>Prompt Prompts</strong> has high standards focused on clarity and brevity. It’s in the name: we deliver prompts <em>fast</em>!</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=e25382270a3b" width="1" height="1" alt=""><hr><p><a href="https://medium.com/prompt-prompts/improve-your-efficiency-in-meetings-with-this-simple-manus-prompt-e25382270a3b">Improve your efficiency in meetings with this simple Manus prompt.</a> was originally published in <a href="https://medium.com/prompt-prompts">Prompt Prompts</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Focus on Bing — Not Google — For AEO, Here Is Why …]]></title>
            <link>https://medium.com/@aldendorosario/focus-on-bing-not-google-for-aeo-here-is-why-42d5ab325e5a?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/42d5ab325e5a</guid>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[analytics]]></category>
            <category><![CDATA[seo]]></category>
            <category><![CDATA[chatgpt]]></category>
            <category><![CDATA[bing]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Fri, 07 Mar 2025 15:47:37 GMT</pubDate>
            <atom:updated>2025-03-07T15:47:37.509Z</atom:updated>
            <content:encoded><![CDATA[<h3>Focus on Bing — Not Google — For AEO, Here Is Why …</h3><p>A few day ago, I published my article <a href="https://medium.com/@aldendorosario/chatgpt-clicks-convert-6-8x-higher-than-google-organic-a457203cfc52?source=friends_link&amp;sk=4b7cd4108c2a89e4cd14e246ba8b889b">“ChatGPT Clicks Convert 6.8X Higher Than Google Organic”</a> here on Medium, and the #1 question flooding my inbox has been:</p><blockquote><strong>How do I optimize my website for AEO (AI Engine Optimization)?</strong></blockquote><p>If you haven’t read the data-backed article yet, do that first. The findings are eye-opening.</p><p><a href="https://medium.com/@aldendorosario/chatgpt-clicks-convert-6-8x-higher-than-google-organic-a457203cfc52">ChatGPT Clicks Convert 6.8X Higher Than Google Organic</a></p><p>But if you want my one-word ADHD response to the AEO question: <strong>Bing</strong>.</p><blockquote>Now, just to be clear — I’m NOT an SEO or AEO specialist. There are far more experienced folks than me in that arena.</blockquote><p>But having analyzed the conversion data and seen the writing on the wall, here’s my perspective:</p><h3>1. Focus on Bing, Not Google</h3><p>Look, I know this sounds counterintuitive. Google has dominated search for so long that “Google it” became a verb. But here’s the reality:</p><ul><li><strong>Bing powers most of the AI answer engines</strong> — ChatGPT, Microsoft Copilot, and influences Perplexity too.</li><li><strong>Google still holds massive market share</strong>, but extracting growth from Google is like trying to juice water from a rock at this point.</li></ul><p>While everyone else is battling for tiny gains in the Google ecosystem — <em>all while Google is rapidly changing its algorithm and juicing every dollar from its paid ads</em> — there’s significantly less competition in optimizing for Bing — which now has outsized importance thanks to its role in powering AI responses.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*h8ZOI2t_gx73OeRu" /></figure><h3>How to Actually Optimize for Bing (Actionable Steps):</h3><ol><li><strong>Exact Keyword Usage</strong>: Unlike Google, Bing places higher emphasis on exact-match keywords. Include your target keywords in titles, H1/H2 tags, meta descriptions, and within the first 100 words of content.</li><li><strong>Focus on High-Quality Backlinks</strong>: Bing prioritizes quality over quantity. Aim for backlinks from older, established domains and authoritative sites (.edu, .gov) rather than chasing volume.</li><li><strong>Optimize for Social Signals</strong>: Unlike Google, Bing actually considers social media signals as ranking factors. Promote your content actively across social platforms to improve engagement and visibility. (<em>PS: The one exception : Lately Google seems to love Reddit</em>)</li><li><strong>Submit Your Sitemap</strong>: Use Bing Webmaster Tools to submit your sitemap directly. This helps Bing crawl and index your content more effectively. (Yeah, I know this sounds basic — but it’s the least you can do)</li><li><strong>Consider Meta Keywords</strong>: While Google ignores meta keywords, Bing still gives them some weight. Include relevant, targeted keywords in your meta tags.</li></ol><p>This isn’t just theory. Our data shows that clicks from ChatGPT (which has ties to Bing’s index) convert at 6.8X the rate of Google organic traffic. That’s not a minor improvement — it’s a major difference for reasons mentioned below.</p><h3>2. Focus on Conversions, Not Traffic</h3><p>The old SEO playbook was simple: more traffic = more success. That’s outdated thinking.</p><p>In our analysis, ChatGPT drove just 4% of our traffic but accounted for 22% of conversions. Think about that — a traffic source that represents a tiny fraction of visitors but delivers over one-fifth of your business outcomes.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*vdEICL_NoUZnsB3j" /></figure><p>And here is the most scary thing: Answer engines like ChatGPT and Perplexity are <strong><em>just getting started </em></strong>— but growing 44% month-over-month. Most of the world hasn’t even heard about these engines yet. In fact, ChatGPT with free search was introduced just a month ago.</p><h3>Conversion-Focused Actions to Take Today:</h3><ol><li><strong>Track AI Source Conversions</strong>: ChatGPT uses <strong>utm_source=chatgpt.com</strong> in its outgoing clicks — start tracking those in your GA4. Perplexity sends the <strong>Referer</strong> header (here is <a href="https://x.com/randomjohnnyh/status/1894522132429734367">my conversation</a> with Perplexity cofounder, Johnny Ho)</li><li><strong>Optimize Landing Pages for AI Visitors</strong>: Create dedicated landing pages that acknowledge the context visitors already have from their AI conversation (e.g., “As you may have discovered in ChatGPT…”). Since ChatGPT sends the utm_source=chatgpt.com param, you could potentially customize the user experience based on that.</li><li><strong>Create Content Assets for AI Citations</strong>: Develop high-value resources specifically designed to be cited by AI engines — comprehensive guides, original research, and <strong>definitive answers to common questions</strong>. While answering questions like “How”, “What”, etc is facing a decline in organic Google search, the content from these is what feeds the AI engines. Do a content gap analysis on your website to see what authoritative questions are missing from your website content.</li></ol><p>This is the quality vs. quantity debate settled with hard data. I’d rather have 100 highly qualified visitors than 10,000 casual browsers any day.</p><h3>3. Focus on Chat, Not Pageviews</h3><p>The conversational nature of AI answer engines fundamentally changes user expectations. When someone clicks through from ChatGPT to your site, they’ve already had a conversation about your topic. They’re primed, informed, and ready to engage.</p><h3>Why You Need a Custom Chatbot on Your Website:</h3><p>Today’s visitors are becoming increasingly “AI-first” in their thinking. They’re comfortable with conversational interfaces and often prefer them to traditional page navigation.</p><blockquote><strong>Remember: AI is the new UI</strong></blockquote><p>Here’s why implementing a custom chatbot on your site is no longer optional:</p><ol><li><strong>Match User Expectations</strong>: Visitors coming from ChatGPT or similar tools are already in a conversational mindset. A custom chatbot provides continuity in their experience.</li><li><strong>Instant Information Access</strong>: Traditional website navigation forces users to hunt through page hierarchies. A custom chatbot lets them simply ask for what they need.</li><li><strong>Personalized Content Delivery</strong>: AI chatbots can tailor responses based on user context, delivering exactly what each visitor needs without overwhelming them.</li><li><strong>Enhanced Data Collection</strong>: Chatbot conversations provide invaluable insights into what your visitors actually want to know — information you can use to improve your content strategy. The chat logs are a <a href="https://medium.com/operations-research-bit/ai-chat-logs-the-hidden-goldmine-your-company-hasnt-discovered-yet-2749b14b7e91">goldmine of information</a> for you to inform your conversion and content strategy.</li><li><strong>24/7 Engagement</strong>: A well-trained chatbot provides immediate responses at any hour — and any language — keeping potential customers engaged when your team is offline. (Tip: This is doubly true if your customers are more active during off hours — like divorce law or plumbing services)</li></ol><blockquote><strong>Tools like ChatGPT or onsite custom chatbots are pre-cooking and pre-selling the user so that they are all ready to buy.</strong></blockquote><h3>Actionable Steps for Chat Optimization:</h3><ol><li><strong>Implement a Custom-Trained AI Chatbot</strong>: Use no-code platforms to create a custom chatbot specifically trained on your content.</li><li><strong>Provide Your Bot with Starter Questions</strong>: Use actual customer inquiries to provide your user with starter questions. You can also easily customize your bot in plain English to act in a certain way by providing it with a custom persona.</li><li><strong>Create Conversational Paths to Conversion</strong>: Design chat flows that naturally guide users toward key conversion points. Using plain English, you can instruct the bot to end responses with a path to your conversion action (e.g. “Signup for our 7-day free trial” or “Contact us and ask about this month’s special offer”)</li><li><strong>Monitor and Improve</strong>: Regularly review chatbot interactions to identify gaps in knowledge or opportunities for improved response.</li></ol><p>Remember: Antiquated page structures and traditional site navigation are increasingly barriers to conversion for AI-native users. A conversational interface removes these friction points.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/957/0*pyPC-lQJeKj5IUGW" /><figcaption>Example conversational agent on a documentation website</figcaption></figure><h3>What This Means for Your Strategy (Actionable Takeaways)</h3><p>If you’re a digital marketer or business owner right now, consider this a rare opportunity. While your competitors continue throwing resources at saturated Google optimization strategies, you can:</p><ol><li><strong>Audit Your Bing Performance Today</strong>: Use Bing Webmaster Tools to assess your current visibility and identify quick optimization opportunities.</li><li><strong>Structure Content for AI Readability</strong>: Implement clear headings, FAQs, structured data, and summary sections to make your content AI-friendly. AND: Answer those <strong>questions</strong> related to your niche.</li><li><strong>Set Up AI Traffic Tracking</strong>: Create conversion tracking and custom reports in GA4 to track traffic and conversions from ChatGPT, Perplexity, and other AI sources.</li><li><strong>Implement a Custom Chatbot Strategy</strong>: Don’t just add a generic antiquated chatbot like Intercom — develop a strategy for how a custom conversational interface can enhance your specific user journey.</li><li><strong>Develop a AEO Plan</strong>: Work with your SEO team or agency to create a specific strategy that prioritizes AEO alongside (not instead of) Google.</li></ol><p>The shift to AI-driven answers isn’t just coming — it’s already here. And based on our data, the conversion potential is too significant to ignore. The websites that adapt fastest will capture disproportionate value in this new ecosystem.</p><p>I’d love to hear if you’ve noticed similar patterns with AEO driven conversions or if you’ve implemented any of these strategies. Drop your experiences in the comments!</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=42d5ab325e5a" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[ChatGPT Clicks Convert 6.8X Higher Than Google Organic]]></title>
            <link>https://ai.plainenglish.io/chatgpt-clicks-convert-6-8x-higher-than-google-organic-a457203cfc52?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/a457203cfc52</guid>
            <category><![CDATA[seo]]></category>
            <category><![CDATA[conversion-optimization]]></category>
            <category><![CDATA[marketing]]></category>
            <category><![CDATA[chatgpt]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Tue, 25 Feb 2025 19:53:05 GMT</pubDate>
            <atom:updated>2025-03-17T01:10:58.123Z</atom:updated>
            <content:encoded><![CDATA[<p>Here’s the deal: I recently dug into some data in GA4 for our website and found that while Google Organic brings in more traffic, ChatGPT clicks convert way better — <strong>6.8X better</strong> for free trial conversions, to be exact.</p><p>If you’re in the trenches of SEO and digital marketing, you know that every conversion counts. Let’s break this down.</p><h3>Digging into GA4.</h3><p>I was checking our traffic numbers in GA4 when something caught my eye. Despite having a fraction of the total visitors, ChatGPT clicks were outperforming Google Organic in converting users into free trials.</p><blockquote>Just to confirm: A free trial activation on our website (<a href="https://customgpt.ai/">customgpt.ai</a>) requires a <strong>work email</strong> AND <strong>credit card</strong> on our platform — so these are not some accidental clicks.</blockquote><p>And yes, we’re talking about a 6.8X conversion multiplier here.</p><p>In this post, I’m going to show you the raw data, dive into each stage of the user funnel, and explain why this matters for your SEO and CRO strategy.</p><h3>The Traditional SEO Playbook</h3><p>For years, we’ve leaned heavily on Google Organic to drive traffic. And given the huge market share Google commands, you probably should.</p><p>The idea is simple: more clicks equal more conversions. But what if quality beats quantity every time?</p><p>But now, there is ChatGPT (and Perplexity!)</p><p>Instead of a broad net, it delivers high-intent traffic. Fewer clicks, but better-qualified leads. This post lays out the numbers behind that claim.</p><p>From a common-sense perspective, it makes perfect sense — instead of TBC (10-Blue-Links), the user is coming to your website highly qualified and probably warmed up and pre-sold on your product and services.</p><p>So it makes intuitive sense that ChatGPT clicks would be more valuable than Google Organic.</p><h3>Data Collection &amp; Methodology</h3><p>I based this analysis on three core metrics:</p><ul><li><strong>Total First-Time Active Users:</strong> The initial touchpoints.</li><li><strong>Total Sign-Ups (Email Required):</strong> The next step in the funnel. [Side note: We require <strong>work email</strong> — so this leaves out fraudulent gmail addresses]</li><li><strong>Total Free Trials (Credit Card Required):</strong> The ultimate conversion goal. [Side note: The credit card is collected using Stripe hosted checkout]</li></ul><p>We also looked at two conversion ratios:</p><ul><li><strong>Free Signup Conversion Ratio</strong></li><li><strong>Free Trial Conversion Ratio</strong></li></ul><p>Data was gathered using GA4 and represents the last couple of days.</p><blockquote>Quick tip: This analysis is now possible because ChatGPT tags its outgoing clicks with a nice <strong>utm_source=chatgpt.com</strong> that can be tracked in GA4. So when your website shows up in the ChatGPT citations, it will have that param appended in its link.</blockquote><h3>Data Overview</h3><p>Take a look at <a href="https://docs.google.com/spreadsheets/d/1SRkUEqlwqZPfjBE2Gk1coyLWsziXWWIAYmsprQ0xCHo/edit?usp=sharing">this table</a> for a quick snapshot:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/970/1*O4EdzvMFF1KHDvT8AVvtAw.png" /></figure><blockquote><strong>In short, while ChatGPT drove just 4% of the traffic, it accounted for 22% of the conversion.</strong></blockquote><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*AJTPEBNKcLi5Ce49" /></figure><p><strong><em>Important note: This traffic distribution is on a log-scale.</em></strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*1qpa5xMlwza0gObE" /></figure><p>Here is the kicker: ChatGPT conversions are 6.8X higher.</p><h3>Detailed Funnel Analysis</h3><p>Let’s slice and dice the funnel:</p><ul><li><strong>First-Time Active Users:</strong> Google Organic brings in 7,079 users, but ChatGPT delivers just 306.</li><li><strong>Sign-Ups:</strong> Google Organic gets 54 sign-ups compared to ChatGPT’s 13.</li><li><strong>Free Trials:</strong> Out of those, only 17 from Google Organic become free trials, versus 5 from ChatGPT.</li></ul><p>Now, here’s the kicker: when you calculate the conversion ratios, ChatGPT users are far more likely to sign up and convert into free trials — even though a full credit card is required.</p><p>The drop-off is far less severe in the ChatGPT funnel, which means these users are high-quality prospects. That’s why I’m using a log scale on the funnel chart — it really lets you see the massive drop-offs and the small, yet significant, ChatGPT cohort.</p><h3>Conversion Ratio Deep Dive</h3><p>Let’s drill down:</p><ul><li><strong>Free Signup Conversion:</strong> Google Organic converts at 0.76% while ChatGPT does it at 4.25%.</li><li><strong>Free Trial Conversion:</strong> The real star — Google Organic converts at 0.24%, but ChatGPT clocks in at 1.63%.</li></ul><p>To be clear: ChatGPT’s free trial conversion is 6.8 times higher than that of Google Organic. It’s not just a fluke; these numbers suggest that ChatGPT clicks come with higher intent. Whether it’s the conversational context or the quality of the referral, something is clearly working in its favor.</p><h3>SEO &amp; Technical Implications</h3><h4>Rethinking Quality vs. Quantity</h4><p>This data challenges the conventional wisdom of chasing massive volumes. Instead, focus on quality traffic that converts.</p><p>Our natural instinct (and that of our bosses) tends to chase the raw traffic numbers. It might be time to look carefully at down-funnel conversions as well.</p><h4>Algorithmic Differences</h4><p>Traditional search engine algorithms aim for volume, but AI-driven platforms like ChatGPT are built to provide highly relevant, context-driven results.</p><p>That difference could be why we see such a high conversion rate.</p><h4>User Experience Matters</h4><p>The landing page experience and the alignment of content with user intent are critical. If you can harness the intent behind ChatGPT clicks, you can optimize your funnel to maximize conversions.</p><blockquote><strong>Disclaimer</strong>: Our product is very tech-oriented, so might get an added boost from the tech-savvy early-adopter ChatGPT crowd. If you are selling pyjamas to grandmas, you wont see similar results.</blockquote><h3>Actionable Insights &amp; Recommendations</h3><p>Here’s what you can do:</p><ul><li><strong>First thing: </strong>Dig into GA4 and see if you can start tracking your incoming traffic from chatgpt.com</li><li><strong>Get your tracking right:</strong> If you haven’t already done so, make sure that all intermediate events in your conversion funnel are being tracked in GA4.</li><li><strong>Optimize for ChatGPT </strong>: While you don’t need to rush with both feet into AEO (or GEO or AISEO or whatever it is called these days!) — do keep an eye out for this changing phenomenon — at least when it comes to conversions. [<em>Side note: If you haven’t already done so, please pay attention to your Bing rankings</em>]</li></ul><h3>Caveats &amp; Considerations</h3><p>And here comes the list of caveats — because its not all milk and honey — a couple of things to keep in mind:</p><ul><li><strong>Sample Size:</strong> ChatGPT traffic numbers are smaller. While the conversion ratios are impressive, broader tests over time are necessary. I did run a statistical significance test, but clearly more data is needed (both on our own website — and other websites)</li><li><strong>Your Results Will Vary : </strong>It’s very much possible that due to the nature of our product, we get a different type of visitor from ChatGPT than we do with Google Organic. Your results will vary drastically depending on what you are selling.</li><li><strong>External Factors:</strong> Various factors like market trends, seasonal effects, or even changes in the algorithms could impact these results. There are literally a million factors at play here — so please consider your own specific conditions.</li><li><strong>No Predictions:</strong> If you notice, I am NOT making any predictions about SEO or market share here (that’s up to pundits to argue over!) — Even though I personally feel ChatGPT, Grok and Perplexity could be nipping at Google’s market share, it’s a fruitless exercise to try and take bets on that.</li></ul><p>Got questions or want to share your own experience? Drop a comment below. Let’s keep the conversation going!</p><h3>Thank you for being a part of the community</h3><p><em>Before you go:</em></p><ul><li>Be sure to <strong>clap</strong> and <strong>follow</strong> the writer ️👏<strong>️️</strong></li><li>Follow us: <a href="https://x.com/inPlainEngHQ"><strong>X</strong></a> | <a href="https://www.linkedin.com/company/inplainenglish/"><strong>LinkedIn</strong></a> | <a href="https://www.youtube.com/channel/UCtipWUghju290NWcn8jhyAw"><strong>YouTube</strong></a> | <a href="https://newsletter.plainenglish.io/"><strong>Newsletter</strong></a> | <a href="https://open.spotify.com/show/7qxylRWKhvZwMz2WuEoua0"><strong>Podcast</strong></a> | <a href="https://differ.blog/inplainenglish"><strong>Differ</strong></a></li><li><a href="https://cofeed.app/"><strong>Check out CoFeed, the smart way to stay up-to-date with the latest in tech</strong></a> <strong>🧪</strong></li><li><a href="https://differ.blog/"><strong>Start your own free AI-powered blog on Differ</strong></a> 🚀</li><li><a href="https://discord.gg/in-plain-english-709094664682340443"><strong>Join our content creators community on Discord</strong></a> 🧑🏻‍💻</li><li>For more content, visit <a href="https://plainenglish.io/"><strong>plainenglish.io</strong></a> + <a href="https://stackademic.com/"><strong>stackademic.com</strong></a></li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=a457203cfc52" width="1" height="1" alt=""><hr><p><a href="https://ai.plainenglish.io/chatgpt-clicks-convert-6-8x-higher-than-google-organic-a457203cfc52">ChatGPT Clicks Convert 6.8X Higher Than Google Organic</a> was originally published in <a href="https://ai.plainenglish.io">Artificial Intelligence in Plain English</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Deepseek for Entrepreneurs : The Reasoning Per Dollar Is All You Need]]></title>
            <link>https://medium.com/@aldendorosario/deepseek-for-entrepreneurs-the-reasoning-per-dollar-is-all-you-need-23df90e8d8bb?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/23df90e8d8bb</guid>
            <category><![CDATA[deepseek-r1]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[deepseek]]></category>
            <category><![CDATA[entrepreneurship]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Mon, 03 Feb 2025 21:25:22 GMT</pubDate>
            <atom:updated>2025-02-03T21:25:22.341Z</atom:updated>
            <content:encoded><![CDATA[<h3>Deepseek for Entrepreneurs : The Reasoning Per Dollar Is All You Need</h3><p>Here’s the deal:</p><p>Every once in a while, a new technology comes along that completely changes the game for entrepreneurs.</p><p>It’s not just about making things a little faster or cheaper — it’s about unlocking opportunities that were previously impossible.</p><p>One of my favorite books of all time is: <strong>On ENTREPRENEURSHIP and IMPACT</strong> by Desh Deshpande (<a href="https://heyzine.com/flip-book/9ed16c66cd.html">you can view it here</a>) — its a beautiful book for Entrepreneurs (because Entrepreneurs are busy and this book covers each chapter in 2 pages)</p><p>Here is the relevant section:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/670/0*Czm5k7UliwsvZ41V" /><figcaption>Credit: <a href="https://www.deshpandefoundation.org/2023/11/06/desh-book/?utm_source=rss&amp;utm_medium=rss&amp;utm_campaign=desh-book">Deshpande Foundation</a></figcaption></figure><p>Deepseek (or similar models like OpenAI’s o3) is one of those technologies. Why? Because it dramatically improves what I like to call <em>“</em><strong><em>reasoning per dollar.</em></strong><em>”</em></p><p>Some others call this “<strong>bending the cost curve</strong>” — you get my point though.</p><h3>What’s “Reasoning Per Dollar”?</h3><p>Think about it this way: How much reasoning power can you get for every dollar you spend?</p><p>This matters because, as entrepreneurs, we’re constantly building things — tools, apps, services — that solve problems.</p><p>But some ideas never make it off the ground because they’re either too expensive to execute or the tech isn’t smart enough to do the job well.</p><p>Deepseek changes that. It makes advanced reasoning affordable.</p><p>Here’s an example: Imagine a “lawyer agent” that reviews contracts for you. Before now, this kind of tool would either cost a fortune to run or be so limited in reasoning ability that it wasn’t worth using. With Deepseek, tools like this become not only possible but practical.</p><h3>My Personal Experience</h3><p>I’ve been working as an AI-powered research writer for three years now. The goal? To create content that’s better than what 99% of humans can write — and certainly better than anything AI has produced so far.</p><blockquote><strong>Hot news</strong>: OpenAI released a similar “<a href="https://openai.com/index/introducing-deep-research/">deep research</a>” agent today. You can do a <a href="https://researcher.customgpt.ai/">side-by-side comparison</a> if you are into deep researchers.</blockquote><p>The journey hasn’t been easy. At first, the output was mediocre. Over time, it got better — “OK” turned into “Really Good.” But it still wasn’t <em>great.</em></p><p>The missing piece? A reasoning model that could deeply understand and synthesize research sources.</p><p>Enter Deepseek. With its high-reasoning capabilities at a reasonable cost, my tool can now deliver results I couldn’t have dreamed of before. And since the tool takes 30+ minutes to run per task (yes, it’s computationally heavy), affordability is a game-changer here.</p><h3>The Big Question You Should Be Asking</h3><p>Here’s what I want every entrepreneur to think about: <em>How does this new capability change what I do?</em></p><p>It’s like 2010 all over again when smartphones hit the mainstream. Back then, we asked ourselves: <em>What happens when people aren’t chained to their desktops anymore?</em> Entire industries were born from that shift.</p><p>Now we’re facing a similar moment: <em>What happens when reasoning power becomes both advanced and affordable?</em> The possibilities are endless — but only if you’re paying attention.</p><h3>Addressing Concerns About “China”</h3><p>Let’s clear something up: Every time someone mentions Deepseek, there’s this knee-jerk reaction about its ties to China. I get it — people are cautious about where their data is hosted and who has access to it.</p><p>But here’s the thing: There are plenty of hosting options available that mitigate these risks. I’ve even put together a quick graphic to explain how this works because education is key here.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/500/0*VMURca1O3CYhr7nl" /></figure><p>To be clear: I have no affiliation with Deepseek — I’m just sharing insights because I think this tech has huge potential for entrepreneurs.</p><p>You should most likely do your own self-assessment — here is a helpful guide I put together for that:</p><p><a href="https://pub.towardsai.net/deepseek-r1-is-it-right-for-you-a-practical-self-assessment-for-businesses-and-individuals-fc0021d85b60">DeepSeek R1 : Is It Right For You? (A Practical Self‑Assessment for Businesses and Individuals)</a></p><h3>Final Thoughts</h3><p>Technologies like Deepseek and o3 are opening up new opportunities. They let us build smarter tools, more efficient services, and entirely new applications that were once out of reach.</p><p>So ask yourself: <strong><em>How can this new capability redefine what I do?</em></strong></p><p>Because if you don’t, someone else will — and they’ll be the ones shaping the future while you’re stuck playing catch-up.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=23df90e8d8bb" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[RAG vs. CAG : Can Cache-Augmented Generation Really Replace Retrieval?]]></title>
            <link>https://pub.towardsai.net/rag-vs-cag-can-cache-augmented-generation-really-replace-retrieval-9078fdbcba2f?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/9078fdbcba2f</guid>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[benchmark]]></category>
            <category><![CDATA[rags]]></category>
            <category><![CDATA[ai]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Thu, 30 Jan 2025 16:58:46 GMT</pubDate>
            <atom:updated>2025-02-05T12:06:30.160Z</atom:updated>
            <content:encoded><![CDATA[<h3>RAG vs. CAG : Can Cache-Augmented Generation Really Replace Retrieval?</h3><p>A recent<a href="https://venturebeat.com/ai/beyond-rag-how-cache-augmented-generation-reduces-latency-complexity-for-smaller-workloads/"> VentureBeat article</a> highlights a new Cache-Augmented Generation (CAG) method that promises no retrieval overhead and even better performance than Retrieval-Augmented Generation (RAG).</p><p>Sounds too good to be true?</p><p>We decided to find out by running our own tests on KV-Cache (a popular CAG implementation) versus RAG.</p><p>Below are our insights on what happens when you apply these methods to real workloads.</p><h3>1. Setting the Stage: RAG vs. KV-Cache (CAG)</h3><h3>RAG</h3><p><strong>What It Is<br></strong>A Retrieval-Augmented Generation approach that uses a retriever to find relevant documents, then passes them to a large language model for final answers.</p><p><strong>Where It Shines</strong></p><ul><li>Handles larger or frequently updated datasets without loading everything at once.</li><li>Avoids massive prompts, which can lead to truncation or context overload.</li></ul><p><strong>Key Limitations</strong></p><ul><li>Adds a retrieval step, which can be slower.</li><li>Often relies on external APIs or indexing overhead.</li></ul><h3>KV-Cache (CAG)</h3><p><strong>What It Is<br></strong>A method that aims for near-zero retrieval time by loading all documents directly into the model’s context window. In principle, it cuts out the retriever entirely.</p><p><strong>Note:</strong> In our benchmarks, we used a <strong>“No Cache”</strong> version of KV-Cache because the model was too large to run locally. Instead, we mimicked the same behavior via an API (OpenRouter) by feeding all documents each time. We’re not comparing retrieval speed here, since KV-Cache would obviously win if run locally on a suitable setup.</p><p><strong>Where It Shines</strong></p><ul><li>If your entire knowledge base easily fits in the model’s context, you get almost instant answers (no retrieval step).</li><li>Best for stable datasets that rarely change.</li></ul><p><strong>Key Limitations</strong></p><ul><li><strong>Context Size:</strong> If you exceed the model’s capacity, you must truncate or compress, killing accuracy.</li><li><strong>Local Requirement:</strong> Real caching needs control over memory, meaning you must run the model on your own infrastructure.</li><li><strong>Frequent Updates:</strong> Reloading the entire knowledge in context is impractical for dynamic data.</li></ul><h3>2. The BIG BUT (and We Cannot Lie)</h3><p>Long-context LLMs (like Google Gemini or Claude with hundreds of thousands of tokens) are emerging, making CAG more appealing for some workloads.</p><p><strong>But there’s a big condition:</strong></p><ul><li><strong>You must run the model locally </strong>and have access to its memory to enable caching. Many high-powered LLMs are hosted, limit context lengths, and obviously, you can’t access the memory for user-level manipulation via an API.</li><li><strong>Once your dataset crosses a threshold</strong> you might exceed the context window. If that happens, the method can break entirely or force you to truncate vital info, tanking accuracy.</li></ul><p>This snippet from one error log says it all:</p><blockquote>“error”:{“message”:”This endpoint’s maximum context length is 131072 tokens. However, you requested about 719291 tokens…”}</blockquote><p><strong>Translation:</strong> You’re out of luck unless you compress or chunk your data which can reduce the performance by a lot.</p><h3>3. Our Benchmark Setup</h3><p>We used the <strong>HotpotQA</strong> dataset (known for multi-hop QA) and ran our tests on the <strong>meta-llama/llama-3.1–8b-instruct</strong> model. We posed <strong>50 questions</strong> each to two knowledge sizes — <strong>50 documents and 500 documents</strong> — to see how each method performs at different scales.</p><p>Because we used an API (OpenRouter) for KV-Cache, there was no actual “cache” or local memory optimization happening; we simply passed all documents in each request.</p><ul><li><strong>top_k=5</strong> for RAG, and <strong>no top_k</strong> for KV-Cache (it loads everything).</li><li><strong>No retrieval time comparison:</strong> Our focus is on semantic accuracy, since KV-Cache would trivially have zero retrieval overhead if it were truly caching locally.</li></ul><h3>4. Results</h3><p>Our benchmark tests on the HotpotQA dataset revealed interesting insights into the performance of RAG and KV-Cache (CAG) under different knowledge sizes.</p><p>Below are the key findings:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*hE0-p4w6ooekGHBf" /><figcaption>Figure 1: Average semantic similarity scores for KV-Cache (No Cache) and RAG across knowledge sizes (k=50 and k=500). Tests were conducted on the HotpotQA dataset using the meta-llama/llama-3.1–8b-instruct model, with 50 questions per knowledge size. KV-Cache used an API (OpenRouter) without local caching, while RAG employed top_k=5 for retrieval.</figcaption></figure><h3>Key Takeaways</h3><ul><li>KV-Cache Struggles with Scale: As the dataset grows, KV-Cache faces context size limits, which require prompt truncation or compression.</li><li>RAG Handles Complexity: RAG’s retrieval mechanism ensures only relevant documents are used, avoiding context overload and maintaining accuracy.</li></ul><p><strong>The Bottom Line</strong></p><p>While KV-Cache shines with small, stable datasets, RAG proves more robust for larger, dynamic knowledge bases, making it a better fit for real-world, enterprise-level tasks.</p><h3>5. KV-Cache (CAG): Pros &amp; Cons</h3><p>CAG can appear unbeatable in early or small-scale tests (e.g. ~50 documents). But scaling up to <strong>500+ documents</strong> reveals some crucial issues:</p><h3>Context Overflow</h3><p>When you exceed the model’s max context window, you risk prompt truncation or outright token-limit errors. <strong>Vital information gets cut, and accuracy suffers.</strong></p><h3>Local Hardware</h3><p>To truly leverage KV-Cache, you need <strong>direct access</strong> to the model’s memory. If you rely on a hosted or API-driven model, there’s no way to manage caching yourself.</p><h3>Frequent Updates</h3><p>Every time your data changes, you have to <strong>rebuild the entire cache</strong>. This overhead can <strong>undermine the supposed “instant” advantage</strong> that KV-Cache promises.</p><h3>6. Quizzing Time: Score Wars — Why ‘Rosie Mac’ is the Winner</h3><p>Not all scores tell the full story. When evaluating model responses, similarity metrics compare generated answers to a reference text. But what happens when one answer is more detailed than the reference? Does it get rewarded — or penalized? Let’s look at a real example from our benchmark.</p><h3>The Question:</h3><p><strong>Q:</strong> <em>Who was the body double for Emilia Clarke playing Daenerys Targaryen in Game of Thrones?</em></p><h3>Two Correct Answers:</h3><h4>Answer A</h4><p><em>“Rosie Mac was the body double for Emilia Clarke in her portrayal of Daenerys Targaryen in Game of Thrones.”</em></p><h4>Answer B</h4><p><em>“Rosie Mac.”</em></p><p>Which one do you think scored higher on our similarity metric? Most people might assume the more detailed answer (<strong>A</strong>) wins. But here are the actual scores:</p><ul><li><strong>Answer A:</strong> 0.60526</li><li><strong>Answer B:</strong> 0.98361</li></ul><p>Yes, the shorter <em>“Rosie Mac.”</em> received the higher score. Why? Because the <strong>ground truth</strong> reference answer was simply <em>“Rosie Mac”</em> — so the more detailed response introduced extra words that lowered the alignment score.</p><p>This doesn’t mean longer answers are worse — often, they provide better context. But it highlights why similarity metrics should be interpreted with caution, especially in nuanced or multi-hop reasoning tasks. Our overall results remain valid, but it’s important to <strong>look beyond raw scores</strong> to gain a comprehensive, unbiased perspective on how these models truly perform.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*D8XQwhbR3KgiFszGEV2s0A.jpeg" /><figcaption>Image Credit: <a href="https://www.linkedin.com/in/kevin-michael-schindler/">Kevin Michael Schindler</a></figcaption></figure><h3>7. Final Thoughts: No Free Lunch</h3><p>Yes, Cache-Augmented Generation can truly offer zero retrieval overhead — <strong>if</strong> your entire knowledge base and context can fit comfortably in your local LLM. But for many <strong>enterprise or multi-hop tasks</strong>, that’s a big <strong>“if.”</strong></p><p>If your data is <strong>large or updates frequently</strong>, RAG approaches like CustomGPT.ai may remain the <strong>more robust and flexible</strong> choice.</p><h3>8. Frequently Asked Questions</h3><h4><strong>What is Retrieval-Augmented Generation (RAG)?</strong></h4><p>It’s a technique that fetches external documents at inference time to enrich a model’s responses, allowing you to handle bigger or changing data sets without overloading the model’s context.</p><h4><strong>How did you measure semantic similarity?</strong></h4><p>We used a BERTScore model (“all-MiniLM-L6-v2”) to compare generated answers with ground-truth references.</p><h4><strong>What does “No Cache” KV-Cache mean in your diagrams?</strong></h4><p>It indicates we didn’t run an actual local caching mechanism. Instead, we replicated the effect by passing all documents via an API request each time, so we could compare its semantic accuracy without focusing on speed.</p><h4><strong>Why was HotpotQA used?</strong></h4><p>HotpotQA requires retrieving multiple documents to answer a single question, making it ideal for testing retrieval methods like RAG and highlighting KV-Cache’s limitations with large knowledge bases.</p><h4><strong>When is multi-hop retrieval needed?</strong></h4><p>When no single document contains the full answer — common in research, legal analysis, and complex reasoning tasks requiring fact linking.</p><h3>Learn More</h3><ul><li><strong>CustomGPT.ai:</strong><a href="https://customgpt.ai/"> RAG-as-a-Service for large or dynamic data</a></li><li><strong>VentureBeat Article:</strong><a href="https://venturebeat.com/ai/beyond-rag-how-cache-augmented-generation-reduces-latency-complexity-for-smaller-workloads/"> How Cache-Augmented Generation Reduces Latency &amp; Complexity</a></li><li><strong>CAG Official Repo:</strong><a href="https://github.com/hhhuang/CAG"> CAG Original</a></li><li><strong>Our Modified Fork (With Results):</strong><a href="https://github.com/KevinMichaelSchindler/CAG"> CAG Fork</a></li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=9078fdbcba2f" width="1" height="1" alt=""><hr><p><a href="https://pub.towardsai.net/rag-vs-cag-can-cache-augmented-generation-really-replace-retrieval-9078fdbcba2f">RAG vs. CAG : Can Cache-Augmented Generation Really Replace Retrieval?</a> was originally published in <a href="https://pub.towardsai.net">Towards AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[DeepSeek R1 : Is It Right For You? (A Practical Self‑Assessment for Businesses and Individuals)]]></title>
            <link>https://pub.towardsai.net/deepseek-r1-is-it-right-for-you-a-practical-self-assessment-for-businesses-and-individuals-fc0021d85b60?source=rss-841a3b5d13d5------2</link>
            <guid isPermaLink="false">https://medium.com/p/fc0021d85b60</guid>
            <category><![CDATA[deepseek]]></category>
            <category><![CDATA[deepseek-r1]]></category>
            <dc:creator><![CDATA[Alden Do Rosario]]></dc:creator>
            <pubDate>Mon, 27 Jan 2025 15:59:08 GMT</pubDate>
            <atom:updated>2025-01-28T12:05:24.424Z</atom:updated>
            <content:encoded><![CDATA[<h3><strong>DeepSeek R1: Is It Right For You? (A Practical Self‑Assessment for Businesses and Individuals)</strong></h3><p>Deepseek just turned the AI world upside down with its new R1 model. It’s all over the news, so I won’t repeat it here.But then, the fears too are justified — as <a href="https://medium.com/the-generator/deepseek-hidden-china-political-bias-5d838bbf3ef9">laid</a> out nicely by <a href="https://medium.com/u/b020b149fc2d">Jim the AI Whisperer</a> .</p><p><a href="https://medium.com/the-generator/deepseek-hidden-china-political-bias-5d838bbf3ef9">AI as political warfare: DeepSeek AI advances the global interests of the Communist Party of China</a></p><p>For example, here is part of the model’s system instructions (laden with Chinese govt interference) extracted by Jim.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*Qey9kr1NXfVjaJ7F" /><figcaption>Credit: <a href="https://medium.com/u/b020b149fc2d">Jim the AI Whisperer</a></figcaption></figure><p>If you’re weighing whether to adopt DeepSeek R1 for personal or business objectives, it’s worthwhile to assess its strengths and concerns against your own needs, values, and requirements.</p><p>Below is a self‑assessment framework you can use to determine whether DeepSeek R1 might be right for you.</p><h3>1. Use Case Clarity: What Do You Want from an AI?</h3><ol><li><strong>Content Generation &amp; Creative Work</strong></li></ol><ul><li><strong>Potential Benefits:</strong> DeepSeek R1’s advanced language modeling capabilities are reportedly strong. If your goal is to produce compelling marketing copy, draft articles, or generate creative content, you might find an upside in the model’s robust generative power.</li><li><strong>Potential Drawbacks:</strong> Some content categories may be censored or skewed by built‑in political or ideological constraints. If you rely on producing uncensored analysis or commentary (especially on certain social or political topics), you might find the outputs constrained.</li><li><strong>Key Questions:</strong> Are you looking for an AI that can handle politically sensitive or regulated topics? Or do you primarily need general, apolitical copy?</li></ul><p><strong>2. Customer Support &amp; Chatbot Integration</strong></p><ul><li><strong>Potential Benefits:</strong> A model with strong, human‑like conversation skills can offer 24/7 customer assistance, handle routine inquiries, and streamline support.</li><li><strong>Potential Drawbacks:</strong> If your brand or organization operates globally, you might worry about built‑in filters or ideological stances that clash with your corporate values, especially if customers raise sensitive questions.</li><li><strong>Key Questions:</strong> Will you be comfortable with possible “official” or heavily filtered replies from your chatbot? Are you willing to monitor or post‑process the AI’s responses to keep them aligned with your business policies?</li></ul><p><strong>3. Business Intelligence &amp; Research</strong></p><ul><li><strong>Potential Benefits:</strong> DeepSeek’s underlying engine appears to handle advanced reasoning tasks well, potentially excelling in summarizing business data, generating data insights, or providing research assistance.</li><li><strong>Potential Drawbacks:</strong> If certain lines of inquiry are restricted by the model’s core alignment, your research may be unintentionally narrowed — particularly on controversial or geopolitically sensitive areas.</li><li><strong>Key Questions:</strong> Does your team rely on neutral, unfiltered access to analyses (especially on internationally relevant topics)? Do you need an AI unencumbered by potential ideological stances?</li></ul><h3>2. Data Governance &amp; Privacy: How Is Your Data Handled?</h3><p><strong>1. Data Sharing &amp; Sovereignty</strong></p><ul><li><strong>Potential Benefits:</strong> DeepSeek R1 may offer attractive hosting options or local‑running instances that keep data in your environment. This is appealing for teams that want more direct control over the AI.</li><li><strong>Potential Drawbacks:</strong> If the model’s underlying infrastructure is physically or contractually tied to government data centers (as hinted by some references), there are open questions about the ultimate destination or usage of your prompts and data.</li><li><strong>Key Questions:</strong> Do you have strict data sovereignty requirements or compliance obligations that could conflict with DeepSeek’s data storage or transfer policies? Does your organization have the legal capacity to evaluate the model’s compliance with international privacy standards?</li></ul><p><strong>2. Security &amp; Encryption</strong></p><ul><li><strong>Potential Benefits:</strong> Advanced security protocols might be in place — there are mentions of quantum encryption or sophisticated channels. If you can verify this, it may be a plus for high‑sensitivity use cases.</li><li><strong>Potential Drawbacks:</strong> Security claims still need to be audited or validated by independent experts. If your sector is highly regulated (finance, healthcare, government), you need a clear chain of trust for your data.</li><li><strong>Key Questions:</strong> Are you comfortable accepting vendor claims of security at face value, or can you audit them? Does your IT security framework allow for third‑party solutions with unknown overseas ties?</li></ul><h3>3. Content Authenticity &amp; Ethical Considerations</h3><p><strong>1. Risk of Political or Ideological Bias</strong></p><ul><li><strong>Potential Benefits:</strong> Not all usage scenarios will trigger political or ideological issues. For straightforward tasks (like summarizing internal documents or providing coding help), the system’s underlying censorship might never surface.</li><li><strong>Potential Drawbacks:</strong> If you or your audience values unfettered discussion of any topic, embedded content restrictions could undermine trust. Moreover, bias in the model’s worldview might seep into subtle but important ways in brand messaging or public‑facing content.</li><li><strong>Key Questions:</strong> Does your brand want to maintain a neutral or independent stance? Could your customers perceive biased or censored outputs as negative, damaging your reputation?</li></ul><p><strong>2. Regulatory and Reputational Exposure</strong></p><ul><li><strong>Potential Benefits:</strong> If you’re operating in a domain where the Chinese market is strategic, using a model aligned with its policies might align you with local regulations.</li><li><strong>Potential Drawbacks:</strong> If your home market or internal policies strongly oppose censorship, or if you must uphold strict standards of freedom of expression, adopting a model with visible or hidden ideological constraints can bring negative PR.</li><li><strong>Key Questions:</strong> Will your stakeholders or clients question your choice if the AI is discovered to have internal “red lines”? Are you comfortable explaining these constraints in board meetings or to the public?</li></ul><h3>4. Technical Versatility &amp; Adaptability</h3><p><strong>1. Integration &amp; Customization</strong></p><ul><li><strong>Potential Benefits:</strong> If DeepSeek R1’s architecture supports customization or fine‑tuning for your in‑house tasks, you may gain advanced capabilities for specialized domains like finance, manufacturing, or biotech.</li><li><strong>Potential Drawbacks:</strong> The ability to “jailbreak” or circumvent censorship reveals an internal conflict in the model’s architecture. This raises questions about reliability: might your integrated solution unexpectedly refuse or alter outputs under certain prompts?</li><li><strong>Key Questions:</strong> How flexible is DeepSeek’s model? Does the vendor share documentation that details the model’s constraints or “break points”? Will you have a fallback method if you need a second model for sensitive tasks?</li></ul><p><strong>2. Scalability &amp; Performance</strong></p><ul><li><strong>Potential Benefits:</strong> Early reports suggest strong performance on reasoning tasks and the potential for quick model updates. At a fraction of the cost.</li><li><strong>Potential Drawbacks:</strong> The official update cycle references “biweekly retraining with new data.” This continuous feed might help keep it fresh, but it might also unpredictably alter its alignment or responses.</li><li><strong>Key Questions:</strong> Do you need a stable environment for your AI integration (with predictable update cadences)? Or do you benefit from an aggressively updated system?</li></ul><h3>5. Cultural, Ethical, and Operational Alignment</h3><p><strong>1. Cultural Fit:</strong></p><ul><li>If your personal or organizational values prioritize open discourse, you’ll want to scrutinize any hidden constraints. However, if your day‑to‑day usage does not brush against politically sensitive topics, these constraints may never surface.</li></ul><p><strong>Governance &amp; Risk Appetite:</strong></p><ul><li>Some organizations have robust compliance frameworks requiring thorough vendor risk assessments. If that’s you, factor in the possibility of state or party influence on the model’s outputs.</li></ul><p><strong>Legal Environment:</strong></p><ul><li>The question of potential oversight or intervention by non‑domestic authorities can raise compliance red flags. For personal use, it’s often less of a concern — though still worth considering if you’re using it on sensitive topics.</li></ul><h3>6. Decision Summary: Is DeepSeek R1 Right for You?</h3><ul><li><strong>Personal/Hobbyist Use:</strong><br>If you’re casually experimenting with AI to write short stories or answer day‑to‑day queries and you have no inclination to discuss sensitive socio‑political topics, DeepSeek R1 might offer advanced capabilities at little or no cost. You’ll want to keep an eye on its data handling if you share private information, though.</li><li><strong>Small Business / Startups:</strong><br>If you want to quickly embed AI in customer support, marketing, or research, DeepSeek R1’s performance could be compelling. However, take note of brand image and the possibility that certain user queries might be “restricted.” Potential backlash might arise if your customers discover ideologically censored content.</li><li><strong>Enterprise / Regulated Sectors:</strong><br>For enterprise usage, especially in sectors like finance, healthcare, defense, or media, you’ll want to conduct due diligence. Validate if the model’s data pipeline meets your privacy/security standards, and weigh reputational risk if a foreign government’s ideological constraints are built in. Some organizations may decide the risks are too high.</li><li><strong>Public Sector / Government:</strong><br>Governments that require AI for public‑facing or internal use will look closely at sovereignty, data compliance, and ideological neutrality. If independence from external political influence is critical, you’ll likely seek a different AI model or a more transparent vendor partnership.</li></ul><h3>7. Action Items and Next Steps</h3><ol><li><strong>Pilot in a Sandbox</strong><br>Test DeepSeek R1 on your own data in a controlled environment. Evaluate censorship triggers and bias in typical use cases before integrating widely.</li><li><strong>Conduct a Thorough Risk Assessment</strong><br>Include legal, security, and compliance teams in your pilot. Identify how potential constraints or hidden influences could harm brand perception or customer trust.</li><li><strong>Explore Alternative or Supplementary Models</strong><br>You might decide to use different models for different tasks — one for general creative tasks and another for sensitive topics. This multi‑model approach can help mitigate the risk that arises from ideological filters. Consider a distilled model (like <a href="https://openrouter.ai/deepseek/deepseek-r1-distill-llama-70b">DeepSeek R1 Distill Llama 70B</a>) that seems to have less censorship and interference built in.</li><li><strong>Monitor Vendor Roadmap</strong><br>If DeepSeek’s leadership can provide clarity on data governance or the possibility of a more neutral “international edition,” it could change your risk profile. Also <strong>monitor other vendors</strong> that could use the techniques implemented in Deepseek to create new models.</li><li><strong>Develop Clear Usage Policies</strong><br>If you adopt DeepSeek R1, craft internal policies about when and how it’s used, clarifying disclaimers for public‑facing interactions.</li></ol><h3>Final Thoughts and Takeaways</h3><p><strong>My thoughts on DeepSeek:</strong> the R1 model is <em>shockingly</em> good in response quality and planning — right up until you peek behind the curtain and see the system instructions saturated with overt government influence.</p><p>For most businesses outside of China, those censorship and propaganda concerns alone are <strong>likely deal-breakers</strong>.</p><p>Yet, there’s a silver lining: the <strong>technical innovations</strong> powering DeepSeek’s strong reasoning will almost certainly appear in other models — <em>minus</em> the political constraints.</p><p>Moreover, DeepSeek’s <a href="https://openrouter.ai/deepseek/deepseek-r1-distill-llama-70b"><strong>distilled version</strong></a> (which leverages LLaMA) is reportedly free from many of these problematic instructions and still offers robust capabilities. (Hint: Searching for “tank man” in this model provides a clear response — whereas it is blocked in the original model)</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/889/0*Ttp8m4sEnvoZ_aX9" /></figure><p><em>Have you tested DeepSeek R1 yet, or are you planning to? Share your experiences below. Let’s keep the conversation going.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=fc0021d85b60" width="1" height="1" alt=""><hr><p><a href="https://pub.towardsai.net/deepseek-r1-is-it-right-for-you-a-practical-self-assessment-for-businesses-and-individuals-fc0021d85b60">DeepSeek R1 : Is It Right For You? (A Practical Self‑Assessment for Businesses and Individuals)</a> was originally published in <a href="https://pub.towardsai.net">Towards AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
    </channel>
</rss>