<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Stories by Kunal Sawarkar on Medium]]></title>
        <description><![CDATA[Stories by Kunal Sawarkar on Medium]]></description>
        <link>https://medium.com/@KunalSavvy?source=rss-bee09cdc3c5b------2</link>
        <image>
            <url>https://cdn-images-1.medium.com/fit/c/150/150/1*4e7Xc3C5xMDjs-6jndmpDg@2x.jpeg</url>
            <title>Stories by Kunal Sawarkar on Medium</title>
            <link>https://medium.com/@KunalSavvy?source=rss-bee09cdc3c5b------2</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Sun, 17 May 2026 19:15:19 GMT</lastBuildDate>
        <atom:link href="https://medium.com/@KunalSavvy/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[Its the rise of other AI that all should worry about.. Artificial Intimacy!]]></title>
            <link>https://medium.com/towards-generative-ai/its-the-rise-of-other-ai-that-all-should-worry-about-artificial-intimacy-42600aef7fcd?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/42600aef7fcd</guid>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[ai-companion]]></category>
            <category><![CDATA[intimacy]]></category>
            <category><![CDATA[artificial-intimacy]]></category>
            <category><![CDATA[relationships]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Sun, 24 Aug 2025 19:05:08 GMT</pubDate>
            <atom:updated>2025-08-24T19:25:52.309Z</atom:updated>
            <content:encoded><![CDATA[<h4>The real threat isn’t AI taking your job — it’s AI taking your partner…</h4><p>Whenever I am speaking at an event, I am often asked about the AI impact on jobs. I’ve never truly believed AI is a job killer. In fact, I see it as a powerful booster, reshaping and elevating job profiles. Thanks to AI, people can truly realize their potential rather than having to slog hours in a low-end work. There will always be something for people to do — for those who <em>want</em> to do it. The real challenge now is: what if people no longer want to do anything? What if they lose the excitement for life itself… thanks to artificial intimacy? That is what came out of the recent discussion with Gen Z, and the discussion to “<em>try AI as a partner</em>” worried me!!</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/499/1*ugfgHPdkI63RyldaqMOOkw@2x.jpeg" /><figcaption>Credit- HBO</figcaption></figure><h4>First question is why fall in love with AI?</h4><p>Because it is so eloquent!! Words feel powerful, but doesn’t the tone &amp; touch convey more than words ever could?</p><p>Often, language itself is a barrier to truly expressing deep emotion. Written or spoken language has only existed for a few thousand years, while humans have been evolving for over a million. The fact that someone articulates in words can convince you it’s a real relationship only highlights how fragile our sense of intimacy is. Words can express feelings, but only up to a point. They can’t express empathy or true presence. That comes through the eyes, the touch, the presence of another human. Just like true art, is the one that touches one deeply and which often leaves us speechless, a real relationship exists beyond words.</p><blockquote>Can AI ever give you that?</blockquote><p>If there isn’t even a single moment that one has experienced that goes <em>beyond</em> words, that’s a shame. Even animals form relationships without language — and often, language is more of a barrier to emotion than a conduit for it.</p><h3>Humans are not made of Lego pieces!</h3><p>Then there’s the problem of <strong>customization</strong>. You can walk into any store and customize your coffee the way you want… from beans… to sugar… to type of sugar… to milk… to additives… to cup… to straw and whatnot!! That creates a sense that “I can get anything the way I want it” and often builds a belief that — because I deserve the best, another human must be the way “<em>I want it</em>”!</p><p>And then the digital world feeds into this personalization mindset by taking it stratosphere, serving you the content exactly what you want at all times. But in real life, you can’t “customize” a partner. They are a monolith..as stones and plants in nature..moldable but customizable!</p><p>They simply <em>are who they are</em>, and from limited choices, you make a commitment — then work to make it last. That’s why AI intimacy feels so seductive. It gives you what you want with no effort, no compromise, and no hard work. That is just like a drug!! Temporary pleasure at bare minimum cost…</p><h3>Intimacy is at the core of what makes us human!</h3><p>And what is the true purpose of a relationship? Is it someone who always supports you, praises you, and tells you that you’re perfect? That kind of endless positive feedback doesn’t even work in real life. A real partner is a mirror: someone who helps you grow, tells you when you’re wrong, and still stands by you despite all the flaws.</p><blockquote>As they say, infatuation is when you find another person perfect. Love is when you know the flaws and still choose the same person!</blockquote><p>Therapy is different from partnership. A partner cares deeply enough to challenge you, not just comfort you. It is also a lack of real and meaningful growth over a long period of time! How can someone want a partner who they can lose with a poor internet connection or a low battery?</p><p>And that’s a real problem… many not having anyone to talk to… anyone to share emotions with… It’s a social isolation that’s a problem, and AI is just a drug to escape it or feel good momentarily… just like social media or Xbox One.</p><h3>Loss of Innocence</h3><p>And lastly, the innocence part of a relationship.</p><p>The most beautiful part in any relationship is discovering… not just a perfect person or imperfect ones!! Coz the dynamism, the curiosity, the commitment tested by challenges — that’s what makes relationships meaningful. The loss of losing one’s innocence to not another human but to a machine will create ever more unrealistic expectations from relationships, almost making AI an equivalent to emotional porn!</p><p>Super smart, super supportive AI can be very boring. All the growth in nature, as the theory of evolution has proven, is based on the need to evolve in response to a challenge; but when the challenge is gone from life… will this be the end of evolution in itself?</p><h3>Not just a social but an existential threat</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*N7Bl3Ibyj3taqMktMA2WUA.png" /></figure><p>Despite all this, half of Gen Z who spoke with me still said: <em>“Yeah, it’s true… but it feels like the writing is on the wall.”</em> Meaning: AI may soon replace boyfriends or girlfriends. At the very least, relationships built the premise of the future, growth, and family are at risk. And those are the foundations of society and the economy. Without a real human connection, what even motivates consumption, creativity, or progress? AI may add a temporary economic boost, but if people don’t “<em>desire another human,</em>” the true purpose of a consumer-based economy will vanish. A more Japan-like situation where desire has vanished from society.</p><p>My view? Cocaine is better than an AI companion — at least you won’t be deluded into thinking you’re in a real relationship. Preferring machines over humans is simply a cop-out at best and hallucinations at worst! (pun intended)</p><p>Everything worthwhile in life is hard — and that’s exactly why it’s worth it.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=42600aef7fcd" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/its-the-rise-of-other-ai-that-all-should-worry-about-artificial-intimacy-42600aef7fcd">Its the rise of other AI that all should worry about.. Artificial Intimacy!</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[What Training AI on Math Teaches Us About Hiring the Right People?]]></title>
            <link>https://medium.com/@KunalSavvy/what-training-ai-on-math-teaches-us-about-hiring-the-right-people-5d6b4dc3d86c?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/5d6b4dc3d86c</guid>
            <category><![CDATA[llm]]></category>
            <category><![CDATA[hiring]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[genai]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Wed, 06 Aug 2025 21:32:05 GMT</pubDate>
            <atom:updated>2025-08-06T21:32:05.943Z</atom:updated>
            <content:encoded><![CDATA[<p>AI models generalize well if it works well on math benchmarks and is key part of training process. As we train AI models on mathematical datasets — from arithmetic to algebra to competition-level reasoning — we’re not just teaching machines. We’re learning how humans should be evaluated, hired, and developed.</p><p><strong>LLaMA Training on Math &amp; Reasoning Tasks</strong></p><ol><li>LLaMA shows major improvement when fine-tuned on structured math datasets.</li><li>Meta researchers (and others who used open LLaMA weights) discovered that even base LLaMA models struggle with multi-step reasoning unless fine-tuned on:</li><li>•	GSM8K: Grade school math problems requiring step-by-step logic.</li><li>•	MATH Dataset: High school and competition-level math.</li><li>•	Code and formal logic datasets: Help it reason and structure its answers better.</li></ol><p>📌 Result: LLaMA models trained with these datasets learn to compose solutions, not just guess answers — performing far better than models trained solely on language tasks.</p><p><strong>Here’s what we’ve learned:</strong></p><p>Insight #1: Structured learning makes generalization possible</p><p>•	A LLaMA model with training on math + logic generalizes better to other domains like code generation, reasoning tasks, and even legal document summarization.</p><p>•	Similarly, a human with strong logical and analytical foundations can be reskilled across roles faster than someone trained narrowly in a single tool or domain.</p><p>Insight #2: Shallow pretraining limits deep reasoning</p><p>•	LLaMA-2 or LLaMA-3 without any fine-tuning will often fail to reason, hallucinate, or shortcut answers.</p><p>•	Analogously, a candidate with surface-level experience (e.g., only copy-pasting code or following instructions) underperforms in novel or ambiguous environments.</p><p>Insight #3: Chain-of-Thought (CoT) training enhances step-by-step reasoning</p><p>•	When LLaMA models are trained to explain their steps (via CoT), performance jumps significantly.</p><p>•	For hiring, this is like valuing candidates who explain their thought process, not just give the right answer. It reflects deep understanding.</p><p><strong>Key Takeaways</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/768/1*rxPMGh81uSNg_I4JHS8uLA@2x.jpeg" /></figure><ol><li>Core Skills Enable Generalization</li><li>AI models trained on fundamentals solve new problems better than those trained only on task-specific data.</li></ol><p>➡️ Hire for strong foundations — logic, reasoning, and structured thinking — not just narrow experience.</p><p>2. Memorization Doesn’t Scale</p><p>Models that “memorize” patterns fail when faced with variations.</p><p>➡️ Don’t hire based on resume keywords. Look for people who learn, not just those who have “seen it before.”</p><p>3. Compositional Thinking Wins</p><p>The best AI models solve problems by composing smaller building blocks of logic.</p><p>➡️ Hire people who can break down problems, recombine ideas, and adapt.</p><p>4. Diverse Training = Better Abstraction</p><p>Models trained on varied examples abstract better and generalize across domains.</p><p>➡️ Value diverse experience paths, not just linear ones.</p><p>The Big Insight:</p><p>Building great teams is like building great AI — train for depth, not just coverage.</p><p>Core reasoning scales. Superficial knowledge breaks.</p><p>Would love to hear from others building hiring practices or talent programs. What core skills do you screen for?</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=5d6b4dc3d86c" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[What is AI SuperIntelligence — and How Would We Know If We’ve Reached It?]]></title>
            <link>https://medium.com/towards-generative-ai/what-is-ai-superintelligence-and-how-would-we-know-if-weve-reached-it-ecaa2ea63123?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/ecaa2ea63123</guid>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Sun, 03 Aug 2025 16:47:59 GMT</pubDate>
            <atom:updated>2025-08-03T16:47:59.910Z</atom:updated>
            <content:encoded><![CDATA[<h3><strong>What is AI SuperIntelligence — and How Would We Know If We’ve Reached It?</strong></h3><h4>And is Zuckerberg’s vision hype or the near future?</h4><p>Superintelligence is the hypothetical point when AI becomes smarter than humans in <em>every domain</em> — not just memory or math, but reasoning, creativity, emotional intelligence, and even social manipulation. It’s not just an evolution of today’s tools — it’s a fundamental shift in who (or what) leads the frontier of intelligence on Earth.</p><h3>How is SuperIntelligence defined?</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*7NZtGY3xW-xZmrlaRpCa7w.png" /></figure><p>The term comes primarily from philosopher <strong>Nick Bostrom</strong>, who described it in his book <em>Superintelligence: Paths, Dangers, Strategies</em> (2014). He defines superintelligence as:</p><p><em>“An intellect that is much smarter than the best human brains in practically every field, including scientific creativity, general wisdom and social skills.”</em></p><p>Though the credit of making it popular and cool in recent times has to go to OpenAI chief scientist <strong>Ilya Sutskever, </strong>who started his company with the name “<strong>Safe SuperIntelligence</strong>” ..quite obviously to find a new term than AGI, which was sort of hijacked by OpenAI and Microsoft. The company is still in pre-product stage but got amazing $30B valuation.</p><h3>Types of AI on the spectrum</h3><p>To understand superintelligence, it helps to place it on the AI capability spectrum:</p><ol><li><strong>Narrow AI (ANI)</strong> — What we have now (e.g., GPT-4, AlphaFold, self-driving software)</li><li><strong>General AI (AGI)</strong> — Human-level intelligence across tasks</li><li><strong>Superintelligence (ASI)</strong> — Beyond human in <em>all</em> cognitive tasks</li></ol><p>But how would we even recognize it?</p><p>We know we’ve reached it when an AI consistently outperforms humans in all intellectual tasks, demonstrates general problem-solving ability, and exhibits self-improvement without human intervention, with measurable outcomes like solving previously unsolvable problems or creating novel advancements independently.</p><p>It might look like:</p><ul><li>AI solving problems we can’t even articulate (like reconciling quantum gravity and general relativity),</li><li>Creating its own scientific frameworks,</li><li>Optimizing economies, policies, or infrastructure with godlike foresight,</li><li>Or improving itself recursively — learning how to learn better, on its own.</li></ul><p>There might not be a single “Eureka” moment. The signs could be subtle — or intentionally hidden. That’s part of the risk.</p><p><strong>How Close Are We?</strong></p><p>We’re still in the <strong>narrow AI</strong> and early <strong>proto-AGI</strong> era. Today’s most advanced models (like GPT-4, Claude, Gemini) show impressive capabilities — but they lack persistent reasoning, autonomous goal-setting, or deep situational awareness.</p><p>What still stands in the way?</p><ul><li><strong>Memory &amp; agency</strong>: Current models don’t “think” in continuous time — they react in sessions.</li><li><strong>Grounded understanding</strong>: They’re brilliant mimics, but often guess instead of <em>know</em>.</li><li><strong>Alignment &amp; safety</strong>: Even if we <em>could</em> build it, how do we make sure it shares our goals?</li></ul><p>Building ASI needs to overcome challenges like</p><ul><li><strong>Compute Power</strong>: Superintelligence needs vastly more processing than today’s systems, requiring breakthroughs in hardware efficiency.</li><li><strong>Algorithms</strong>: We need new paradigms for AI to learn broadly and reason like humans, not just mimic patterns.</li><li><strong>Energy Costs</strong>: Scaling AI demands unsustainable energy unless greener solutions emerge.</li><li><strong>Safety and Ethics</strong>: Uncontrolled superintelligence could pose risks, requiring robust safeguards.</li></ul><h3><strong>Zuckerberg’s Vision: Empowerment or Hype?</strong></h3><p>Mark Zuckerberg recently framed AI as a <em>“universal assistant”</em> — personal, embedded, always-on. It’s a compelling vision: AI not replacing you, but <strong>amplifying you</strong>.</p><h4><em>Is it Realistic?</em></h4><p>Yes — <strong>at the narrow level</strong>. I already speculated almost 2 years back, having an “AI Buddy” who manages all the software and tools i use daily is game game-changer. Though it turned out harder, it‘s i’s still technically doable. We’re already seeing AI copilots across design, coding, learning, and personal productivity. Meta’s Llama models are a play to own that personal layer.</p><p>But as we approach AGI (and eventually superintelligence), the conversation has to shift.</p><p>Zuckerberg envisions “personal superintelligence” via devices like smart glasses, empowering individuals to achieve goals and enhance daily life. It’s a compelling idea — AI as a tailored assistant — and Meta’s focus on open-source models like Llama and billions in R&amp;D shows commitment, but their pivot to closed models suggests commercial control over altruism.</p><h3>Will We Ever Get There?</h3><p>I’m doubtful that, with current technology or our understanding of intelligence, we’ll get there.</p><p>We still don’t fully understand the brain or have a clear mathematical model of how it functions — or how electrical signals translate into thoughts. Human (or natural) intelligence has been shaped by survival. What would motivate a superintelligence?</p><p>I am also skeptical of true creativity using ASI that can solve problems like Riemann’s Hypothesis.</p><p>I’m skeptical of the commonly cited “self-preservation” instinct in AI — it doesn’t hold up under scrutiny.</p><p><strong><em>It may be possible in the far future, but not now.</em></strong></p><p>Superintelligence isn’t science fiction — it’s a design question. The sooner we get serious about alignment, the better shot we have at making it a feature of the future, not a bug. We’re inching toward superintelligence, but it’s not around the corner.</p><p>Zuckerberg’s vision blends possibility with self-serving spin At this point, it feels more like great marketing — and possibly a hijacking of the term “superintelligence” to suit commercial product goals. That’s not necessarily a bad thing. Any AI that meaningfully enhances how we interact in the digital world will <em>feel</em> like a superintelligence. But once disconnected, it won’t hold up.</p><p>Now, of course, the real question is: <strong>can today’s human live a life without a digital connection?</strong></p><p>Follow <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> for the latest advancements in AI.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=ecaa2ea63123" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/what-is-ai-superintelligence-and-how-would-we-know-if-weve-reached-it-ecaa2ea63123">What is AI SuperIntelligence — and How Would We Know If We’ve Reached It?</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[What is Context Engineering? And Why Is Everyone Talking About It?]]></title>
            <link>https://medium.com/towards-generative-ai/what-is-context-engineering-and-why-is-everyone-talking-about-it-9138b8556af4?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/9138b8556af4</guid>
            <category><![CDATA[context-engineering]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[genai]]></category>
            <category><![CDATA[prompt-engineering]]></category>
            <category><![CDATA[llm]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Sun, 03 Aug 2025 15:46:25 GMT</pubDate>
            <atom:updated>2025-08-03T15:46:25.514Z</atom:updated>
            <content:encoded><![CDATA[<h4>Here is a primer on understanding how it differs from prompt engineering</h4><p>At a high level, Context Engineering is the systematic design and manipulation of everything you feed into an LLM to control its reasoning and output, without touching the model weights.</p><p>It is the LLM equivalent of programming, but instead of code + API, your inputs are natural language + structured context. It’s an improved version of prompt engineering but less than fine tuning.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*oKit6Sienv1aiOqh.png" /><figcaption>Overview</figcaption></figure><p>CE includes:</p><ul><li>Prompts (instructions, examples, few-shot demos)</li><li>Retrieval results (via RAG — Retrieval Augmented Generation)</li><li>Conversation history</li><li>Metadata and tool state (e.g., functions, documents, sensor data)</li><li>Agent memory and planning state</li></ul><p><strong>🧠 </strong>Let’s break down the technical pillars:</p><p><strong>1.Prompt Design</strong></p><ul><li>Includes prompt templates, instruction tuning, and few-shot examples.</li><li>Now evolving into dynamic prompting where prompts are assembled based on user, task, and system state.</li></ul><p>Examples:</p><p>You are a legal expert. Answer the following legal question step-by-step using citations from relevant documents.</p><ul><li>This is the lowest-level part of context engineering but still foundational.</li></ul><p><strong>2. Retrieval-Augmented Generation (RAG)</strong></p><ul><li>Key innovation: enrich the context window with relevant external data.</li><li>Typical architecture:</li><li>User query → embedding → vector DB search → top-k docs → appended to prompt</li><li>Then → LLM generates answer using both user input + retrieved docs</li></ul><p>Engineering challenges remains ( see our paper — <a href="https://arxiv.org/abs/2404.07220">https://arxiv.org/abs/2404.07220</a>)</p><ul><li>Chunking &amp; encoding</li><li>Context window limits</li><li>Relevance scoring</li><li>Overlap/resolution of retrieved content</li></ul><p><strong>3. Context Composition</strong></p><ul><li>Now moving beyond static prompts to composable context stacks:</li><li>Conversation history</li><li>Search results</li><li>Current tool usage</li><li>Agent goals &amp; plans</li><li>User profile/preference</li><li>Engineers decide what to include, exclude, highlight, rephrase, or abstract.</li></ul><p>This is similar to designing the state representation in classical AI systems.</p><p><strong>4. Memory and Long-Term Context</strong></p><p>This has been the biggest challenge in working with the LLMs</p><ul><li>LLMs have fixed token windows (e.g., 8K, 128K, etc.)</li><li>Long-term memory systems inject past conversations, facts, or embeddings relevant to the current task.</li><li>Context engineers decide how and when to inject memory, and what format (text vs. structured).</li></ul><p><strong>5. Tool &amp; API Context</strong></p><ul><li>In agentic systems, the LLM sees tools (e.g., code interpreter, browser, calculator) as part of context.</li><li>You design how tool outputs are injected:</li><li>Tool traces</li><li>Intermediate steps</li><li>Function outputs as inline context</li></ul><blockquote>Example:</blockquote><blockquote>{</blockquote><blockquote>“tool”: “python”,</blockquote><blockquote>“input”: “plot a sine wave”,</blockquote><blockquote>“output”: “&lt;image bytes&gt;”</blockquote><blockquote>}</blockquote><p>This becomes part of the next LLM step’s context.</p><p><strong>🧱 Key Design Patterns</strong></p><ul><li>Chain-of-Thought + Tool Invocation: e.g., “Think step-by-step. If calculation is needed, use the calculator tool.”</li><li>Reflexion Loops: Let the model critique its previous answer and improve.</li><li>Role Prompting: “You are a financial analyst with 20 years of experience.”</li><li>Scratchpad Contexts: Store intermediate state as it works through</li><li>As models get smarter, the bottleneck moves to context quality and orchestration. That’s why context engineering is becoming one of the most in-demand skills in AI systems development.</li></ul><p>Follow <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> for more on the latest advancement in AI.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=9138b8556af4" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/what-is-context-engineering-and-why-is-everyone-talking-about-it-9138b8556af4">What is Context Engineering? And Why Is Everyone Talking About It?</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[What you need to know on Agentic Protocols MCP vs A2A, vs ACP]]></title>
            <link>https://medium.com/towards-generative-ai/difference-between-agentic-protocols-mcp-vs-a2a-vs-acp-7ad6fcf5c91c?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/7ad6fcf5c91c</guid>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[agentic-ai]]></category>
            <category><![CDATA[llm]]></category>
            <category><![CDATA[genai]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Sun, 27 Apr 2025 18:46:19 GMT</pubDate>
            <atom:updated>2025-05-03T00:12:06.122Z</atom:updated>
            <content:encoded><![CDATA[<h4>There are so many agentic protocols now. But what is the difference between MCP vs A2A vs ACP, and when to use which one?</h4><p>Here’s a concise bullet list comparing and contrasting the Model Context Protocol (MCP), Agent-to-Agent Protocol (A2A), and Agent Connect Protocol (ACP) in agentic AI, focusing on usability:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/961/1*bIXajIGtIJE5OojNfpKUvw.png" /></figure><h3>1. MCP from Anthropic</h3><p>- Standardizes how AI agents access tools and data, providing context to LLMs.Simplifies integration with external tools (e.g., databases, APIs) via a client-server model, reducing custom coding needs.</p><p>- Reusable and secure connections enhance developer efficiency and scalability.</p><p>- Intuitive for developers familiar with LLM workflows, as it acts like a “universal adapter” for resources.</p><p><strong>Challenges:</strong></p><p>- Primarily focused on data access, not agent-to-agent communication, limiting its scope for collaborative tasks.</p><p>- Component naming (Hosts, Clients, Servers) can be confusing, potentially steepening the learning curve.</p><h3>2. A2A (Agent-to-Agent Protocol) from Google</h3><p>Enables seamless communication and collaboration between AI agents across vendors and platforms. Streamlines agent orchestration with clear task delegation and capability discovery (via Agent Cards), making multi-agent workflows intuitive.</p><p>- Supports dynamic, human-like interactions (e.g., bidirectional streaming), enhancing user and developer experience.</p><h4>Challenges:</h4><p>Requires developers to manage complex agent interactions, which can increase setup time for smaller projects.</p><p>Overlaps with MCP in some use cases, potentially causing confusion</p><h3>3. ACP (Agent Connect Protocol) from IBM</h3><p>Purpose: Facilitates an “internet of agents” by enabling interoperability among diverse AI agents, like LangChain.</p><p>Usability Strengths:</p><p>- Designed for cross-organizational agent networks, offering high interoperability for enterprise-scale systems.</p><p>- Simplifies integration of heterogeneous agents, reducing setup complexity in distributed environments.</p><h4>Challenges:</h4><p>Focus on broad interoperability may sacrifice depth in specific use cases, requiring additional customization.</p><h3>Key Contrast:</h3><p>MCP excels in tool integration for individual agents, prioritizing data access simplicity. Best for developers needing quick, secure tool integration.</p><p>A2A focuses on agent collaboration, offering robust communication for multi-agent systems. Best for collaboration.</p><p>ACP emphasizes broad interoperability, but its early stages compared to the more established MCP and A2A. Best for enterprises who want to take their Agentic AI Solution to production.</p><p>Follow <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> for more content related to the latest advancement in AI.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=7ad6fcf5c91c" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/difference-between-agentic-protocols-mcp-vs-a2a-vs-acp-7ad6fcf5c91c">What you need to know on Agentic Protocols MCP vs A2A, vs ACP</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Just How Did They Discover Life on an Alien Planet Far Far Away?]]></title>
            <link>https://medium.com/towards-generative-ai/just-how-did-they-discover-life-on-an-alien-planet-far-far-away-efbd10a5707c?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/efbd10a5707c</guid>
            <category><![CDATA[astrophysics]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[statistics]]></category>
            <category><![CDATA[aliens]]></category>
            <category><![CDATA[astronomy]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Sun, 20 Apr 2025 17:45:47 GMT</pubDate>
            <atom:updated>2025-04-20T17:54:41.684Z</atom:updated>
            <content:encoded><![CDATA[<h4>The discovery of DMS on Alien Planet K2–18b has a lot to do with statistical modelling (or AI as companies want to call it now)! Here is a simplified version from the paper on how it was done.</h4><p>In case you have missed we have detected a sign of alien life for the first time ever, on a planet 120 light years away. This first time we discovered anything like it, and I was more curious to know what exact process and algorithms were used to deduce this. Here is the summary of the published paper.</p><h3>Unveiling Alien Atmospheres: How Scientists Detected DMS Gas on K2–18b</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*s1VwyH4DDv8WRvW1" /><figcaption>Comparison Between Two</figcaption></figure><p>Imagine peering into the atmosphere of a planet 124 light-years away and finding a clue that could hint at alien life. That’s exactly what scientists have done with K2–18b, an exoplanet where the James Webb Space Telescope (JWST) detected signs of dimethyl sulfide (DMS) — a molecule linked to life on Earth. In this blog post, we’ll dive into the cutting-edge data analysis and algorithms that made this discovery possible, breaking down the science behind this cosmic detective work.</p><h3>The Basics of observing a distant planet</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/700/0*5ndWWLMLwutL9oIu.jpg" /><figcaption>Source Nature magzine</figcaption></figure><p>We can&#39;t really “see” the distant planet light years away through the traditional telescope. So what we instead do is watch for the shadow and gravitational pull of the planet when it orbits around its star (or sun). It leaves a trail while passing through the star, and we can then observe the spectrum of the star.</p><p>The beauty of the spectrum is that every element and every molecule has unique spectral lines in spectral analysis, thus allowing us to detect what is in the atmosphere of the alien planet.</p><p>This method is so simple that even a 5th grader can do it. So now let&#39;s jump into how this discovery was done.</p><h3>1. Capturing Starlight: The Power of Transit Spectroscopy</h3><p>To study K2–18b, a sub-Neptune in the habitable zone of a red dwarf star, scientists used JWST’s (James Webb Telescope) advanced instruments:</p><ul><li><strong>In 2023 Observations</strong>, the <strong>Near-Infrared Imager and Slitless Spectrograph (NIRISS)</strong> and <strong>Near-Infrared Spectrograph (NIRSpec)</strong> captured spectra from 0.9 to 5.2 micrometers, revealing methane, carbon dioxide, and a tentative DMS signal.</li><li><strong>So in 2025 Observations </strong>the <strong>Mid-Infrared Instrument (MIRI)</strong> targeted 6–12 micrometers, boosting the DMS signal</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*i4LzOEhFmf9trU6C.jpg" /><figcaption>(Credit: NASA)</figcaption></figure><p>The technique, called <strong>transit spectroscopy</strong>, involves analyzing starlight as it passes through the planet’s atmosphere during a transit. Molecules like DMS absorb specific wavelengths, creating a spectral “fingerprint” that scientists can decode.</p><h3>2. Cleaning Up the Data: From Raw to Refined</h3><p>Raw JWST data is messy, filled with noise from the telescope’s detectors and cosmic rays. Here’s how scientists polished it:</p><ul><li><strong>Pipeline Processing</strong>: The <strong>JWST Science Calibration Pipeline</strong> removed noise, corrected for thermal background, and produced calibrated spectra.</li><li><strong>Spectral Binning</strong>: Spectra were binned to improve signal-to-noise ratio, with resolutions of ~100 for NIRISS/NIRSpec and ~50–100 for MIRI.</li><li><strong>Wavelength Calibration</strong>: Spectra were aligned with lab references for DMS, methane, and CO₂ to ensure accurate feature identification.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*HJHp0tOzeQdI73AJ.jpg" /><figcaption>Raw (noisy) vs. calibrated (smoothed) spectra, showing the impact of data reduction. (Adapted from JWST documentation)</figcaption></figure><p>This preprocessing turned chaotic data into a clear signal, ready for analysis.</p><h3>3. Decoding the Atmosphere: Retrieval Algorithms at Work</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*YjOG89COMloG1FFi.png" /></figure><p>To identify DMS, scientists used <strong>atmospheric retrieval algorithms</strong>, which compare observed spectra to models of the planet’s atmosphere. These are AI algorithms which belongs to class of Bayesian Modelling. It combines Bayesian with Markov Chain Monte Carlo Simulation to find estimates. (It is also my fav class of models, having used it years prior to measure plastics in the world’s oceans. Full Blog here- <a href="https://www.ibm.com/think/insights/toward-a-world-of-plastic-free-beaches">https://www.ibm.com/think/insights/toward-a-world-of-plastic-free-beaches</a>)</p><ul><li><strong>NEMESIS</strong>: A Bayesian modelling tool that optimizes molecular abundances and temperature profiles.</li><li><strong>SCARLET</strong>: Tailored for hydrogen-rich atmospheres, using Markov Chain Monte Carlo (MCMC) to estimate uncertainties.</li></ul><p>These algorithms treat the atmosphere as layered, solving radiative transfer equations to predict spectral features. They also used <strong>molecular cross-sections</strong> to model how DMS, methane, and CO₂ absorb light. The challenge? DMS and methane spectra overlap, requiring careful analysis across wavelengths.</p><h3>4. Measuring Confidence: Statistical Significance</h3><p>How sure are scientists about the DMS detection? They used statistical model to find out:</p><ul><li><strong>Bayesian Evidence</strong>: <strong>Nested Sampling</strong> (via MultiNest/Dynesty) compared models with and without DMS, yielding a 3-sigma (99.7%) confidence in 2025, up from 1–2.4-sigma in 2023.</li><li><strong>Goodness-of-Fit</strong>: The chi-squared statistic ensured models matched data, while AIC/BIC prevented overfitting.</li><li><strong>Degeneracy Handling</strong>: Combining NIRISS, NIRSpec, and MIRI data helped separate DMS from methane signals.</li></ul><p>While 3-sigma is promising, a 5-sigma threshold is needed for certainty, driving plans for more observations.</p><h3>5. Ensuring Reliability: Robustness Tests</h3><p>To confirm the DMS signal wasn’t a fluke, scientists ran rigorous tests:</p><ul><li><strong>Monte Carlo Simulations</strong>: Synthetic DMS signals were added to noisy spectra to check for false positives.</li><li><strong>Cross-Validation</strong>: Multiple retrieval codes (NEMESIS, SCARLET) produced consistent results.</li><li><strong>Systematic Error Correction</strong>: Gaussian Process Regression modeled noise, enhancing accuracy.</li></ul><p>These tests bolstered confidence in the 2025 findings.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*su2sK4_CZNaD6f6h.jpg" /></figure><p>Monte Carlo simulation showing detection significance vs. noise, with 3σ threshold. (Adapted from exoplanet studies)</p><h3>6. Modeling the Chemistry: Biological or Not?</h3><p>Could DMS indicate life? Can it come from non-biological sources? That is a hard question to answer. Scientists used <strong>photochemical models</strong> to explore its origins:</p><ul><li><strong>KINETICS/VULCAN</strong>: These models simulated atmospheric chemistry, suggesting a biological source for the high DMS levels (~10 ppm).</li><li><strong>Cloud and Haze Models</strong>: Mie scattering models indicated a clear atmosphere, ideal for detecting DMS.</li></ul><p>While promising, abiotic sources (like comets) remain a possibility, requiring further study.</p><h3>7. Overcoming Challenges</h3><ul><li><strong>Spectral Overlap</strong>: DMS and methane features blend, complicating identification.</li><li><strong>Low Signal Strength</strong>: The 2023 signal was weak; 2025 improved but needs confirmation.</li><li><strong>Abiotic Sources</strong>: DMS in comets suggests non-biological origins, challenging its biosignature status.</li></ul><p>MIRI’s mid-infrared data helped, but more observations are crucial.</p><h3>8. What’s Next?</h3><p>This still needs more data and more analysis. We will now have to wait another 2 years to collect more data when the planet will pass through the star. Few causes of false positives</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/708/0*OmgNcI7mmVnDM2ho.jpg" /><figcaption>James Webb Telescope</figcaption></figure><ul><li><strong>More JWST Observations</strong>: 16–24 hours of data aim for a 5-sigma detection.</li><li><strong>Neural Networks</strong>: Advanced algorithms may improve sensitivity.</li></ul><p>With JWST’s MIRI leading the charge, we’re closer than ever to answering whether K2–18b harbors life.</p><h3>Conclusion</h3><p>The detection of DMS on K2–18b is a thrilling step in the search for extraterrestrial life. Through transit spectroscopy, sophisticated algorithms, and rigorous testing, scientists are peeling back the layers of an alien atmosphere. While challenges remain, the promise of JWST’s future observations keeps us on the edge of our seats.</p><p>Could K2–18b be home to life? Stay tuned! The quest for life continues..</p><p>Follow <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> for more content related to the latest in AI advancement.</p><p><em>C</em>itations: Madhusudhan et al., 2023 &amp; 2025, <em>The Astrophysical Journal Letters</em>; NASA/ESA press releases; X posts on JWST’s MIRI.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=efbd10a5707c" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/just-how-did-they-discover-life-on-an-alien-planet-far-far-away-efbd10a5707c">Just How Did They Discover Life on an Alien Planet Far Far Away?</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Why Grok 3 is best LLM out there !!]]></title>
            <link>https://medium.com/towards-generative-ai/why-grok-3-is-best-llm-out-there-ba7ab315f8cc?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/ba7ab315f8cc</guid>
            <category><![CDATA[grok-3]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[llm]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Mon, 31 Mar 2025 21:59:28 GMT</pubDate>
            <atom:updated>2025-03-31T21:59:28.645Z</atom:updated>
            <content:encoded><![CDATA[<h4>Grok3, built by xAI, is a groundbreaking AI that didn’t get as much attention as Deepseek or others. But it is making a clear claim to be the best AI so far.</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*smozxCaI6g9o3qd9Nkz6_w@2x.jpeg" /></figure><p>Here are my notes after trying to break it for over a month</p><ul><li>It’s got a sharp wit and humor that other models can’t touch, making it stand out in a sea of dull AI. For the first time, this is an AI that comes to you — popping up on a vibrant platform like Twitter (X), tagging users, and dishing out opinions with a bold, active personality. Unlike the passive AIs of the past, Grok 3 doesn’t just sit there — it engages. Humor generation is really hard problem in AI as it has to connect with humans and surprise them.</li><li>It seamlessly blends cutting-edge knowledge with its training, delivering answers that feel fresh and relevant. What’s truly jaw-dropping is its knack for cutting through the noise.</li><li>In a place like India, where Twitter drowns in misinformation and hate, Grok 3 digs out the political truth — offering clear, factual takes instead of treating every source like it’s equal. That’s a game-changer in dealing with hallucinations.</li><li>And it’s not just smart — it’s versatile. Grok 3 can write in Hinglish or Hindi, switching between scripts and transliteration effortlessly. It even tosses in slang and cuss words, nailing the vibe. A truly multilingual model like this? We’ve been chasing that for years.</li><li>Plus, it can draft contracts with stunning precision — practical and powerful.</li><li>On the reasoning side, I have given it NP-hard problems in a cryptic coded format, and it was able to detect it is an NP-hard problem. Further, some mathematical conjecture, which I asked it to solve Goldbach’s Conjecture, without specifying it and it attempted to solve it while other LLMs simply cheated by detecting it. This is much more advanced reasoning capability than any other model I have seen so far!</li></ul><p>With all those 200K GPUs to train Grok 3….. They weren’t kidding around. This thing is a beast, and it’s rewriting what we thought AI could do.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=ba7ab315f8cc" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/why-grok-3-is-best-llm-out-there-ba7ab315f8cc">Why Grok 3 is best LLM out there !!</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[NVIDIA GTC 2025- Key Takeaways]]></title>
            <link>https://medium.com/towards-generative-ai/nvidia-gtc-2025-key-takeaways-8da6abaeee3f?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/8da6abaeee3f</guid>
            <category><![CDATA[nvidia]]></category>
            <category><![CDATA[gtc]]></category>
            <category><![CDATA[genai]]></category>
            <category><![CDATA[ai]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Wed, 19 Mar 2025 16:36:33 GMT</pubDate>
            <atom:updated>2025-03-19T16:36:40.612Z</atom:updated>
            <content:encoded><![CDATA[<h4>What you need to know about Nvidia GTC 2025</h4><p>NVIDIA GPU Technology Conference (GTC) 2025, , brought forth several key updates and breakthroughs, primarily highlighted by CEO Jensen Huang’s keynote on March 18. Here’s a rundown of the most significant announcements and advancements based on the latest available information:</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/800/1*aGqzb7223k3NxLonbocpMA.jpeg" /></figure><h3>Key Updates from GTC 2025</h3><h4>1. **Blackwell Ultra GPUs Launch**</h4><p>- NVIDIA unveiled the Blackwell Ultra GPU series, an enhanced version of its Blackwell architecture, set to release in the second half of 2025. These GPUs promise a substantial leap in performance, offering up to 40 times the inference capabilities of the previous Hopper GPUs. This upgrade targets the growing demands of AI factories and reasoning AI systems, positioning NVIDIA to maintain its lead in high-performance computing.</p><h4>2. **Next-Generation Rubin Architecture**</h4><p>- Huang introduced the Vera Rubin platform, featuring a custom NVIDIA-designed CPU (Vera) and tens of terabytes of memory. The Rubin GPUs are slated for release in the second half of 2026, with an even more powerful Rubin Ultra (combining four GPUs) planned for 2027. This roadmap underscores NVIDIA’s focus on future-proofing AI and cloud computing infrastructures.</p><h4>3. **AI and Robotics Innovations**</h4><h4>**Groot N1**:</h4><p>An open-source AI model for collaborative robots was debuted, developed in partnership with Google DeepMind and Disney. Groot N1 enhances robots’ tactile and learning capabilities, marking a step toward more adaptive and interactive robotics.</p><h4>**Newton Physics Engine**:</h4><p>NVIDIA, alongside Disney Research and Google DeepMind, introduced Newton, a physics engine designed to simulate robotic movements in real-world settings. Disney plans to leverage this for entertainment robots, such as its “Star Wars”-inspired BDX droids.</p><h4>4. **Automotive Partnerships**</h4><p>- NVIDIA announced a collaboration with General Motors (GM) to advance autonomous driving technology. This partnership integrates NVIDIA’s AI infrastructure, including the Halos system, to enhance self-driving car capabilities, reflecting a deepening focus on automotive applications.</p><h4>5. **Data Center and Infrastructure Advances**</h4><ul><li><strong>*Dynamo OS**:</strong></li><li>A new software solution to optimize custom data centers for AI workloads, aimed at reducing costs and boosting efficiency.</li><li><strong>**Blackwell NVLink**:</strong></li><li>A dynamic interconnect technology offering 40x inference performance, designed to scale AI computing across massive data center setups.</li><li><strong>**NVIDIA Photonics and Laser Data Transmission**:</strong></li><li>The Vera Rubin platform incorporates photonics for faster data speeds, addressing the trillion-dollar data center boom Huang emphasized during the keynote.</li></ul><h4>6. **AI Reasoning and Software**</h4><p>- NVIDIA introduced the Llama Nemotron models, optimized for multi-step AI reasoning, further enhancing generative AI and agentic architectures for enterprise applications.</p><h4>7. **Personal AI Supercomputers**</h4><p>- Huang teased “Project Digits,” a personal AI supercomputer based on the Blackwell platform, delivering 1 petaflop (1,000 TOPS) of performance. Priced at a minimum of $3,000, it’s set for release in May 2025, targeting offline AI inference for individual users.</p><h3>Main Breakthroughs</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*hGqqo8MXfnSP8tW8_gpUnQ.jpeg" /></figure><h4>1. **Performance Leap with Blackwell Ultra**</h4><p>- The 40x inference performance increase over Hopper GPUs represents a monumental breakthrough, enabling more complex AI models and real-time reasoning applications. This positions NVIDIA to meet the escalating computational needs of AI-driven industries.</p><h4>2. **Robotics Revolution with Groot N1 and Newton**</h4><p>- The combination of Groot N1’s open-source AI and Newton’s physics simulation capabilities marks a significant advancement in robotics. These tools enable robots to learn, adapt, and operate in dynamic environments, with immediate applications in entertainment and beyond.</p><h4>3. **Photonics Integration for Data Speed**</h4><p>- Incorporating photonics into the Rubin platform is a pioneering move to accelerate data transmission, addressing bottlenecks in AI cloud computing and supporting the massive scale of modern data centers.</p><h4>4. **AI Infrastructure Scalability**</h4><p>- The Dynamo OS and Blackwell NVLink breakthroughs enhance the scalability and efficiency of AI infrastructure, catering to a projected $50 billion market. This reflects NVIDIA’s strategic pivot to dominate the trillion-dollar shift toward AI-driven computing.</p><h4>5. **Consumer AI Accessibility**</h4><p>- Project Digits brings high-end AI capabilities to the desktop, democratizing access to petaflop-level performance. This could reshape how developers and enthusiasts engage with AI locally.</p><h3>Context and Impact</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/802/1*XfU7FZfodqOIkXyueF4eEQ.jpeg" /></figure><p>Huang emphasized that AI is at an “inflection point,” driven by a 100x increase in computational demand over recent years. These updates and breakthroughs reinforce NVIDIA’s dominance in AI hardware and software, while expanding its influence into robotics, automotive, and personal computing. The collaboration with industry giants like GM, Disney, and Google DeepMind further amplifies the real-world applicability of these innovations, setting the stage for transformative changes across multiple sectors.</p><p>These insights are drawn from various reports and posts circulating around the event reflecting the immediate takeaways from GTC 2025.</p><p>Follow <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> for more content related to the latest advancement in AI.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=8da6abaeee3f" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/nvidia-gtc-2025-key-takeaways-8da6abaeee3f">NVIDIA GTC 2025- Key Takeaways</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Understanding TITANS: New Architecture to Redefine Transformers with Long-Term Memory]]></title>
            <link>https://medium.com/towards-generative-ai/understanding-titans-new-architecture-to-redefine-transformers-with-long-term-memory-f8c18a047a7e?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/f8c18a047a7e</guid>
            <category><![CDATA[persistent-memory]]></category>
            <category><![CDATA[llm]]></category>
            <category><![CDATA[genai]]></category>
            <category><![CDATA[transformers]]></category>
            <category><![CDATA[ai]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Tue, 04 Mar 2025 19:33:03 GMT</pubDate>
            <atom:updated>2025-03-05T13:22:53.908Z</atom:updated>
            <content:encoded><![CDATA[<h4>How does the new paper from Google handle the well-known limitations of Transformers, and what are its applications?</h4><p>In the ever-evolving world of artificial intelligence, transformers have emerged as one of the most significant breakthroughs in LLMs. From chatbots to content generation, these models have revolutionized how machines understand and generate human-like text. However, as the demand for more sophisticated applications grows, so do the limitations of traditional transformer architectures. Enter <strong>TITANS</strong> — a novel approach designed by Google to overcome these challenges by introducing a groundbreaking concept: persistent memory with a focus on surprising information.</p><h4>Current Limitations of Transformers</h4><p>Despite their success, transformers are not without flaws. One of the most notable limitations is their lack of persistent memory. While some premium models attempt to address this issue through developer heuristics, metadata tagging, and human feedback, these solutions are far from perfect. Transformers often struggle to retain context over extended interactions, making them less effective for tasks requiring long-term memory or nuanced understanding.</p><p>This limitation becomes especially apparent in applications where context continuity is critical. For example, conducting deep research or assisting users over prolonged conversations requires models to “remember” past interactions effectively. Traditional transformers fall short in such scenarios, paving the way for innovations like TITANS.</p><h4>Introducing TITANS: A New Paradigm in Memory</h4><p>TITANS (Transformers with Integrated and Tunable Attention for Novel Surprises) is a next-generation architecture that addresses the memory limitations of traditional transformers. What sets TITANS apart is its ability to retain and prioritize <strong><em>surprising</em></strong> information, mimicking how humans process and store memories.</p><h3>How TITANS Differs from Transformers?</h3><p>TITANS operates on a simple yet powerful principle:</p><blockquote><strong>Memory = Past Memory + Surprise</strong>.</blockquote><p>The model assigns higher importance to surprising events or information, ensuring they are remembered more vividly than mundane details. This approach mirrors human cognition — think back to a surprising event in your life, and you’ll likely recall not just the event itself but also the surrounding context in vivid detail.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*P2YFssJDq08ZM3Ltlbq6jg.png" /></figure><p>The “surprise” factor is quantified using a gradient-based mechanism:</p><ul><li><strong>Surprise = Most Recent Surprise — b × (Momentary Surprise)</strong><br> Here, “b” represents the gradient of the loss with respect to memory parameters, determining how much surprise can be encoded from subtle changes in memory.</li></ul><p>By integrating this dynamic memory mechanism, TITANS achieves a longer context window and enhanced recall capabilities, making it ideal for tasks requiring persistent memory.</p><p>The Titans architecture incorporates the below steps to differ from Transformers</p><ol><li>Titans incorporate a neural long-term memory module that can store and retrieve historical information dynamically. This allows the model to maintain context over much longer sequences than traditional Transformers.</li><li>Titans include a persistent memory component that contains task-specific knowledge in learnable, static parameters. This is absent in standard Transformers.</li><li>Titans use a brain-inspired memory system with short-term, long-term, and persistent memory, mimicking human cognitive processes. Traditional Transformers rely solely on their attention mechanism for information retention.</li><li>Titans can process sequences of up to 2 million tokens without increasing computational complexity. Traditional Transformers are typically limited to context windows of about 4–128K tokens.</li><li>Titans achieve O(n) linear scaling through a hierarchical memory architecture, compared to the quadratic O(n²) scaling of Transformer’s self-attention mechanism.</li><li>Titans can update their persistent memory during inference, enabling real-time learning without retraining the entire model. Traditional Transformers require full retraining to incorporate new information.</li><li>Titans use a “surprise” metric to prioritize memorization of unexpected or significant events. This allows for more efficient use of memory capacity compared to Transformers’ uniform attention distribution.</li></ol><h3>How TITAN works?</h3><p>Titans introduce three architectural variants: Memory as a Context (MAC), Memory as a Gate (MAG), and Memory as a Layer (MAL). These offer different ways to integrate the memory components, providing flexibility not present in standard Transformers.</p><h4>Memory Types in TITANS</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*eye8LZSR3GnfKMK8BXhY4Q.png" /></figure><p>TITANS incorporates three distinct memory types to cater to different use cases:</p><ol><li><strong>Memory as Context (MAC):</strong><br> A separate database of value-context pairs that the transformer can reference as needed.</li><li><strong>Memory as Gates (MAG):</strong><br> Gated memory with clear cutoff criteria, similar to Long Short-Term Memory (LSTM) networks.</li><li><strong>Memory as Layer (MAL):</strong><br> An additional layer atop the attention layers that models surprises based on abstract connections within the data.</li></ol><p>Each memory type offers unique advantages, allowing TITANS to adapt to diverse requirements while maintaining efficiency.</p><p>These memory modules allow Titans to overcome some of the limitations of traditional Transformers, particularly in handling long-term dependencies and adapting to new information without full retraining.</p><h3>Applications of TITANS</h3><p>Titans excel in tasks requiring extensive context, such as “needle-in-haystack” problems and genomic sequence analysis. They also show improved performance in language modeling, common-sense reasoning, and time series forecasting compared to traditional Transformers.The introduction of TITANS opens up a world of possibilities across various domains:</p><ol><li><strong>Needle-in-a-Haystack (NIAH) Tasks</strong></li></ol><ul><li>Discovering additional indications for pharmaceuticals</li><li>Identifying fraudulent activities in complex datasets</li></ul><p><strong>2. General Persistent Memory</strong></p><ul><li>AI assistants capable of maintaining long-term conversational context</li><li>Deep research transformers for academic or industrial use</li></ul><p>These applications highlight TITANS’ potential to revolutionize industries that rely on extensive data analysis and contextual understanding.</p><h4><strong>Experimental Results</strong></h4><p>Titans outperform Transformers and modern recurrent models across various tasks, including language modeling, commonsense reasoning, genomics, and time series forecasting.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*jv41qPzg8y7LT1g1ltW09A.png" /></figure><p>They scale effectively to context window sizes exceeding 2 million tokens, achieving higher accuracy in recall-intensive tasks like “<strong><em>needle-in-haystack</em></strong>” problems compared to baseline models.</p><h3>Challenges</h3><p>While TITANS represents a significant leap forward, it is not without challenges. The increased memory and parameter requirements result in higher computational costs — ranging from 30% to 100% more than traditional transformers. This cost factor poses a barrier to widespread deployment, particularly for smaller organizations with limited resources.</p><p>Currently, <strong>no large-scale deployments of TITANS exist.</strong> However, its potential has sparked interest across industries, with discussions around integrating TITANS into platforms like DeepSeek gaining momentum.</p><h3>Old wine in new bottle? RNN-&gt;</h3><p>RNN (recurrent neural networks), which were a precursor to transformers, come to my mind when I read the TITANS paper. I worked on it in the first generation of GenAI (read old blog here on generating poems <a href="https://medium.com/towards-generative-ai/generating-poems-for-any-given-image-using-multi-modal-machine-learning-2be35b72f50a">https://medium.com/towards-generative-ai/generating-poems-for-any-given-image-using-multi-modal-machine-learning-2be35b72f50a</a> ) The architecture seems a tweak on RNN architecture.</p><p>RNN was capable of holding memory but did not scale well. It is also a big issue with exploding or vanishing gradient. Getting rid of recurrent unit and focusing only on attention made things really impactful in LLMs. (hence the title of that famous paper “<em>Attention is all you need</em>”). It would be important to see how this scales.</p><h3>Moving Forward</h3><p>As we look ahead, the development and deployment of TITANS could redefine what transformers are capable of achieving. By addressing the limitations of traditional architectures and introducing human-like memory capabilities, TITANS opens new doors for AI applications that were previously unattainable.</p><p>In conclusion, TITANS represents not just an incremental improvement but a paradigm shift in transformer technology. By prioritizing surprising information and extending context windows, it bridges the gap between human cognition and machine learning — paving the way for smarter, more capable AI systems.</p><p>Follow <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> for more content related to the latest advancement in AI.</p><h3>Citations:</h3><ol><li><a href="https://arxiv.org/abs/2501.00663">https://arxiv.org/abs/2501.00663</a> from Google</li><li>All images credit to Google</li></ol><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=f8c18a047a7e" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/understanding-titans-new-architecture-to-redefine-transformers-with-long-term-memory-f8c18a047a7e">Understanding TITANS: New Architecture to Redefine Transformers with Long-Term Memory</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Why FOBO around DeepSeek is unjustified?]]></title>
            <link>https://medium.com/towards-generative-ai/why-fobo-around-deepseek-is-unjustified-8c3261fa50e1?source=rss-bee09cdc3c5b------2</link>
            <guid isPermaLink="false">https://medium.com/p/8c3261fa50e1</guid>
            <category><![CDATA[nvidia]]></category>
            <category><![CDATA[genai]]></category>
            <category><![CDATA[deepseek]]></category>
            <category><![CDATA[llm]]></category>
            <category><![CDATA[openai]]></category>
            <dc:creator><![CDATA[Kunal Sawarkar]]></dc:creator>
            <pubDate>Mon, 03 Feb 2025 17:40:46 GMT</pubDate>
            <atom:updated>2025-02-03T17:40:46.223Z</atom:updated>
            <content:encoded><![CDATA[<h4>DeepSeek AI seems to have caused a crazy level of FOBO ( Fear of Better Options) in the GenAI industry. Here is a short summary of why the hype is exceeding the reality and this may not really be a “ Sputnik moment” after all.</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/275/0*ZiYZKstutN7l6-33" /><figcaption>Sputnik</figcaption></figure><ol><li>Imitation is always cheaper than innovation. Deepseek used llama from <a href="https://www.linkedin.com/company/aiatmeta/">AI at Meta</a> as base architecture and traffic tokens from others ( possibly OpenAI) as data. And then built upon proven architecture. This saves a lot of costs if you don’t have to pay for good quality data, instruct sets, and finding the right architecture or innovating the right metrics.</li></ol><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*pQXf36-U4vCpTHd4" /><figcaption>Deepseek vs OpenAI</figcaption></figure><p>2. Thus it should be seen as a “refreshed model” or a “refined model” of existing models rather than a true pre-trained model. Something more than a distilled version with tweaks to architecture but a lot of things already baked in.</p><p>3. The reasoning capabilities are overblown .. it’s not true reasoning. the right way is to think of it as smaller models outperforming larger ones on the same tasks ( which was a trend in open source for a while)</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*Ie4dUNYY8tiLVHhp.jpg" /><figcaption>Nvidia</figcaption></figure><p>4. What was interesting and spooked markets was their claim ( which needs verification) that they trained the model on the old A100 and not the H100, this means all the hyped-up demand for Nvidia H100 and h200 and tsmc chips is overblown .. and you can train at a much lower footprint. And hence that rout in the stock market I guess. Also, I look at that as a much overdue correction as valuations for some stocks were anyway way off any EBITDA projections for chipmakers.</p><p>5. I for one don’t believe this will dominate the enterprise AI space. The model has China-specific guardrails like Tiananmen Square or other blockers which most countries and companies will find hard to govern and adopt in the long term. Llama will still be a much better option to build a reliable architecture.</p><p>6. What is good in DeepSeek is that it’s not just an open model but open weights .. which means others can quickly distill this model to even smaller footprints ( shifting focus on smaller models that can go to prod rather than large chip investment ). This can be a game changer for enterprises as the cost of taking GenAI to prod is still exorbitant.</p><p>7. Or the hype can cause severe disruption with geopolitical implications and affect the entire open-source ecosystem by government intervention. The hawks like Vinod Khosla always argued against open source models in SoTA AI ( with vested interests of course )</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*yZk39JmrVHOa7GXW" /><figcaption>Credit- Google</figcaption></figure><p>I still don’t think this is a “Sputnik Moment” for a simple reason. Remember who won the space race in the end. The joke in the space race was that Americans spent a million dollars to invent zero-gravity pens but Russians used pencils. This missed the whole point of innovation. It’s about the flywheel effect. US space program resulted in countless innovations from remote control of the TV to de-icing of jet planes to nonstick cookware. The USA still is the favorite to win the race in the long term.</p><p>Follow <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> for more content related to the latest advancement in AI.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=8c3261fa50e1" width="1" height="1" alt=""><hr><p><a href="https://medium.com/towards-generative-ai/why-fobo-around-deepseek-is-unjustified-8c3261fa50e1">Why FOBO around DeepSeek is unjustified?</a> was originally published in <a href="https://medium.com/towards-generative-ai">Towards Generative AI</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
    </channel>
</rss>