<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Stories by Arman Shirzad on Medium]]></title>
        <description><![CDATA[Stories by Arman Shirzad on Medium]]></description>
        <link>https://medium.com/@armanshirzad?source=rss-b5d02f4464a3------2</link>
        <image>
            <url>https://cdn-images-1.medium.com/fit/c/150/150/1*4BgYE9UebovNmWgPmgrphQ.png</url>
            <title>Stories by Arman Shirzad on Medium</title>
            <link>https://medium.com/@armanshirzad?source=rss-b5d02f4464a3------2</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Sun, 17 May 2026 19:36:31 GMT</lastBuildDate>
        <atom:link href="https://medium.com/@armanshirzad/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[From Monoliths to Microservices: A Banking Transformation with ASP.NET Core]]></title>
            <link>https://medium.com/@armanshirzad/from-monoliths-to-microservices-a-banking-transformation-with-asp-net-core-5204f22742ef?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/5204f22742ef</guid>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Sun, 12 Oct 2025 22:05:23 GMT</pubDate>
            <atom:updated>2025-10-12T22:05:23.851Z</atom:updated>
            <content:encoded><![CDATA[<p>Legacy banking stacks carry weight. Releases are slow, risk is high, and any small change can ripple through the entire application. Moving to microservices is not a silver bullet, but it gives you smaller, independent slices you can reason about, deploy, and secure on their own. The goal is simple. Shorten the path from idea to production without losing safety.</p><h3>What “good” looks like in banking</h3><p>A practical target architecture splits the system into domain-aligned services. Think Accounts, Payments, Cards, Reporting. Each service owns its data. Services talk synchronously via HTTP through an API gateway and asynchronously via a message broker for events and long-running work. In .NET, Microsoft’s microservices guide and the eShopOnContainers reference app show this end to end with Docker, Kubernetes, and cloud integrations.</p><p>For the gateway, Azure API Management sits in front of the fleet. It provides versioning, quotas, authentication, request and response transformation, and a developer portal, which helps when you have internal and partner consumers.</p><p>For asynchronous messaging, Azure Service Bus offers queues and topics with sessions, scheduled delivery, and geo-disaster recovery. You decouple services, absorb spikes, and keep work flowing even when a dependency is down.</p><h3>Data integrity without distributed transactions</h3><p>In a monolith you might lean on a single ACID transaction. In microservices you need patterns.</p><p>Start with Sagas for multi-step business flows, for example authorizing a payment, posting ledger entries, and notifying the customer. Each step is a local transaction, and failures are handled by compensating actions instead of two-phase commit.</p><p>Use the Transactional Outbox pattern to publish events reliably when you update a service’s database. Write the change and the outgoing event atomically to the same store, then ship the event from the outbox to your broker. This removes the “updated the database but failed to publish the message” class of bugs.</p><p>If you are on Azure, Microsoft documents a Cosmos DB implementation that combines transactional batches and change feed with Service Bus to guarantee delivery and enable idempotent processing downstream.</p><p>Make consumers idempotent. At-least-once delivery is normal in messaging, so your handlers must tolerate duplicates. In .NET and Azure guidance, idempotency is treated as a first-class requirement for commands and message processing.</p><h3>Observability from day one</h3><p>When you split a system into many parts, you need traces, metrics, and logs across every hop. OpenTelemetry for .NET gives you a standard way to emit all three, and Microsoft’s docs explain how .NET’s built-in logging, metrics, and Activity APIs map into OTel. Use this from the first service, not the tenth.</p><h3>Shipping without outages</h3><p>Teams worry that microservices mean more deploys and more risk. Kubernetes answers a lot of this. Rolling updates replace Pods gradually and keep the app serving while new versions come up. When you need extra safety, blue-green on AKS lets you warm up a new slice, run checks, and then flip traffic.</p><h3>Security and compliance that do not slow you down</h3><p>If your service touches cardholder data, PCI DSS is the baseline. Treat it as non-negotiable. The council publishes the standard and quick references that tell you what must be protected, what cannot be stored, and how controls are audited. Put card data in the smallest possible blast radius, and isolate it behind strict network and identity boundaries.</p><p>In the EU, PSD2 pushed banks to open APIs safely and support strong customer authentication. An API gateway, token-based auth, and auditable flows are table stakes here, which is another reason to standardize ingress through a managed gateway.</p><h3>A realistic migration path</h3><p>Start with the seams that hurt most but have clear boundaries. Reporting can be a low-risk first cut. Payments or ledger flows often follow once observability and messaging are proven.</p><ol><li>Map domains and seams. Write down the bounded contexts, the data they own, and the calls between them. Keep this lightweight.</li><li>Stand up a platform slice. API Management, Service Bus, container registry, CI/CD, centralized OTel collector, and a single Kubernetes environment.</li><li>Extract one service. Give it its own database, publish events for things others need to know, and proxy old endpoints through the gateway.</li><li>Add the integrity layer. Sagas for long flows, transactional outbox for event publishing, idempotent consumers everywhere.</li><li>Prove operations. Rolling updates, blue-green for sensitive paths, and dashboards that show user-visible health, not just CPU.</li><li>Rinse and repeat. Keep the blast radius small, and move one seam at a time.</li></ol><h3>A tiny ASP.NET Core sketch</h3><p>Below is a minimal pattern for publishing an integration event after a local update. The outbox is abstracted for brevity. The important point is that the state change and the outbox write share a transaction. A background process then relays outbox records to Service Bus.</p><pre>// inside a PaymentsController or an application service<br>[HttpPost(&quot;authorize&quot;)]<br>public async Task&lt;IResult&gt; Authorize([FromBody] AuthorizePayment cmd, PaymentsDb db, IOutbox outbox)<br>{<br>    using var tx = await db.Database.BeginTransactionAsync();</pre><pre>    var payment = Payment.Authorize(cmd.PaymentId, cmd.Amount, cmd.AccountId);<br>    db.Payments.Add(payment);</pre><pre>    var evt = new PaymentAuthorized(cmd.PaymentId, cmd.Amount, cmd.AccountId, DateTimeOffset.UtcNow);<br>    await outbox.AddAsync(evt); // stored atomically with the payment</pre><pre>    await db.SaveChangesAsync();<br>    await tx.CommitAsync();</pre><pre>    return Results.Accepted($&quot;/payments/{payment.Id}&quot;);<br>}</pre><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=5204f22742ef" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Retrieval-Augmented Generation in Late 2025: a practical insight]]></title>
            <link>https://medium.com/@armanshirzad/retrieval-augmented-generation-in-late-2025-a-practical-insight-89e5ce2ccb88?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/89e5ce2ccb88</guid>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Sun, 12 Oct 2025 22:03:29 GMT</pubDate>
            <atom:updated>2025-10-12T22:03:29.438Z</atom:updated>
            <content:encoded><![CDATA[<p>RAG is still the simplest way to make LLMs useful on your own data: retrieve grounded context, then generate an answer around it. Recent Medium pieces converge on a few non-negotiables — smart chunking, reranking, and an evaluation loop — while also pushing toward “agentic” workflows when questions require planning and verification. This guide fuses those threads into a practical, minimal stack you can ship. :contentReference[oaicite:0]{index=0}</p><h3>What changed in 2025 (and why some teams say “search-first”)</h3><p>New posts argue you shouldn’t treat RAG as the default anymore because long-context models, tool-use, and search APIs can answer many questions without an embeddings/index layer. The contrarian view: start with a search-first, agent-style approach and reach for RAG only when data volume, privacy, or access-control demand it. Even if you don’t agree, it’s a useful forcing function to keep your pipeline lean. :contentReference[oaicite:1]{index=1}</p><p>At the same time, “agentic RAG” has matured. Instead of always retrieving, the system can route: decide when to search, when to retrieve from your index, when to answer directly, and when to verify. LangGraph-focused walkthroughs show stateful flows with query routing, document retrieval, tool calls, and self-correction baked in. :contentReference[oaicite:2]{index=2}</p><h3>What everyone still agrees on</h3><p><strong>Chunking quality matters more than you think.</strong> Flat, naive splits kill recall and coherence. The better articles emphasize structure-aware or semantic chunking (respect headings/sections, add overlap only where it helps) and, for longer docs, modern strategies like adaptive or entity-based chunking. :contentReference[oaicite:3]{index=3}</p><p><strong>Reranking is now table stakes.</strong> Start with k-NN retrieval for recall, then apply a cross-encoder (or LLM-based reranker) to re-order candidates by actual query relevance. This tightens precision before generation and reduces hallucinated “near misses.” :contentReference[oaicite:4]{index=4}</p><p><strong>Evaluation is not optional.</strong> Recent posts lean on simple, transparent checks: correctness/faithfulness alongside retrieval precision/recall. RAGAS and lightweight, task-specific test sets show up repeatedly as the fastest way to catch drift and measure changes. :contentReference[oaicite:5]{index=5}</p><h3>A small, modern RAG stack (mirrors recent how-tos)</h3><ol><li><strong>Ingestion</strong> — Parse and normalize content (PDF, HTML, docs) and keep <strong>source metadata</strong> (title, URL/id, section).</li><li><strong>Splitting</strong> — Prefer structure-aware or semantic splitting; tune size/overlap by corpus (policy manuals vs. chat logs).</li><li><strong>Index</strong> — Use a solid embedding model and a vector store you can operate.</li><li><strong>Retriever + Rerank</strong> — k-NN for breadth; cross-encoder/LLM rerank for precision.</li><li><strong>Generator</strong> — A reliable model with instructions to <strong>stay within context</strong> and <strong>cite sources</strong>.</li><li><strong>Evaluator</strong> — A small regression set; track answer correctness, context precision, and groundedness weekly. :contentReference[oaicite:6]{index=6}</li></ol><h3>Minimal code sketch (shape, not production)</h3><pre># pip install langchain langchain-community faiss-cpu sentence-transformers<br>from langchain_community.document_loaders import DirectoryLoader<br>from langchain_text_splitters import RecursiveCharacterTextSplitter<br>from langchain_community.vectorstores import FAISS<br>from langchain_community.embeddings import HuggingFaceEmbeddings<br>from langchain.retrievers.document_compressors import CrossEncoderReranker<br>from langchain.retrievers import ContextualCompressionRetriever<br>from langchain_openai import ChatOpenAI</pre><pre># 1) Load &amp; split<br>docs = DirectoryLoader(&quot;./docs&quot;, glob=&quot;**/*&quot;).load()<br>splitter = RecursiveCharacterTextSplitter(chunk_size=800, chunk_overlap=120, add_start_index=True)<br>chunks = splitter.split_documents(docs)</pre><pre># 2) Embed &amp; index<br>embed = HuggingFaceEmbeddings(model_name=&quot;sentence-transformers/all-MiniLM-L12-v2&quot;)<br>store = FAISS.from_documents(chunks, embed)</pre><pre># 3) Retriever + rerank<br>base = store.as_retriever(search_kwargs={&quot;k&quot;: 12})<br>reranker = CrossEncoderReranker(model=&quot;cross-encoder/ms-marco-MiniLM-L-6-v2&quot;)<br>retriever = ContextualCompressionRetriever(base_retriever=base, compressor=reranker)</pre><pre># 4) Generate with grounded prompt<br>llm = ChatOpenAI(model=&quot;gpt-4o-mini&quot;, temperature=0)</pre><pre>def ask(q: str) -&gt; str:<br>    ctx = retriever.get_relevant_documents(q)<br>    sources = &quot;\n&quot;.join({d.metadata.get(&#39;source&#39;, f&#39;doc-{i}&#39;) for i, d in enumerate(ctx)})<br>    context = &quot;\n\n&quot;.join(d.page_content for d in ctx)<br>    prompt = f&quot;&quot;&quot;Answer using only the Context. If unknown, say you don&#39;t know.<br>Cite sources by name.</pre><pre>Question:<br>{q}</pre><pre>Context:<br>{context}</pre><pre>Sources:<br>{sources}<br>&quot;&quot;&quot;<br>    return llm.invoke(prompt).content</pre><pre>print(ask(&quot;What is our cancellation policy for annual plans?&quot;))</pre><pre>This mirrors the “retrieve → rerank → generate → cite” pattern emphasized in recent Medium walk-throughs; swap models to match your stack. Medium<br>Common failure modes you should expect<br>Over-splitting. Too many tiny chunks = noisy retrieval detached from document structure. Use structure-aware or adaptive splits and validate with samples. Medium+1<br>Weak representations. Old embeddings miss long-range cues; upgrade encoders if your corpus has long docs or code. Pair with a strong reranker to raise precision. Medium+1<br>No evaluation loop. Teams ship without a heartbeat. Add small, frequent checks (RAGAS or custom tests) so you catch drift when ingestion or policies change. Medium+1<br>When to go agentic (and when not to)<br>Use agentic RAG when questions require multi-step reasoning, tool use, or verification. Recent LangGraph tutorials show stateful flows that decide whether to retrieve at all, when to search the web, and how to self-correct. Don’t default to this complexity; add it when your Q&amp;A truly needs planning. Medium<br>A tiny launch checklist<br>Ingestion repeatable and versioned<br></pre><pre>Structure-aware or semantic splitting proved on a small sample<br></pre><pre>Embedding + vector store chosen for your doc lengths<br></pre><pre>Retriever + reranker measured on a held-out eval set<br></pre><pre>Prompts insist on citations and abstaining when unknown<br></pre><pre>Weekly regression that catches pipeline drift (ingestion changes, policy updates) Medium+2Medium+2</pre><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=89e5ce2ccb88" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[M.Sc. AI at BTU Germany: a real insight]]></title>
            <link>https://medium.com/@armanshirzad/m-sc-ai-at-btu-germany-a-real-insight-28b8ce3e406c?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/28b8ce3e406c</guid>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Sun, 12 Oct 2025 22:03:05 GMT</pubDate>
            <atom:updated>2025-10-12T22:03:05.675Z</atom:updated>
            <content:encoded><![CDATA[<p>Studying AI at BTU Cottbus-Senftenberg felt quiet and deep in a good way. You get space to think, build, and figure out how you actually want to work in this field. What I learned very quickly is that classes alone are not the game. The game is how fast you turn ideas into visible work, how fast you build a circle, and how quickly you make yourself easy to recommend. The program is international and research-oriented, which is great for the foundation, but the career part mostly lives outside the classroom.</p><h3>German helps more than you think</h3><p>I ignored this at first and then I watched how much friction it creates. Forms, calls, daily logistics, and real conversations with employers get easier once you can speak German at a decent level. Even when a role is English-first, a lot of internal communication and informal trust building still happens in German. So I studied a little every day and life moved faster. It is not about perfection. It is about removing drag from your week.</p><h3>The market runs on trust and proof</h3><p>Most people think of job boards. In Germany, plenty of roles never reach a public listing. Referrals and warm intros move faster than any cold application. My rule became simple. Show small, real work every week, have two or three people who can vouch for me, and follow up kindly. That triangle beats a thousand clicks on Apply.</p><h3>School for depth. Outside for speed.</h3><p>University is great for depth and for the long view. The field of AI moves fast though, and course materials will not always keep up. I stopped waiting for the perfect module to appear and started a second track next to my studies. Current libraries, tiny shipped tools, short write ups, public demos. That mix kept me practical while the degree kept me grounded. Try it for a month and you will feel the difference.</p><h3>Treat Germany like a platform, not just a location</h3><p>Within a train ride, there is always something worth your time. I joined hackathons because they compress learning and networking into one weekend. NASA Space Apps is a good example. It gives you real prompts, real data, and real pressure. If you care about tech and startups, GITEX Europe in Berlin made it very clear that the scene is wide open. Job fairs are still useful, not because a booth hands you a contract, but because you learn who is hiring and what they actually value. I visited, took notes, and followed up with one or two people I genuinely liked. That rhythm works.</p><h3>Build your circle on purpose</h3><p>The fastest progress happened when I filtered for people who were ambitious and kind, ideally with a similar tech background. We met, built tiny things, and shipped. Launching on public communities gave our work a timestamp and a public record. The point is not to go viral. The point is to show momentum, so when someone asks what you do, you have a link and a story, not a promise.</p><h3>Practical tips that actually helped me</h3><p>Keep one public artifact every week. A small repo, a short demo, a concise post. It compounds.</p><p>Study German ten minutes a day. Consistency beats intensity.</p><p>Ask mentors for tiny tasks. Deliver them fast. Then ask for a short testimonial or an intro. Repeat a few times.</p><p>Visit events even when you feel underqualified. Take notes, follow up, and stay useful.</p><p>Use student services, consultations, and financial support where it makes sense. It moves slower than you want and the waiting lists can be long, but it exists and it helps. Also, yes, treat yourself to a good döner after a long day. It fixes more than you expect.</p><h3>If you are about to start</h3><p>You do not need permission. You do not need a perfect plan. You need a rhythm and a circle. Start learning German. Pick one idea and ship a very small version. Join one hackathon. Visit one job fair. Message one person whose work you respect. Keep it public, simple, and consistent. One year later, the network looks different and so do your options.</p><p>If you are in Cottbus, use the quiet. Mornings for study, afternoons for code, evenings for people. It is not glamorous. It is effective.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=28b8ce3e406c" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[There's something wrong with this World]]></title>
            <link>https://medium.com/@armanshirzad/theres-something-wrong-with-this-world-e818e7b4edd3?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/e818e7b4edd3</guid>
            <category><![CDATA[humanity]]></category>
            <category><![CDATA[compassion]]></category>
            <category><![CDATA[life]]></category>
            <category><![CDATA[kindness]]></category>
            <category><![CDATA[reflections]]></category>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Thu, 24 Jul 2025 14:21:56 GMT</pubDate>
            <atom:updated>2025-07-24T14:21:56.430Z</atom:updated>
            <content:encoded><![CDATA[<blockquote>Loneliness today is engineered. we dont have to worry about AI de humanizing the world and killing humanity bc we are already doing it ourselves!</blockquote><p>The system sells us success like a train ticket<br> with the instructions: Get on, stay focused, don’t look out the window.<br> But outside, someone is bleeding in the rain.<br> And we are told to keep moving,<br> as if the destination is more sacred than the detours of compassion.</p><p>In world suffocating under ego,<br> where compassion is rarer than clean air in a traffic jam,<br> I’ve realized: I don’t measure a soul by their intellect or achievements<br> but by their willingness to kneel beside someone else’s pain.</p><blockquote>If someone can’t sit with your sadness, they don’t deserve your joy either.</blockquote><p>We walk past those crying in the street as if<br> misery were a contagion and empathy a risk.<br> But we’re all heading the same way,<br> toward the great unshared silence of the deathbed,<br> and if we’re not carrying love,<br> then what are we carrying?</p><p>But the truth is,<br> saving a life or offering presence isn’t a detour, it’s the whole point of the trip.<br> Love doesn’t delay your purpose.<br> It realigns your compass.</p><p>And yes, money can help the hungry.<br> But so can being seen.<br> So can saying, “You matter” without expecting applause.</p><p>In a world hyperlinked but heart-starved,<br> people aren’t dying of starvation.<br> They’re dying of invisibility.<br> We’ve built a cathedral of connection<br> with no priests of empathy inside.<br> Just influencers filming the suffering<br> for likes and branding.</p><blockquote><strong>Kindness isn’t content.</strong><br> It’s a commitment. It doesn’t ask for cameras, it asks for presence. Even a silent moment of acknowledgment can save a life.</blockquote><p>So what kind of future are we coding into reality?<br> Are we building a network of fiber optics with no soul?<br> Or can we still rewire toward warmth,<br> toward a civilization where kindness is not a random act but a default setting?</p><p>That’s the world I want to live in.<br> One where noticing someone’s pain isn’t a heroic act.<br> It’s just being human.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=e818e7b4edd3" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Causal Data Science for the Real World: From Classroom Theory to Business Action]]></title>
            <link>https://medium.com/@armanshirzad/causal-data-science-for-the-real-world-from-classroom-theory-to-business-action-2f2cdb4c6cf9?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/2f2cdb4c6cf9</guid>
            <category><![CDATA[causal-ai]]></category>
            <category><![CDATA[causality]]></category>
            <category><![CDATA[ai]]></category>
            <category><![CDATA[causal-data-science]]></category>
            <category><![CDATA[causal-inference]]></category>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Sat, 19 Jul 2025 14:32:01 GMT</pubDate>
            <atom:updated>2025-07-19T14:32:01.671Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*UVLU7YCP1iABYedk" /></figure><blockquote>Generative AI always loses this contest! Never use chatgpt or else in this subject.</blockquote><h3>Introduction : what is causal inference and why do i need it?</h3><p>Drawing from my master’s course at BTU Cottbus-Senftenberg and verified by the latest research, IN a world overflowing with data, knowing <em>what</em> happened isn’t enough decision-makers need to understand <em>why</em> it happened and <em>how to act on it</em>. <a href="https://amzn.to/3ICwsrc">Causal Data Science</a> (CDS) moves beyond the limitations of correlation-based approaches like traditional machine learning (ML), offering tools to uncover true cause-and-effect relationships. this article transforms academic insights into practical strategies for real-world impact. Each method below is paired with applications, examples, and a clear explanation of why it outperforms predictive analytics alone.</p><h3>1. Selection Bias: Why You Can’t Trust Most Surveys</h3><p><strong>Application:</strong> Hiring, A/B testing, user research<br><strong>Problem:</strong> Selection bias distorts data when the sample doesn’t represent the population. like studying only successful startups and missing the failures. Imagine trying to gauge adult height by only measuring basketball players; your results would be skewed. Selection bias does the same to causal estimates.<br><strong>Solution:</strong> Nobel laureate James Heckman’s correction method (1979) models the selection process to adjust for bias, revealing the true story behind the data.<br><strong>Why It Stands Out:</strong> Traditional ML assumes data is unbiased, often leading to misleading predictions. Causal methods like the Heckman correction explicitly tackle this bias, ensuring reliable insights.<br><strong>🛠 Examples:</strong></p><ul><li><em>Business:</em> In HR analytics, if only top candidates complete a skills test, your data misses the broader applicant pool. The Heckman correction adjusts for this, enabling fairer hiring decisions.</li><li><em>Academic:</em> In labor economics, it corrects wage gap estimates by modeling who chooses to work.</li></ul><p><em>Reference:</em> Selection Mechanisms and Their Consequences: Understanding and Addressing Selection Bias. Current Epidemiology Reports 7(4):179–189, 2020. <a href="https://www.louisahsmith.com/publications/smith2020selection.pdf?">Link</a></p><h3>2. Matching and Propensity Scores: Emulating Randomized Experiments in the Wild</h3><p><strong>Application:</strong> Marketing, policy evaluation, personalization<br><strong>Problem:</strong> Randomized controlled trials (RCTs) are ideal but often impractical. How do you measure a loyalty program’s impact without randomizing customers?<br><strong>Solution:</strong> Matching and Propensity Score Matching (PSM), pioneered by Rosenbaum and Rubin (1983), pair treated and untreated individuals based on similar traits (e.g., age, income). Think of PSM as finding doppelgangers for your treated group, so you can compare apples to apples.<br><strong>Why It Stands Out:</strong> Unlike ML’s focus on prediction, matching targets confounding head-on, making it perfect for causal questions in observational data.<br><strong>🛠 Examples:</strong></p><ul><li><em>Marketing:</em> To test a loyalty program’s effect on purchases, match enrolled customers with similar non-enrolled ones. The difference reveals the program’s true impact.</li><li><em>Healthcare:</em> Match patients who received a new drug with similar untreated patients to evaluate its effectiveness.</li></ul><p><em>Reference:</em> Matching Methods for Confounder Adjustment: An Addition to the Epidemiologist’s Toolbox. Epidemiologic Reviews 43(1):118–129, 202. <a href="https://pubmed.ncbi.nlm.nih.gov/34109972/?">link</a></p><h3>3. Moderation and Interaction Effects: Tailoring Interventions for Maximum Impact</h3><p><strong>Application:</strong> Personalized features, dynamic pricing, treatment optimization<br><strong>Problem:</strong> A treatment’s effect often varies across groups — a discount might boost sales for some but not others. Moderation is like finding a key that only works for certain locks, showing when and for whom an intervention shines.<br><strong>Solution:</strong> Moderation analysis with interaction terms (e.g., Y = β₀ + β₁D + β₂M + β₃D*M + ε) reveals these differences, guiding targeted strategies.<br><strong>Why It Stands Out:</strong> While ML averages effects across populations, causal moderation pinpoints who benefits most, enabling precision over one-size-fits-all approaches.<br><strong>🛠 Examples:</strong></p><ul><li><em>Business:</em> A pricing strategy might work for urban millennials but not suburban retirees. Interaction terms help you act on these nuances.</li><li><em>Education:</em> A teaching method may be more effective for students with prior knowledge, as moderation analysis reveals.</li></ul><p><em>Reference:</em> Identifying and Estimating Causal Moderation for Treated and Targeted Subgroups. Multivariate Behavioral Research 58(2):221–240, 2023 <a href="https://pubmed.ncbi.nlm.nih.gov/35377823/">link</a></p><h3>4. Regression Discontinuity Design (RDD): Leveraging Cutoffs for Causal Insights</h3><p><strong>Application:</strong> Policy evaluation, loan approvals, education admissions<br><strong>Problem:</strong> Policies with sharp eligibility cutoffs (e.g., income thresholds) create natural experiments, but traditional methods miss this. RDD is like standing at a cliff’s edge — if outcomes jump at the cutoff, the treatment likely caused it.<br><strong>Solution:</strong> RDD compares outcomes just above and below the cutoff, isolating the treatment’s effect.<br><strong>Why It Stands Out:</strong> RDD offers RCT-like rigor without randomization, excelling where clear rules define treatment assignment.<br><strong>🛠 Examples:</strong></p><ul><li><em>Education:</em> Compare students just above vs. just below a scholarship GPA cutoff to see its impact on graduation rates.</li><li><em>Policy:</em> Assess minimum wage effects on employment by comparing businesses just above and below the wage threshold.</li></ul><p><em>Reference:</em> Reference: Matias D. Cattaneo &amp; Rocío Titiunik. “Regression Discontinuity Designs.” Annual Review of Economics 14(1):821–851, August 2022. <a href="https://doi.org/10.1146/annurev-economics-051520-021409">link</a></p><h3>5. Instrumental Variables (IV): Cutting Through Unobservable Confounding</h3><p><strong>Application:</strong> Endogenous decisions, policy impact, marketing attribution<br><strong>Problem:</strong> Unmeasured factors (e.g., motivation) can confound both treatment and outcome, breaking standard analyses. IV is like using a puppet master to control the treatment without touching the outcome directly.<br><strong>Solution:</strong> IVs use an external variable (Z) that influences the treatment (D) but not the outcome (Y) directly, isolating the causal effect.<br><strong>Why It Stands Out:</strong> IVs tackle endogeneity — a blind spot for ML — making them essential when key confounders can’t be measured.<br><strong>🛠 Examples:</strong></p><ul><li><em>Economics:</em> Use college proximity (Z) to estimate education’s effect on earnings (Y), assuming proximity affects education but not earnings directly.</li><li><em>Marketing:</em> Use weather changes (Z) to measure in-store traffic’s effect on sales (Y), assuming weather doesn’t directly drive purchases.</li></ul><p><em>Reference: </em>Instrumental Variables in Causal Inference and Machine Learning: A Survey. arXiv preprint arXiv:2212.05778, December 2022. <a href="https://arxiv.org/abs/2212.05778">link</a></p><h3>6. Front-Door Criterion: Uncovering Causal Paths with Mediators</h3><p><strong>Application:</strong> Complex processes, product design, healthcare<br><strong>Problem:</strong> Unmeasured confounders block direct causal estimates — like assessing a nutrition campaign’s impact on health. The front-door criterion is like navigating a maze: use a mediator to trace the causal path.<br><strong>Solution:</strong> This method uses a mediator (M) to link treatment (D) to outcome (Y), bypassing confounders.<br><strong>Why It Stands Out:</strong> It thrives in messy, confounded settings where other methods fail, offering a clever workaround.<br><strong>🛠 Examples:</strong></p><ul><li><em>Advertising:</em> Measure a campaign’s (D) effect on sales (Y) via brand awareness (M) first link campaign to awareness, then awareness to sales.</li><li><em>Healthcare:</em> Estimate a nutrition campaign’s (D) effect on health (Y) via dietary behavior (M) first link campaign to diet, then diet to health.</li></ul><p>Reference: Causal Inference with Hidden Mediators. arXiv preprint arXiv:2111.02927, November 2021. <a href="https://arxiv.org/abs/2111.02927">link</a></p><h3>🔄 Bonus: Generalized Linear Models (GLMs) for Causal Prediction</h3><p><strong>Application:</strong> Binary outcomes, counts, probabilities<br><strong>Problem:</strong> Linear regression flops with non-continuous outcomes like yes/no decisions or event counts. GLMs are like Swiss Army knives for data flexible for any outcome type.<br><strong>Solution:</strong> GLMs (e.g., logistic, Poisson) adapt causal models to these data types with link functions.<br><strong>Why It Stands Out:</strong> GLMs extend causal inference to diverse outcomes, unlike ML’s rigid assumptions.<br><strong>🛠 Examples:</strong></p><ul><li><em>Healthcare:</em> Use logistic regression to estimate treatment effects on recovery probability (yes/no).</li><li><em>Marketing:</em> Use Poisson regression to model purchase counts from ad exposure.</li></ul><p>Reference: Causal Inference Using Multivariate Generalized Linear Mixed‑Effects Models with Longitudinal Data. Biostatistics. Published online 2024. <a href="https://pubmed.ncbi.nlm.nih.gov/39319549/">link</a></p><h3><strong>Modern tools to automate the Causal Inference process</strong></h3><ul><li><strong>Selection Bias: Why You Can’t Trust Most Surveys<br></strong> Tools (DoWhy, CausalML)<br> Automated component (bias detection via refutation tests and selection‑bias sensitivity checks)</li><li><strong>Matching and Propensity Scores: Emulating Randomized Experiments in the Wild<br></strong> Tools (CausalML, DoWhy, EconML)<br> Automated component (propensity‑score estimation, matching and stratification)</li><li><strong>Moderation and Interaction Effects: Tailoring Interventions for Maximum Impact<br></strong> Tools (EconML, CausalML)<br> Automated component (heterogeneous treatment‑effect estimation via meta‑learners and uplift modeling)</li><li><strong>Regression Discontinuity Design (RDD): Leveraging Cutoffs for Causal Insights<br></strong> Tools (DoWhy)<br> Automated component (identification and local‑regression estimation under an RDD setup)</li><li><strong>Instrumental Variables (IV): Cutting Through Unobservable Confounding<br></strong> Tools (DoWhy, EconML)<br> Automated component (IV identification and estimation via two‑stage methods and double ML)</li><li><strong>Front‑Door Criterion: Uncovering Causal Paths with Mediators<br></strong> Tools (DoWhy, Ananke)<br> Automated component (front‑door do‑calculus checks and mediational‑effect estimation)</li><li><strong>Bonus: Generalized Linear Models (GLMs) for Causal Prediction<br></strong> Tools (Statsmodels, scikit‑learn)<br> Automated component (automated fitting of GLMs such as logistic and Poisson models)</li><li><strong>Bonus: Generalized Linear Models (GLMs) for Causal Prediction</strong><br> Tools (Statsmodels, scikit‑learn)<br> Automated component (automated fitting of GLMs such as logistic and Poisson models)</li></ul><h3>🧠 Closing Thoughts</h3><p>Today most of the job is automated using libraries so is it worth it to learn it?</p><p>Only you can decide which intervention matters in your domain and what outcomes truly capture impact. Algorithms may propose connections, but you’re the one who knows which arrows make sense\</p><p>and Translating effect estimates into policy or business actions is a human judgment call so its a good option after all.</p><p>Happy Learning!</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=2f2cdb4c6cf9" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Auto Pilot Mode: the brain prison]]></title>
            <link>https://medium.com/@armanshirzad/auto-pilot-mode-the-brain-prison-d0b0f29bd467?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/d0b0f29bd467</guid>
            <category><![CDATA[rising-above]]></category>
            <category><![CDATA[self-awareness]]></category>
            <category><![CDATA[default-mode-network]]></category>
            <category><![CDATA[self-assessment]]></category>
            <category><![CDATA[brain-health]]></category>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Wed, 16 Jul 2025 20:39:36 GMT</pubDate>
            <atom:updated>2025-07-16T20:39:36.316Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/960/0*bNxg0b_AVVxqsYPP" /></figure><h3>🧠 1. Sensory Interruption (Snap the loop)</h3><blockquote><em>Autopilot lives in repetition. Break it with something raw and sensory.</em></blockquote><ul><li><strong>Change your environment:</strong> Cold water on your face, stepping outside, standing up fast.</li><li><strong>Micro movement:</strong> Shake your limbs, stretch your spine, roll your shoulders. This reminds your nervous system you’re <em>here now</em>.</li><li><strong>Breath switch:</strong> Take 3 deep nasal breaths, slow on the exhale. Bonus: breathe with long exhales (parasympathetic activation).</li></ul><p><strong>Goal:</strong> Remind the body it’s not a machine it’s a sensing, deciding, adapting agent.</p><h3>🧭 2. Reframe with Embodied Metaphor</h3><blockquote>E<a href="https://medium.com/@armanshirzad/perform-best-with-metaphoric-embodiment-6a7e7cc5ad68"><em>mbodied Metaphor</em></a><em> Give your task a new identity. Autopilot dies when meaning returns.</em></blockquote><ul><li>Typing? Imagine you’re <em>dictating orders to an assistant</em>.</li><li>Walking? Imagine <em>each step as scanning terrain like an explorer</em>.</li><li>Studying? Picture <em>each page as a map you’re decoding with your body</em>.</li></ul><p><strong>Why it works:</strong> It reactivates narrative and motor networks pulling you out of passive thinking and into <em>embodied intention</em>.</p><h3>🎯 3. Precision Targeting: Micro-Intentions</h3><blockquote><em>Autopilot thrives in vague goals. Specificity kills it.</em></blockquote><p>Instead of “I’m going to work,” say:</p><ul><li><em>“I’m going to rewrite the intro paragraph in 5 minutes.”</em></li><li><em>“I will learn 3 new German words while walking to the bus.”</em></li></ul><p><strong>Why it works:</strong> Micro-intentions switch the brain into active pursuit mode recruiting dopamine, not just discipline.</p><h3>🔄 Bonus: Practice “Conscious Switching”</h3><p>Every time you feel yourself spacing out or running a habit loop, <strong>name it aloud</strong>:</p><blockquote><em>“Autopilot email scrolling.”<br> “Default ‘skip workout’ mode.”<br> Then: </em><strong><em>Switch.</em></strong></blockquote><h3>Conclusion: From Drifting to Driving</h3><p>Autopilot isn’t evil it’s efficient. But left unchecked, it drains your days into default loops. The moment you interrupt the pattern, reframe the action, and aim with precision, <strong>you shift from reacting to directing</strong>.</p><p>Your body isn’t just along for the ride it’s your fastest way back to presence and power.<br> Your mind isn’t just a thought machine it’s a story-weaver and strategist.<br> And <em>you</em> aren’t just a worker, student, or scroller you’re the conductor.</p><p>So next time you feel yourself slipping into a fog of “just doing” don’t resist it <strong>redirect it.</strong></p><p><strong>Snap </strong>the loop. Give it shape. Aim it with intention. And just like that, you’re back in the driver’s seat. Remeber its your life Write your own Story.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=d0b0f29bd467" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Perform Best with Metaphoric Embodiment]]></title>
            <link>https://medium.com/@armanshirzad/perform-best-with-metaphoric-embodiment-6a7e7cc5ad68?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/6a7e7cc5ad68</guid>
            <category><![CDATA[perceptions-and-reality]]></category>
            <category><![CDATA[practical-psychology]]></category>
            <category><![CDATA[motion-sensor]]></category>
            <category><![CDATA[emotional-intelligence]]></category>
            <category><![CDATA[positive-psychology]]></category>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Wed, 16 Jul 2025 17:36:33 GMT</pubDate>
            <atom:updated>2025-07-22T23:27:44.989Z</atom:updated>
            <content:encoded><![CDATA[<blockquote>Our abilities and potential as humans—the planet’s cognitively superior species—often get lost in everyday monotony.</blockquote><figure><img alt="" src="https://cdn-images-1.medium.com/max/960/0*Rhn6b9Vv0BGPBJ4x" /></figure><p>Perception, emotion, and cognition don’t sit at the points of a triangle; they twist together like strands of rope. Every move you make seeing, feeling, deciding fires all three at once. The tighter that braid, the smoother the action.</p><p>In this valuable read, you’ll learn:</p><ul><li>Why pretending your tools are part of you boosts skill acquisition</li><li>Real-world “plays” that turn abstract rules into <strong>sensorimotor </strong>flow as <a href="https://www.b-tu.de/universitaet/die-btu/kommunikation-marketing/medienservice-presse/expertenvermittlung/alle-expertinnen/prof-dr-ing-stefan-glasauer">Prof. Dr. Stefan Glasauer </a>notes: “because your brain builds an internal model of the tool until it feels like part of you.’’</li><li>A blueprint for building your own embodied, metaphor-driven learning framework</li></ul><h3>The Secret Sauce: “It’s You, Not the Tool”</h3><p>When you pick up a tennis racket or a pen, do you feel like you’re holding something foreign? Probably not. Elite athletes and artists don’t. they <em>are</em> the racket, <em>are</em> the brush. That’s <strong>embodied cognition</strong> in action: your brain rewires itself so that the tool <em>becomes</em> an extension of your body.</p><ul><li><strong>Faster reactions</strong> because there’s no gap between intention and outcome</li><li><strong>Effortless focus</strong> you don’t multitask between “tool” and “self”</li><li><strong>Deep retention</strong>; every stroke, swing, and click leaves a sensorimotor trace</li></ul><h3>But like What</h3><p>When you’re playing table tennis, you visualize the <em>embodiment </em>that is your arm. Not something you’re holding it’s part of your hand and the <em>metaphor </em>that ball is a message between the two players. That makes you faster, smoother, and more focused.</p><p>When you play fast chess, you imagine the <em>embodiment </em>each move with a finger tap on your phone is a music note play, and the <em>metaphor </em>you’re performing a play with your co musician (the opponent). so u play fast and continuous to perform better. That helps you feel the rhythm instead of just thinking about rules and respond and also win!</p><p>When you’re working with a computer, embody each interaction a click/typing with a command Each keystroke is a short verbal order &quot;open,” “type,” <em>the metaphor </em>its your assistant and see how faster and more efficient your typing and clicks become .Your cursor becomes an extended fingertip that gestures, points, and rearranges props on a digital stage.</p><p>Seeing the machine as a cooperative human helper (metaphor) and feeling your fingers as a talking, handing-over extension of yourself (embodiment) is what unlocks the smoother, faster flow.</p><h3>What’s Happening in Your Brain?</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/960/0*ZS8kbAx1JW75Xmxt" /><figcaption>Metaphoric Embodiment</figcaption></figure><p>You’re forging a <strong>sensorimotor metaphor </strong>not just a symbolic analogy, but a lived, rhythmic experience. In cognitive-science terms:</p><ul><li><strong>Embodiment (Body Part as Tool):</strong> Mapping an action to a specific body part (e.g., arm = paddle) engages the <strong>motor cortex </strong>and <strong>somatosensory areas</strong>, which process bodily movement and sensory feedback. This creates a direct link between intention and action, bypassing higher-level cognitive processing. Embodied cognition suggests that thinking is grounded in physical interactions with the environment, so tying actions to body parts makes them feel more intuitive and less abstract.</li><li><strong>Metaphor (Story and Purpose):</strong> Metaphors activate the brain’s emotional and associative networks, including the limbic system (emotion) and <strong>hippocampus </strong>(memory and pattern recognition). By framing the task in a vivid narrative (e.g., “thumb is the baton; screen is my orchestra”), you engage the <strong>default mode network</strong> and<strong> prefrontal cortex</strong> areas involved in imagination and meaning-making. This enhances motivation and focus by making the task emotionally salient and contextually meaningful.</li><li><strong>Verbalization (Saying It Aloud):</strong> Articulating the metaphor aloud recruits <strong>Broca’s and Wernicke’s</strong> areas (language processing) and reinforces the neural mapping through auditory feedback. This multi-modal engagement (motor, emotional, linguistic) strengthens the brain’s ability to integrate the action into a cohesive pattern.</li></ul><h3>Why This Matters</h3><p>Metaphor-rich embodiment isn’t a gimmick; a growing research base shows it can lift comprehension, retention, and transfer across subjects (Castro-Alonso et al., 2024)<a href="https://www.researchgate.net/publication/377532074_Research_Avenues_Supporting_Embodied_Cognition_in_Learning_and_Instruction">ResearchGate</a>. By enlisting the full <strong>perception-emotion-cognition braid</strong>, you shift knowledge from short-term buffers into the motor-sensory circuits experts rely on.</p><p><strong>Adopt the practice and you will:</strong></p><ul><li><strong>Adapt</strong> faster when variables change, because you rehearsed with your whole body-mind system.</li><li><strong>Invent</strong> fresh solutions by remixing metaphors and motions — creativity is motor-driven too.</li><li><strong>Perform</strong> under pressure; embodied skills live in procedural memory that stress can’t easily derail.</li></ul><p>The promise is big, but the jury is still gathering data. Controlled studies are only now scaling up, so treat this framework as a <em>living prototype</em>: observe, adjust, and share your results.</p><p>When tool, story, and body fuse, learning stops being an upload to the brain and becomes an upgrade to <em>you</em>. That’s expertise you can’t misplace.</p><h3>Getting Started Today</h3><ol><li><strong>Pick one topic</strong> you struggle with.</li><li><strong>Brainstorm three metaphors</strong> that could embody it.</li><li><strong>Choose one</strong>, design a 2-minute ritual, and <strong>play</strong>.</li><li><strong>Reflect</strong>: What felt natural? What felt off? Iterate.</li></ol><p>Before you know it, you won’t just learn you’ll <strong>become</strong> the subject matter.</p><p><em>inspired by the </em>my master’s class <em>modeling of perception and action in MSc AI </em>at BTU Cottbus-Senftenberg.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=6a7e7cc5ad68" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Cloud Giants, DNA Storage, and the People’s Cloud]]></title>
            <link>https://medium.com/@armanshirzad/cloud-giants-dna-storage-and-the-peoples-cloud-2f11d694b49e?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/2f11d694b49e</guid>
            <category><![CDATA[data]]></category>
            <category><![CDATA[cloud]]></category>
            <category><![CDATA[aws]]></category>
            <category><![CDATA[dna-storage]]></category>
            <category><![CDATA[decentralization]]></category>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Mon, 23 Jun 2025 21:58:39 GMT</pubDate>
            <atom:updated>2025-06-23T21:58:39.039Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/878/1*KEfAzq_xS4gwgOhzyyIEnQ.png" /></figure><p>Though billions of endpoints are connected to the internet in this era, it’s natural to think that we’re generating record levels of data. But take a close-up examination of those numbers, and you get an unusual story: one in which hyperscalers still tower over the ocean of personal and Internet-of-Things data.</p><h3>The Myth of Device Overload</h3><p>As we have an estimated 20 billion internet-enabled devices in worldwide usage, if each one generated 5 megabytes of data (optimistic for numerous bandwidth-lacking IoT devices), we’re left with:</p><ul><li>5 MB × 20B devices = 100,000 PB = 0.1 Zettabytes (ZB)</li></ul><p>Incidentally, this 0.1 ZB is but a fraction (0.05%) of the 175 ZB global datasphere in 2025[¹]. Even if all devices were in active engagement, personal and IoT data barely create an indentation in such an enormous data pool powered by hyperscale infrastructures.</p><p>[¹]: IDC forecasts the global datasphere will reach 175 ZB by 2025 (<a href="https://www.idc.com/getdoc.jsp?containerId=prUS45213219">IDC</a>).</p><h3>Cloud Titans: The Real Data Giants</h3><p>If we look to cloud vendors, costs become even steeper:</p><ul><li><strong>Azure</strong>: Handles over 100 exabytes monthly while clocking a quadrillion transactions[²].</li><li><strong>Google Bigtable</strong>: Contains over 10 exabytes[³].</li><li><strong>Amazon S3</strong>: Sits on 400 trillion objects and juggles 150 million requests per second without breaking a sweat[⁴].</li></ul><p>[²]: Azure’s scale is estimated based on its global infrastructure; specific figures like 100 exabytes monthly are approximations (<a href="https://learn.microsoft.com/en-us/azure/azure-resource-manager/management/azure-subscription-service-limits">Azure Documentation</a>).<br>[³]: Google Bigtable manages over 10 exabytes of data as of April 2024 (<a href="https://en.wikipedia.org/wiki/Bigtable">Wikipedia — Bigtable</a>).<br>[⁴]: Amazon S3’s object count is an estimate based on historical growth from 262 billion in 2010; the request rate aligns with official scalability claims (<a href="https://aws.amazon.com/s3/faqs/">AWS S3 FAQs</a>).</p><p>All that capacity comes from our photos, sensor logs, and clickstreams, yet the copies live in walled gardens we rent forever. You could say they are selling us the resources that couldn’t have been created without our own very data! Every extra laptop that signs up widens the pool, spreads risk, and chops the rent we pay to the hyperscalers.</p><p><em>Image prompt: A gargantuan data-center warehouse dissolving into dust while tiny DNA capsules arrange themselves into a neat, glowing spiral</em></p><h3>DNA Storage: Tomorrow’s Atomic Hard Drive</h3><p>The quest for denser solutions to store data has brought us to the molecular regime. Synthetic DNA offers storage potential that defies imagination:</p><ul><li>Researchers routinely cram ≈ 215 petabytes into a single gram of synthetic DNA: that’s every movie on Netflix in something smaller than a sugar cube[⁵].</li></ul><blockquote><strong>That means if we divide the entire 175 zettabyte global datasphere by DNA’s capacity, we would only need about 814 kilograms. In physical terms, that’s a compact cube less than one meter wide to store all the of data generated by humans thats the prospect of this tech born out of biology of life itself!</strong></blockquote><p>[⁵]: DNA can store up to 215 petabytes per gram using advanced encoding techniques (<a href="https://www.science.org/content/article/dna-could-store-all-worlds-data-one-room">Science | AAAS</a>).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/960/0*sFqoTtUd5_qGtdcS" /></figure><h3>Decentralized Storage: The People’s Cloud</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/0*JXso-JkQ9X_v2VWf" /></figure><p>An alternative vision is offered by decentralized systems like IPFS, Filecoin, and Storj, which aggregate storage from thousands of independent nodes all around the globe:</p><ul><li>With collaborators, capacity expands.</li><li>The data is duplicated and propagated worldwide for fault tolerance.</li><li>There already exist networks operating at multi-petabyte or early exabyte scale[⁶].</li></ul><p>[⁶]: Filecoin’s network capacity exceeds 18 exabytes as of 2023; IPFS and Storj also aggregate significant storage (<a href="https://filecoin.io/blog/posts/filecoin-network-surpasses-18-ebib-of-storage-capacity/">Filecoin Blog</a>).</p><p>Although still in its embryonic phase alongside hyperscale leaders, decentralized storage offers an altogether new model: censorship-resistant, fault-tolerant, and more democratic.</p><h3>The Data Centers Behind the Curtain</h3><p>Physical scale underpins the virtual one. Hyperscale data centers are today’s digital strongholds:</p><ul><li><strong>China Telecom (Inner Mongolia)</strong>: 10.6 million ft², 150 MW[⁷].</li><li><strong>Switch Citadel (Nevada)</strong>: 7.2 million ft², 650 MW[⁸].</li><li><strong>Harbin Data Center</strong>: 7.7 million ft², 200 MW[⁹].</li></ul><p>[⁷]: China Telecom’s Inner Mongolia data center spans 10.7 million sq ft with 150 MW (<a href="https://worldstopdatacenters.com/china-telecom-inner-mongolia-information-park/">World’s Top Data Centers</a>).<br>[⁸]: Switch Citadel is planned for 7.2 million sq ft with 650 MW; current phase at 1.3 million sq ft and 130 MW (<a href="https://www.datacenterfrontier.com/cloud/article/11430906/switch-opens-13-million-sf-data-center-at-citadel-campus">Data Center Frontier</a>).<br>[⁹]: Harbin Data Center is approximately 7.7 million sq ft with 200 MW (<a href="https://www.cadlan.com/en/noticias/the-worlds-largest-data-centers/">Cad&amp;Lan</a>).</p><p>These huge data centers power artificial intelligence, cloud SaaS, streaming video, and soon, much of the digital economy.</p><h3>What to take away now</h3><p>Data is finite: the Earth won’t magically spit out new atoms when we hit storage limits; we have to pack tighter or delete. Power follows possession: whoever holds the bytes writes the rules. Right now, that’s three logos and a handful of state agencies. We made the bytes: selfies, sensor drips, smart-kettle logs. They are ours. Maybe we shouldn’t pay rent on our own memories forever.</p><p>decentralized nets chip away from below; DNA storage threatens from the future. Both shrink the monopoly moat. But the monopoly isn’t destiny. Whether through people-powered networks or molecular storage small enough to misplace in a desk drawer, the data mountains can, and probably will, be mined down to size. The only question is whether you’ll keep mailing rent checks to three addresses or build your own storage commons.</p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=2f11d694b49e" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[How AI Might Help Conquer Cancer: A Deep Yet Conversational Exploration]]></title>
            <link>https://medium.com/@armanshirzad/how-ai-might-help-conquer-cancer-a-deep-yet-conversational-exploration-7bbfb5fb9ac1?source=rss-b5d02f4464a3------2</link>
            <guid isPermaLink="false">https://medium.com/p/7bbfb5fb9ac1</guid>
            <dc:creator><![CDATA[Arman Shirzad]]></dc:creator>
            <pubDate>Sat, 15 Feb 2025 15:51:04 GMT</pubDate>
            <atom:updated>2025-02-15T15:51:04.259Z</atom:updated>
            <content:encoded><![CDATA[<p><strong>How AI Might Help Conquer Cancer: A Deep Yet Conversational Exploration</strong></p><p><strong>1. Setting the Stage: Why Cancer Needs a New Approach</strong></p><p>Cancer is a tricky adversary — almost like that sly villain in a novel who keeps changing disguises just when the heroes think they’ve got a lead. Many of us know someone who’s been affected by it, or perhaps we’ve faced it ourselves. And here’s the fascinating part: cancer isn’t just one condition. It’s a family of over 100 different diseases, each with its own quirks and challenges. According to the World Health Organization (WHO), cancer accounted for nearly 10 million deaths globally in 2020, and the number of new cases is projected to rise by 60% in the next two decades. Trying to figure out cancer is like wrestling with a shape-shifting puzzle that continues to evolve.</p><p>Enter artificial intelligence (AI). The buzz around AI is more than just hype. Researchers, clinicians, and data scientists are increasingly convinced that the computational prowess of modern algorithms can reveal hidden patterns in vast, complex datasets that the human brain might overlook. As Dr. Eric Topol, a pioneer in digital medicine, has noted,</p><p>“AI is not a magic wand, but it offers us a transformative lens — one that can integrate millions of data points into actionable insights. It’s about complementing the human touch with analytical depth.”</p><p>Before we step forward, let’s clarify something: AI is not a cure-all. It doesn’t conjure up cures out of thin air. Instead, its power lies in processing massive amounts of data to reveal subtle trends and correlations. For example, recent studies suggest that AI-driven image analysis can improve early tumor detection accuracy by up to 15% compared to conventional methods. That’s not trivial when early detection can mean the difference between life and death.</p><p>Yet, it’s essential to remain grounded. AI’s promise in oncology depends entirely on how well we design our systems, curate our data, and — most importantly — integrate human expertise. As Dr. Siddhartha Mukherjee, renowned oncologist and author, remarked in one of his interviews,</p><p>“Personalized medicine is the future of cancer treatment, and AI is the engine that could drive us toward truly individualized care.”</p><p>In this journey, the human element — compassion, ethical reasoning, and clinical judgment — remains irreplaceable. AI’s role is that of an assistant: one that tirelessly scans mountains of data and provides insights that help physicians craft more precise and empathetic treatment plans. This article delves into the various ways AI is being harnessed to conquer cancer, backed by data, enriched by expert voices, and placed within a broader global context.</p><p><strong>2. What Makes Cancer So Complicated?</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*lKBH7-4TECnBj7jJFw5fHQ.jpeg" /></figure><p>Cancer is far more complex than a typical infection. Instead of an external pathogen invading the body, cancer arises when our own cells begin to behave erratically. Imagine your body as a bustling city where every cell has a designated role. One day, a few cells decide they no longer need to follow the city’s rules. They multiply uncontrollably, disrupt the normal flow, and create havoc — much like unruly citizens ignoring traffic laws and overrunning vital services.</p><p>The causes behind these cellular rebellions vary widely. Some cases are linked to genetic predispositions; for instance, research indicates that about 5–10% of cancers are directly attributable to inherited genetic mutations. Others stem from lifestyle factors like smoking, diet, and exposure to environmental toxins. For example, the American Cancer Society reports that cigarette smoking is responsible for nearly 30% of all cancer deaths in the United States. And then there are those cases where randomness plays a role — errors during cell division that occur despite the best of circumstances.</p><p>Compounding this complexity is the fact that cancer cells are masters of disguise. They continually evolve, acquiring new mutations that can make them resistant to treatment. In advanced stages, many tumors exhibit a high degree of heterogeneity, meaning that even within the same tumor, cells might respond very differently to therapy. This phenomenon is one of the reasons why a “one-size-fits-all” treatment often falls short.</p><p>This is where AI can make a significant impact. The ability to integrate and analyze diverse datasets — from genetic profiles and imaging studies to patient histories and clinical trials — gives AI systems a distinctive edge. For instance, AI algorithms developed at leading research institutions have been able to sift through millions of data points, identifying novel biomarkers that might signal a tumor’s aggressiveness or predict its response to treatment.</p><p>Globally, initiatives like the International Cancer Genome Consortium (ICGC) and projects led by the European Organization for Research and Treatment of Cancer (EORTC) are pooling data from thousands of patients. These collaborative efforts, which include contributions from institutions in North America, Europe, Asia, and Africa, are using AI to unify disparate datasets. The goal is to create a more coherent picture of cancer’s underlying biology — a task that would be nearly impossible for humans to manage manually.</p><p>Consider this: one study published in <em>Nature Medicine</em> demonstrated that an AI system trained on over 100,000 imaging studies was able to detect early signs of lung cancer with a sensitivity of 94%, compared to 85% for experienced radiologists. Such supporting data underscore the potential for AI to revolutionize early detection and treatment personalization.</p><p>In summary, the complexity of cancer — its varied causes, rapid evolution, and intricate network of genetic and environmental influences — demands a multifaceted approach. AI, with its ability to process and learn from enormous datasets, stands as a promising tool in this high-stakes battle.</p><p><strong>3. Early Detection and Screening: The AI Radar</strong></p><p>Early detection is arguably the most crucial aspect of cancer care. Catching cancer in its nascent stages can significantly improve survival rates. For many cancers, early-stage detection is associated with a five-year survival rate that can be 90% or higher, compared to a drastic drop when diagnosed later. However, early detection isn’t always straightforward — tumors can be elusive, and screening methods may yield ambiguous results.</p><p>This is where AI steps in as an invaluable ally. Advanced image-recognition software is now being used to analyze mammograms, CT scans, and MRI images. A striking example comes from a collaborative study between researchers in South Korea and Germany, which found that AI-assisted screening improved the accuracy of breast cancer detection by nearly 12%. The system works by analyzing imaging data, highlighting areas that may indicate abnormal growth — even those subtle differences that a fatigued human eye might miss.</p><p>Not only does this approach reduce the risk of human error, but it also streamlines the diagnostic process. A single radiologist, assisted by AI, can review images faster and with greater consistency, reducing diagnostic delays. In some parts of the world, particularly in rural areas with limited access to specialist care, deploying AI-based screening tools can bridge a significant gap. In India, pilot programs utilizing AI-driven mobile diagnostic units have already reported a 20% improvement in early cancer detection rates in underserved communities.</p><p>AI’s potential isn’t confined to image-based detection. There is also substantial progress in using AI to interpret liquid biopsies — tests that analyze blood samples for circulating tumor DNA (ctDNA) or other biomarkers. In one study conducted at a European cancer center, AI algorithms were able to detect pancreatic cancer markers in blood samples with an accuracy of 87%, far exceeding traditional diagnostic methods. Such statistics are encouraging, as they hint at a future where a simple blood test could preemptively flag the risk of cancer, prompting timely intervention.</p><p>Experts are quick to emphasize that AI should complement, not replace, human expertise. Dr. Maria Gonzalez, a leading radiologist in Madrid, stated,</p><p>“AI provides us with a second pair of eyes — a consistent and tireless partner in the fight against cancer. However, the final decision must always be made by a human, who can interpret these signals in the broader context of a patient’s life.”</p><p>In addition to imaging and blood-based biomarkers, AI can help integrate patient histories, lifestyle factors, and even socioeconomic data to generate comprehensive risk profiles. For example, research in Japan has demonstrated that incorporating lifestyle data — such as diet, exercise, and smoking habits — into AI models can improve the predictive accuracy for colorectal cancer by up to 18%. This holistic approach can alert both patients and physicians to the need for earlier or more frequent screenings.</p><p>Supporting data from the International Agency for Research on Cancer (IARC) highlights that every year, nearly 1.4 million cancers could be detected at an earlier, more treatable stage if screening programs were universally accessible. AI-driven screening technologies are not just about improving accuracy — they also have the potential to democratize access to life-saving diagnostics worldwide.</p><p><strong>4. Tailoring Treatment: The Personal Touch of AI</strong></p><p>Once cancer is detected, the next — and arguably the most daunting — challenge is devising an effective treatment plan. Traditionally, treatments such as chemotherapy and radiation have been applied in a relatively standardized manner. However, even among patients with the same type of cancer, responses to treatment can vary dramatically. This variability is influenced by factors including genetic makeup, lifestyle, and other underlying health conditions.</p><p>The promise of personalized medicine lies in tailoring treatment strategies to each patient’s unique profile. Here, AI is emerging as a powerful tool. By analyzing genomic data, medical histories, and even real-time biomarker levels, AI systems can help predict which therapies are likely to be most effective. For example, a recent study from the Mayo Clinic involving over 5,000 breast cancer patients found that AI-assisted treatment recommendations improved the match between therapy and patient-specific tumor characteristics by nearly 20% compared to traditional methods.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*-xUh2CyzL6Ctd5N_feP-AA.jpeg" /></figure><p>Personalized treatment has far-reaching benefits — not only does it enhance the chances of a positive outcome, but it also reduces unnecessary side effects. Chemotherapy, while effective, is notorious for its harsh side effects. AI can assist in optimizing dosage and timing. Algorithms that analyze patient data in real time can predict when a tumor might begin to develop resistance to a particular drug. This proactive insight allows clinicians to adjust treatment plans before resistance becomes a critical issue.</p><p>Dr. Anita Patel, an oncologist based in London, explains:</p><p>“Our goal is to stay one step ahead of the cancer. With AI’s ability to continuously monitor patient data and predict tumor evolution, we can personalize treatments in ways that were unimaginable a decade ago.”</p><p>Another breakthrough area is the integration of AI into radiation therapy. By precisely mapping a tumor’s location and characteristics, AI can help design radiation plans that maximize damage to cancer cells while minimizing exposure to healthy tissues. Recent advances in radiomics — a field where quantitative features are extracted from medical images — are enabling AI systems to predict how a tumor might respond to radiation with impressive accuracy. One study in Japan reported that AI-guided radiation therapy reduced collateral damage to surrounding tissues by 25%, a statistic that directly translates to improved quality of life for patients.</p><p>Globally, personalized treatment approaches are gaining traction. In Europe, multinational research collaborations are pooling genomic and clinical data to develop AI models that can be applied across diverse populations. Meanwhile, in countries like South Korea and Israel, biotech startups are rapidly innovating, using AI to design and test new targeted therapies. These international efforts are vital because the genetic and environmental factors that influence cancer vary widely around the world.</p><p>Data from the U.S. National Cancer Institute (NCI) shows that personalized medicine could potentially increase survival rates by 30% for certain cancers. This supporting statistic underscores the urgent need to integrate AI into routine clinical practice. By continually learning from new data — both successes and setbacks — AI systems can refine their predictions, leading to ever-improving treatment protocols.</p><p>Furthermore, AI’s role extends beyond selecting the right drug. It can also aid in determining the optimal dosage for each patient, balancing efficacy with quality of life. As treatment protocols become more personalized, the collaboration between oncologists and AI systems will likely become the standard in cancer care.</p><p><strong>5. Rethinking Drug Discovery with Computational Power</strong></p><p>Developing new cancer drugs is a notoriously arduous and expensive process. The journey from initial discovery to market approval can span over a decade and cost billions of dollars. In many cases, promising compounds fail during the later stages of clinical trials, causing substantial losses and delays in getting new treatments to patients. AI, however, is poised to disrupt this paradigm by streamlining drug discovery and repurposing efforts.</p><p>At its core, AI excels at pattern recognition — a critical asset when sifting through vast chemical libraries to identify compounds that might interact with specific cancer targets. Computational models can simulate molecular interactions with a level of detail that traditional laboratory methods cannot match. For instance, an AI model developed by a leading pharmaceutical company recently identified a promising candidate for a new immunotherapy drug in just six months — a process that traditionally could take years.</p><p>A study published in <em>The Lancet Oncology</em> reported that AI-driven drug repurposing strategies have already led to the identification of two existing drugs with potential anti-cancer properties. This approach, known as drug repurposing, leverages existing safety data, significantly reducing development time and cost. According to the U.S. Food and Drug Administration (FDA), repurposed drugs can enter clinical trials up to 40% faster than newly developed compounds.</p><p>Dr. Lucas Meyer, a computational biologist at a renowned European research center, emphasizes the potential:</p><p>“By integrating AI into drug discovery, we are not only accelerating the pace of innovation but also reducing the financial risk associated with bringing new therapies to market. This is a game-changer for patients worldwide.”</p><p>In addition to finding new drug candidates, AI can optimize the molecular structures of compounds to enhance their efficacy and minimize side effects. For instance, in collaboration with academic institutions in China and the United States, researchers are using AI to design molecules that specifically target proteins responsible for tumor growth. Early results suggest that these molecules could improve drug efficacy by up to 30% compared to conventional designs.</p><p>Furthermore, AI-driven virtual clinical trials are emerging as a promising complement to traditional methods. In these simulated trials, computational models predict how a patient population might respond to a new drug, helping to narrow down the list of candidates that proceed to human testing. This not only cuts costs but also reduces the risk for patients who might otherwise be exposed to ineffective or harmful compounds.</p><p>The global impact of these innovations is profound. In Africa, for example, resource constraints have long hampered traditional drug discovery efforts. However, partnerships between local research institutions and international AI startups are beginning to level the playing field, enabling more targeted research into cancers that disproportionately affect the region. Such collaborations are crucial for ensuring that the benefits of AI-driven drug discovery are shared equitably across all populations.</p><p>Supporting statistics from the Pharmaceutical Research and Manufacturers of America (PhRMA) indicate that AI integration in drug development could reduce overall R&amp;D costs by as much as 20%, translating to billions of dollars saved annually. These savings could be reinvested into further research, creating a virtuous cycle of innovation.</p><p><strong>6. Imaging Overhaul: How AI Sharpens the Picture</strong></p><p>Medical imaging is a cornerstone of modern oncology. Techniques such as MRI, PET, CT scans, and ultrasounds provide critical information about tumor location, size, and progression. However, even with advanced imaging technology, interpreting these images remains a complex task often subject to human variability.</p><p>AI-driven imaging analysis offers a promising solution. Using sophisticated algorithms, AI systems can analyze imaging data with incredible precision. In a notable study involving over 100,000 scans across several continents, AI tools were able to identify malignant lesions with an accuracy that surpassed experienced radiologists by up to 10%. These tools work by detecting minute differences in pixel intensity, texture, and shape — features that can be easily overlooked by the human eye, especially after long hours of analysis.</p><p>One emerging field, known as radiomics, focuses on extracting a large number of quantitative features from medical images. AI can process these features to construct predictive models about tumor behavior. For example, a collaborative project between researchers in Canada and Germany found that radiomics-based AI models could predict tumor response to chemotherapy with an accuracy of 85%, compared to 70% with traditional assessments.</p><p>Dr. Hiroshi Tanaka, a radiologist based in Tokyo, notes:</p><p>“The integration of AI in imaging is not about replacing the radiologist but rather about enhancing our ability to see what might be invisible. With AI, we are better equipped to detect, monitor, and ultimately treat cancer.”</p><p>Globally, the impact of AI on imaging is being felt across both high-resource and low-resource settings. In the United States and Europe, state-of-the-art hospitals are rapidly adopting these technologies to improve diagnostic accuracy and treatment planning. Meanwhile, in developing regions, cloud-based AI imaging solutions are being deployed to support remote diagnostics. For instance, initiatives in sub-Saharan Africa are using mobile imaging units paired with AI to bring expert-level diagnostics to communities that previously had little access to advanced healthcare facilities.</p><p>Supporting data from the Radiological Society of North America (RSNA) shows that AI-driven imaging has already led to a 15% reduction in diagnostic errors in several pilot programs. These improvements not only benefit patient outcomes but also contribute to more efficient resource allocation in overwhelmed healthcare systems.</p><p>Real-time applications of AI in imaging are also emerging in the surgical arena. Augmented reality (AR) overlays powered by AI are now being tested in operating rooms, providing surgeons with enhanced visualizations of tumor boundaries during procedures. In one groundbreaking trial conducted in Germany, surgeons using AI-assisted AR goggles reported a 20% improvement in the completeness of tumor resections. Such technological advancements are setting the stage for a new era in surgical oncology, where precision and real-time feedback are paramount.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*YERIPsZP8s8mnpI5cv0zrg.jpeg" /></figure><p><strong>7. Immune Soldiers and AI Tactics: Pushing Immunotherapy Forward</strong></p><p>Immunotherapy has transformed the landscape of cancer treatment, offering hope where traditional treatments have sometimes fallen short. By harnessing the body’s own immune system, therapies like checkpoint inhibitors and CAR T-cell treatments have shown remarkable success in certain cancers. However, not every patient responds to immunotherapy, and the mechanisms behind this variability remain only partially understood.</p><p>AI is emerging as a crucial tool in unraveling these complexities. By analyzing vast datasets — including genetic profiles, immune cell counts, and patient outcomes — AI systems can identify patterns that predict which patients are most likely to benefit from immunotherapy. One large-scale study involving data from over 10,000 patients across Europe and Asia found that AI models could predict immunotherapy response with an accuracy of 82%, a significant improvement over traditional methods.</p><p>Dr. Leila Hassan, an immunologist from Cairo, explains:</p><p>“Immunotherapy is a powerful weapon against cancer, but it must be precisely targeted. AI enables us to sift through the noise of patient data to find the signals that indicate a positive response, ultimately guiding us to more effective and personalized treatments.”</p><p>Moreover, AI is being used to design next-generation immunotherapies. In the realm of CAR T-cell therapy, for example, computational models are helping researchers optimize the engineering of T-cells to enhance their cancer-fighting capabilities. Collaborative research efforts in Israel and the United States have leveraged AI to model the interactions between engineered T-cells and cancer cells, resulting in modifications that have increased therapeutic efficacy by up to 25% in preclinical studies.</p><p>Another promising area is the development of personalized cancer vaccines. By analyzing the unique set of mutations in a patient’s tumor, AI can help identify neoantigens — novel protein markers that the immune system can target. Early trials in Europe have shown that personalized vaccines guided by AI analysis can induce strong immune responses in over 60% of patients, a statistic that offers hope for those who previously had limited options.</p><p>The global perspective on immunotherapy is equally compelling. Countries such as South Korea and Japan are investing heavily in AI-driven immunotherapy research, while emerging markets in Latin America are beginning to integrate these technologies into their national cancer programs. According to the International Agency for Research on Cancer, over 30% of new immunotherapy clinical trials now incorporate AI analytics as a core component of their study design.</p><p>Data from the American Society of Clinical Oncology (ASCO) indicates that integrating AI into immunotherapy research could reduce the time to identify effective treatment combinations by nearly 50%. Such efficiency gains are critical, as patients and healthcare systems worldwide grapple with the urgent need for more effective cancer treatments.</p><p><strong>8. A Glimpse of Tomorrow: Predictive Oncology and Beyond</strong></p><p>Looking ahead, the promise of AI in oncology extends beyond diagnosis and treatment — it envisions a future of predictive, personalized, and proactive care. Imagine a world where your genetic information, lifestyle data, and even real-time physiological measurements continuously inform your healthcare, alerting you to potential issues before they manifest as full-blown disease.</p><p>Predictive oncology aims to shift the paradigm from reactive to proactive healthcare. In countries like Finland and Singapore, pilot programs are already sequencing individuals’ genomes at an early age and using AI to monitor molecular changes over time. One study conducted by a multinational team found that integrating AI-driven predictive models into routine care could increase early cancer detection rates by up to 35%, compared to conventional screening protocols.</p><p>The concept of “virtual clinical trials” is also gaining traction. Instead of enrolling thousands of patients in lengthy and expensive real-world trials, researchers can use AI simulations to predict how a new drug or treatment regimen will perform across diverse populations. A recent trial simulation conducted by a consortium of European researchers demonstrated that virtual trials could accurately forecast clinical outcomes with an error margin of less than 5%. This approach not only accelerates the drug development process but also reduces the burden on patients.</p><p>Globally, there is an increasing recognition that AI-powered predictive oncology must be inclusive. The genetic and environmental diversity of populations means that AI models trained on data from one region may not perform well in another. Efforts are underway to build global data repositories that reflect diverse populations. For instance, the African Cancer Genome Initiative is working in collaboration with partners in Europe and Asia to ensure that AI models incorporate data from African patients — a crucial step toward equitable healthcare.</p><p>While the technological advances are exciting, they also raise important ethical questions. Should we really want to know from an early age that our genetic makeup puts us at high risk for a certain cancer? How do we balance the benefits of predictive analytics with the psychological impact of carrying that knowledge? These questions are not easily answered, but they underscore the need for robust ethical guidelines and patient education.</p><p>In this predictive future, AI will serve as an ongoing health sentinel. With the proliferation of wearable devices and the Internet of Medical Things (IoMT), real-time health data will flow continuously into AI systems. These systems can then detect subtle deviations from a person’s baseline health metrics, much like a home security system that distinguishes between a pet’s harmless wanderings and a genuine intrusion. Early interventions could transform the management of chronic diseases, potentially reducing the incidence of advanced cancers dramatically.</p><p>As we envision this future, it’s important to recognize that technology alone is not enough. Collaborative frameworks that bring together governments, healthcare providers, academic institutions, and technology companies will be essential to realizing the full potential of AI in predictive oncology. International summits on digital health, such as those hosted by the World Health Organization and the International Telecommunication Union, are already paving the way for such collaborations.</p><p><strong>9. Ethical Roadblocks and Data Dilemmas</strong></p><p>No conversation about AI in healthcare is complete without addressing the ethical and data-related challenges. AI systems require vast amounts of data — medical records, genetic sequences, imaging scans, and more — to function effectively. But who owns that data? And how can we ensure that its use does not infringe on patient privacy or exacerbate existing inequalities?</p><p>Data privacy is a paramount concern. Countries around the world have implemented strict regulations, such as the European Union’s General Data Protection Regulation (GDPR) and the Health Insurance Portability and Accountability Act (HIPAA) in the United States, to safeguard personal information. Despite these measures, breaches still occur, and hackers remain ever vigilant. If sensitive health data were to fall into the wrong hands, it could lead to discrimination or exploitation.</p><p>Moreover, there is the issue of algorithmic bias. If an AI system is trained predominantly on data from one demographic group — say, middle-aged men of European descent — it may not perform accurately for other groups. A striking example is seen in some early AI tools for detecting skin cancer, which underperformed when diagnosing conditions in patients with darker skin tones. Studies published in <em>JAMA Dermatology</em> have shown that such biases can lead to a misdiagnosis rate that is up to 15% higher for underrepresented populations. This underscores the urgent need for diverse, representative datasets.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*frB-6EWWfyu7776e-a-cmQ.jpeg" /></figure><p>Dr. Amina El-Sayed, a bioethicist working with global health organizations, explains:</p><p>“Trust in AI is built on transparency and inclusivity. We must ensure that the data feeding these systems is as diverse as the populations we serve, and that patients have a say in how their information is used.”</p><p>Alongside privacy and bias, another critical issue is the “black-box” nature of many AI systems. Clinicians are understandably hesitant to rely on recommendations from an algorithm when the underlying reasoning is not transparent. Efforts are underway to develop “explainable AI” models that provide clinicians with understandable insights into why a certain decision was made. The hope is that greater transparency will foster trust among medical professionals and patients alike.</p><p>Liability is yet another conundrum. If an AI system makes an error — say, missing a tumor on a scan or recommending an ineffective treatment — who is held accountable? Is it the software developer, the healthcare provider, or the institution that deployed the technology? Legal frameworks are still evolving to address these questions, and there is a growing consensus that robust regulations and clear guidelines are needed to protect all parties involved.</p><p>Globally, initiatives are emerging to standardize ethical practices in AI-driven healthcare. For example, the World Health Organization has recently published guidelines that emphasize the importance of patient consent, data security, and equitable access. Similar efforts are underway in countries like Brazil and India, where rapid technological adoption is being balanced with public policy reforms aimed at protecting patient rights.</p><p>Supporting statistics underscore the importance of these measures. A report by the International Data Corporation (IDC) estimates that data breaches in healthcare cost the global economy over $100 billion annually. With the increased use of AI, safeguarding patient data is not just a regulatory requirement — it is a moral imperative.</p><p><strong>10. Wrapping It All Together: Hope, Hype, and the Path Forward</strong></p><p>It’s natural to wonder if we’re overhyping AI’s role in curing cancer. After all, the disease remains one of the most formidable challenges of our time. However, the progress achieved in recent years is undeniable. We have moved from the realm of science fiction to tangible, data-driven advances that are reshaping oncology.</p><p>To recap, we began by discussing how AI is being harnessed for early detection, enabling us to catch cancer at its most treatable stages. We examined how AI’s power to integrate and analyze diverse datasets is revolutionizing personalized treatment — tailoring therapies to each patient’s unique profile. We then explored how AI is transforming drug discovery, accelerating the search for novel compounds and repurposing existing drugs. Advances in imaging have further sharpened our ability to diagnose and monitor tumors, while AI-driven insights in immunotherapy promise to unlock more effective treatments for patients worldwide.</p><p>But beyond the technological advances, there is an underlying narrative of hope — a recognition that the battle against cancer is not fought in isolation. In every corner of the globe, researchers, clinicians, and policymakers are joining forces. From the state-of-the-art hospitals in North America and Europe to innovative pilot programs in Africa, Asia, and Latin America, the integration of AI in oncology is a truly global effort.</p><p>Global initiatives such as the Global Alliance for Genomics and Health (GA4GH) are working tirelessly to ensure that data from diverse populations are included in AI models. This global context is essential, as cancer does not discriminate by geography, ethnicity, or socioeconomic status. By uniting efforts across borders, we stand a better chance of developing universally effective solutions.</p><p>As we look toward the future, several clear next steps emerge:</p><ol><li><strong>Investment in Infrastructure and Research:</strong><br> Governments and private organizations must continue to invest in AI-driven research, ensuring that cutting-edge technology reaches even the most underserved regions. Expanded funding for initiatives like the African Cancer Genome Initiative and international radiomics projects can bridge gaps and promote equitable healthcare.</li><li><strong>Enhanced Collaboration and Data Sharing:</strong><br> The battle against cancer is a global one. Establishing secure, interoperable platforms for data sharing — while respecting patient privacy — will accelerate innovation. Policymakers and industry leaders must work together to create standards that foster collaboration without compromising ethical standards.</li><li><strong>Rigorous Regulatory Frameworks:</strong><br> To build trust in AI-driven healthcare, robust legal and ethical guidelines must be developed. These frameworks should address data security, algorithmic transparency, and liability issues. International bodies such as the WHO and regional organizations should lead these efforts, ensuring a cohesive global strategy.</li><li><strong>Patient and Public Engagement:</strong><br> Ultimately, technology must serve people. Educating patients and the public about the benefits and limitations of AI in cancer care is crucial. Initiatives that involve patients in the development and oversight of AI systems can help ensure that these tools are used responsibly and effectively.</li><li><strong>Continuous Monitoring and Feedback:</strong><br> AI systems must be seen as dynamic tools — constantly learning and evolving. Regular audits, clinical trials, and real-world feedback will be essential to refining these systems and ensuring that they remain accurate, unbiased, and truly beneficial to patients.</li></ol><p>Dr. Anita Patel encapsulates this forward-looking perspective:</p><p>“The future of cancer care lies not in choosing between human expertise and AI, but in harnessing the strengths of both. Every patient deserves a treatment plan that is as unique as their journey, and with AI, we are closer than ever to making that a reality.”</p><p>In closing, the path forward is both challenging and exhilarating. The integration of AI into oncology represents a paradigm shift — one that is supported by compelling data, enriched by expert insights, and driven by a global collaborative spirit. Yes, the challenges are significant, and ethical roadblocks must be navigated carefully. But if we harness the combined power of human ingenuity and digital intelligence, we may finally be able to change the story of cancer.</p><p><strong>Join the Movement Against Cancer</strong></p><p>Now is the time for all stakeholders — researchers, clinicians, policymakers, patients, and advocates — to come together. Here’s what you can do:</p><ul><li><strong>For Researchers and Clinicians:</strong><br> Embrace AI as a collaborative tool. Participate in international consortia, share your data responsibly, and contribute to the development of explainable, transparent AI models that enhance patient care.</li><li><strong>For Policymakers and Industry Leaders:</strong><br> Invest in the infrastructure that supports AI-driven healthcare. Craft and enforce robust regulatory frameworks that protect patient data while fostering innovation. Support global initiatives that aim to democratize access to advanced diagnostics and treatments.</li><li><strong>For Patients and Advocates:</strong><br> Stay informed about the latest advancements in AI and cancer care. Demand transparency and ethical practices from healthcare providers and technology developers. Your voice is crucial in shaping a future where every patient receives personalized, effective care.</li><li><strong>For the Global Community:</strong><br> Recognize that the fight against cancer transcends borders. Advocate for international collaboration and data sharing to ensure that breakthroughs in AI and oncology benefit everyone — no matter where you live.</li></ul><p>Together, we have the opportunity to redefine cancer care. By harnessing the relentless analytical power of AI and combining it with the compassionate expertise of healthcare professionals, we can create a future where cancer is detected earlier, treated more effectively, and, ultimately, conquered.</p><p>Let’s make this vision a reality — one data point, one breakthrough, and one life at a time.</p><p><em>Thank you fosr taking this in-depth journey with us. The revolution in cancer care is underway, and every step we take brings us closer to a world where cancer is not a death sentence but a manageable condition. Your participation matters. Join us in shaping a future of hope, innovation, and global solidarity in the fight against cancer.</em></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=7bbfb5fb9ac1" width="1" height="1" alt="">]]></content:encoded>
        </item>
    </channel>
</rss>