PinnedJAIGANESANExploring LLM: A Collection of My Articles❤️Dive into the intricate world of large language models with in-depth articles on their architectures, MoE, and RAG. Discover more by…Jun 25Jun 25
JAIGANESANinTowards AIQuantization: Post Training Quantization, Quantization Error, and Quantization Aware TrainingEfficient Inference in AI Models3d ago3d ago
JAIGANESANinTowards AIVisualizing Low-Rank Adaptation (LoRA) 👀Exploring Singular Value Decomposition (SVD), Feed-Forward Networks (FFN), and LoRAJul 13Jul 13
JAIGANESANinTowards AIKAN (Kolmogorov-Arnold Networks): A Starter Guide 🐣Exploring MLP, the Universal Approximation theorem, the Kolmogorov-Arnold Representation theorem, the Bezier Curve, B-splines, and KAN.Jul 4Jul 4
JAIGANESANinTowards AIBERT: In-depth exploration of Architecture, Workflow, Code, and Mathematical FoundationsDelving into Embeddings, Masked Language Model Tasks, Attention Mechanisms, and Feed-Forward Networks: Not Just Another BERT ArticleJun 26Jun 26
JAIGANESANinTowards AIA Visual Walkthrough of DeepSeek’s Multi-Head Latent Attention (MLA) 🧟♂️Exploring Bottleneck in GPU Utilization and Multi-head Latent Attention Implementation in DeepSeekV2Jun 20Jun 20
JAIGANESANinTowards AIRetrieval Augmented Generation (RAG): A Comprehensive Visual Walkthrough 🧠📖🔗🤖A Simple Illustrated GuideJun 192Jun 192
JAIGANESANinTowards AIFrom Pixels to Words: How Model Understands? 🤝🤝From the pixels of images to the words of language, explore how multimodal AI models bridge diverse data types through sophisticated…Jun 17Jun 17
JAIGANESANinTowards AIRevolutionizing AI with DeepSeekMoE: Fine-grained Expert, and Shared Expert isolation 🧞♂️Optimizing MoE with Fine-Grained and shared expert isolation for enhanced precision and efficiency in Large Language models.Jun 10Jun 10
JAIGANESANinTowards AIThe architecture of Mistral’s Sparse Mixture of Experts (S〽️⭕E)Exploring Feed Forward Networks, Gating Mechanism, Mixture of Experts (MoE), and Sparse Mixture of Experts (SMoE).Jun 4Jun 4