Hyunjong LeeDissecting BERT: A Comprehensive Exploration of the Inner Workings of Transformer-Based NLP ModelsExplore BERT’s code-level mechanics, from tokenization to self-attention layers and model architecture6h ago
Najib Sharifi, Ph.D.inAI AdvancesGenerative AI: Transformers For Molecular DesignBuilding a transformer model for generating molecules with desired physical properties. A pytorch implementationAug 20
Vipra SinghLLM Architectures Explained: NLP Fundamentals (Part 1)Deep Dive into the architecture & building of real-world applications leveraging NLP Models starting from RNN to the Transformers.Aug 1510Aug 1510
Shravan Kumar📣 HUGE NEWS! …Llama 3.2 is here! 🦙Meta has released Llama 3.2, bringing multimodal capabilities and tiny Llamas for on-device usage! 🎉5h ago5h ago
Sascha KirchinTowards Data ScienceTowards Mamba State Space Models for Images, Videos and Time SeriesPart 1Aug 14Aug 14
Hyunjong LeeDissecting BERT: A Comprehensive Exploration of the Inner Workings of Transformer-Based NLP ModelsExplore BERT’s code-level mechanics, from tokenization to self-attention layers and model architecture6h ago
Najib Sharifi, Ph.D.inAI AdvancesGenerative AI: Transformers For Molecular DesignBuilding a transformer model for generating molecules with desired physical properties. A pytorch implementationAug 20
Vipra SinghLLM Architectures Explained: NLP Fundamentals (Part 1)Deep Dive into the architecture & building of real-world applications leveraging NLP Models starting from RNN to the Transformers.Aug 1510
Shravan Kumar📣 HUGE NEWS! …Llama 3.2 is here! 🦙Meta has released Llama 3.2, bringing multimodal capabilities and tiny Llamas for on-device usage! 🎉5h ago
Sascha KirchinTowards Data ScienceTowards Mamba State Space Models for Images, Videos and Time SeriesPart 1Aug 14
Vipra SinghLLM Architectures Explained: Encoder-Decoder Architecture (Part 4)Deep Dive into the architecture & building real-world applications leveraging NLP Models starting from RNN to Transformer.Sep 17
Pavan SaishFrom Generic to Genius: Personalizing AI with Your Data — Part 1Fine-Tuning Llama 3.1 8B model with Your Own Dataset: A Step-by-Step Guide; Transformers; Quantization; PEFT techniques: LoRA and QLoRA;21h ago1
Anindya Dey, PhDinTowards Data ScienceSpeeding Up the Vision Transformer with BatchNormHow integrating Batch Normalization in an encoder-only Transformer architecture can lead to reduced training time and inference time.Aug 6