Shravan KumarMeta Llama 3: The most capable openly available LLM to date and Itโs Application๐ฅ ๐๐จ๐จ๐ฆ ๐ข๐ง ๐ญ๐ก๐ ๐๐ ๐๐จ๐ซ๐ฅ๐: ๐๐ฅ๐๐ฆ๐ ๐ ๐ข๐ฌ ๐๐๐ซ๐! ๐ฅ5 min readยทApr 26, 2024----
Shravan KumarA deep dive into TokenizationWe recall that the language modelling involves computing probabilities over a sequence of tokens9 min readยทMar 20, 2024----
Shravan KumarGemma: Introducing new state-of-the-art open model by GoogleGemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Geminiโฆ6 min readยทFeb 23, 2024--1--1
Shravan KumarBidirectional Encoder Representations from Transformers (BERT)In my previous blogs we studied about entire overview on Generative Pretrained Transformer a blog on Generative Pretrained Transformerโฆ13 min readยทDec 17, 2023--1--1
Shravan KumarDecoding Strategies of all Decoder only Models (GPT)In my previous blogs we studied about entire overview on Generative Pretrained Transformer and also a blog on Generative Pretrainedโฆ11 min readยทDec 10, 2023----
Shravan KumarGenerative Pretrained Transformer (GPT)โ Pre-trainingย , Fine Tuning & Different Use CaseโฆIn the previous blog we studied about entire overview on Generative Pretrained Transformer. Now let us look at super important topics onโฆ6 min readยทNov 27, 2023----
Shravan KumarGenerative Pretrained Transformer (GPT)A primer into the Decoder only ModelโโโCausal Langauge Modelling10 min readยทNov 25, 2023----
Shravan KumarIntroduction to Language ModellingIn my previous blog we learned about the components of the transformer architecture in the context of machine translation.7 min readยทNov 22, 2023----
Shravan KumarTransformers: Attention is all you needโโโLayer NormalizationThere are two major concepts which we are going to discuss here are6 min readยทNov 16, 2023----
Shravan KumarTransformers: Attention is all you needโโโPositional EncodingPlease refer to below blogs before reading this:6 min readยทNov 9, 2023----