Shravan KumarMeta Llama 3: The most capable openly available LLM to date and It’s Application💥 𝐁𝐨𝐨𝐦 𝐢𝐧 𝐭𝐡𝐞 𝐀𝐈 𝐖𝐨𝐫𝐥𝐝: 𝐋𝐥𝐚𝐦𝐚 𝟑 𝐢𝐬 𝐇𝐞𝐫𝐞! 💥Apr 26Apr 26
Shravan KumarA deep dive into TokenizationWe recall that the language modelling involves computing probabilities over a sequence of tokensMar 20Mar 20
Shravan KumarGemma: Introducing new state-of-the-art open model by GoogleGemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini…Feb 231Feb 231
Shravan KumarBidirectional Encoder Representations from Transformers (BERT)In my previous blogs we studied about entire overview on Generative Pretrained Transformer a blog on Generative Pretrained Transformer…Dec 17, 20231Dec 17, 20231
Shravan KumarDecoding Strategies of all Decoder only Models (GPT)In my previous blogs we studied about entire overview on Generative Pretrained Transformer and also a blog on Generative Pretrained…Dec 10, 2023Dec 10, 2023
Shravan KumarGenerative Pretrained Transformer (GPT)— Pre-training , Fine Tuning & Different Use Case…In the previous blog we studied about entire overview on Generative Pretrained Transformer. Now let us look at super important topics on…Nov 27, 2023Nov 27, 2023
Shravan KumarGenerative Pretrained Transformer (GPT)A primer into the Decoder only Model — Causal Langauge ModellingNov 25, 2023Nov 25, 2023
Shravan KumarIntroduction to Language ModellingIn my previous blog we learned about the components of the transformer architecture in the context of machine translation.Nov 22, 2023Nov 22, 2023
Shravan KumarTransformers: Attention is all you need — Layer NormalizationThere are two major concepts which we are going to discuss here areNov 16, 2023Nov 16, 2023
Shravan KumarTransformers: Attention is all you need — Positional EncodingPlease refer to below blogs before reading this:Nov 9, 2023Nov 9, 2023