PinnedGPT - Generative Pre-trained TransformerUnderstanding the GPT family, architectures and limitationsMay 17May 17
PinnedBERT - Bidirectional Encoder Representations from TransformersExploring the State Of The Art NLP Encoder ModelMay 11May 11
PinnedUnlocking Transformer Power: A Deep Dive into Attention MechanismsExploring Self-Attention and Pre-Training for Enhanced Sequence ModelingApr 29Apr 29
Speeding-up Transformers[ποΈ βπ]Optimizing Transformer architecture for faster inferences ππMay 25May 25