PinnedYacine BouaouniQuantization Techniques for Large Language Models (Part 1): A Comprehensive Guide for Machine…Quantization is a critical technique for optimizing deep learning models, particularly when deploying them in resource-constrained…Aug 4Aug 4
Yacine BouaouniMixture of ExpertsMixture of Experts (MoE) is a machine learning technique that combines multiple expert models to improve the performance. This technique…Dec 14, 2023Dec 14, 2023
Yacine BouaouniFrom RNNs to Transformers: A Journey Through the Evolution of Attention Mechanisms in NLPModels based on Recurrent Neural Networks (RNNs), Long Short-Term Memory (LSTM), and Gated Recurrent Units (GRUs) played a significant role…Aug 8, 2023Aug 8, 2023