InGenerative AIbyFabio MatricardiWhy Small Language Models are so good?They are over-trained, but they work just fine. What’s the secret behind it?Oct 132
Sai Chaitanya PachipulusuQuantization, Distillation and PruningAdvancing LLMs Through Efficient Model Compression TechniquesNov 21
InHuggingFacebyVictor Sanh🏎 Smaller, faster, cheaper, lighter: Introducing DilBERT, a distilled version of BERTYou can find the code to reproduce the training of DilBERT along with pre-trained weights for DilBERT here.Aug 28, 201920Aug 28, 201920
David PollingtonEnabling AI at the EdgeCutting-edge AI/ML models, and especially large language and multimodal models (LLMs; LMMs), are capable of a wide-range of sophisticated…Nov 12Nov 12
Aaditya uraQuantization vs Distillation in Neural Networks: A ComparisonA dive into the techniques of quantizing and distilling deep learning models: What are they and how do they differ?Nov 11, 2023Nov 11, 2023
InGenerative AIbyFabio MatricardiWhy Small Language Models are so good?They are over-trained, but they work just fine. What’s the secret behind it?Oct 132
Sai Chaitanya PachipulusuQuantization, Distillation and PruningAdvancing LLMs Through Efficient Model Compression TechniquesNov 21
InHuggingFacebyVictor Sanh🏎 Smaller, faster, cheaper, lighter: Introducing DilBERT, a distilled version of BERTYou can find the code to reproduce the training of DilBERT along with pre-trained weights for DilBERT here.Aug 28, 201920
David PollingtonEnabling AI at the EdgeCutting-edge AI/ML models, and especially large language and multimodal models (LLMs; LMMs), are capable of a wide-range of sophisticated…Nov 12
Aaditya uraQuantization vs Distillation in Neural Networks: A ComparisonA dive into the techniques of quantizing and distilling deep learning models: What are they and how do they differ?Nov 11, 2023
InIt’s My Life 2.3byObinna MortonDistillation Toward, Godwilling, AlchemyOur default to mediocrity, living in reality while still believing it is possible? I am really taking a minute to breathe and think when I…Oct 191
InTowards AIbyLouis-François BouchardSmaller, Faster, Smarter: The Power of Model DistillationWhy OpenAI’s New Approach Challenges the Open-Source AI CommunityOct 101
Bijit GhoshSelf-Improving LLMs & Enhancing ReasoningAlphaLLM-CPL Revolutionizing LLM Reasoning with MCTS Behavior Distillation and Adaptive Learning StrategiesOct 16