Rahul KumarModel Quantization : GPT-Q, AWQ, and BeyondDo you know how model quantization works? 🤔19h ago
InTowards Data SciencebyMichio SuginooRisk Implications of Excessive Multiple Local Minima during Hyperparameter TuningOur Epistemological Limitation and Illusion of KnowledgeOct 12, 20221
InTowards Data SciencebyNicolas LupiFeature Selection with OptunaA versatile and promising approach for the feature selection taskMay 91May 91
Ethan HenleyTensor Shapes Tell Stories: Following Neural Networks Through CompilationA primer on speaking the language of neural networks5d ago5d ago
Eleventh Hour EnthusiastScaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model ParametersPaper ReviewOct 5Oct 5
Rahul KumarModel Quantization : GPT-Q, AWQ, and BeyondDo you know how model quantization works? 🤔19h ago
InTowards Data SciencebyMichio SuginooRisk Implications of Excessive Multiple Local Minima during Hyperparameter TuningOur Epistemological Limitation and Illusion of KnowledgeOct 12, 20221
InTowards Data SciencebyNicolas LupiFeature Selection with OptunaA versatile and promising approach for the feature selection taskMay 91
Ethan HenleyTensor Shapes Tell Stories: Following Neural Networks Through CompilationA primer on speaking the language of neural networks5d ago
Eleventh Hour EnthusiastScaling LLM Test-Time Compute Optimally can be More Effective than Scaling Model ParametersPaper ReviewOct 5
InTowards AIbyAchyut BoggaramMachine Learning Model PruningThe Art and Science of Efficient Machine Learning SeriesDec 41
DeeperAndCheaper[YoloV9][Model Optimization][Knowledge Distillation] #1 — Why Knowledge Distillation for Object…Why knowledge distillation ?May 9