Md Monsur aliHow to Use Ollama with GGUF Models from Hugging Face Hub: A Step-by-Step GuideLearn how to easily run GGUF quantized models from Hugging Face using Ollama, customize quantization, chat templates, and sampling…6h ago
DThe easiest way to convert a model to GGUF and Quantizedocker run ghcr.io/ggerganov/llama.cpp. The models will be in the same folder with .bin extension. That’s it!Jun 182
Wei-Meng LeeinAI AdvancesServing LLMs using LM StudioLearn how to run and serve a LLM Locally on your computer2d ago2d ago
Nicolas MorenoLocal LLMs on iOSExploring the implementation of open-source ML and LLMs models locally on a iOS mobile device with GGUF formats.May 20May 20
Lada HangRunning ComfyUI — Flux-Upscaler-GGUF-Workflow on MimicPCWhen I initially created a Flux-dev-Upscaler-workflow on MimicPC, some users reported issues running it on Medium configurations (T4 16GB…Sep 30Sep 30
Md Monsur aliHow to Use Ollama with GGUF Models from Hugging Face Hub: A Step-by-Step GuideLearn how to easily run GGUF quantized models from Hugging Face using Ollama, customize quantization, chat templates, and sampling…6h ago
DThe easiest way to convert a model to GGUF and Quantizedocker run ghcr.io/ggerganov/llama.cpp. The models will be in the same folder with .bin extension. That’s it!Jun 182
Wei-Meng LeeinAI AdvancesServing LLMs using LM StudioLearn how to run and serve a LLM Locally on your computer2d ago
Nicolas MorenoLocal LLMs on iOSExploring the implementation of open-source ML and LLMs models locally on a iOS mobile device with GGUF formats.May 20
Lada HangRunning ComfyUI — Flux-Upscaler-GGUF-Workflow on MimicPCWhen I initially created a Flux-dev-Upscaler-workflow on MimicPC, some users reported issues running it on Medium configurations (T4 16GB…Sep 30
Plaban NayakinThe AI ForumInstruction Fine-Tuning Gemma-2B on Medical Reasoning and Convert the finetuned model into GGUF…Gemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini…Mar 102
Pierre MesureImport embeddings model with OllamaAre you struggling like I did to import your Pytorch embeddings model to Ollama? Here’s how I did it.Sep 28
kirouane AyoubinGoPenAIExploring Bits-and-Bytes, AWQ, GPTQ, EXL2, and GGUF Quantization Techniques with Practical Examples1. Bits-and-Bytes QuantizationAug 22