PinnedPublished inIntel Analytics SoftwarePersonalized Stable Diffusion with Few-Shot Fine-TuningCreate Your Own Stable Diffusion on a Single CPUNov 1, 20221Nov 1, 20221
Published inIntel Analytics Software10 Tips for Quantizing LLMs and VLMs with AutoRoundAutoRound V0.4 has been released, featuring major updates to experimentally support Vision-Language Models (VLMs). In this article, we will…Nov 22Nov 22
Published inIntel Analytics SoftwareQuantization on Intel Gaudi Series AI AcceleratorsIntel Neural Compressor v3.0 Supports Quantization across Intel HardwareAug 16Aug 16
Published inIntel Analytics SoftwareAccelerating Qwen2 Models with Intel Extension for TransformersHigh Performance WOQ INT4 Inference on Intel Xeon ProcessorsJun 6Jun 6
Published inIntel Analytics SoftwareAccelerating GGUF Models with TransformersImproving Performance and Memory Usage on Intel PlatformsMay 31May 31
Published inIntel Analytics SoftwareLow-Bit Quantized Open LLM LeaderboardA New Tool to Find High-Quality Models for a Given ClientMay 11May 11
Published inIntel Analytics SoftwareThe AutoRound Quantization AlgorithmWeight-Only Quantization for LLMs Across Hardware PlatformsApr 2Apr 2
Published inIntel Analytics SoftwareRun LLMs on Intel GPUs Using llama.cppTaking Advantage of the New SYCL BackendMar 221Mar 221
Published inIntel Analytics SoftwareEfficient Quantization with Microscaling Data Types for Large Language ModelsNew Quantization Recipes Using Intel Neural CompressorMar 1Mar 1
Published inIntel Analytics SoftwareEfficient Natural Language Embedding Models with Intel Extension for TransformersMaking Retrieval-Augmented Generation More EfficientFeb 8Feb 8