PinnedIntel(R) Neural CompressorinIntel Analytics SoftwarePersonalized Stable Diffusion with Few-Shot Fine-TuningCreate Your Own Stable Diffusion on a Single CPUNov 1, 20221Nov 1, 20221
Intel(R) Neural CompressorinIntel Analytics SoftwareQuantization on Intel Gaudi Series AI AcceleratorsIntel Neural Compressor v3.0 Supports Quantization across Intel HardwareAug 16Aug 16
Intel(R) Neural CompressorinIntel Analytics SoftwareAccelerating Qwen2 Models with Intel Extension for TransformersHigh Performance WOQ INT4 Inference on Intel Xeon ProcessorsJun 6Jun 6
Intel(R) Neural CompressorinIntel Analytics SoftwareAccelerating GGUF Models with TransformersImproving Performance and Memory Usage on Intel PlatformsMay 31May 31
Intel(R) Neural CompressorinIntel Analytics SoftwareLow-Bit Quantized Open LLM LeaderboardA New Tool to Find High-Quality Models for a Given ClientMay 11May 11
Intel(R) Neural CompressorinIntel Analytics SoftwareThe AutoRound Quantization AlgorithmWeight-Only Quantization for LLMs Across Hardware PlatformsApr 2Apr 2
Intel(R) Neural CompressorinIntel Analytics SoftwareRun LLMs on Intel GPUs Using llama.cppTaking Advantage of the New SYCL BackendMar 221Mar 221
Intel(R) Neural CompressorinIntel Analytics SoftwareEfficient Quantization with Microscaling Data Types for Large Language ModelsNew Quantization Recipes Using Intel Neural CompressorMar 1Mar 1
Intel(R) Neural CompressorinIntel Analytics SoftwareEfficient Natural Language Embedding Models with Intel Extension for TransformersMaking Retrieval-Augmented Generation More EfficientFeb 8Feb 8
Intel(R) Neural CompressorinIntel Analytics SoftwareAdvancing Large Language Models on Intel PlatformsThe Evolution of Intel NeuralChat-7B LLMDec 19, 20231Dec 19, 20231