Published inTDS ArchiveThe Journey of RAG Development: From Notebook to MicroservicesConverting a Colab notebook to two microservices with support for Milvus and NeMo GuardrailsFeb 21, 20243Feb 21, 20243
Published inTDS ArchiveNeMo Guardrails, the Ultimate Open-Source LLM Security ToolkitExploring NeMo Guardrails’ practical use casesFeb 9, 2024Feb 9, 2024
Published inTDS Archive12 RAG Pain Points and Proposed SolutionsSolving the core challenges of Retrieval-Augmented GenerationJan 30, 202416Jan 30, 202416
Published inTDS ArchiveJump-start Your RAG Pipelines with Advanced Retrieval LlamaPacks and Benchmark with Lighthouz AIExploring robust RAG development with LlamaPacks, Lighthouz AI, and Llama GuardJan 29, 20242Jan 29, 20242
Published inTDS ArchiveExploring mergekit for Model Merge, AutoEval for Model Evaluation, and DPO for Model Fine-tuningMy observations from experimenting with model merge, evaluation, and two model fine-tuning techniquesJan 19, 20243Jan 19, 20243
Published inTDS ArchiveDemocratizing LLMs: 4-bit Quantization for Optimal LLM InferenceA deep dive into model quantization with GGUF and llama.cpp and model evaluation with LlamaIndexJan 15, 20243Jan 15, 20243
Published inTDS ArchiveDeploying LLM Apps to AWS, the Open-Source Self-Service WayA step-by-step guide on deploying LlamaIndex RAGs to AWS ECS fargateJan 8, 20243Jan 8, 20243
Published inTDS ArchiveSafeguarding Your RAG Pipelines: A Step-by-Step Guide to Implementing Llama Guard with LlamaIndexHow to add Llama Guard to your RAG pipelines to moderate LLM inputs and outputs and combat prompt injectionDec 27, 20232Dec 27, 20232
Published inLevel Up Coding10+ Ways to Run Open-Source Models with LlamaIndexLlamaIndex’s open-source model integration with Hugging Face, vLLM, Ollama, Llama.cpp, liteLLM, Replicate, Gradient, and moreDec 19, 20233Dec 19, 20233