Published inTowards Data ScienceThe Journey of RAG Development: From Notebook to MicroservicesConverting a Colab notebook to two microservices with support for Milvus and NeMo GuardrailsFeb 213Feb 213
Published inTowards Data ScienceNeMo Guardrails, the Ultimate Open-Source LLM Security ToolkitExploring NeMo Guardrails’ practical use casesFeb 9Feb 9
Published inTowards Data Science12 RAG Pain Points and Proposed SolutionsSolving the core challenges of Retrieval-Augmented GenerationJan 3016Jan 3016
Published inTowards Data ScienceJump-start Your RAG Pipelines with Advanced Retrieval LlamaPacks and Benchmark with Lighthouz AIExploring robust RAG development with LlamaPacks, Lighthouz AI, and Llama GuardJan 292Jan 292
Published inTowards Data ScienceExploring mergekit for Model Merge, AutoEval for Model Evaluation, and DPO for Model Fine-tuningMy observations from experimenting with model merge, evaluation, and two model fine-tuning techniquesJan 193Jan 193
Published inTowards Data ScienceDemocratizing LLMs: 4-bit Quantization for Optimal LLM InferenceA deep dive into model quantization with GGUF and llama.cpp and model evaluation with LlamaIndexJan 152Jan 152
Published inTowards Data ScienceDeploying LLM Apps to AWS, the Open-Source Self-Service WayA step-by-step guide on deploying LlamaIndex RAGs to AWS ECS fargateJan 83Jan 83
Published inTowards Data ScienceSafeguarding Your RAG Pipelines: A Step-by-Step Guide to Implementing Llama Guard with LlamaIndexHow to add Llama Guard to your RAG pipelines to moderate LLM inputs and outputs and combat prompt injectionDec 27, 20232Dec 27, 20232
Published inLevel Up Coding10+ Ways to Run Open-Source Models with LlamaIndexLlamaIndex’s open-source model integration with Hugging Face, vLLM, Ollama, Llama.cpp, liteLLM, Replicate, Gradient, and moreDec 19, 20233Dec 19, 20233