Published inTDS ArchivePreference Alignment for Everyone!Frugal RLHF with multi-adapter PPO on Amazon SageMakerNov 8, 2024A response icon2Nov 8, 2024A response icon2
Published inTDS ArchiveA Deep Dive into Fine-TuningStepping out of the “comfort zone” — part 3/3 of a deep-dive into domain adaptation approaches for LLMsJun 3, 2024Jun 3, 2024
Published inTDS ArchiveA Deep Dive into In-Context LearningStepping out of the “comfort zone” — part 2/3 of a deep-dive into domain adaptation approaches for LLMsMay 31, 2024May 31, 2024
Published inTDS ArchiveIntroduction to Domain Adaptation- Motivation, Options, TradeoffsStepping out of the “comfort zone” — part 1/3 of a deep-dive into domain adaptation approaches for LLMsMay 28, 2024May 28, 2024
LLM domain adaptation using continued pre-training — Part 4/4Exploring domain adaptation via continued pre-training for large language models (LLMs)? This 4-part series answers the most common…May 21, 2024May 21, 2024
LLM domain adaptation using continued pre-training — Part 2/4Exploring domain adaptation via continued pre-training for large language models (LLMs)? This 4-part series answers the most common…May 21, 2024A response icon1May 21, 2024A response icon1
Published inTDS ArchiveLeveraging qLoRA for Fine-Tuning of Task-Fine-Tuned Models Without Catastrophic ForgettingLearn how to infuse knowledge into purpose-fine-tuned models while keeping their task-specific natureSep 7, 2023Sep 7, 2023
Measuring and improving UX metrics of LLM-powered chatbots: latency benchmarking with LangChain…What is this about and why is this relevant?Aug 29, 2023A response icon1Aug 29, 2023A response icon1