AIGuys Digest | Jan 2025
š Welcome to the AIGuys Digest Newsletter, where we cover State-of-the-Art AI breakthroughs and all the major AI newsš. Donāt forget to check my new book on AI, it covers a lot of AI optimizations and hands-on code:
Ultimate Neural Network Programming with Python
š Inside this Issue:
- š¤ Latest Breakthroughs: This month itās all about DeepSeek, Agentic Framework, and RAG.
- š AI Monthly News: Discover how these stories revolutionize industries and impact everyday life.
- š Editorās Special: This covers the interesting talks, lectures, and articles we came across recently.
Letās embark on this journey of discovery together! šš¤š
Follow me on Twitter and LinkedIn at RealAIGuys and AIGuysEditor.
Latest Breakthroughs
Recently, post-training has emerged as an important component of the full training pipeline. It has been shown to enhance accuracy on reasoning tasks, align with social values, and adapt to user preferences, all while requiring relatively minimal computational resources against pre-training.
In the context of reasoning capabilities, OpenAIās o1 series models were the first to introduce inference-time scaling by increasing the length of the Chain-of-Thought reasoning process. This approach has significantly improved in various reasoning tasks, such as mathematics, coding, and scientific reasoning.
Several previous works have explored various approaches, including process-based reward models, reinforcement learning, and search algorithms such as Monte Carlo Tree Search and Beam Search. However, none of these methods has achieved general reasoning performance comparable to OpenAIās o1 series models. So, letās see what Deepseek has cooked to challenge the leader in reasoning.
DeepSeek R1 Beating OpenAI In Reasoning
2025 is bringing a big change in AI ā the rise of agent frameworks. While weāve made great progress with Large Language Models (LLMs), just making them bigger and training them on more data isnāt giving us major improvements anymore. Even OpenAIās O3 model shows that weāre hitting the limits of what we can achieve through pre-training alone.
So whatās next? The answer lies in reinforcement learning (RL) and agents ā AI systems that can actively think through problems and work towards goals. Instead of just responding to prompts, these systems use RL to learn how to reason and make decisions.
Letās look at the different frameworks that are making this possible and how theyāre changing the way AI works.
Understanding Agentic Framework
Retrieval-augmented generation (RAG) has gained traction as a powerful approach for enhancing language models by integrating external knowledge sources. However, RAG introduces challenges such as retrieval latency, potential errors in document selection, and increased system complexity.
With the advent of large language models (LLMs) featuring significantly extended context windows, cache-augmented generation (CAG) bypasses real-time retrieval. It involves preloading all relevant resources into the LLMās extended context and caching its runtime parameters, especially when the documents or knowledge for retrieval are limited and manageable.
Donāt Do RAG, Itās Time For CAG
AI Monthly News
DeepSeekās AI Advancements
Chinese AI startup DeepSeek has released its AI Assistant, utilizing the V3 model, which has quickly become the highest-rated free app on the U.S. iOS App Store. Notably, DeepSeek achieved this with significantly fewer resources than its competitors, training its model with approximately 2,000 GPUs over 55 days at a cost of $5.58 million ā about one-tenth of Metaās recent AI expenditures. This efficiency has led to concerns about the U.S. maintaining its lead in AI development.
Wiki: Source
Financial Impact News: Source
Metaās Continued Investment in AI
Despite DeepSeekās rapid progress, Meta remains committed to substantial AI investments. The company plans to allocate hundreds of billions of dollars to AI initiatives, aiming to solidify its market position within the year. Metaās strategy includes enhancing AI infrastructure, integrating AI into platforms like Facebook and Instagram, and improving ad targeting. CEO Mark Zuckerberg acknowledges DeepSeekās advancements but emphasizes Metaās focus on establishing an American standard in open-source AI.
News: Source
Predictions for AIās Future
Yann LeCun, Metaās chief AI scientist, predicts a significant AI revolution within the next five years. He emphasizes the need for breakthroughs that enable AI systems to understand and interact with the physical world, which is essential for developing domestic robots and fully autonomous vehicles. LeCun notes that while current AI excels at language manipulation, it lacks comprehension of the physical environment.
News: Source
Editorās Special
- GRPO, the algorithm behind DeepSeekās success: Click here
- Gaussian Splatting: Click here
- Understanding Artificial Super Intelligence: Click here
š¤ Join the Conversation: Your thoughts and insights are valuable to us. Share your perspectives, and letās build a community where knowledge and ideas flow freely. Follow us on Twitter and LinkedIn at RealAIGuys and AIGuysEditor.
Thank you for being part of the AIGuys community. Together, weāre not just observing the AI revolution; weāre part of it. Until next time, keep pushing the boundaries of whatās possible. šš
Your AIGuys Digest Team