Skip to main content

RAG in 2025

Research Paper on MoA (Mixture of Agents)

Despite one-year setback...

MoA is All You Need: Building LLM Research Team using Mixture of Agents


My first attempt at using NotebookLM to create the podcast from research papers.

YouTube url on the research paper.


Currently looking for ways to improve my consumption of the relevant research papers.

Anyone know of a good platform that can turn the .wav files to realistic video?



Comments

Popular posts from this blog

2024 Progress...

My team has made considerable advancements in applying various emerging technologies for IMG (Investment Management Group). Predictive Models We have transitioned from conventional methods and refined our approach to using alternative data to more accurately predict the CPI numbers. Our initial approach has not changed by using 2 models (top-down & bottoms-up) for this prediction.   So far we have outperformed both our larger internal team and major banks and dealers in accurately predicting the inflation numbers. Overall roughly 80% accuracy with the last 3 month prediction to be right on the spot.  We have also developed predictive analytics for forecasting prepayment on mortgage-backed securities and predicting macroeconomic regime shifts. Mixed Integer Programming  / Optimization Another area of focus is on numerical optimization to construct a comprehensive portfolio of fixed-income securities for our ETFs and Mutual Funds. This task presents ...

Gemma 3 - Quick Summary & Why this matters

Introduction Despite being labeled the laggard in the language model race behind OpenAI and Anthropic, Google holds two decisive advantages in 2025's evolving AI landscape: unparalleled high-quality data reserves and compute infrastructure that dwarfs even Meta's formidable 600,000 H100 GPUs. As pre-training scaling laws plateau, these assets become critical differentiators. This is especially important in 2025 when everyone is looking for the killer application that can legitimize the research on language models. Combined with DeepMind's elite research talent and visionary leadership, Google possesses a power that competitors ignore at their peril. Gemma is a family of open-weight large language models (LLMs) developed by Google DeepMind and other teams at Google, leveraging the research and technology behind the Gemini models. Released starting in February 2024, Gemma aims to provide state-of-the-art performance in lightweight formats, making advanced AI accessible f...

LLM as an Operating System ?

LLM as an Operating System? Since 2023, researchers have been exploring the concept of LLMs functioning as operating system. This analogy makes intutive sense when we consider how traditional operating systems serve as intermediaries between users and computer resources. I remember encountering visualization that mapped out this transformation. In the traditional OS model, we have layers like the kernel, system calls, and user interface sitting between hardware and applications. With LLM as an OS, we can reimagine these layers, positoning language models and agentic components as the new intermediaries between user and their digital resources - whether that's a data repositories, computational tools, or planning system.  What makes this vision particularly compelling is the role of multimodal interfaces in this "compressed intellgience". Voice and vision capabilities fundamentally reshape how humans interact with thiis "cognitive OS". Instead of typing comm...