Skip to main content

Iterated Distillation and Amplification

 This is a quick summary of an interesting paper I read today. Supervising  strong learners by amplifying weak experts

Iterated Distillation and Amplification (IDA) is a proposed scheme for training machine learning systems that can be robustly aligned with complex human values. The approach draws inspiration from AlphaGoZero's training methodology and is notably similar to expert iteration.

The core concept involves two key processes: amplification and distillation. In the amplification phase, a learned model serves as a subroutine in a more powerful decision-making process, similar to how AlphaGoZero uses Monte Carlo Tree Search (MCTS) to improve upon its policy network's choices. The distillation phase then involves training the model to directly predict the results of this amplified process, effectively compressing the improved capabilities into a faster system.

IDA aims to address AI safety problems by creating a powerful AI that never intentionally optimizes for something harmful to humans and remains correctable after deployment. Rather than proposing a specific implementation, it presents a design framework where capabilities are safely scaled up through iteration: a safe but slow method of amplification is distilled into a faster but slightly weaker AI, which can then be amplified again, with this process repeating until a sufficiently capable system is developed.

A key aspect of IDA is its use of an "overseer" (typically a human) who guides the process. The goal is to produce an agent that does what the overseer would want it to do, with the definition of "what the overseer would want" being determined through repeated application of the amplification procedure.



Comments

Popular posts from this blog

2024 Progress...

My team has made considerable advancements in applying various emerging technologies for IMG (Investment Management Group). Predictive Models We have transitioned from conventional methods and refined our approach to using alternative data to more accurately predict the CPI numbers. Our initial approach has not changed by using 2 models (top-down & bottoms-up) for this prediction.   So far we have outperformed both our larger internal team and major banks and dealers in accurately predicting the inflation numbers. Overall roughly 80% accuracy with the last 3 month prediction to be right on the spot.  We have also developed predictive analytics for forecasting prepayment on mortgage-backed securities and predicting macroeconomic regime shifts. Mixed Integer Programming  / Optimization Another area of focus is on numerical optimization to construct a comprehensive portfolio of fixed-income securities for our ETFs and Mutual Funds. This task presents numer...

What matters?

 What matters? Six things that matter in LLM in July 2024. 1) Scale of the model, number of parameters: Scale with brute force alone won't work. But the scale does matter depending on the overall goal and the purpose of what the LLM is trying to solve.   2) Compute matters: Even more than ever, we need to look at the infrastructures around LLMs. Infrastructure is also one of the main constraints for the near term and strategically provides an advantage to a few Middle East countries. 3) Data, quality & quantity. It remains true that high-quality data with extensive (longer) training is the way. Quantity of the data also matters. 4) Loss function matters: If your loss function isn't sophisticated or incentivizes the "right" thing, you will have limited improvement. 5) Symmetry or architecture: Do you have the correct architecture around your model(s) and data? Inefficient engineering can be costly to the overall performance and output. There are inherent structural...

Research Paper on MoA (Mixture of Agents)

Despite one-year setback... MoA is All You Need: Building LLM Research Team using Mixture of Agents My first attempt at using NotebookLM to create the podcast from research papers. YouTube  url on the research paper. Currently looking for ways to improve my consumption of the relevant research papers. Anyone know of a good platform that can turn the .wav files to realistic video?