The podcast explores AI alignment and safety, focusing on ensuring advanced technologies are used responsibly. David Shapiro, an AI safety researcher, discusses his approach to aligning AI with human values through "heuristic imperatives." These imperatives include reducing suffering, increasing prosperity, and increasing understanding in the universe, which serve as linguistic functions rather than mathematical formulas. Shapiro argues against single objective functions, advocating for a framework of multiple objectives to create a self-stabilizing system. He emphasizes that intelligence is a linguistic problem, not solely a math problem, and that AI alignment should consider the broader tech stack beyond individual models. The conversation also touches on the challenges of incentivizing alignment through competition and the importance of global consensus.
Sign in to continue reading, translating and more.
Continue