In this episode of "The Deep Dive," the hosts explore OpenAI's release of new OpenWeight LLMs, GPT-AOLF120B and GPT OSS 20B, marking a shift towards open-source accessibility. They discuss the evolution of these models from GPT-2, highlighting architectural changes like the adoption of ROPE, SWIJ-LU, MOE, and GQA, which enhance efficiency and performance. The hosts compare GPT-OSS with QUIN-3, analyzing trade-offs between model depth and width, expert configurations, and the impact of MXFP4 optimization on local deployment. They also touch on licensing, training specifics, reasoning control features, and potential for tool use, ultimately emphasizing the democratization of AI and its impact on innovation.
Sign in to continue reading, translating and more.
Continue