This podcast features an interview with Andrew Feldman, co-founder and CEO of Cerebras, a company challenging Nvidia in the AI inference market. The conversation covers Cerebras's innovative wafer-scale architecture, its advantages over GPU-based systems in terms of speed and efficiency, and the future of AI, including the potential shift away from transformer models. Feldman highlights the inefficiency of current AI algorithms and the vast potential for improvement, emphasizing the importance of speed in interactive AI applications. A key takeaway is that the cost of inference will decrease due to more efficient algorithms and hardware, leading to the emergence of new AI applications.
Sign in to continue reading, translating and more.
Continue