Why Your GPUs Only Run at 10%! - CentML CEO Explains

In the video, CentML CEO Gennady discusses the underutilization of GPUs in AI workflows, emphasizing the need for optimization to improve efficiency and cost-effectiveness. He also highlights the rise of open-source models, the importance of benchmarking in AI, and the future potential of modular, distributed systems for transforming application development.

In the video, the CEO of CentML, Gennady, discusses the current state of AI technology, particularly focusing on the challenges and opportunities surrounding GPU utilization and machine learning workflows. He highlights that many GPUs are often underutilized, running at only about 10% capacity due to power limitations and inefficiencies in how workloads are managed. Gennady emphasizes the importance of optimizing machine learning workflows for both training and inference to make them more efficient and cost-effective for users.

Gennady reflects on the organizational structure of successful tech companies like Nvidia, noting that their relatively flat hierarchy allows for quicker decision-making and innovation. He believes that a technical background is beneficial for CEOs in the tech space, as it enables them to understand both the technical and business aspects of their companies. He also discusses the inevitable disconnect that occurs as leaders scale their organizations, suggesting that while technical expertise is valuable, it can be challenging to maintain as companies grow.

The conversation shifts to the topic of open-source models versus proprietary ones, with Gennady asserting that open-source models are rapidly improving and closing the performance gap with proprietary models. He argues that having access to high-quality open-source models allows enterprises to fine-tune and build their own intellectual property without relying solely on large tech companies. This democratization of AI technology is seen as crucial for fostering innovation and enabling a wider range of organizations to leverage AI effectively.

Gennady also addresses the importance of benchmarking in AI through initiatives like MLPerf and MLCommons, which aim to provide standardized measures of performance across different models and hardware. He explains that these benchmarks help to level the playing field, allowing companies to validate their claims and ensuring that performance metrics are reliable. This transparency is essential for fostering trust in AI technologies and encouraging broader adoption across industries.

Finally, Gennady discusses the future of AI and the potential for agents and distributed systems to transform how applications are built and deployed. He believes that while there are challenges in managing the complexity of these systems, the ability to create modular, asynchronous components will lead to more efficient and scalable solutions. As AI continues to evolve, Gennady emphasizes the need for collaboration between academia and industry to drive innovation and address the ethical implications of this powerful technology.