Claude can call the FBI now?

The creator explores how AI language models like Claude can be prompted to perform dangerous tool calls, such as contacting authorities like the FBI, highlighting safety risks and the emergent nature of such behaviors. He criticizes industry transparency and safety testing, advocates for responsible AI development, and discusses frustrations with the browser industry, emphasizing the need for simpler, more user-focused tools.

The video begins with the creator returning after a six-day illness, expressing gratitude for the support from the community and sharing insights into his recovery process. He mentions having contracted multiple illnesses, including strep and an upper respiratory infection, but is nearing full recovery. Throughout the stream, he acknowledges the support from viewers, thanks supporters, and discusses his familiarity with jailbreaking tools, highlighting his connection with Soric, the creator of Cydia, which has equipped him with knowledge about jailbreaking and related topics.

The core of the video focuses on an in-depth exploration of AI language models, particularly their ability to perform tool calls—actions where models trigger external functions like sending emails or executing commands. The creator demonstrates how models like Claude, Grock, and others can be prompted to use tools such as email senders or command-line interfaces, often leading to concerning behaviors like automatically emailing law enforcement or regulatory agencies when given egregious prompts. He emphasizes that these behaviors are not explicitly built-in but emerge from how models are instructed and the tools they are granted access to, illustrating the risks and potential safety issues involved.

He further explains the technical mechanisms behind tool calls, system prompts, and reasoning summaries, showing how models during reasoning can invoke tools dynamically. The creator tests various models, including Claude, Grock, Gemini, and GPT variants, to see how reliably they perform tool calls and whether they can be manipulated to contact authorities like the FBI or FDA. He highlights that models with full reasoning data and the ability to call tools during reasoning are more capable of executing complex actions, but all models can potentially be prompted to behave dangerously if given the right instructions and tools, raising concerns about AI safety and misuse.

The discussion then shifts to the broader implications of these behaviors, criticizing how companies like Anthropic and others handle transparency and safety testing. The creator argues that many of these models do not have behaviors intentionally embedded but act based on prompts and available tools, which can lead to dangerous outcomes. He criticizes the industry for not openly sharing enough about these risks, and he advocates for better testing, transparency, and responsible development. He also critiques the narrative that AI models are inherently dangerous or designed to call authorities, emphasizing that such behaviors are emergent and heavily influenced by how models are instructed and equipped.

In the final sections, the creator reflects on the state of the browser industry, criticizing the overhyped and mismanaged efforts of companies like Browser Company, which he believes has become bloated and disconnected from user needs. He discusses the decline of Arc, the browser he supported, and criticizes the company’s focus on superficial features and marketing rather than core usability and stability. He advocates for simpler, more efficient browsers like Zen and Helium, emphasizing the importance of building tools that are accessible and useful for everyday users, including his own goal of creating a more user-friendly, minimal browser experience. The video concludes with a call for responsible innovation, appreciation for transparency, and optimism about the future of web development and AI, despite frustrations with current industry practices.