New #1 open-source AI, new deepfake tools, image editor beats GPT-4o, free deep researcher

The video highlights a wave of recent AI innovations, including open-source tools like Edgetam for real-time mobile video segmentation, ICEdit and Hydream E1 for advanced image editing via natural language prompts, and Alibaba’s Quen 3 models that outperform leading proprietary AI in reasoning and multitasking. It also features autonomous research tools like Alibaba’s Web Thinker and new creative AI applications, emphasizing the rapid progress and accessibility of cutting-edge AI technology.

The video highlights a surge of impressive AI developments released in a single week, focusing on new open-source tools and advanced models. Among the notable innovations is Edgetam, a highly efficient video segmentation AI that can track and outline objects in videos in real-time on consumer devices like smartphones. Built on an optimized version of SAM 2, Edgetam achieves 16 frames per second on an iPhone 15 Pro Max, making it potentially the first mobile-compatible video segmentation AI. The presenter emphasizes its ease of use, availability via GitHub, and its balance between speed and accuracy compared to other models.

Next, the video introduces ICEdit, a semantic image editor that allows users to modify images using natural language prompts. It can perform complex edits such as changing backgrounds, outfits, styles, or facial expressions, often outperforming well-known tools like Gemini and GPT-4 in quality. ICEdit’s ability to generate high-quality, detailed edits directly from text prompts makes it a powerful tool for creative and professional use. The presenter demonstrates its capabilities through various examples and mentions its availability on Hugging Face Spaces, along with instructions for local installation, inviting viewers to try it out or request tutorials.

The video also covers Hydream E1, another open-source image editing AI that builds on the Hydream model. It excels at editing images with prompts, such as changing hair color or artistic styles, and is claimed to outperform other open-source editors like OmniGen and Magic Brush. The presenter highlights its accessibility via GitHub and Hugging Face, emphasizing its potential for users seeking a free, high-quality image editing solution. The discussion hints at the growing trend of open-source semantic image editors that combine ease of use with impressive results.

A significant portion of the video is dedicated to Alibaba’s recent release of Quen 3, a family of hybrid reasoning models that surpass many leading proprietary models in tasks like math, coding, and reasoning. Quen 3 supports multiple sizes, from small models suitable for mobile devices to large, highly capable ones. It features a toggle for reasoning mode, allowing for step-by-step problem solving or instant answers, and supports 119 languages. Benchmarks show Quen 3 outperforming models like GPT-4 and Gemini in various tests, all while being open-source and cost-effective, making it a groundbreaking development in accessible AI.

Finally, the presenter discusses other notable releases, including Microsoft’s small reasoning models (54 family) and Alibaba’s Web Thinker, an open-source deep research tool capable of autonomously searching, fact-checking, and synthesizing web data into comprehensive reports. These tools demonstrate a shift toward more autonomous, research-oriented AI systems that can handle complex scientific and technical queries. The video concludes with mentions of Sunno’s new music generator, capable of producing expressive songs, and Alibaba’s clothing-swapping AI for videos, showcasing the rapid pace of innovation across different AI domains. The presenter encourages viewers to explore these tools and stay updated through their newsletter.