This week’s AI news highlights significant advancements including Real Restore for photo repair, Skywork AI’s interactive video worlds, Meta’s brain activity prediction model, and Google’s efficient TurboQuant compression technique, alongside innovative tools for 3D animation, motion generation, and deepfake technology. Additionally, OpenAI discontinued its Sora app to focus on robotics, while Google’s Gemini 3.1 Flash Live offers real-time AI voice, and the Origin F1 humanoid robot demonstrates hyperrealistic emotional expressions, marking progress across AI research, applications, and human-robot interaction.
This week in AI news has been packed with groundbreaking developments across various domains. One standout is Real Restore, an open-source AI model designed to repair damaged or low-quality photos by removing noise, sharpening details, and restoring colors, rivaling top closed-source editors. In video generation, Skywork AI’s Matrix Game 3.0 impresses with its ability to create interactive, real-time 720p video worlds that maintain long-term memory and consistency, while Da Vinci Maggie Human offers a 15 billion parameter model that generates video with synchronized audio in multiple languages, outperforming existing open models like LTX 2.3. Complementing these, Prism Audio generates realistic, perfectly timed sound effects for silent videos, making it the best open-source tool for audio-video synchronization.
In 3D animation and brain simulation, Retime GS produces smooth full-body 3D animations from 2D videos by creating continuous time 3D representations, outperforming competitors in consistency and quality. Meta’s Tribe V2 AI predicts human brain activity in response to stimuli, trained on extensive fMRI data, and can generalize to new content and individuals, offering a digital twin of human perception with potential applications in neuroscience and healthcare. Comfy UI’s dynamic VRAM management now allows users with mid-range GPUs to run larger models more efficiently by loading parts of AI models only when needed, significantly reducing memory errors and speeding up generation times.
Several innovative AI tools have also emerged for motion and environment generation. Action Plan enables real-time human motion generation from text prompts, with applications in humanoid robot control. World Agents and a related project demonstrate that 2D image models can be guided to build consistent 3D worlds and reconstruct 3D scenes from AI-generated videos, respectively, pushing the boundaries of 3D content creation without additional training. Lumos X advances deepfake technology by generating consistent multi-person videos using relational attention mechanisms, while Real Master transforms plastic-looking game footage into photorealistic videos without altering original geometry, useful for autonomous driving training.
On the AI research front, OpenAI announced the shutdown of its Sora video generation app to focus resources on robotics and real-world physical tasks, citing high costs and safety concerns. Google introduced TurboQuant, a technique that compresses large AI models up to six times while maintaining performance, enabling faster and more efficient AI deployment. ZAI released GLM 5.1, a fast and cost-effective model excelling in agentic coding, and ARC AGI 3 benchmark revealed that current top AI models struggle with real-time learning and adaptation in new environments, highlighting a key area for future improvement.
Finally, Google’s Gemini 3.1 Flash Live offers a natural, low-latency real-time AI voice capable of fluid conversation and task execution, available via API and AI Studio. Other notable releases include Coher Transcribe, a top-performing open-source transcription tool supporting multiple languages; Logger NVS, which generates new scene views from few images to create 3D environments; Pulse of Motion, which corrects video timing for realistic motion; Mega Flow, an optical flow model tracking fast and chaotic pixel movements; and CUA Suite, a massive dataset for training AI to use computers like humans. Additionally, the Origin F1 humanoid robot showcases hyperrealistic facial expressions and emotional responsiveness, signaling advances in AI-driven robotics and human interaction.