Offset: 0.0s
Space Play/Pause

ChatGPT's NEW Voice Mode, Group Chats, Shopping Research & More!

The AI landscape is constantly evolving, and this week has seen a flurry of exciting developments from major players. Here’s a breakdown of the latest advancements you can use.

5 min read

The AI landscape is constantly evolving, and this week has seen a flurry of exciting developments from major players. Here’s a breakdown of the latest advancements you can use.

[00:16]

The week started with Claude Opus 4.5 being introduced as “the best model in the world for coding, agents, and computer use.” This is a significant step forward, previewing what’s to come in terms of AI capabilities.

[00:22]

Following this, OpenAI announced the rollout of GPT-5.1 to all ChatGPT users. Described as “smarter, more reliable, and a lot more conversational,” this update promises enhanced user experience.

[00:27]

The announcements continued with Claude Opus 4.5 being officially introduced. This marks a significant advancement, positioning it as a leader in AI development.

[00:32]

ChatGPT has also launched new features, including a voice mode. This allows for a more conversational interaction, with users able to speak their prompts and receive spoken responses. This is a major step towards more natural human-AI interaction.

[00:39]

Transitioning to a more practical topic, the video also touches upon selecting the best TVs for bright rooms. Key criteria include brightness (at least 500 nits, ideally 1000+), anti-glare coatings, backlighting (full-array local dimming), optimal placement, and panel technology (QLED/Mini-LED often outperform OLED in bright conditions). If you provide your budget, screen size, and brand preferences, a curated list of 3-5 fits can be generated.

[01:15]

The new voice mode in ChatGPT is a significant upgrade. Previously, using voice involved a separate interface, but now it’s integrated directly into the chat.

[01:26]

“You hit this button, and you could speak into it.” This allows for a more seamless conversation flow.

[01:47]

The AI can now pick up on what you say once you start talking to it. Earlier, it only processed input after a specific action. This change improves the conversational experience, making it more intuitive.

[02:41]

The AI’s ability to remember past conversations is managed through memory settings, specifically “Reference saved memories” and “Reference chat history.” This allows ChatGPT to learn and adapt based on previous interactions.

[03:04]

Previously, pressing the voice button would lead to a separate interface. Now, the experience is more integrated, allowing for direct conversation within the chat.

[03:46]

The ability to dictate directly into ChatGPT, and have it transcribe and respond, is a significant improvement. This feature makes it easier for users to interact with the AI at the speed of thought.

[04:45]

JetBrains has introduced Matter, an AI-powered software assistant for product teams. This tool aims to streamline the development process by allowing teams to prototype on the real app codebase without breaking existing workflows. It’s designed to assist both designers and developers in bringing their ideas to life through natural language interaction.

[05:15]

The Matter Playground offers a browser-based environment to experiment with AI-powered image editing. It supports features like contour lines, spotlight effects, and bobbleheads, enabling users to create fun and engaging visuals. Users can also upload their own videos to apply these effects.

[05:50]

Matter allows users to take a screenshot of a website and then prompt the AI to match the color scheme. This is a powerful feature for maintaining design consistency across projects.

[06:22]

Matter integrates directly with GitHub, allowing users to create pull requests from the interface. This streamlines the process of sharing code changes and collaborating with team members.

[06:43]

Matter facilitates seamless collaboration between product managers, designers, and developers by providing a shared context for their work. It helps bridge the gap between different roles and ensures everyone is on the same page.

[07:25]

ChatGPT has introduced a “Shopping research” feature. This allows users to input product queries, and the AI will provide relevant alternatives and comparison tables, assisting in making informed purchasing decisions.

[08:13]

The “Shopping research” feature is highly effective, providing unbiased and comprehensive product recommendations, which is particularly useful during major sales events like Black Friday.

[08:55]

ChatGPT has also rolled out “Group chats,” enabling users to collaborate with multiple people within the platform. This feature allows for shared conversations, idea generation, and collective problem-solving.

[09:11]

In a demonstration, a user requested ChatGPT to generate “10 progressively clickbaity youtube thumbnails for a video on new chatgpt features like group chats and the updated voice feature.” The AI successfully generated a variety of creative thumbnails, showcasing its image generation capabilities within a collaborative chat environment.

[10:35]

Black Forest Labs has announced FLUX.2, an open-source model for image generation and editing. This model is lauded for its advanced capabilities, including the ability to edit images from text and multiple references, and to generate photorealistic images with infographics.

[11:36]

The FLUX Playground offers users free credits to experiment with the latest FLUX models, including FLUX.2 and FLUX.1 Kontexte. This provides a hands-on experience with the AI’s image generation and editing features.

[12:02]

FLUX.2 excels in various editing tasks such as “blurring faces,” “blurring license plates,” and “pixelating screens.” It can also be used to “track anything,” including motion tails, bounding boxes, and simple masks. The “Just for fun” templates offer creative effects like “clone squad,” “contour lines,” and “bobbleheads.”

[13:03]

The “Clone” effect in FLUX.2 is particularly impressive, allowing for the seamless replication of objects within a video. This feature is demonstrated by cloning a wakeboarder multiple times in a single video clip.

[14:00]

OpenAI has also released GPT-5.1-Codex-Max, a new frontier agentic coding model. This model is an update to their foundational reasoning model and is trained for specific tasks in software engineering, math, and research. It’s recommended for “agentic coding tasks” and offers advanced capabilities in Excel.

[14:21]

Claude for Excel is now available in beta as a research preview. This add-on allows users to leverage Claude’s capabilities directly within Excel, assisting with formula explanations and workbook analysis.

This week has showcased significant progress in AI, from enhanced conversational abilities and collaborative features to powerful image generation and specialized coding models. These advancements promise to further revolutionize how we interact with technology and create content.