Microsoft Copilot’s New Features: Screen Reading, Deep Thinking, and Voice Interaction

Discover how Microsoft's latest Copilot update transforms AI-powered productivity. With features like screen reading, advanced reasoning, and real-time voice interaction, Copilot becomes your intelligent companion for every task.

Mendy Berrebi
By Mendy Berrebi
6 Min Read

On October 1, 2024, Microsoft announced a significant update for Copilot, introducing features that make it more than a traditional AI assistant. These new capabilities allow Copilot to read your screen, think deeply, and speak to you. This update follows OpenAI’s Advanced Voice Mode, which was launched on September 24, 2024. Let’s explore how these new features reshape the way users interact with their AI assistants.

Copilot Vision: A Visual Assistant at Your Fingertips

One of the most exciting additions to Copilot is Copilot Vision, which lets the AI assistant see and analyze what’s on your screen, especially when using Microsoft Edge. This feature enables Copilot to read text, recognize images, and suggest next steps based on what it sees, allowing users to get answers or assistance without disrupting their workflow.

By simply typing @copilot into the Edge address bar, users can activate Copilot Vision. This feature ensures user privacyno data is stored or used for training once a session ends. Initially, Copilot Vision is available for Copilot Labs users in the United States and works only on pre-approved websites, prioritizing safety and privacy.

Enhancing User Experience with Copilot Vision

Copilot Vision elevates how users interact with content on websites. For example, if you’re shopping for furniture, Copilot Vision can help you compare colors, recommend matching items, and suggest layout options. This feature aims to assist with real-time decision-making, offering suggestions directly based on the page’s content.

Unlike similar tools from Google, Microsoft emphasizes that Copilot Vision focuses on privacy and does not interact with sensitive or paywalled content, preventing potential legal challenges that other AI tools have faced.

Think Deeper: Advanced Problem-Solving with AI

Microsoft also introduced Think Deeper, a feature designed to handle complex queries. Think Deeper enhances Copilot’s reasoning abilities, allowing it to analyze and compare options in detail, providing step-by-step answers for more challenging problems.

For example, if you’re deciding between moving to two different cities, Think Deeper can break down all the factors to consider. The technology behind this feature comes from the latest OpenAI models, fine-tuned specifically for Copilot’s capabilities.

Think Deeper is currently available only to Copilot Labs users, with plans for a broader rollout soon. This feature is ideal for users who need comprehensive, thoughtful answers rather than quick responses.

Copilot Voice: Natural Conversations with AI

Alongside visual and reasoning improvements, Copilot Voice was also introduced, allowing users to engage with their assistant through spoken conversations. This feature enables Copilot to speak aloud its responses, making the interaction more dynamic. Available in English-speaking countries like the U.S., U.K., and Canada, Copilot Voice recognizes tone and responds accordingly.

Whether you’re brainstorming on the go or simply want to hear the latest updates, Copilot Voice makes interactions feel more personal and intuitive. This feature is part of Microsoft’s broader goal to make AI interactions as natural as possible.

How Copilot Voice Compares to OpenAI’s Advanced Voice Mode

The release of Copilot Voice comes shortly after OpenAI’s Advanced Voice Mode. While both systems offer advanced voice synthesis, Copilot Voice integrates more deeply with Windows, enabling users to interact not just with the AI but also with their desktop environment. Copilot Voice also features four synthetic voices, allowing for personalized experiences.

One important note: Copilot Voice has usage limits. Users with Copilot Pro accounts have more time, but availability can vary based on demand. This feature is particularly useful for people who prefer voice-driven interactions.

What’s Next for Copilot?

The new features like Copilot Vision, Think Deeper, and Copilot Voice mark a significant step forward in making AI more interactive and user-friendly. These updates reflect Microsoft’s goal of turning Copilot into a true digital companion that goes beyond basic assistance, helping users navigate complex tasks and make informed decisions.

Microsoft is taking a user-first approach, gathering feedback through Copilot Labs to continuously improve the product while keeping privacy and security at the forefront. As these features evolve, they promise to further enhance both professional productivity and personal decision-making.

Conclusion: A New Era of AI Interaction

Microsoft’s latest update to Copilot is a game-changer. By adding the ability to read your screen, think deeply, and speak aloud, the AI becomes more than just an assistant—it transforms into a trusted companion for both work and personal life.

These features are only the beginning. With continued feedback and refinement, Copilot will continue to evolve, helping users tackle everything from day-to-day tasks to significant life decisions.


Have you tried these new features? Let us know your thoughts in the comments!

Share This Article
Follow:
Hi, I’m Mendy BERREBI, a seasoned e-commerce director and AI expert with over 15 years of experience. My passion lies in driving innovation and harnessing the power of artificial intelligence to transform the way businesses operate. I specialize in helping e-commerce companies seamlessly integrate AI into their processes, unlocking new levels of efficiency and performance. Join me on this blog as we explore the future of digital transformation and how AI can elevate your business to new heights. Welcome aboard!
Leave a comment

Leave a Reply