DeepSeek Begins Limited Testing of New Vision Mode
  • News
  • Asia

DeepSeek Begins Limited Testing of New Vision Mode

Select users can now access the new feature on both web and app versions, marking its multimodal debut.

5/1/2026
Ali Abounasr El Alaoui
Back to News

AI company DeepSeek has initiated a significant expansion of its platform by launching a limited test for its first multimodal vision feature. Starting April 29, select users gained access to the new capability, which allows the AI to interpret and analyze images. This strategic move marks the company's entry into the competitive multimodal AI arena, moving beyond its established text-only models.


A Glimpse into New Visual Capabilities

The new "vision mode" is now visible to a select group of users on both the web and mobile app versions of the platform, appearing alongside the existing Fast and Expert modes. Early user reports and screenshots demonstrate the model's ability to accurately process visual information when an image is uploaded. For instance, the AI can precisely identify and describe various elements within a picture, including individuals, backgrounds, specific actions, and colors.

Strategic Rollout and Social Media Teasers

DeepSeek has opted for a quiet, phased rollout, with the feature currently in a grayscale testing stage without a formal public announcement. The company has not yet released any official technical reports, model weights, or updates on its official channels like its website or GitHub. This controlled launch allows the company to gather user feedback and refine the technology before a wider, more public release.

Anticipation for the launch was subtly built through social media posts from Chen Xiaokang, the company's core multimodal technology lead. He shared a cryptic image on the platform X with the caption "Now we see you," confirming the feature's availability for some users in the comments. This followed a now-deleted post from the previous day with a similar image captioned "Soon we will be able to see you," hinting at the imminent release.

Expanding Beyond Text-Based AI

This development represents a pivotal evolution for DeepSeek, whose models were previously limited to processing and generating text. The inability to understand visual data from screenshots, documents, or photographs was a significant constraint that the new vision mode directly addresses. By integrating multimodal functionality, the company is substantially enhancing the platform's utility and creating more versatile and intuitive user interactions.

Industry observers speculate that the initial application of DeepSeek's multimodal technology will focus on analytical tasks rather than generative ones. The most likely use cases include advanced image comprehension, sophisticated optical character recognition (OCR), and the analysis of complex documents and charts. For now, capabilities such as image or video generation are not expected to be part of the company's short-term roadmap.


DeepSeek's cautious yet deliberate entry into the multimodal space with its new vision mode is a clear signal of its ambition to innovate and compete at a higher level. While the feature remains in limited testing, its initial capabilities demonstrate a significant technological leap forward for the company. The broader tech community now awaits a formal announcement that will provide more details on the model's performance, underlying architecture, and eventual public availability.