DeepSeek-VL, Janus, and JanusFlow
The field of artificial intelligence is rapidly evolving, with significant strides being made in models that can understand and process information from multiple modalities, such as vision and language. DeepSeek AI has emerged as a key player in this domain, releasing a series of influential research papers and open-source models focused on advancing the state-of-the-art in vision-language (VL) understanding and unified multimodal capabilities. This blog post provides an overview of their key contributions, categorizing their work into distinct model families and highlighting the core innovations of each.
Published in March 2024, the paper "DeepSeek-VL: Towards Real-World Vision-Language Understanding" laid the groundwork for DeepSeek AI's contributions to the field. The central theme of DeepSeek-VL was the development of an open-source VL model specifically designed for practical, real-world applications.
A key focus of DeepSeek-VL was the diversity and scalability of its training data*. The researchers aimed to create a model that could effectively understand a wide range of visual and textual information encountered in everyday scenarios, including web screenshots, PDFs, OCR content, charts, and general knowledge. To enhance the model's usability, the team also developed a *taxonomy of real-world user scenarios and used this to create an instruction-tuning dataset. This approach aimed to improve the model's ability to function effectively as a vision-language chatbot in practical settings.
Technically, DeepSeek-VL incorporated a hybrid vision encoder* designed to efficiently process high-resolution images (1024 x 1024) without incurring excessive computational costs. Furthermore, the research emphasized the importance of *preserving strong language abilities during the pretraining process. The pretraining strategy was carefully designed to integrate language model training from the outset, managing the interaction between visual and linguistic information to ensure that the model maintained proficiency in language-centric tasks.
The DeepSeek-VL family included both 1.3 billion and 7 billion parameter models, both of which were made publicly accessible. The paper reported that these models achieved competitive or even state-of-the-art performance on various visual-language benchmarks while also maintaining robust language capabilities. The community engagement on the paper's Hugging Face page indicates interest in the model and its potential applications.
Building upon the success of DeepSeek-VL, DeepSeek AI introduced "DeepSeek-VL2: Mixture-of-Experts Vision-Language Models for Advanced Multimodal Understanding" in December 2024. The primary goal of DeepSeek-VL2 was to significantly improve upon its predecessor by incorporating a Mixture-of-Experts (MoE) architecture.
The MoE approach was primarily applied to the language component of the model, leveraging DeepSeekMoE models with a Multi-head Latent Attention mechanism*. This mechanism is designed to compress the key-value cache into latent vectors, thereby enabling more efficient inference and higher throughput. For the vision component, DeepSeek-VL2 introduced a **dynamic tiling vision encoding strategy**. This innovation was specifically aimed at improving the model's ability to process high-resolution images with varying aspect ratios, a common challenge in real-world visual inputs. The model was also trained on an *improved vision-language dataset, further contributing to its enhanced performance.
DeepSeek-VL2 was released in three variants with different activated parameter sizes: Tiny (1.0B), Small (2.8B), and Base (4.5B). The research demonstrated that these models achieved superior capabilities across a range of multimodal tasks, including visual question answering, optical character recognition, document, table, and chart understanding, and visual grounding. Notably, DeepSeek-VL2 was reported to achieve competitive or state-of-the-art performance with a similar or even fewer number of activated parameters compared to other open-source dense and MoE-based models. Like its predecessor, the code and pretrained models for DeepSeek-VL2 were made publicly accessible, fostering further community research and development. The comments on the paper's Hugging Face page show users engaging with the model and its capabilities.
In parallel to the advancements in the DeepSeek-VL series, DeepSeek AI also introduced the Janus family of models, focusing on the ambitious goal of unified multimodal understanding and generation within a single framework. This family includes Janus, JanusFlow, and Janus-Pro, each building upon the previous work to create more versatile and capable multimodal AI systems.
The first in this series, "Janus: Decoupling Visual Encoding for Unified Multimodal Understanding and Generation," was published in October 2024. The core idea behind Janus was to address the potential conflict arising from using a single visual encoder for both understanding and generation tasks, which often have different requirements in terms of information granularity.
Janus introduced an autoregressive framework* that unified these two functionalities by **decoupling the visual encoding pathways** into separate streams, one optimized for understanding and the other for generation. Despite this decoupling, Janus still utilized a *single, unified transformer architecture for processing the information. This design allowed each component (understanding and generation) to independently select the most suitable encoding methods, enhancing the overall flexibility and performance of the model. Experiments reportedly showed that Janus surpassed previous unified models and matched or even exceeded the performance of task-specific models on various benchmarks. The authors highlighted the simplicity, flexibility, and effectiveness of Janus as key strengths, positioning it as a promising architecture for next-generation unified multimodal models. Community interest in Janus is evident from the comments and model citations on its Hugging Face page.
Published shortly after in November 2024, "JanusFlow: Harmonizing Autoregression and Rectified Flow for Unified Multimodal Understanding and Generation" took a different approach to unified multimodal capabilities. JanusFlow aimed to integrate autoregressive language models with rectified flow*, a state-of-the-art generative modeling technique, within a *minimalist architectural framework.
A key finding of this research was that rectified flow could be effectively trained within a large language model framework without necessitating complex architectural modifications. To further enhance the performance of their unified model, the researchers adopted two key strategies: decoupling the understanding and generation encoders* (similar to the original Janus) and *aligning their representations during unified training. This alignment process helped to ensure that the information learned by the understanding encoder was effectively utilized by the generation encoder. Extensive experiments reportedly demonstrated that JanusFlow achieved comparable or even superior performance to specialized models in their respective domains while significantly outperforming existing unified approaches across standard benchmarks. This work was presented as a significant step towards developing more efficient and versatile vision-language models. The Hugging Face page for JanusFlow shows community engagement and links to the model's code.
The latest evolution in the Janus family, "Janus-Pro: Unified Multimodal Understanding and Generation with Data and Model Scaling," was published in January 2025. Janus-Pro was explicitly presented as an advanced version of the original Janus model.
The significant advancements achieved by Janus-Pro were attributed to three main factors: an optimized training strategy, expanded training data, and scaling to a larger model size*. Through these improvements, Janus-Pro reportedly achieved significant progress in both **multimodal understanding and text-to-image instruction-following capabilities**, while also enhancing the *stability of text-to-image generation. The researchers expressed hope that this work would inspire further exploration in the field, and like the other DeepSeek AI models, the code and models for Janus-Pro were made publicly available. The Hugging Face page for Janus-Pro lists several models citing the paper, indicating community adoption and further development.
Collectively, the research papers from DeepSeek AI showcase a strong and consistent effort to push the boundaries of multimodal AI. Several overarching themes emerge from their work:
The progression from specialized vision-language models like DeepSeek-VL to the more unified and advanced capabilities demonstrated by the Janus family highlights a clear trajectory towards more versatile, efficient, and practically applicable multimodal AI systems. DeepSeek AI's contributions are not only advancing the technical frontiers of the field but also enabling broader access and utilization of these powerful technologies through their open-source initiatives.
DeepSeek-VL, Janus, and JanusFlow
Multimodal AI