HyperCLOVA X 8B Omni

Reading time: 38 minute
...

šŸ“ Original Paper Info

- Title: HyperCLOVA X 8B Omni
- ArXiv ID: 2601.01792
- Date: 2026-01-05
- Authors: NAVER Cloud HyperCLOVA X Team

šŸ“ Abstract

In this report, we present HyperCLOVA X 8B Omni, the first any-to-any omnimodal model in the HyperCLOVA X family that supports text, audio, and vision as both inputs and outputs. By consolidating multimodal understanding and generation into a single model rather than separate modality-specific pipelines, HyperCLOVA X 8B Omni serves as an 8B-scale omni-pathfinding point toward practical any-to-any omni assistants. At a high level, the model unifies modalities through a shared next-token prediction interface over an interleaved multimodal sequence, while vision and audio encoders inject continuous embeddings for fine-grained understanding and grounding. Empirical evaluations demonstrate competitive performance against comparably sized models across diverse input-output combinations spanning text, audio, and vision, in both Korean and English. We anticipate that the open-weight release of HyperCLOVA X 8B Omni will support a wide range of research and deployment scenarios.

šŸ’” Summary & Analysis

1. **Multimodal Integration**: HyperCLOVA X 8B Omni can process various modalities like text, images, and audio within a single model. This is akin to someone who excels at writing, drawing, and singing all at once.
  1. Interleaved Multimodal Sequence Processing: OMNI processes tokens from different modalities in an interleaved manner using a single Transformer backbone. It’s similar to mixing various ingredients in one pot for cooking.

  2. Efficient Learning: The model efficiently learns across multiple modalities with competitive performance, particularly bilingual capabilities in Korean and English. This is like someone who can fluently write in both Korean and English simultaneously.

šŸ“„ Full Paper Content (ArXiv Source)

# Introduction
style="width:100.0%" />
Comparison of multimodal capabilities across text, vision, and audio for both generation and understanding tasks. The results highlight the unified end-to-end design of HyperCLOVA X 8B Omni, which supports any-to-any multimodal understanding and generation within a single model.

The tight integration of AI systems into real-world contexts necessitates their ability to understand and generate across multiple modalities, such as text, audio, and vision. This requirement arises in part because specific applications inherently involve multimodal inputs and outputs. Moreover, human-generated text is projected to accumulate at a rate that cannot keep up with the rapid scaling of large language modelsĀ (LLMs;). Even if it were the case, text alone cannot capture the full spectrum of multimodal dimensions of realityĀ .

One strategy for developing multimodal models extends existing LLMs by sequentially incorporating encoders and decoders for various modalities. While such modality extension enables a cost- and time-efficient transformation of a text-based model into a multimodal one, multimodal training often incurs catastrophic forgetting of knowledge within the LLM backboneĀ . This challenge calls for a joint training across multiple modalities in a unified framework.

In response, we introduce HyperCLOVA X 8B Omni(OMNI), an omnimodal model that supports text, audio, and vision modalities as both inputs and outputs, as shown in FigureĀ 1. OMNI is a decoder-only Transformer jointly modeling an interleaved multimodal sequence of tokens and embeddings. Modality-specific tokens and embeddings share a common next-token prediction interface, thereby facilitating semantic composition across modalities.

We compare the performance of OMNI against that of comparably sized models on benchmarks spanning diverse combinations of input and output modalities, including text-to-text, vision-to-text, text-to-vision, speech-to-text, audio-to-text, and speech-to-speech. In addition, we present a human preference study on text-to-speech conversion. For most modality combinations, evaluations are carried out in both Korean and English to assess bilingual ability. The results demonstrate the competitive performance of OMNI across the board, despite it being the only model capable of handling all combinations of input and output modalities.

OMNI is released as an open-weight model under a custom license that permits commercial use subject to specified conditions. Given its compact size and competitive performance across diverse input and output modalities, we present OMNI as a valuable resource for academic and industry partners in both the Korean and global research community.

HyperCLOVA X 8B Omni

style="width:100.0%" />
Overall architecture of HyperCLOVA X 8B Omni. Text, vision, and audio inputs are encoded into continuous embeddings and discrete tokens via modality-specific encoders and tokenizers, which are interleaved and jointly processed by a single decoder-only Transformer backbone. Modality-specific decoders reconstruct visual and auditory outputs from the shared sequence representations, enabling end-to-end any-to-any multimodal generation.

Design Motivation and Pathfinding

Recent multimodal systems span a wide design space, ranging from late-fusion integration to modality-specific generation pipelinesĀ . In our approach, the guiding hypothesis is that multimodal capabilities can be effectively realized when modality-specific tokens and embeddings share a common next-token prediction interface, enabling semantic composition across modalities. As illustrated in FigureĀ 2, text is represented as discrete tokens, while vision and audio are represented with both discrete tokens and continuous embeddings; these representations are interleaved and jointly processed by a single decoder-only Transformer backbone.

We instantiate the backbone as a 36-layer auto-regressive Transformer with a hidden size of 4,096, closely following the architectural and implementation choices of HyperCLOVA X 32B Think(THINK,Ā ). Following THINK, the text tokenization pipeline combines a morphology-preserving pretokenizer and a subword tokenizer and applies low-probability StoChasTok to mitigate token-boundary bias while preserving token efficiency. For subword tokenization, we adapt an English-centric tokenizer via a three-stage vocabulary modification, which significantly improves Korean token efficiency without degrading performance on English, code, or math tasks.

Operationally, we unify multimodal generation by treating each modality tokenizer’s discrete codebook entries as additional vocabulary items of the language model, thereby extending next-token prediction from text to a shared multimodal token space. For understanding and fine-grained grounding, we additionally attach modality encoders that produce continuous embeddings projected into the backbone embedding space. Modality decoders subsequently convert predicted non-text tokens into their native signal domains (pixels and waveforms).

The following sections provide detailed specifications of the vision and audio tokenizers/encoders and the associated decoders.

Vision Modality

OMNI processes visual information through a synergistic integration of three components: a continuous vision encoder for perceptual understanding, a discrete semantic tokenizer for generative representation, and a diffusion-based decoder for pixel synthesis. This tripartite architecture is designed to natively handle interleaved multimodal sequences within a unified framework, in which each component plays a functional role.

First, a continuous vision encoder extracts dense features that are directly aligned with the LLM backbone to support overall vision understanding. Second, to support vision generation, OMNI incorporates a vision tokenizer that quantizes visual features into discrete semantic tokens. This choice is closely tied to the auto-regressive (AR) nature of our Transformer backbone, which is inherently well-suited to modeling discrete tokens . Unlike models such as Janus-Pro or Emu 3 that rely on low-level VAE-style tokenizers, our tokenizer operates at the semantic level to maximize cross-modal synergy with text embeddings —a critical advantage for our compact 8B backbone, where efficient and semantically aligned features are essential.

Finally, vision generation proceeds by decoding these discrete tokens into pixels using a diffusion-based vision decoder. Because semantic tokenization introduces unavoidable information loss by discarding fine-grained visual details, the diffusion model acts as a complementary component that stochastically recovers missing details. It synthesizes high-frequency textures and fine structures through a channel-concatenation-based architecture, which enables significantly faster convergence and supports near-native aspect ratios.

Encoder.

Architecturally, the vision-understanding component of OMNI follows THINK, adopting the Vision Transformer (ViT) architecture from Qwen2.5-VLĀ  for unified image and video modeling. For architectural stability, we utilize a streamlined linear adapter to align visual features with the LLM backboneĀ . A primary design objective is computational efficiency; by optimizing the visual token allocation, we reduce training costs by approximately 53% in GPU-hours compared to standard settings. Static images and 120-frame videos are compressed into efficient budgets of 3K and 11K tokens, respectively. Notably, the encoder remains unfrozen throughout training to establish Korean-centric multimodal capabilities, essential for internalizing Korean-specific visual contexts, cultural landmarks, and high-density OCR.

Tokenizer.

We reuse a pretrained text-aligned tokenizer, TA-TokĀ , and keep it fully frozen during training. TA-Tok fine-tunes SigLIPĀ 2Ā  to quantize its output—patch-wise visual features—into discrete tokens and reconstruct the original visual features from these tokens. One practical limitation of TA-Tok is its fixed input resolution of 384Ɨ384. While the loss of resolution is largely compensated by the diffusion-based vision decoder, non-square images must be resized, which may introduce geometric distortion. We empirically evaluated this issue in advance and found that it does not lead to severe degradation in practice (see Figure 3). The degradation is further mitigated by training our own decoder from scratch, in which such resizing scheme is directly integrated.

(Left) Reconstruction test of TA-TokĀ  using its accompanying decoder. The reconstruction is imperfect due to unavoidable information loss from semantic abstraction and quantization (see the eye and the feather pattern of the bird, and the tonal difference in both cases). A non-square image shown at the bottom is ā€œresized to square → tokenized → decoded as a square image → resized back to the original aspect ratio.ā€ We observe that the distortion level is not critical and presume that it could be compensated for by training a new decoder reflecting this process. (Right) Convergence of validation loss for the classic attention-based architecture (green) and our channel-concatenation-based architecture (blue).

Decoder.

Our vision decoder is similar to the decoders released alongside the TA-Tok modelĀ , but it differs in two key aspects. First, it uses a channel-concatenation-based conditioning architecture that enables significantly faster convergence than attention-based. Second, it supports near-native aspect ratios, avoiding the strict square-image constraint imposed by TA-Tok decoders.

The model adopts a diffusion transformer composed exclusively of single-stream blocks of MMDiTĀ , 2B parameters in total. It operates on the latent space of FLUX.1 VAE Ā  with patch size 1. Importantly, our model does not use any text-conditioning; the only conditioning signal is vision tokens, which are injected via channel-wise concatenation with the noisy latents. Concretely, discrete vision tokens produced by TA-Tok have a fixed spatial resolution of 27Ɨ27; these tokens are first reconstructed into continuous feature vectors and then resized to match the shape of the latents (e.g., 116Ɨ78 for a 928Ɨ624 image) before concatenation. We empirically observe that this design significantly improves the convergence speed (see FigureĀ 3). Moreover, by avoiding attention–based conditioning, the overall computational cost of the model is reduced significantly. Detailed descriptions of the decoder model training and inference are provided in AppendixĀ 7.

Audio Modality

OMNI is designed to support both audio understanding and generation within a unified language modeling framework. The audio module consists of a continuous audio encoder, a discrete audio tokenizer, and a neural audio decoder. Continuous acoustic embeddings and discrete audio tokens are provided as separate input streams to the language model, enabling joint processing of audio and text within a single Transformer backbone. For speech synthesis, discrete audio tokens predicted by the language model are passed to the audio decoder, which reconstructs the time-domain waveform.

Encoder.

For continuous audio representation, we adopt a pretrained audio encoder , which is initialized from the Whisper-large-v3 model . The input audio is resampled to 16 kHz and transformed into a 128-channel log-mel spectrogram using a 25 ms window size and a 10 ms hop size. A pooling layer with a stride of two is applied to reduce the temporal resolution, such that each output frame approximately corresponds to a 40 ms segment of the original audio. As a result, the encoder produces continuous audio embeddings at an effective frame rate of 25 Hz. Subsequently, the encoder outputs are mapped to the dimension of the language model embeddings via a two-layer MLP adapter consisting of a Linear-GELU-Linear structure. To handle audio within video sequences efficiently, we implement an additional token compression mechanism following . Specifically, we incorporate a single-layer MambaMia moduleĀ  after the MLP adapter to further downsample the audio representations from 25 Hz to 1 Hz. This architectural refinement significantly enhances token efficiency, allowing the model to process long-form video-interleaved audio while maintaining a manageable context budget. Throughout the training process, the audio encoder remains frozen to fully leverage the robust acoustic representations learned during large-scale pretraining.

Tokenizer.

In addition to continuous embeddings, we employ a pretrained audio tokenizer to represent speech as discrete units. This tokenizer inserts a finite scalar quantization (FSQ) module into the encoder of a pretrained SenseVoice-Large Automatic Speech Recognition (ASR) model . The input speech is first processed by a stack of Transformer blocks to obtain intermediate representations, which are then projected into a low-rank space and quantized using bounded rounding in the FSQ module. The quantized representations are subsequently projected back to the original dimensionality, and discrete audio tokens are obtained by indexing the quantized low-rank vectors in a (2K+1)-ary system. This process yields a codebook of size 6,561 tokens. The resulting audio tokens are generated at a fixed rate of 25 tokens per second, perfectly aligning with the temporal resolution of the continuous audio embeddings.

This dual-encoding design allows the model to exploit the complementary advantages of both representations. Continuous audio embeddings preserve fine-grained acoustic information and rich prosodic details, while discrete audio tokens provide a compact and generation-friendly representation that is well-suited for autoregressive modeling and waveform synthesis.

Decoder.

To reconstruct time-domain waveforms from discrete audio tokens, we propose an audio decoder named Unit-BigVGAN. The decoder is architecturally derived from BigVGAN-v2 , but is adapted to consume discrete audio tokens generated by the LLM rather than continuous mel-spectrogram features. As the decoder directly operates on symbolic unit sequences, which encode limited speaker identity information, the model conditions the generator on an explicit speaker embedding. A reference speech signal is processed by an ECAPA-TDNN to extract a fixed-dimensional speaker embedding that captures speaker-specific characteristics. The embedded discrete tokens are concatenated with the speaker embedding along the channel dimension and used as input to the generator.

Given this combined representation, the decoder follows a BigVGAN-style upsampling and residual processing pipeline to progressively increase temporal resolution and generate the final waveform. The generator consists of an initial convolution layer followed by multiple upsampling stages with residual dilated convolution blocks. Within these residual blocks, a filtered Snake nonlinearity is applied to obtain an anti-aliased representation of discrete-time one-dimensional signals, an approach known as anti-aliased multi-periodicity composition (AMP). Under this decoding pipeline, each discrete token represents a fixed temporal span of 40 ms, corresponding to a token rate of 25 Hz.

Pre‑Training

style="width:100.0%" />
Overview of the training process. The model is first trained using discrete modality tokens for text, vision, and audio, establishing a unified symbolic token interface across modalities. Continuous vision and audio encoders are then integrated and jointly trained alongside the discrete tokens, enabling richer multimodal perception within the same Transformer backbone.

The pre-training consists of two phases to train a text-centric foundation and progressively instill multimodal capabilities. First, the model learns discrete modality tokens for text, vision, and audio, establishing a unified symbolic token interface that enables joint sequence modeling across modalities. Next, continuous modality encoders for vision and audio are integrated and jointly optimized with the existing discrete representations, allowing the model to incorporate rich perceptual signals while preserving a unified token-based processing paradigm.

As shown in FigureĀ 4, the training recipe is designed to progressively build up the omnimodal capabilities of the model. Training begins with text-only pre-training to establish a strong sequence-modeling foundation, followed by omnimodal pre-training that extends the backbone to vision and audio through discrete token learning and the integration of continuous-modality encoders.

Text Pre-training

Data.

The data preparation pipeline follows the scalable preprocessing framework used in THINK, implemented with a hybrid processing stack based on Datatrove and NeMo-Curator. Raw data are collected and normalized into a unified schema, annotated with document-level quality signals (including PII masking), filtered using a combination of heuristic and model-based criteria, expanded with synthetic data, and finally serialized into sharded files for efficient streaming-based training. Data filtering combines heuristic rules and model-based signals to remove low-quality samples while minimizing unnecessary data loss. This consideration is crucial for Korean corpora, where overall data availability is relatively limited. Filtering policies are pre-defined and applied consistently within each training run. Synthetic data generation leverages two complementary approaches: seed-based generation and document rewriting. Reasoning-oriented synthetic data for STEM, code, and mathematics domains are incorporated during mid-training rather than post-training. The proportion of synthetic data is controlled to preserve training stability while maintaining sufficient diversity.

Backbone.

The text backbone is pretrained using a multi-stage curriculum with progressively increasing context lengths of $`4`$K, $`8`$K, and $`32`$K tokens. Later stages place greater emphasis on long-form, high-quality, and reasoning-oriented data, with batch sizes adjusted accordingly to accommodate longer contexts. To improve training efficiency under the limited parameter capacity of the $`8`$B-scale backbone, we employ multi-token predictionĀ . Specifically, an auxiliary prediction head with a single additional layer is introduced and weighted by a scaling factor of 0.2. This design increases supervision density per token while preserving the original next-token prediction objective, resulting in more effective utilization of training signals without altering the primary optimization target.

Training Discrete Modality Tokens

Stage 1: Multimodal Vocabulary Expansion.

In this stage, the discrete codebooks produced by modality-specific tokenizers (vision & audio) are incorporated into the text vocabulary, effectively expanding the model’s token space to support multimodal symbolic representations. To prevent degradation of text capabilities during the initial multimodal expansion, the text-token portions of the token embedding matrix and the LM head, along with the decoder layers, are frozen, whereas the embeddings for the newly introduced non-text modality tokens are trained for alignment. This stage is trained for 36K steps, corresponding to 302B tokens, using primarily image–text and audio–text paired data, with text-only data kept at a minimal ratio. To account for modality token scale differences, the input mixture is controlled with a fixed Image:Audio ratio of 3:1.

Stage 2: Full-Parameter Multimodal Pre-training.

All parameters are made trainable to enable cross-modal fusion and multimodal reasoning. This stage performs large-scale end-to-end multimodal pre-training over 2.3T tokens, with modality ratios and loss masking carefully controlled to mitigate text degradation caused by the large vision token budget. A curriculum-based loss masking strategy is applied to stabilize training. Specifically, during the initial phase spanning the first 1T tokens, the modality mixture is set to Text:Image:Audio = 2:6.5:1.5, with a vision-token loss masking factor of 0.5. For the second phase, spanning from 1T to 2.3T tokens, the same mixture ratio is maintained while vision loss masking is restored to 1.0.

Stage 3: Long-Context Adaptation for Multimodal.

A short and focused long-context adaptation is performed to support downstream vision-interleaved and high-difficulty reasoning data under an extended context. This stage continues from the Stage 2 checkpoint with a 32K context length and a reduced global batch size to improve stability on long sequences, and is trained on approximately 20B tokens.

Integrating Continuous Modality Encoders

In this phase, we integrate continuous modality encoders for both vision and audio to strengthen perceptual modeling and to align continuous and discrete modality representations within a unified sequence modeling framework. While both encoders are incorporated into the architecture, only the vision encoder is actively optimized to enhance visual perception and to align its representations with those produced by the vision tokenizer.

Stage 1: Vision Encoder Alignment.

In the first stage, we align visual features with the language model’s embedding space. We keep both the language model backbone and the vision encoder frozen, and we train only a lightweight linear adapter. The training data predominantly consists of image–caption pairs (75.0%), basic OCR tasks (20.0%), and VQA samples (5.0%), establishing a foundational mapping between visual tokens and linguistic representations.

Stage 2: Vision-Centric Full-Parameter Pre-training.

In the second stage, we train all model parameters to enhance Korean-specific visual perception, including cultural entities, local landmarks, and high-density Korean-script OCR. To preserve previously acquired capabilities across other modalities, we train on a total of 1.5T tokens spanning text (12.1%), visual understanding (38.5%), visual generation (34.4%), and audio (15.0%). The visual understanding data primarily comprises interleaved in-house and public datasets that capture general visual knowledge, along with text-rich OCR data. The visual generation data includes not only text-to-image samples but also image-editing data. Since image editing utilizes both the vision encoder and the vision tokenizer, this stage further promotes representation alignment between the two components.

Stage 3: Audio Encoder Alignment.

In the final step of architectural integration, we incorporate the continuous audio encoder to complete the omnimodal framework. While the preceding stages leveraged discrete tokens to accelerate training, this stage introduces a continuous encoder in parallel to process dense acoustic information. To achieve this, we focus exclusively on ASR tasks, training only a lightweight adapter to bridge the audio encoder and the language model backbone. This stage finalizes the unified architecture, establishing a stable foundation for the subsequent post-training pipeline.

Post-Training

The post-training is designed to transform the pre-trained omnimodal backbone into a Korean-centric AI assistant capable of seamless, instruction-following interaction across text, audio, and vision modalities. Our primary philosophy centers on a staged curriculum that progressively transitions from foundational conversational alignment to complex, intent-aware reasoning. While reinforcement learning is recognized as a potent tool for preference alignment, this work primarily focuses on the architectural and data-driven strengths of a supervised fine-tuning (SFT) framework. By establishing this high-fidelity functional baseline, we provide a stable platform capable of seamless cross-modal reasoning, which we identify as the primary milestone for our omnimodal assistant.

Data Composition and Strategy

The post-training data is meticulously curated to balance general linguistic intelligence with specialized multimodal capabilities. We utilize a mix of high-quality human-annotated dialogues and synthetic reasoning traces designed to enhance logical depth.

A central pillar of our strategy is the prevention of catastrophic forgetting; hence, the curriculum starts with a high concentration of text-based SFT to maintain a stable linguistic foundation before transitioning to complex omnimodal tasks. Specifically, in the initial stage, text-only data constitutes the majority of the training volume ($`50.2\%`$), ensuring that OMNI preserves its core reasoning and multilingual abilities.

For vision and audio modalities, our strategy emphasizes a progressive increase in task-oriented instructions. This phased, omnimodal-aware approach ensures that OMNI handles diverse any-to-any scenarios without degrading its fundamental capabilities. The detailed distribution of these datasets across the four training stages is visualized in FigureĀ 5.

Training Recipe

The SFT process for OMNI is organized into four sequential stages, each targeting specific functional milestones.

style="width:100.0%" />
Distribution of the post-training datasets across four training stages. Stage 1 focuses on foundational conversational alignment; Stage 2 expands to task-oriented multimodal instructions; Stage 3 introduces temporal and long-context understanding; and Stage 4 refines user-intent reasoning through integrated reasoning paths.

Stage 1: Foundational Omni Alignment.

The initial stage is dedicated to establishing foundational instruction-following capabilities by adapting the pre-trained backbone to a conversational framework. As this stage serves as the primary transition point toward a dialogue-centric assistant, it required the most substantial investment of computational resources and training time within the post-training pipeline. Central to this phase is the prioritization of text-based SFT, which constitutes the majority of the training volume ($`50.2\%`$). This high concentration of linguistic data is critical for ensuring that OMNI maintains a stable linguistic foundation and robust reasoning capabilities before the introduction of more complex omnimodal tasks. This core text-only data is complemented by foundational omnimodal tasks, such as image captioning, ASR, Text-to-Speech (TTS), image generation, and image editing tasks. By allocating the largest portion of the total computational budget to this stage, we ensure a reliable alignment between the model’s perceptual outputs and the nuances of human dialogue.

Stage 2: Task-Oriented Omni Specialization.

The second stage represents a pivotal expansion of the model’s functional repertoire, characterized by an exponential increase in both the volume and diversity of instructional tasks. With the linguistic foundation firmly established in Stage 1, the training focus strategically shifts toward large-scale multimodal instruction tuning. During this phase, the proportion of text-only SFT is significantly reduced ($`8.3\%`$) to accommodate a vast ecosystem of task-oriented omnimodal data, with a primary emphasis on complex image understanding. The objective is to cultivate omnimodal synergy by exposing OMNI to heterogeneous scenarios. By navigating these interleaved tasks, the model learns to synthesize cross-modal evidence, enabling it to solve complex queries that require simultaneous processing of text, audio, and vision.

Stage 3: Long-Context and Video SFT.

Stage 3 is primarily dedicated to temporal modeling and long-context management, with a central focus on video understanding. We integrate a substantial volume of video understanding data ($`41.3\%`$) and incorporate instruction samples that feature extensive internal reasoning traces to bolster model’s logical depth. This recipe instills OMNI with the ability to maintain semantic coherence over long multimodal sequences and to perform reasoning over temporal events. To handle the high-resolution audio-visual streams within video efficiently, we introduce a dedicated audio token compressor for video inputs. By preceding this stage with a brief alignment phase trained exclusively on video data, with only the compressor module set as trainable, we ensure stable integration of the compressor module into the omnimodal pipeline.

Stage 4: Intent-Aware Multistep Reasoning.

The final stage of our post-training curriculum is designed to instill OMNI with the capacity for both high-level intent parsing and sustained, multistep logical reasoning. We internalize a structured reasoning mechanism—the <think> block—to serve as the model’s cognitive workspace. In this framework, the reasoning trace typically incorporates an initial intent classification step, where OMNI identifies the task category and orchestrates the necessary modality-specific modules. For high-complexity tasks, such as STEM problem-solving or cross-modal integration, this initial mapping can naturally transition into an extended deductive reasoning process. By systematically breaking down instructions into intermediate logical steps within the latent space, the model can navigate complex problem domains while maintaining strict adherence to user-defined constraints. This versatile reasoning paradigm ensures that OMNI not only selects the correct functional path but also executes deep, context-aware analysis when the task demands it. Examples illustrating the foundational intent parsing and task orchestration are provided in AppendixĀ 9.

Evaluation

Unified Benchmark Results across Text, Vision, and Audio Modalities. For speech-to-text benchmarks, both clean (c) and other (o) splits are used. The symbol ā€˜-’ indicates that the model does not support the corresponding modality or task.
Language Dataset
8B Omni
Omni 7B
8B
7B
7B
Mini 8B
7B Instruct
Flamingo3 7B
Korean KMMLU-pro 64.9 31.1 18.7 16.4 17.7 38.6 23.8 31.8
HAERAE 75.3 51.0 20.0 14.9 21.3 59.9 40.8 46.0
KoBALT 27.7 17.7 10.3 5.9 9.6 17.1 8.9 12.9
Flores+ (En→Ko) 29.2 23.7 0.5 5.7 22.7 19.2 10.5
English MMLU 75.7 71.6 32.4 48.2 40.0 64.9 42.9 58.7
MMLU-pro 54.2 50.5 10.8 20.0 19.3 40.2 17.5 30.8
GSM8K 87.3 87.0 3.0 43.5 61.7 75.3 39.5 62.2
Flores+ (Ko→En) 27.8 28.6 0.9 12.1 28.2 21.6 18.4
Korean KoNET 33.0 14.7 0.6 0.3 11.3
K-MMBench 80.2 76.5 15.7 36.3 38.9
K-DTCBench 78.8 88.8 31.7 29.2 48.3
English SEED-IMG 80.3 77.0 69.0 72.4 74.0
LLaVA-W 93.8 88.5 51.0 78.2 74.2
TextVQA 80.3 84.4 62.9 58.7 77.5
DocVQA 90.7 94.9 74.1 43.4 88.1
English GenEval 0.64 0.39 0.78 0.67
ImgEdit 3.83 1.28 1.30
Korean KsponSpeech-c 28.74 34.96 73.20 54.30
KsponSpeech-o 33.09 36.76 83.92 52.59
Fleurs-ko 15.33 16.23 46.20 36.86
English LibriSpeech-c 1.93 4.13 11.34 3.56 1.41
LibriSpeech-o 4.47 5.67 15.57 6.15 3.02
Fleurs-en 7.00 5.53 15.20 6.42 3.99
English Clotho-v1 0.259 0.051 0.238 0.138 0.296
En → Ko Fleurs-en2ko 24.70 0.00 0.09
Ko → En Fleurs-ko2en 22.91 17.76 14.96

To evaluate the performance of OMNI1, we select a set of open-source multimodal models as baselines, covering text, vision, and audio modalities. The baseline models are chosen based on similarity in model scale, reproducibility of reported evaluation results, and representativeness within each modality. All comparisons are conducted only on the modalities that each model explicitly supports.

Qwen2.5-Omni-7BĀ  is an Omni model that jointly supports text, vision, and audio, and serves as the most directly comparable baseline to OMNI across all modalities. Both models are based on a general-purpose Omni architecture, enabling consistent comparisons on text understanding and generation, vision–language reasoning, and speech recognition and generation tasks. Accordingly, Qwen2.5 Omni is used as the primary comprehensive baseline throughout the evaluation.

Models such as Emu3 8BĀ , Janus-Pro 7BĀ , and X-Omni 7BĀ  are text–vision–centric multimodal models. While they support vision–language understanding and generation, they do not provide audio-related capabilities. Therefore, these models are included as baselines only for text and vision benchmarks.

For audio-related tasks, we include Step-Audio2-Mini-8BĀ  and Qwen2-Audio-7B InstructĀ  models as baselines. These models can be evaluated on all audio benchmarks. In addition, the Audio Flamingo3 7BĀ  is selected due to its reported strengths in English-centric audio and speech-to-text tasks.

Overall, we carefully align each baseline model with the modalities it supports, and apply this principle consistently across both vision and audio evaluations. This evaluation setup allows us to analyze the performance of OMNI as a general-purpose Omni model under realistic and well-defined comparison settings, without relying on unsupported assumptions

Text-only Results

We evaluate the text-to-text performance of OMNI on both Korean and English benchmarks, and the results are reported in TableĀ 1.

Korean Text-to-Text benchmarks.

Across Korean benchmarks, OMNI shows a clear performance advantage over all comparison models. On KMMLU-ProĀ , HAERAE-1.0Ā , and KoBALTĀ , OMNI outperforms other large-scale models by a large margin. These results indicate that OMNI effectively learns Korean-based multi-domain knowledge and reasoning abilities, and that the Korean-focused data composition and training strategy directly contributes to performance gains.

English Text-to-Text benchmarks.

OMNI also achieves strong results on English benchmarks, outperforming comparison models across all evaluation metrics. The MMLU benchmarkĀ  series, including MMLU and MMLU-Pro, evaluates broad English multi-task knowledge and reasoning ability. OMNI consistently achieves top-level performance on all benchmarks in this series. In addition, OMNI attains the highest score on the GSM8KĀ  math reasoning benchmark, showing that numerical reasoning and step-by-step problem-solving abilities are also well learned.

Translation benchmarks.

We evaluate 1-shot translation performance on the Flores+ benchmarkĀ  using the BLEU metric for the English-Korean pair. For the English-to-Korean direction, we apply Ko-Mecab pre-tokenization to the generated text to compute the scores. As shown in TableĀ 1, OMNI achieves the best performance in English-to-Korean translation and delivers performance comparable to the top-performing models in Korean-to-English translation. These results demonstrate that OMNI has strong cross-lingual capabilities between Korean and English compared to other baseline models.

Overall, OMNI demonstrates strong text-to-text performance on both Korean and English benchmarks compared to existing models. Notably, the performance gap is larger on Korean benchmarks than on English benchmarks, which shows that OMNI provides stable and robust text understanding and reasoning performance in the Korean language setting.

Vision & Text Results

We evaluate the visual–language understanding and visual generation capabilities of OMNI using a diverse set of public vision benchmarks. The evaluation covers Vision-to-Text tasks, which include image-based question answering and visual reasoning, as well as Text-to-Vision tasks, which focus on text-conditioned image generation and editing. The quantitative results are summarized in TableĀ 1. Furthermore, we extend our evaluation to the temporal domain by assessing the model’s video understanding performance on both public benchmarks and specialized internal datasets. These assessments, detailed in the following paragraph, underscore OMNI’s capacity for high-fidelity reasoning across the full visual spectrum.

Vision-to-Text Benchmarks.

On Korean Vision-to-Text benchmarks, OMNI achieves the best performance on KoNETĀ  and K-MMBenchĀ , and the second-best performance on K-DTCBenchĀ . KoNET is a multimodal visual–language reasoning benchmark constructed from a broad range of Korean national educational tests, spanning elementary, middle, high school, and college-level curricula. It evaluates not only the integration of visual and textual information but also diverse forms of Korean linguistic knowledge and educational reasoning across subjects and difficulty levels. The large performance margin observed on KoNET suggests that the Korean-focused data composition and training strategy of OMNI are effective for robust visual–language understanding in the Korean educational context.

OMNI also shows strong performance on English Vision-to-Text benchmarks. It achieves the best results on SEED-IMGĀ  and LLaVA-WĀ , and the second-best results on TextVQAĀ  and DocVQAĀ . SEED-IMG and LLaVA-W evaluate general visual question answering and holistic multimodal reasoning, and the strong performance on these benchmarks demonstrates that OMNI effectively aligns visual representations with textual semantics. On benchmarks that emphasize diagram- and document-level understanding, such as AI2D, DocVQA, and ChartQA, OMNI performs slightly below the top model but consistently remains among the top-performing systems, indicating stable and robust visual–language understanding.

Text-to-Vision Benchmarks.

For Text-to-Vision tasks, we evaluate image generation quality and the accuracy of the text-conditioning capability. OMNI achieves the best performance on ImgEditĀ , which focuses on image editing under text constraints, and ranks third on GenEvalĀ , which evaluates general text-to-image generation quality. These results suggest that OMNI is particularly strong in preserving semantic intent while performing localized image edits.

An important observation is that OMNI delivers stable performance across both Vision-to-Text and Text-to-Vision tasks while consistently supporting bidirectional text and vision inputs and outputs. In contrast, Qwen2.5 Omni 7B, which performs well on Vision-to-Text benchmarks, does not support Text-to-Vision generation. Janus-Pro and X-Omni support bidirectional text–vision interaction similar to OMNI, but show noticeably lower performance across benchmarks.

In addition to quantitative evaluations, we provide qualitative examples to illustrate the text-to-vision capabilities of OMNI. FigureĀ 6 demonstrates that the model generates semantically consistent images from prompts expressed in different languages (English and Korean), indicating robust cross-lingual alignment in text-to-image generation. FigureĀ 7 further highlights the model’s ability to incorporate Korean cultural attributes into generated images, reflecting effective grounding in culturally specific visual concepts. Finally, FigureĀ 8 showcases the model’s image editing abilities, including style change, object removal, and background replacement, which qualitatively supports the strong performance observed on the ImgEdit benchmark.

Furthermore, OMNI supports not only text and vision but also audio inputs and outputs within a single unified model. This broader multimodal capability provides an additional advantage beyond visual–language tasks and enables more complex multimodal application scenarios. The evaluation of audio-related tasks is presented in the following SectionĀ 5.3.

Video Benchmarks.

As an omnimodal model built on a unified framework, OMNI natively supports video understanding through its integrated vision–language processing pipeline. We evaluate our model on two distinct benchmarks: Video-MME, a comprehensive multi-modal video evaluation suite, and an internal benchmark assessing NAVER TVĀ 2 comprehensionĀ , designed to assess the comprehension of real-world Korean video contents.

The quantitative results are summarized in TableĀ 2. In evaluations on Video-MMEĀ  (conducted without subtitles), OMNI achieves a score of 58.2. While this is lower than the 64.3 of Qwen2.5 Omni 7B, it remains highly competitive for an 8B-scale model, comparable to GPT-4VĀ  and exceeding LLaVA-NeXTĀ  and LLaVA-OneVision. Furthermore, on the NAVER TV benchmarkĀ , OMNI scores 69.7, which is a substantial improvement over GPT-4V’s 50.0. These results indicate that the model’s joint training on interleaved Korean-centric multimodal data effectively internalizes the complex temporal and cultural nuances required for specialized video understanding applications.

Language Benchmark
8B Omni GPT-4V
Omni 7B
OneVision 7B
NeXT 7B
English Video-MME 58.2 59.9 64.3 57.6 33.8
Korean NAVER TV Content 69.7 50.0

Performance comparison on video understanding benchmarks. Video-MME results are reported without subtitles. NAVER TV Content is an internal benchmark consisting of real-world Korean video content to evaluate temporal and cultural context understanding.

Consistency of generated images from the same semantics, different languages (English and Korean).
Generated images incorporating Korean cultural attributes.
Image editing ability of our model. Input images are marked with green borders. Style change, object removal, background replacement are shown, respectively.
Language
8B Omni
30B-A3B
English 3.94 ($`\pm`$ 0.14) 3.96 ($`\pm`$ 0.15) 4.44 ($`\pm`$ 0.14) 4.11 ($`\pm`$ 0.15) 4.08 ($`\pm`$ 0.14)
Korean 4.22 ($`\pm`$ 0.11) 3.40 ($`\pm`$ 0.12) 4.20 ($`\pm`$ 0.12) 4.05 ($`\pm`$ 0.12) 3.43 ($`\pm`$ 0.12)

Human Evaluation Results on Text-to-Speech Benchmark. A total of 30 participants evaluated 20 samples per model across 5 models, resulting in 100 evaluations per listener. Values in parentheses represent the 95% confidence intervals.

Audio & Text Results

We evaluate the audio-related performance of OMNI across automatic speech recognition, speech translation, audio captioning, and text-to-speech tasks. Quantitative evaluations based on public benchmarks are reported in TableĀ 1, while human evaluations of speech synthesis quality under real-world commercial settings are presented in TableĀ 3.

Speech/Audio-to-Text Benchmarks.

On ASR benchmarks, OMNI shows competitive performance in both English and Korean. On Korean ASR benchmarks, including KsponSpeechĀ  and Fleurs-koĀ , OMNI achieves state-of-the-art word error rates (WER), establishing strong recognition performance for Korean speech. On English datasets such as LibriSpeechĀ  and FleursĀ , OMNI performs slightly below Audio Flamingo3 7B, but remains competitive with other English speech models.

For audio-to-text tasks, OMNI also achieves the second-highest SPIDEr score on the Clotho-v1Ā  audio captioning benchmark, following only Audio Flamingo3 7B, which indicates its ability to effectively summarize acoustic events and semantic information in text form. We attribute part of these gains to the unified multimodal architecture, which supports generalization across ASR and audio-to-text generation tasks.

Speech-to-Speech Benchmarks.

We conduct a Speech-to-Speech (S2S) evaluation using a speech-to-speech translation (S2ST) benchmark to evaluate the model’s ability to directly convert spoken language from a source tongue into spoken language in a target tongue while maintaining semantic integrity and naturalness. Unlike traditional cascaded systems, OMNI aims to streamline this process, thereby reducing latency and potential error propagation. To rigorously assess these capabilities, we conducted cross-lingual translation tasks between English and Korean using a curated dataset of 270 translation pairs for both En $`\rightarrow`$ Ko and Ko $`\rightarrow`$ En directions.

To quantitatively assess the translation accuracy of the generated audio, we utilize the ASR-BLEU metric. This methodology involves transcribing the model’s speech output into text, which is then compared against the ground-truth reference via the BLEU score. To ensure a fair and consistent comparison, we employed gpt-4o-mini-transcribe as the unified ASR engine for all candidate models. BLEU scores were evaluated using sacrebleuĀ . For English-to-Korean translation evaluation, we applied Ko-Mecab pre-tokenization before score computation. Our results demonstrate that OMNI shows superior performance on speech translation tasks; specifically, it achieves the highest performance in both English-to-Korean (En $`\rightarrow`$ Ko) and Korean-to-English (Ko $`\rightarrow`$ En) translation directions among all evaluated models.

Text-to-Speech Human Evaluation.

To evaluate the performance of Text-to-Speech capabilities, we conducted a Mean Opinion Score (MOS) test focusing on the naturalness of the synthesized speech. A total of 30 human listeners participated in the evaluation. The test set comprised 20 distinct utterances, consisting of 10 English sentences and 10 Korean sentences, to assess the model’s proficiency across different linguistic contexts.

The evaluation primarily focused on how closely the synthesized speech resembles human-like pronunciation, intonation, and rhythm. Participants were instructed to rate each audio sample on a 5-point Likert scale, where a score of 1 indicates “Bad: Due to severe artifacts and lack of naturalness, the audio is nearly unintelligible.” and 5 indicates “Excellent: The speech is nearly indistinguishable from a real human voice, with natural pronunciation, intonation, and rhythm.”. Details of the evaluation protocol, participant setup, and scoring criteria are provided in AppendixĀ 8.

TableĀ 3 reports the results of human evaluations (MOS) conducted on an internal dataset, comparing OMNI with widely used commercial text-to-speech systems under real-world service conditions. OMNI achieves competitive MOS scores in both English and Korean in terms of naturalness and pronunciation clarity. In particular, OMNI receives higher scores than comparison models for Korean text-to-speech.

Overall, OMNI achieves balanced performance across a wide range of audio tasks, including speech recognition, audio understanding, speech translation, and text-to-speech. OMNI supports a unified interface in which both inputs and outputs are represented as either audio or text. This design allows the model to be easily extended to multiple audio-related tasks without requiring task-specific model architectures. As a result, heterogeneous tasks such as speech recognition, audio understanding, speech translation, and speech synthesis can be handled in a consistent manner within a single model.

In addition to quantitative evaluations on public benchmarks, human evaluation results conducted under real-world commercial settings show that OMNI achieves quality comparable to existing commercial speech models. In particular, the human evaluation results for text-to-speech indicate that OMNI provides audio quality suitable not only for research settings but also for practical service environments. These results suggest that OMNI, as a general-purpose Omni model, can be effectively applied to a wide range of audio-centric application scenarios.

Conclusion

In this work, we presented HyperCLOVA X 8B Omni, the first omnimodal model in the HyperCLOVA X family that supports text, audio and vision modalities as both inputs and outputs. HyperCLOVA X 8B Omni is trained with a unified autoregressive objective that extends next-token prediction beyond text by incorporating discrete vision and audio codebook entries as additional vocabulary items in interleaved sequences. On top of this symbolic interface, continuous vision/audio encoders inject richer perceptual embeddings projected into the same backbone space, while modality-specific decoders translate the shared sequence representations back to pixels and waveforms, compensating for information lost in semantic tokenization.

Empirical evaluations show that HyperCLOVA X 8B Omni achieves competitive performance against comparably sized models across diverse combinations and input and output modalities: text-to-text, vision-to-text, text-to-vision, speech-to-text, audio-to-text, speech-to-speech, and text-to-speech. We expect that its open-sourcing will benefit researchers and practitioners seeking a compact yet versatile model.

The $`8`$B-scale HyperCLOVA X 8B Omni model serves as the first pathfinding point of the design in which a unified auto-regressive backbone supports both interleaved multimodal understanding and any-to-any generation when paired with modality-specific encoders and decoders. While the performance of HyperCLOVA X 8B Omni is strong relative to its size, we anticipate that increasing its size will yield considerable performance gains. A larger and more advanced variant would be particularly valuable in situations with sufficient computational resources where higher performance is desired. Therefore, scaling up the model represents an important avenue for our future research toward developing a collection of robust omnimodal models that can accommodate the needs and restrictions of diverse scenarios.

Contributions and Acknowledgments

Within each role, names are listed in alphabetical order by first name, followed by the last name.

2

Technical Writing

Cheonbok Park
Dongyoon Han
Geewook Kim
Hwiyeol Jo
Jeonghoon Kim
Jin-Hwa Kim
Jiseob Kim
Joosung Lee
Sangdoo Yun
Sanghyuk Choi
Sungwook Jeon
Taeho Kil
Yoonsik Kim

Model Research and Training

Bado Lee
Cheonbok Park
Daehee Kim
Geewook Kim
Gichang Lee
Hangyeol Yu
Heesu Kim
Hodong Lee
Jeonghoon Kim
Jinbae Im
Jinhyeon Kim
Jiseob Kim
Jungwhan Kim
Ka Yeon Song
Kyeongseok Jeong
Moonbin Yim
Nako Sung
Ohsung Kwon
Sang Hee Park
Sanghyuk Choi
Seongjin Shin
Seunggyu Chang
Soyoon Kim
Suk Min Seo
Taeho Kil
Taehwan Yoo
Yeontaek Oh
Yoonsik Kim

Model Evaluation and Analysis

Daehee Kim
Dongjin Lee
Gayoung Lee
Hagyeong Lee
Hangyeol Yu
Heesu Kim
Hwiyeol Jo
Hyunhoon Jung
Hyunsoo Ha
Jeonghyun Lee
Jieun Lee
Jieun Shin
Jonghak Kim
Joosung Lee
Jungwhan Kim
Ka Yeon Song
Kiyoon Moon
Minkyoung Kim
Munhyong Kim
MyungIn You
Saerim Cho
Soyoon Kim
Suk Min Seo
Taemin Lim
Taeyong Kim
Woobin Choi
Yehbin Lee
Yelim Jeong
Yeonsun Ahn
Yeontaek Oh
Youngjin Kwon
Zoo Hyun Lee

Data

Byoungeul Kim
Byungwook Lee
Chan-Ho Song
Chansong Jo
Chiheon Ham
Donghyeon Ko
Dongjin Lee
Eunwoo Song
Hanbyul Kim
Hoyeon Lee
Hyun-Wook Yoon
Hyunsoo Ha
Injae Lee
Jaehong Lee
Jaemin Han
Jaeuk Byun
Jahyeong Lee
Jeongmin Liu
Jieun Lee
Jin-Seob Kim
Jinbae Im
Jingu Kang
Jisung Wang
Jong-Hwan Kim
Juncheol Kim
Kang Lae Jung
Kiyoon Moon
Kyeongseok Jeong
Min Young Lee
Min-Seok Choi
Minjae Lee
Minkyoung Kim
Minseong Choi
Moonbin Yim
Munhyong Kim
MyungIn You
Ohsung Kwon
Sangkil Lee
Seongjin Shin
Seunggyu Chang
Shinyoung Joo
Soo-Whan Chung
Sookyo In
Soyeon Choe
Suhyeon Oh
Sung Ae Lee
Sungju Kim
Sungjun Choi
Sunmi Rim
Taehong Min
Taehwan Yoo
Taeyong Kim
Yeguk Jin
Yehbin Lee
You Jin Kim
Youna Ji
Youngjun Kim
Youngki Hong

Model Serving and Inference

Hanbae Seo
Hodong Lee
Hyunjoon Jeong
Jaeeun Kil
Jaegwang Lee
Jeongtae Lee
Jinhyeon Kim
Joonghoon Kim
Junhee Yoo
Minjung Jo
Minsub Kim
Sang Hee Park
Sungjae Lee
Sungju Kim
Yeonsun Ahn

Model Planning

Gayoung Lee
Hagyeong Lee
Hyunhoon Jung
Jeonghyun Lee
Jieun Shin
Jonghak Kim
Saerim Cho
Taemin Lim
Woobin Choi
Yelim Jeong
Youngjin Kwon
Zoo Hyun Lee

Business and Brand Strategy

Dukmin Jung
Kyungmin Lee
Hyojin Park
Sujin Roh
Misuk Park

Residency Program

Bumkyu Park
Byung Hyun Lee
Doohyuk Jang
Geeho Kim
Hyewon Jeon
Hyunbin Jin
Hyungwook Choi
Ijun Jang
Inju Ha
Jewon Yeom
Jihwan Kim
Jihwan Kwak
Joonki Min
Juan Yeo
Junbeom Kim
Junyeob Kim
Kunhee Kim
Kyubyung Chae
Kyudan Jung
Minha Jhang
Sangyoon Lee
Sehyun Lee
Seunghee Kim
Song-ha Jo
Suho Ryu
Yokyung Lee

Internships

Dong-Jae Lee
Jihwan Moon
Jinho Heo
Jisu Jeon
Minsik Choi
Seulbi Lee
Singon Kim
Sumin Cho
Woojin Chung

Appendix

Implementation Details of the Vision Decoder

Training Curriculum.

The training of the vision decoder is organized into four sequential phases to stabilize optimization and ensure high-fidelity synthesis across various scales:

  1. Low-resolution crop training: Initial training at 0.25 Megapixels ($`\sim`$512$`\times`$512) focusing on cropped regions.

  2. Full-resolution crop training: Training at 0.6 Megapixels ($`\sim`$768$`\times`$768) using 1:2 cropped image regions.

  3. Full-resolution full training: Training on the entire image area at 0.6 Megapixels.

  4. Refinement stage: A final stage with a reduced learning rate to stabilize the model and polish fine visual details.

Crop training is particularly effective in our framework because each cropped region is tightly coupled to its corresponding vision token grid. This allows the diffusion model to unambiguously reference the correct conditioning tokens even when operating on partial image patches.

Inference and Autoguidance.

During inference, we adopt autoguidanceĀ , which leads to a substantial improvement in visual quality (see FigureĀ 9). Given that our decoder relies on dense semantic conditioning, the model can occasionally exhibit overly local patterns, resulting in degraded small-scale textures. Autoguidance amplifies the conditioning signal, helping the decoder preserve fine structures—such as typography and intricate patterns—more consistently across the generated output. To apply autoguidance, we train a smaller model (470M) briefly ($`\sim`$1/20 steps of the main model) and use guidance scale of 1.75.

AutoguidanceĀ  significantly improves the overall quality of our vision decoder (see the enhancement in hockey helmets, typography, and fingers).

Text-to-Speech Evaluations

Participants were presented with synthesized speech from five TTS systems, anonymized and randomly shuffled to prevent model identification. For each text script, annotators listened to all five audio samples and assigned Mean Opinion Scores (1–5) based on fluency and pronunciation clarity. The quality of the synthesized speech was evaluated using a 5-point Likert scale, measuring naturalness and intelligibility. Each score corresponds to the following descriptive criteria:

1 (Bad):
Due to severe artifacts and lack of naturalness, the audio is nearly unintelligible.

2 (Poor):
The audio sounds heavily robotic, and issues with pronunciation or intonation make it uncomfortable to listen to.

3 (Fair):
While there is a noticeable awkwardness, there are no major issues in understanding the content.

4 (Good):
The speech is generally natural, though slight awkwardness or a robotic feel may be heard occasionally.

5 (Excellent):
The speech is nearly indistinguishable from a real human voice, with natural pronunciation, intonation, and rhythm.

Annotation interface used for MOS evaluation.

Examples of User-Intent Aware Reasoning

As discussed in Section 4.1, OMNI incorporates an internal reasoning process to classify user intent and determine the optimal modality-specific modules for response generation. This section provides three illustrative examples of the internalized reasoning (the <think> block), demonstrating the model’s ability to parse complex instructions and manage cross-modality constraints.

Visual Question Answering

    <|im_start|>assistant
    <think>
    This is a VQA task requiring high-level visual reasoning. 
    The input modality consists of an image and a textual query. 
    I will leverage the continuous vision encoder to extract 
    perceptual features and map them to linguistic 
    representations, providing an objective and detailed 
    factual response.
    </think>
    
    [Actual Text Response]
    <|im_end|>

Image Editing

    <|im_start|>assistant
    <think>
    This is a request to edit specific visual attributes of 
    the input image. The input modality is image-text
    interleaved. I will utilize the synthesis functions of 
    the diffusion-based vision decoder to modify the requested 
    elements while maintaining the semantic consistency of 
    unchanged objects.
    </think>
    
    [Generated Image Output]
    <|im_end|>

Audio-Visual Understanding with Speech Response

    <|im_start|>assistant
    <think>
    This is a complex omnimodal task involving an 
    audio-based user query regarding visual content. 
    I will first process the continuous audio 
    embedding to parse the user's intent, analyze 
    the image features via the vision encoder, 
    and finally generate a synchronized speech response 
    using discrete audio tokens for the neural audio decoder.
    </think>
    
    [Actual Multimodal/Audio Response]
    <|im_end|>

šŸ“Š 논문 ģ‹œź°ģžė£Œ (Figures)

Figure 1



Figure 2



Figure 3



Figure 4



Figure 5



Figure 6



Figure 7



Figure 8



Figure 9



Figure 10



Figure 11



A Note of Gratitude

The copyright of this content belongs to the respective researchers. We deeply appreciate their hard work and contribution to the advancement of human civilization.

  1. In these experiments, we use the checkpoint released under the TAG-2025-12-31, tag: https://huggingface.co/naver-hyperclovax/HyperCLOVAX-SEED-Omni-8B/tree/TAG-2025-12-31 ↩︎

  2. https://tv.naver.com/  ↩︎

Start searching

Enter keywords to search articles

↑↓
↵
ESC
⌘K Shortcut