Thanks to visit codestin.com
Credit goes to github.com

Skip to content

KD-TAO/OmniZip

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

OmniZip: Audio-Guided Dynamic Token Compression for Fast Omnimodal Large Language Models

Keda Tao, Kele Shao, Bohan Yu, Weiqiang Wang, Jian liu, Huan Wang, "OmniZip: Audio-Guided Dynamic Token Compression for Fast Omnimodal Large Language Models"

[Paper]

🔥🔥🔥 News

  • 2025-11-19: The paper is released.
  • 2025-11-18: This repo is released.

overview

Abstract: Omnimodal large language models (OmniLLMs) have attracted increasing research attention of late towards unified audio-video understanding, wherein processing audio-video token sequences creates a significant computational bottleneck, however. Existing token compression methods have yet to accommodate this emerging need of jointly compressing multimodal tokens. To bridge this gap, we present OmniZip, a training-free, audio-guided audio-visual token-compression framework that optimizes multimodal token representation and accelerates inference. Specifically, OmniZip first identifies salient audio tokens, then computes an audio retention score for each time group to capture information density, thereby dynamically guiding video token pruning and preserving cues from audio anchors enhanced by cross-modal similarity. For each time window, OmniZip compresses the video tokens using an interleaved spatio-temporal scheme. Extensive empirical results demonstrate the merits of OmniZip - it achieves 3.42 $\times$ inference speedup and 1.4 $\times$ memory reduction over other top-performing counterparts, while maintaining performance with no training.

⚒️ TODO

  • Release code
  • Release paper
  • Release evaluation script for all benchmarks
  • Support more models

Install

1. Clone this repository and navigate to the LLaVA folder:
git clone https://github.com/KD-TAO/OmniZip.git
cd OmniZip
2. Install the inference package:
conda create -n omnizip python=3.10 -y
conda activate omnizip
pip install --upgrade pip
bash setup.sh

cd lmms-eval
pip install -e .

# Recommend
# pip install torch==2.6.0 torchvision==0.21.0
pip install flash-attn --no-build-isolation

Input Video Setting

You can adjust the number of input frames and the maximum pixel size according to your own computing resources.

VIDEO_MIN_PIXELS = 128 * 28 * 28
VIDEO_MAX_PIXELS = 768 * 28 * 28 # <- 1 (We use 128*28*28 in the paper.)
FRAME_FACTOR = 2
FPS = 2.0
FPS_MIN_FRAMES = 4
FPS_MAX_FRAMES = 768 # <- 2

Quick Start

To have a quick demo of OmniZip with Qwen2.5-Omni, please run

python demo.py --omnizip

You can set the relevant parameters by

python demo.py --omnizip --rho_audio 0.4 --rho_video 0.7

Evaluation

For VideoMME and WorldSense (Benchmarks in lmms-eval)

  • We use the lmms-eval toolkit to evaluate our models. It's worth noting that you can specify OmniZip Settings via parameters in eval.sh, such as:
export WRAPPER=OmniZip
OMNIZIP_RHO_AUDIO=0.3
OMNIZIP_RHO_VIDEO=0.6
OMNIZIP_G=3
OMNIZIP_CONTEXTUAL_RATIO=0.05
...
  • Then you can run for evaluation
bash eval.sh

For Other Benchmark (AVUT, ShorVid-Bench)

  • We will sort out the code as soon as possible.

👀 Results on Audio-Video Understanding Task

overview

overview

Acknowledgement

This project is based on Qwen2.5-Omni. Thanks for their awesome work.

Contact

If you have any questions, please feel free to contact me at [email protected]

Citation

If you find this work useful for your research, please consider citing our paper:

@article{omnizip,
  title={OmniZip: Audio-Guided Dynamic Token Compression for Fast Omnimodal Large Language Models}, 
  author={Keda Tao and Kele Shao and Bohan Yu and Weiqiang Wang and Jian liu and Huan Wang},
  journal={arXiv preprint arXiv:2511.14582},
  year={2025}
}

About

OmniZip: Audio-Guided Dynamic Token Compression for Fast Omnimodal Large Language Models

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published