▲ 2 VLLM-Omni: A framework for efficient model inference with Omni-modality models (github.com) by zyh888 | Dec 2, 2025 | 1 comments on HN Visit Link