Open Issues Need Help
View All on GitHubA framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
AI Summary: The GitHub issue proposes adding support for CosyVoice 2/3, a popular LLM-based streaming Text-to-Speech (TTS) model. It leverages a Qwen2.5-0.5B backbone to generate speech tokens via Finite Scalar Quantization (FSQ), followed by causal flow matching for audio synthesis. The integration aims to benefit from `vllm-omni`'s existing transformer optimizations, offering low-latency streaming and multi-language support.
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
AI Summary: This RFC proposes enhancing vLLM-Omni's support for state-of-the-art DiT (Diffusion Transformer) models, encompassing image, video, and any-to-any generation. It aims to extend the existing Qwen-Image implementation and actively seeks community contributions for this rapidly evolving domain.
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
AI Summary: This GitHub issue requests support for the PyTorch profiler in vLLM-Omni, as its current absence makes performance bottleneck analysis difficult for multimodal/diffusion workloads. Users have confirmed that setting `profile=True` or configuring `VLLM_TORCH_PROFILER_DIR` does not activate the profiler. Maintainers acknowledge this as a planned but unscheduled feature.
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models
A framework for efficient model inference with omni-modality models