You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, recently vLLM-Omni has been released to support non-text outputs and non-AR arch. Is it possible to get this fantastic work integrated with vLLM-Omni which finally will speed up the dLLM inference seamlessly aligned with vLLM?