Skip to content

intel/llm-scaler

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

231 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

LLM Scaler

LLM Scaler is an GenAI solution for text generation, image generation, video generation etc. running on Intelยฎ Arcโ„ข Pro B60 GPUs. LLM Scalar leverages standard frameworks such as vLLM, ComfyUI, SGLang Diffusion, Xinference etc and ensures the best performance for State-of-Art GenAI models running on Arc Pro B60 GPUs.


Latest Update

  • ๐Ÿ”ฅ [2026.03] We released intel/llm-scaler-vllm:0.14.0-b8.1 to support Qwen3.5-27B, Qwen3.5-35B-A3B and Qwen3.5-122B-A10B (FP8/INT4 online quantization, GPTQ)
  • ๐Ÿ”ฅ [2026.03] We released intel/llm-scaler-omni:0.1.0-b6 for ComfyUI to support CacheDiT and torch.compile(), ComfyUI-GGUF, and more model workflows, and support FP8 for SGLang Diffusion.
  • ๐Ÿ”ฅ [2026.03] We released intel/llm-scaler-vllm:0.14.0-b8 for vLLM 0.14.0 and PyTorch 2.10 support, various new models support and performance improvement.
  • [2026.01] We released intel/llm-scaler-vllm:1.3 (or, intel/llm-scaler-vllm:0.11.1-b7) for vLLM 0.11.1 and PyTorch 2.9 support, various new models support and performance improvement.
  • [2026.01] We released intel/llm-scaler-omni:0.1.0-b5 for Python 3.12 and PyTorch 2.9 support, various ComfyUI workflows and more SGLang Diffusion support.
  • [2025.12] We released intel/llm-scaler-vllm:1.2, same image as intel/llm-scaler-vllm:0.10.2-b6.
  • [2025.12] We released intel/llm-scaler-omni:0.1.0-b4 to support ComfyUI workflows for Z-Image-Turbo, Hunyuan-Video-1.5 T2V/I2V with multi-XPU, and experimentially support SGLang Diffusion.
  • [2025.11] We released intel/llm-scaler-vllm:0.10.2-b6 to support Qwen3-VL (Dense/MoE), Qwen3-Omni, Qwen3-30B-A3B (MoE Int4), MinerU 2.5, ERNIE-4.5-vl etc.
  • [2025.11] We released intel/llm-scaler-vllm:0.10.2-b5 to support gpt-oss models and released intel/llm-scaler-omni:0.1.0-b3 to support more ComfyUI workflows, and Windows installation.
  • [2025.10] We released intel/llm-scaler-omni:0.1.0-b2 to support more models with ComfyUI workflows and Xinference.
  • [2025.09] We released intel/llm-scaler-vllm:0.10.0-b3 to support more models (MinerU, MiniCPM-v-4.5 etc), and released intel/llm-scaler-omni:0.1.0-b1 to enable first omni GenAI models using ComfyUI and Xinference on Arc Pro B60 GPU.
  • [2025.08] We released intel/llm-scaler-vllm:1.0.

LLM Scaler vLLM

llm-scaler-vllm supports running text generation models using vLLM, featuring:

  • CCL support (P2P or USM)
  • INT4 and FP8 quantized online serving
  • Embedding and Reranker model support
  • Multi-Modal model support
  • Omni model support
  • Tensor Parallel, Pipeline Parallel and Data Parallel
  • Finding maximum Context Length
  • Multi-Modal WebUI
  • BPE-Qwen tokenizer

Please follow the instructions in the Getting Started to use llm-scaler-vllm.

Supported Models

Category Model Name FP16 Dynamic Online FP8 Dynamic Online Int4 MXFP4 Notes
Language Model openai/gpt-oss-20b โœ…
Language Model openai/gpt-oss-120b โœ…
Language Model deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B โœ… โœ… โœ…
Language Model deepseek-ai/DeepSeek-R1-Distill-Qwen-7B โœ… โœ… โœ…
Language Model deepseek-ai/DeepSeek-R1-Distill-Llama-8B โœ… โœ… โœ…
Language Model deepseek-ai/DeepSeek-R1-Distill-Qwen-14B โœ… โœ… โœ…
Language Model deepseek-ai/DeepSeek-R1-Distill-Qwen-32B โœ… โœ… โœ…
Language Model deepseek-ai/DeepSeek-R1-Distill-Llama-70B โœ… โœ… โœ…
Language Model deepseek-ai/DeepSeek-R1-0528-Qwen3-8B โœ… โœ… โœ…
Language Model deepseek-ai/DeepSeek-V2-Lite โœ… โœ… export VLLM_MLA_DISABLE=1
Language Model deepseek-ai/deepseek-coder-33b-instruct โœ… โœ… โœ…
Language Model Qwen/Qwen3-8B โœ… โœ… โœ…
Language Model Qwen/Qwen3-14B โœ… โœ… โœ…
Language Model Qwen/Qwen3-32B โœ… โœ… โœ…
Language Model Qwen/Qwen3.5-27B โœ… โœ… โœ…
Language MOE Model Qwen/Qwen3-30B-A3B โœ… โœ… โœ…
Language MOE Model Qwen/Qwen3-235B-A22B โœ…
Language MOE Model Qwen/Qwen3-Coder-30B-A3B-Instruct โœ… โœ… โœ…
Language MOE Model Qwen/Qwen3-Coder-Next โœ… โœ… โœ…
Language MOE Model Qwen/Qwen3.5-35B-A3B โœ… โœ… โœ…
Language MOE Model Qwen/Qwen3.5-122B-A10B โœ… โœ…
Language Model Qwen/QwQ-32B โœ… โœ… โœ…
Language Model mistralai/Ministral-8B-Instruct-2410 โœ… โœ… โœ…
Language Model mistralai/Mixtral-8x7B-Instruct-v0.1 โœ… โœ… โœ…
Language Model meta-llama/Llama-3.1-8B โœ… โœ… โœ…
Language Model meta-llama/Llama-3.1-70B โœ… โœ… โœ…
Language Model baichuan-inc/Baichuan2-7B-Chat โœ… โœ… โœ… with chat_template
Language Model baichuan-inc/Baichuan2-13B-Chat โœ… โœ… โœ… with chat_template
Language Model THUDM/CodeGeex4-All-9B โœ… โœ… โœ… with chat_template
Language Model zai-org/GLM-4-9B-0414 โœ… use bfloat16
Language Model zai-org/GLM-4-32B-0414 โœ… use bfloat16
Language MOE Model zai-org/GLM-4.5-Air โœ… โœ…
Language MOE Model zai-org/GLM-4.7-Flash โœ… โœ…
Language Model ByteDance-Seed/Seed-OSS-36B-Instruct โœ… โœ… โœ…
Language Model miromind-ai/MiroThinker-v1.5-30B โœ… โœ… โœ…
Language Model tencent/Hunyuan-0.5B-Instruct โœ… โœ… โœ… follow the guide in here
Language Model tencent/Hunyuan-7B-Instruct โœ… โœ… โœ… follow the guide in here
Multimodal Model Qwen/Qwen2-VL-7B-Instruct โœ… โœ… โœ…
Multimodal Model Qwen/Qwen2.5-VL-7B-Instruct โœ… โœ… โœ…
Multimodal Model Qwen/Qwen2.5-VL-32B-Instruct โœ… โœ… โœ…
Multimodal Model Qwen/Qwen2.5-VL-72B-Instruct โœ… โœ… โœ…
Multimodal Model Qwen/Qwen3-VL-4B-Instruct โœ… โœ… โœ…
Multimodal Model Qwen/Qwen3-VL-8B-Instruct โœ… โœ… โœ…
Multimodal MOE Model Qwen/Qwen3-VL-30B-A3B-Instruct โœ… โœ… โœ…
Multimodal Model openbmb/MiniCPM-V-2_6 โœ… โœ… โœ…
Multimodal Model openbmb/MiniCPM-V-4 โœ… โœ… โœ…
Multimodal Model openbmb/MiniCPM-V-4_5 โœ… โœ… โœ…
Multimodal Model OpenGVLab/InternVL2-8B โœ… โœ… โœ…
Multimodal Model OpenGVLab/InternVL3-8B โœ… โœ… โœ…
Multimodal Model OpenGVLab/InternVL3_5-8B โœ… โœ… โœ…
Multimodal MOE Model OpenGVLab/InternVL3_5-30B-A3B โœ… โœ… โœ…
Multimodal Model rednote-hilab/dots.ocr โœ… โœ… โœ…
Multimodal Model ByteDance-Seed/UI-TARS-7B-DPO โœ… โœ… โœ…
Multimodal Model google/gemma-3-12b-it โœ… use bfloat16
Multimodal Model google/gemma-3-27b-it โœ… use bfloat16
Multimodal Model THUDM/GLM-4v-9B โœ… โœ… โœ… with --hf-overrides and chat_template
Multimodal Model zai-org/GLM-4.1V-9B-Base โœ… โœ… โœ…
Multimodal Model zai-org/GLM-4.1V-9B-Thinking โœ… โœ… โœ…
Multimodal Model zai-org/Glyph โœ… โœ… โœ…
Multimodal Model opendatalab/MinerU2.5-2509-1.2B โœ… โœ… โœ…
Multimodal Model baidu/ERNIE-4.5-VL-28B-A3B-Thinking โœ… โœ… โœ…
Multimodal Model zai-org/GLM-4.6V-Flash โœ… โœ… โœ… pip install transformers==5.0.0rc0 first
Multimodal Model PaddlePaddle/PaddleOCR-VL โœ… โœ… โœ… follow the guide in here
Multimodal Model deepseek-ai/DeepSeek-OCR โœ… โœ… โœ…
Multimodal Model deepseek-ai/DeepSeek-OCR-2 โœ… โœ… โœ… There may be accuracy issues when using --quantization fp8
Multimodal Model moonshotai/Kimi-VL-A3B-Thinking-2506 โœ… โœ… โœ…
omni Qwen/Qwen2.5-Omni-7B โœ… โœ… โœ…
omni Qwen/Qwen3-Omni-30B-A3B-Instruct โœ… โœ… โœ…
audio openai/whisper-medium โœ… โœ… โœ…
audio openai/whisper-large-v3 โœ… โœ… โœ…
Embedding Model Qwen/Qwen3-Embedding-8B โœ… โœ… โœ…
VL Embedding Model Qwen3-VL-Embedding-2B/8B โœ… โœ… โœ… follow the guide in here
Embedding Model BAAI/bge-m3 โœ… โœ… โœ…
Embedding Model BAAI/bge-large-en-v1.5 โœ… โœ… โœ…
Reranker Model Qwen/Qwen3-Reranker-8B โœ… โœ… โœ…
VL Reranker Model Qwen3-VL-Reranker-2B/8B โœ… โœ… โœ… follow the guide in here
Reranker Model BAAI/bge-reranker-large โœ… โœ… โœ…
Reranker Model BAAI/bge-reranker-v2-m3 โœ… โœ… โœ…

LLM Scaler Omni (experimental)

llm-scaler-omni supports running image/voice/video generation etc., featuring Omni Studio mode (using ComfyUI) and Omni Serving mode (via SGLang Diffusion or Xinference).

Please follow the instructions in the Getting Started to use llm-scaler-omni.

Omni Demos

Qwen-Image Multi B60 Wan2.2-T2V-14B
Qwen Image Demo Wan2.2 T2V Demo

Omni Studio (ComfyUI WebUI interaction)

Omni Stuido supports Image Generation/Edit, Video Generation, Audio Generation, 3D Generation etc.

Model Category Model Type
Image Generation Qwen-Image, Qwen-Image-Edit Text-to-Image, Image Editing
Image Generation Stable Diffusion 3.5 Text-to-Image, ControlNet
Image Generation Z-Image-Turbo Text-to-Image
Image Generation Flux.1, Flux.1 Kontext dev Text-to-Image, Multi-Image Reference, ControlNet
Image Generation FireRed-Image-Edit-1.1 Image Editing
Video Generation Wan2.2 TI2V 5B, Wan2.2 T2V 14B, Wan2.2 I2V 14B Text-to-Video, Image-to-Video
Video Generation Wan2.2 Animate 14B Video Animation
Video Generation HunyuanVideo 1.5 8.3B Text-to-Video, Image-to-Video
Video Generation LTX-2 Text-to-Video, Image-to-Video
3D Generation Hunyuan3D 2.1 Text/Image-to-3D
Audio Generation VoxCPM1.5, IndexTTS 2 Text-to-Speech, Voice Cloning
Video Upscaling SeedVR2 Video Restoration and Upscaling

Please check ComfyUI Support for more details.

Omni Serving (OpenAI-API compatible serving)

Omni Serving supports Image Generation, Audio Generation etc.

  • Image Generation (/v1/images/generations): Stable Diffusion 3.5, Flux.1-dev
  • Text to Speech (/v1/audio/speech): Kokoro 82M
  • Speech to Text (/v1/audio/transcriptions): whisper-large-v3

Please check Xinference Support for more details.


Releases


Get Support

  • Please report a bug or raise a feature request by opening a Github Issue

About

No description, website, or topics provided.

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Packages

 
 
 

Contributors