Skip to content

docs: add ATOM vLLM plugin guide#347

Closed
wuhuikx wants to merge 2 commits intomainfrom
hattie/plugin_doc
Closed

docs: add ATOM vLLM plugin guide#347
wuhuikx wants to merge 2 commits intomainfrom
hattie/plugin_doc

Conversation

@wuhuikx
Copy link
Contributor

@wuhuikx wuhuikx commented Mar 17, 2026

Motivation

Add comprehensive documentation for ATOM's vLLM OOT plugin mode covering
architecture, supported models, configuration translation, attention
integration, model wrapper, environment variables, and troubleshooting.

wuhuikx added 2 commits March 17, 2026 03:20
Add comprehensive documentation for ATOM's vLLM OOT plugin mode covering
architecture, supported models, configuration translation, attention
integration, model wrapper, environment variables, and troubleshooting.

Signed-off-by: wuhuikx <hattie.wu@amd.com>
@wuhuikx wuhuikx requested a review from zejunchen-zejun March 17, 2026 10:14
> | Component | Import / Entry Point | Purpose |
> |-----------|---------------------|---------|
> | `register_platform` | `atom.plugin.vllm.register:register_platform` | vLLM platform plugin — returns `ATOMPlatform` |
> | `register_model` | `atom.plugin.vllm.register:register_model` | vLLM general plugin — overrides model registry |
Copy link
Contributor

@zejunchen-zejun zejunchen-zejun Mar 17, 2026

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

For the function register_platform and register_model, we need to add more descriptions for both functions because they are the entrance of the ATOM OOT plugin of vLLM. https://github.com/ROCm/ATOM/blob/main/pyproject.toml#L33-L39

## 1. Overview

When ATOM is installed alongside vLLM, it automatically registers itself via
Python entry points. vLLM discovers ATOM at startup and delegates model
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Python entry points. vLLM discovers ATOM at startup and delegates model
Python entry points when launching the vLLM server. vLLM discovers ATOM at startup and delegates model

| `Qwen3MoeForCausalLM` | `atom.models.qwen3_moe.Qwen3MoeForCausalLM` | `ATOMMoEForCausalLM` | MoE |
| `DeepseekV3ForCausalLM` | `atom.models.deepseek_v2.DeepseekV3ForCausalLM` | `ATOMMoEForCausalLM` | MoE |
| `GptOssForCausalLM` | `atom.models.gpt_oss.GptOssForCausalLM` | `ATOMMoEForCausalLM` | MoE |
| `Glm4MoeForCausalLM` | `atom.models.glm4_moe.Glm4MoeForCausalLM` | `ATOMMoEForCausalLM` | MoE |
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Kimi is also supported but shares the same architecture with DS. Here we can tell to users that Kimi-K2 is supported.

If you see `Fallback to original vLLM attention backend`, the attention plugin
is disabled.

### Illegal memory access in MoE
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We can remove this section about illegal mem access issue because the issue is gone

@zejunchen-zejun
Copy link
Contributor

Do we need to add the SVG illustration in doc?
image

@zejunchen-zejun
Copy link
Contributor

Hi, @wuhuikx

I have already added docs in PR: #343
Lingpeng suggested us to merge 2 PRs.

@wuhuikx
Copy link
Contributor Author

wuhuikx commented Mar 18, 2026

Hi, @wuhuikx

I have already added docs in PR: #343 Lingpeng suggested us to merge 2 PRs.

great. I will close this one.

@wuhuikx wuhuikx closed this Mar 18, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants