Skip to content

[pull] main from inclusionAI:main#22

Merged
pull[bot] merged 1 commit intoaxistore80-coder:mainfrom
inclusionAI:main
Mar 31, 2026
Merged

[pull] main from inclusionAI:main#22
pull[bot] merged 1 commit intoaxistore80-coder:mainfrom
inclusionAI:main

Conversation

@pull
Copy link
Copy Markdown

@pull pull bot commented Mar 31, 2026

See Commits and Changes for more details.


Created by pull[bot] (v2.0.0-alpha.4)

Can you help keep this open source service alive? 💖 Please sponsor : )

…1118)

Improve FP8 robustness: extend shard alignment validation to
GroupedExperts, fix DTensor handling in dense FP8 linear forward,
add early checkpoint compatibility checks, and clean up config API.

Key changes:
- Validate GroupedExperts w1/w2/w3 shapes in post-parallelism check
- Convert DTensor to local tensor in FP8 linear forward for TP>1
- Restrict FP8 dequant to float8_e4m3fn matching prepare path
- Fail fast on Shard(1) FP8 checkpoints before DCP I/O
- Add ArchonFP8Config.enabled property to centralize mode checks
- Document exclude_modules default list in YAML example
@pull pull bot locked and limited conversation to collaborators Mar 31, 2026
@pull pull bot added the ⤵️ pull label Mar 31, 2026
@pull pull bot merged commit 0ee8562 into axistore80-coder:main Mar 31, 2026
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant