Skip to content

feat: Add merged vLLM rollout weights#631

Open
vivekkalyan wants to merge 5 commits intomainfrom
feat/merged-inference
Open

feat: Add merged vLLM rollout weights#631
vivekkalyan wants to merge 5 commits intomainfrom
feat/merged-inference

Conversation

@vivekkalyan
Copy link
Collaborator

@vivekkalyan vivekkalyan commented Mar 25, 2026

Enable ART to serve merged LoRA weights through dedicated vLLM so Qwen3.5-MoE training works on the current vLLM build.

Changes

  • Add rollout_weights_mode: "lora" | "merged" with dedicated-only validation and a merged-mode requirement for Qwen/Qwen3.5-35B-A3B and Qwen/Qwen3.5-397B-A17B
  • Push merged weights into dedicated vLLM with native weight transfer while keeping LoRA checkpoints for training and persistence
  • Update dedicated server wiring, validation, and Qwen3.5 smoke scripts/tests for the merged-inference path

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant