Skip to content

Conversation

@aaarkai
Copy link

@aaarkai aaarkai commented Jan 7, 2026

Purpose

to fix issues: #31889

  • pass indexer_rotary_emb into MLAModules for KimiMLAAttention to match the updated constructor
  • do the same for OpenPangu MLA to prevent the same init-time crash
  • keep MLA init aligned across models that don’t use sparse indexing

Test Plan

vllm serve moonshotai/Kimi-Linear-48B-A3B-Instruct --trust-remote-code --host 0.0.0.0 --port 8999  --tensor-parallel-size 4 --max-model-len 1048576 --no-enable-prefix-caching

Test Result

server launch success


Essential Elements of an Effective PR Description Checklist
  • The purpose of the PR, such as "Fix some issue (link existing issues this PR will resolve)".
  • The test plan, such as providing test command.
  • The test results, such as pasting the results comparison before and after, or e2e results
  • (Optional) The necessary documentation update, such as updating supported_models.md and examples for a new model.
  • (Optional) Release notes update. If your change is user facing, please update the release notes draft in the Google Doc.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request addresses an initialization crash in KimiMLAAttention and OpenPanguMLAAttention by passing the newly required indexer_rotary_emb argument to the MLAModules constructor. The changes are correct and necessary to align with the updated MLAModules interface. The fix is straightforward, well-contained, and I have no further comments.

@aaarkai aaarkai force-pushed the fix-mla-indexer-rotary-emb branch from f961095 to cc6d7d8 Compare January 7, 2026 10:45
Copy link
Collaborator

@jeejeelee jeejeelee left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@aaarkai
Copy link
Author

aaarkai commented Jan 8, 2026

The fix has already been applied in e3fbb6f
, so this pr can be closed.

@aaarkai aaarkai closed this Jan 8, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants