Skip to content

Conversation

@yannicks1
Copy link
Collaborator

@yannicks1 yannicks1 commented Nov 13, 2025

changes:

  • increasing number of blocks from 2080 to 8192

Note: do not merge yet until...

@github-actions
Copy link

👋 Hi! Thank you for contributing to vLLM support on Spyre.
Just a reminder: Make sure that your code passes all the linting checks, otherwise your PR won't be able to be merged. To do so, first install the linting requirements, then run format.sh and commit the changes. This can be done with uv directly:

uv sync --frozen --group lint --active --inexact

Or this can be done with pip:

uv pip compile --group lint > requirements-lint.txt
pip install -r requirements-lint.txt
bash format.sh

Now you are good to go 🚀

Signed-off-by: Yannick Schnider <[email protected]>
@joerunde
Copy link
Collaborator

joerunde commented Nov 13, 2025

@yannicks1 I posted this on our internal issue as well but we need to have some way to try to check the version of the spyre runtime stack so that we can set these values appropriately. We wouldn't want a newer version of vllm-spyre to set these expanded limits when its installed alongside an older version of the spyre runtime that doesn't support them.

(This would go away if we actually had APIs to call to get this data which we originally thought we would, but here we are 🤷 )

@yannicks1
Copy link
Collaborator Author

yeah, this makes sense! did not mean to merge this as is, but rather to have a branch to test this.
we can add the checks here once we know how to do it.

@yannicks1
Copy link
Collaborator Author

note: tkv x batchsize constraint will not be increased. number of blocks will be increased (for prefix caching)

@yannicks1 yannicks1 changed the title [DO NOT MERGE] bumping spyre card limits [DO NOT MERGE] increase number of blocks on spyre cards Nov 20, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants