Skip to content

[Feature] Adding test-deployment-baremetal to the vllm_performance actuator#784

Draft
DanteNiewenhuis wants to merge 3 commits intoIBM:mainfrom
DanteNiewenhuis:baremetal-experiment
Draft

[Feature] Adding test-deployment-baremetal to the vllm_performance actuator#784
DanteNiewenhuis wants to merge 3 commits intoIBM:mainfrom
DanteNiewenhuis:baremetal-experiment

Conversation

@DanteNiewenhuis
Copy link
Copy Markdown

Started with creating a bare-metal experiment. At the moment, it correctly checks if an endpoint is available and otherwise throws a VLLMBenchmarkError. However, when trying to serve a model, It is giving a lot of errors.

…ctly checkes if an endpoint is available, and otherwise throws an VLLMBenchmarkError. However, when trying to serve a model, It is giving a lot of errors.
@christian-pinto christian-pinto marked this pull request as draft April 2, 2026 14:33
…en not running already. Currently it is not yet correctly checking if serving is done.
… after runs benchmarks. Still not working:

- No alignment with input parameters.
- No checking if the correct model is being served (only if any is being served).
- vLLM instance is not yet correctly released after the benchmark.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant