As an engineer,
I would like to have all load test results for each model server in one place so that I can have a reference to compare results whenever I need to make changes to see if latencies are increased.
A good place could be a Markdown file committed in the inference-services repository along with the corresponding input data and commands. That way it will also act as a runbook one can just follow before/after deploying a model server if needed.