BenchmarkConfigΒΆ
class BenchmarkConfig ( VeekshaCommand )
Run a benchmark against an LLM inference server.
Fields:
output_dirstr ="benchmark_output"Base directory for all benchmark outputs (traces, metrics, logs)
seedint =42Seed for the random number generator.
session_generatorBaseSessionGeneratorConfig = SyntheticSessionGeneratorConfigThe session generator configuration for the benchmark.
Available types:
synthetic: SyntheticSessionGeneratorConfiglmeval: LmevalSessionGeneratorConfigtrace: TraceSessionGeneratorConfig
traffic_schedulerBaseTrafficConfig = RateTrafficConfigThe traffic scheduler configuration for the benchmark.
Available types:
rate: RateTrafficConfigconcurrent: ConcurrentTrafficConfigsequential_launch: SequentialLaunchTrafficConfig
evaluatorslist [ BaseEvaluatorConfig ] =[<PerformanceEvaluatorConfig>]List of evaluators to run.
Available types:
performance: PerformanceEvaluatorConfigaccuracy_lmeval: LMEvalAccuracyEvaluatorConfig
clientBaseClientConfig = OpenAIChatCompletionsClientConfigThe client configuration for the benchmark.
Available types:
openai_chat_completions: OpenAIChatCompletionsClientConfigopenai_completions: OpenAICompletionsClientConfigopenai_router: OpenAIRouterClientConfig
runtimeRuntimeConfig = RuntimeConfigThe runtime configuration for the benchmark.
trace_recorderTraceRecorderConfig = TraceRecorderConfigTrace recorder configuration. Records dispatched requests (unlike the evaluator, which records them after completion).
serverOptional [ BaseServerConfig ] =NoneServer configuration for managed servers. If set, client.model, client.api_key and client.api_base will be overwritten.
Available types:
vllm: VllmServerConfigvajra: VajraServerConfigsglang: SglangServerConfig
wandbWandbConfig = WandbConfigWeights & Biases logging configuration.