Inference batcher#
Inference batchers can be accessed from the Predictor metadata objects.
predictor.inference_batcher
InferenceBatcher #
Configuration of an inference batcher for a predictor.
| PARAMETER | DESCRIPTION |
|---|---|
enabled | Whether the inference batcher is enabled or not. The default value is TYPE: |
max_batch_size | Maximum requests batch size. TYPE: |
max_latency | Maximum latency for request batching. TYPE: |
timeout | Maximum waiting time for request batching. TYPE: |
| RETURNS | DESCRIPTION |
|---|---|
|
|