Batching
Overview
Automated Batching
Configuration During Capability Registration
from gllm_datastore.data_store import MilvusDataStore
from gllm_inference.em_invoker import OpenAIEMInvoker
em_invoker = OpenAIEMInvoker(model_name="text-embedding-3-small")
store = (
MilvusDataStore(
collection_name="my_collection",
uri="http://localhost:19530",
)
.with_vector(
em_invoker=em_invoker,
default_batch_size=100,
)
)
await store.vector.create(huge_list_of_chunks)Environment-specific Examples
Per-Call Batching Override
Best Practices
Environment
Recommended Batch Size
Why?
Last updated
Was this helpful?