Skip to content
This repository has been archived by the owner on Oct 16, 2023. It is now read-only.

How to use dynamic batch features #199

Open
hudengjunai opened this issue Mar 1, 2023 · 1 comment
Open

How to use dynamic batch features #199

hudengjunai opened this issue Mar 1, 2023 · 1 comment

Comments

@hudengjunai
Copy link

Hello, I have launched the opt-125M inference, and send request to server with locust. but what ever config the max_batch_size, the InferenceEngine always run in batch_size =1. how can i use the dynamic batch feature in Batch_server_manager?

@ver217
Copy link
Member

ver217 commented Mar 20, 2023

It will make batch only when a sequence of input can be batched, e.g. when they have same generation steps.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants