Make the app more scalable.

#2
by xianbao HF staff - opened

Increase the concurrent limit to shorten the queue. The inference api should be quite scalable.

deleted changed pull request status to closed

Sign up or log in to comment