Make the app more scalable.
#2
by
xianbao
HF staff
- opened
Increase the concurrent limit to shorten the queue. The inference api should be quite scalable.
deleted
changed pull request status to
closed
Increase the concurrent limit to shorten the queue. The inference api should be quite scalable.