Skip to content

Commit

Permalink
Add FAQ on ASGI vs WSGI
Browse files Browse the repository at this point in the history
  • Loading branch information
faph committed Apr 11, 2024
1 parent c8ce5c7 commit 90e6c30
Showing 1 changed file with 9 additions and 0 deletions.
9 changes: 9 additions & 0 deletions docs/faq.rst
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,15 @@ For example, when using Gunicorn, this could be done from a post-fork Gunicorn h
Does **inference-server** support async/ASGI webservers?
--------------------------------------------------------

No.

**inference-server** is a WSGI application to be used by synchronous webservers.

For most ML models that will be the correct choice as model inference is typically CPU-bound.
Therefore, a multi-process based WSGI server is a good choice whereby the number of workers is equal to the number of CPU cores available.

For more details see :ref:`deployment:Configuring Gunicorn workers`.


My model is leaking memory, how do I address that?
--------------------------------------------------
Expand Down

0 comments on commit 90e6c30

Please sign in to comment.