-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Threaded workers accept connections even when no ready thread exist #908
Comments
what do you mean by "When using threads (with sync worker class), worker accepts connection even when all threads is busy. " ? With the |
yes, (actually |
thanks for your answer. will work on it during the week. Expect a fix for the coming release this we. |
Hrm, I am not sure how to reproduce that issue. When I look at the code of the threaded workers it appears, that the worker should crash if the number of active connections is greater than the number of accepted worker connections: https://github.com/benoitc/gunicorn/blob/master/gunicorn/workers/gthread.py#L174-L178 When it happen and no futures have returned in the timeout the connections breaks. Are you sure the connections are not handled by another worker? |
Steps to regenerate
import time
import os
def application(env, start):
t = time.time()
print('---- start in ', os.getpid())
if env['QUERY_STRING'] == 'wait':
while time.time() - t < 100: time.sleep(1)
print('--- end in', os.getpid())
start('200 OK', []);
return [b'Hello World']
3 of 4 threads are now busy, and one is idle.
|
We're having the same problem, settings workers to 2+ threads causes exact same issue as explained above. |
on which OS? kernel? |
Description: Ubuntu 12.04.4 LTS |
I've run the example provided above and I think I understood the problem. So the issue is that the threaded worker is not designed to accept long connections right now. It is mostly working like the sync worker. Using 2 threads / workers mean that only 4 simultaneous connections can be handled. Trying to wait like this will rapidly lock exhaust the number of threads. However since the accepting loop is async and handled in its own loop, you may have more connections accepted that will all wait for a free worker, I need to think more about that design. I am thinking to remove any async accepting loop and handle the pool synchronously instead. |
👍 better to handle the pool synchronously. That is better for upstream load balancers because they will see a slow connect and might fail over. |
Could it be that this error is also present even if you have 1 threaded worker ? Seems that we have a odd situation that even after commenting threads=2 in our config we're able to reproduce state when a deadlocked worker accepts connections. EDIT: The interesting thing is that the deadlock is during the requests initialization phase. user - > nginx -> gunicorn(webapp) -> rpcserver the deadlock happens on connecting to rpcserver, and when that happens we have gunicorn accept connections and wait... |
Listening socket is always open, and connecting clients will be queued in TCP backlog (even gunicorn does not accept a new connection). That's not related to this issue. But there is another issue: After sending stop signal, while it waits for current requests to complete, gunicorn does not close listen socket. |
@tahajahangir this actually by desig. But why would you want it? Though this is not related to this ticket, can you open a new ticket about it? |
I have raised this issue before and tried to solve it. I forget why I 👍 opening a new issue for it.
|
@tilgovi If I recall correctly, the reason we are not doing it is for the case where we d a USR2 and stop the old master. In that case we don't want to close socket so the master can still listen on it. I am actually wondering if the system doesn't take care of that for us.... (also if it's portable) |
That was indeed the reason. I'll follow up on the other issue.
|
When the max_requests option is in use, and there are pending requests queued up for the worker, but the worker is shutting down, the queued up requests are closed off forcely 5xx error. |
i am slightly revisiting the gthreads worker. The new workflow is quite more simple:
I will have a patch available tomorrow. |
this change makes sure that a worker don't handle more requests than it can achieved. The new workflow is quite more simple: listeners are put in the poller. On read we try to accept on them. When a connection is accepted it is put in the execution queue When a request is done and the socket can be kept alived, we put it in the poller, on read event we will try to handle the new request. If it is not put out of the poller before the keepalive timeout the socket will be closed. if all threads are busy we are waiting until one request complet. If it doesn't complete before the timeout we kill the worker. fix #908
please review #962 which should solve the issue. |
Hi, Switching to gevent workers solved the issue for us. |
When using threads (with sync worker class), worker accepts connection even when all threads is busy. In our environment, sometimes (because a deadlock in cairocffi library) all threads of a worker are locked, but worker continues accepting connections until limit of open-files is reached.
I looked at the code in
gthread.py
and it regards theworker_connections
setting (despite what the documentation says, that it only affects Eventlet and Gevent workers).I think in threaded workers,
worker_connections
always should be set to number of threads.The text was updated successfully, but these errors were encountered: