[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: trunk proxy_http failures

> Am 19.09.2018 um 17:12 schrieb Yann Ylavic <ylavic.dev@xxxxxxxxx>:
> On Wed, Sep 19, 2018 at 4:46 PM Stefan Eissing
> <stefan.eissing@xxxxxxxxxxxxx> wrote:
>> Thanks, Yann, this helped me pin the problem down further:
>> - with disablereuse=on everything works fine
>> - with ttl=1 the problem is still there
> Is the KeepAliveTimeout on the backend side above 1 (second)?

Yes, upped it to 20, no difference.

>> and then:
>> - with mpm_worker, the problem also disappears (no disable/ttl needed)
> Hmm, something (new?) which does not respect KA timeout on MPM event...
> Can you confirm this (for instance with tcpdump or alike)?

I will debug more tomorrow. As usual, timing seems to play a role. Basically there
is a sequence of 3 requests in play which repeat with different content:
1. POST (no expect), small request body
2. POST (expect-100-cont), upload.py body is 1k/10k/100k/ file
3. GET on files/data-1k etc. 

The requests are done in this order, not parallel. All on a new connection.

If a request fails, it is always 3 and always with proxy re-using a dead connection.
So, assuming it is the same proxy connection that gets re-used, what may cause the
connection to close after request 2?

That is runs on mpm_worker *may* point to mpm_event. But with "LogLevel core:trace8"
it seems to disappear in event, also. With core:trace6 it still happens...

Strange thing.

Btw. Solving this is not urgent for me. I see this only in trunk.

Let's see if I gain more insight tomorrow on this.



>> These tests were running since the dawn of h2 time and are still running in 2.4.x. Since the problem also goes away on worker, this looks like a new problem with mpm_event and connection closing (keepalive)?
> I'm having a look at it, will come back if something pops up.