-
Notifications
You must be signed in to change notification settings - Fork 18k
net/http: setting a timeout on http requests that use TLS can result in excessive requests to server #59017
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Comments
cc @neild |
This comment was marked as spam.
This comment was marked as spam.
same question when using nginx as the server |
We shouldn't be using an individual request's context to dial a connection which may be reused for many requests, to avoid problems exactly like this. I suspect that someone is relying on passing values down to the dial via the context, however, so we likely need to decouple the request context's values from its cancelation. (Conveniently, #40221 has been implemented now, so this is feasible.) I'm not certain what the correct cancelation behavior is. Perhaps cancel a dial once there are no requests blocked on it? There might still be some access patterns where we repeatedly cancel an in-progress dial, but this would at least avoid the scenario where we cancel a dial call that a request was hoping to make use of. |
This comment was marked as off-topic.
This comment was marked as off-topic.
Using the Those graphed metrics are directly out of net/http/httptrace. One thing that's not clear to me at the moment is if those dialed connections are pooled as idle connections, because metrics on the connection pool are not readily accessible. |
The transport itself uses it for client trace (but it could be propagated from transportRequest if necessary) Lines 1629 to 1639 in 27f41bb
|
Using I have one question which I'm not sure on the answer to: How, if at all, should we bound the lifetime of dials? If a user creates a Transport, makes a request, and cancels the request, should the dial call continue indefinitely? Or should there be some way of cleaning up stale in-progress dials? |
Change https://go.dev/cl/576555 mentions this issue: |
Maybe we can introduce an optional |
What version of Go are you using (
go version
)?go version go1.20.2 darwin/amd64
Does this issue reproduce with the latest release?
Yes
What operating system and processor architecture are you using (
go env
)?GOARCH="amd64"
GOOS="darwin"
What did you do?
When running a large number of concurrent requests to Amazon S3 with a request timeout set (either in the
http.Client
instance or a context directly set on the request), we noticed an unexpectedly large amount of long running requests. When we removed the timeout, the number of long running requests dropped. The long running requests were not directly caused by the timeout being hit - all requests completed in under the timeout.We created a standalone program to reproduce the problem and added logging via
httptrace
. In the output fromhttptrace
we observed a large number of requests that were reported with the errorcontext cancelled
to theTLSHandshakeDone
callback in our trace. These requests did not cause failed requests as reported by the http client.Digging into the http Transport code, it appears that when a connection is not immediately available for use in the connection pool, the runtime starts a race between obtaining a connection returned to the pool and dialing a new connection. In our case, the "obtain connection returned to the pool" was generally winning the race. The behavior on the losing side of the race differed depending on whether the request used a timeout or not. On requests without a timeout, the losing leg of the connection continued through the TLS handshake, and was then placed into the pool as a valid connection. On requests with a timeout, the losing leg was aborted mid-TLS-handshake due to the cancellation of the request context as the request completed using the connection that was returned to the pool.
The net result of this behavior was that whenever a request legitimately required a new connection to be established, it was often queued up (probably at the server end) behind a large number of TLS handshakes that would be cancelled in flight. This manifested as excessive time to complete the request and noticeably lower throughput.
What did you expect to see?
Client does not produce a large volume of aborted TLS handshakes to server.
What did you see instead?
Slowness caused by excessive TLS handshakes to server
I created a gist that reproduces this at https://gist.github.com/mpoindexter/f6bc9dac16290343efba17129c49f9d5. If you uncomment the timeout on line 56 you can see the throughput of the test drop and stall periodically, but if the timeout remains commented throughput remains steady.
To test the analysis above, I implemented a custom
DialTLSContext
function that did the TLS dial using a new context if a deadline was set on the original context. This resolved the problem.The text was updated successfully, but these errors were encountered: