Uh oh!
There was an error while loading.Please reload this page.
- Notifications
You must be signed in to change notification settings - Fork32k
gh-91555: disable logger while handling log record#131812
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to ourterms of service andprivacy statement. We’ll occasionally send you account related emails.
Already on GitHub?Sign in to your account
Uh oh!
There was an error while loading.Please reload this page.
Conversation
Prevent the possibility of re-entrancy and deadlock or infinite recursioncaused by logging triggered by logging by disabling logging while the logger ishandling log messages.
Most changes to Pythonrequire a NEWS entry. Add one using theblurb_it web app or theblurb command-line tool. If this change has little impact on Python users, wait for a maintainer to apply the |
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
return | ||
maybe_record = self.filter(record) | ||
if not maybe_record: | ||
if self._is_disabled(): |
graingertMar 29, 2025 • edited
Loading Uh oh!
There was an error while loading.Please reload this page.
edited
Uh oh!
There was an error while loading.Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
rather than disabling the logging, can we instead append the record to aself._reentrant_records = collections.deque()
, and then process all of the pending records:
maybe_record=self.filter(record)ifnotmaybe_record:returnifisinstance(maybe_record,LogRecord):record=maybe_recordwas_calling_handlers=set_calling_handlers()try:ifnotwas_calling_handlers:self.callHandlers(record)whileTrue:try:record=self._reentrant_records.popleft()exceptIndexError:returnself.callHandlers(record)else:self._reentrant_records.append(record)finally:set_not_calling_handlers()
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
This will still produce a stack overflow if handling the deferred log message itself logs another message
graingertMar 30, 2025 • edited
Loading Uh oh!
There was an error while loading.Please reload this page.
edited
Uh oh!
There was an error while loading.Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
I'm not following how this would cause a stack overflow, if handling the log message logs another message it would go onto the _reentrant_records queue, and then be processed later once the stack returns all the way back to whereset_calling_handlers()
is first called.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
I'm not following how this would cause a stack overflow, if handling the log message logs another message it would go onto the _reentrant_records queue, and then be processed later once the stack returns all the way back to where
set_calling_handlers()
is first called.
Sorry, I should have said deadlock with the current example. The stack overflow is from a different way of triggering this (see the second unit test added).
The trouble is that when the first recursive logging call exits thefinally
block it clears the "calling handlers" flag, which means a subsequent (still recursive) one takes the wrong path and deadlocks/overflows.
That can be avoided for the original triggering example by only clearing the "handling" flag if it was initially unset (the deferred records collection also needs to be TLS not a member variable). It ends up looking something like this:
ifnothasattr(self._tls,'reentrant_records'):self._tls.reentrant_records=deque()deferred=self._tls.reentrant_recordswas_calling_handlers=self.set_calling_handlers()try:ifnotwas_calling_handlers:self.callHandlers(record)whiledeferred:self.callHandlers(deferred.popleft())else:deferred.append(record)finally:ifnotwas_calling_handlers:self.set_not_calling_handlers()
This fixes the two bugs, which only log thefirst time they try to process a log record (and means those recursive log messages are logged and not silently ignored, which is nice). However a different example which logsevery time (such as the second unit test) will still live-lock and never exit thatwhile
loop.
graingertMar 30, 2025 • edited
Loading Uh oh!
There was an error while loading.Please reload this page.
edited
Uh oh!
There was an error while loading.Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
will still live-lock and never exit that while loop.
Does the system keep logging forever instead?
This seems acceptable as you'd easily track down this bug just by looking at the logs
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
Does the system keep logging forever instead?
Yep, it will just sit in a tight loop spamming the log forever, or at least until/unless that exhausts disk space or wherever the logs are actually going.
This seems acceptable as you'd easily track down this bug just by looking at the logs
IMO it is not agreat failure mode, but it will certainly be obvious!
FWIW I think I prefer ignoring them: the code is much simpler and it prevents the issue in non-trivial handler implementations like Sentry's (that would otherwise trigger the live-lock failure). I was hoping this fix would mean they would be able to remove that nasty monkey-patching on supported versions.
OTOH it is definitely nice to actually handle instead of drop the recursive log messages, in cases where they don't always continue to recurse.
graingertMar 30, 2025 • edited
Loading Uh oh!
There was an error while loading.Please reload this page.
edited
Uh oh!
There was an error while loading.Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
we could have alogging.(_)N_RECURSIVE_CALLS
constant to limit this so it's not forever
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
We could. Another alternative would be Victor Stinner's suggestion in the discourse discussion to raise an exception. That would bring it to the user's attention and force them to deal with it.
Ultimately, though, the way they will have to deal with it is by preventing, disabling, or otherwise intercepting and ignoring all such logging. That will be difficult to do reliably outside the core, will likely be overlooked unless/until it bites, and have to be done in every susceptible log handler or application that uses such.
IMO it would be better for us to do this once, centrally, with a small, simple, and robust fix.
Uh oh!
There was an error while loading.Please reload this page.
TODO list:
|
Most changes to Pythonrequire a NEWS entry. Add one using theblurb_it web app or theblurb command-line tool. If this change has little impact on Python users, wait for a maintainer to apply the |
Would a news entry be appropriate for this? |
yes this needs a news entry |
bedevere-bot commentedApr 14, 2025
🤖 New build scheduled with the buildbot fleet by@vsajip for commit7b68d12 🤖 Results will be shown at: https://buildbot.python.org/all/#/grid?branch=refs%2Fpull%2F131812%2Fmerge If you want to schedule another build, you need to add the🔨 test-with-buildbots label again. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
@duaneg, from the issue:
It disables logging on a per-thread basis and only while the thread runs the message handlers (+filters). Which is to say, it is disabled only while running code that would trigger this bug if it logged a message. Or at least, that is my intent: if I've overlooked anything or there is a bug, please let me know!
It doesn't stop other threads from logging to the same logger and/or handlers.
I think a test with logging from multiple threads and other handlers should be added to confirm this.
A Python core developer has requested some changes be made to your pull request before we can consider merging it. If you could please address their requests along with any other requests in other reviews from core developers that would be appreciated. Once you have made the requested changes, please leave a comment on this pull request containing the phrase |
…andles amessage does not block a different thread from handling a message on the samelogger.
Good idea, will add, thanks! |
bedevere-bot commentedApr 18, 2025
🤖 New build scheduled with the buildbot fleet by@vsajip for commit99788f7 🤖 Results will be shown at: https://buildbot.python.org/all/#/grid?branch=refs%2Fpull%2F131812%2Fmerge If you want to schedule another build, you need to add the🔨 test-with-buildbots label again. |
2561e14
intopython:mainUh oh!
There was an error while loading.Please reload this page.
GH-133898 is a backport of this pull request to the3.13 branch. |
GH-133899 is a backport of this pull request to the3.14 branch. |
…GH-133898)Co-authored-by: Duane Griffin <duaneg@dghda.com>
…GH-133899)Prevent the possibility of re-entrancy leading to deadlock or infinite recursion (caused by logging triggered by logging), by disabling logging while the logger is handling log messages.(cherry picked from commit2561e14)Co-authored-by: Duane Griffin <duaneg@dghda.com>
Uh oh!
There was an error while loading.Please reload this page.
Prevent the possibility of re-entrancy and deadlock or infinite recursion caused by logging triggered by logging by disabling logging while the logger is handling log messages.