Browse Source

Various clean-ups to the logging context code (#8935)

tags/v1.25.0rc1
Patrick Cloke 3 years ago
committed by GitHub
parent
commit
1619802228
No known key found for this signature in database GPG Key ID: 4AEE18F83AFDEB23
11 changed files with 20 additions and 39 deletions
  1. +1
    -1
      changelog.d/8916.misc
  2. +1
    -0
      changelog.d/8935.misc
  3. +1
    -1
      synapse/config/logger.py
  4. +1
    -2
      synapse/http/site.py
  5. +5
    -19
      synapse/logging/context.py
  6. +3
    -4
      synapse/metrics/background_process_metrics.py
  7. +1
    -2
      synapse/replication/tcp/protocol.py
  8. +3
    -3
      tests/handlers/test_federation.py
  9. +2
    -5
      tests/logging/test_terse_json.py
  10. +1
    -1
      tests/test_federation.py
  11. +1
    -1
      tests/test_utils/logging_setup.py

+ 1
- 1
changelog.d/8916.misc View File

@@ -1 +1 @@
Improve structured logging tests.
Various clean-ups to the structured logging and logging context code.

+ 1
- 0
changelog.d/8935.misc View File

@@ -0,0 +1 @@
Various clean-ups to the structured logging and logging context code.

+ 1
- 1
synapse/config/logger.py View File

@@ -206,7 +206,7 @@ def _setup_stdlib_logging(config, log_config_path, logBeginner: LogBeginner) ->
# filter options, but care must when using e.g. MemoryHandler to buffer
# writes.

log_context_filter = LoggingContextFilter(request="")
log_context_filter = LoggingContextFilter()
log_metadata_filter = MetadataFilter({"server_name": config.server_name})
old_factory = logging.getLogRecordFactory()



+ 1
- 2
synapse/http/site.py View File

@@ -128,8 +128,7 @@ class SynapseRequest(Request):

# create a LogContext for this request
request_id = self.get_request_id()
logcontext = self.logcontext = LoggingContext(request_id)
logcontext.request = request_id
self.logcontext = LoggingContext(request_id, request=request_id)

# override the Server header which is set by twisted
self.setHeader("Server", self.site.server_version_string)


+ 5
- 19
synapse/logging/context.py View File

@@ -203,10 +203,6 @@ class _Sentinel:
def copy_to(self, record):
pass

def copy_to_twisted_log_entry(self, record):
record["request"] = None
record["scope"] = None

def start(self, rusage: "Optional[resource._RUsage]"):
pass

@@ -372,13 +368,6 @@ class LoggingContext:
# we also track the current scope:
record.scope = self.scope

def copy_to_twisted_log_entry(self, record) -> None:
"""
Copy logging fields from this context to a Twisted log record.
"""
record["request"] = self.request
record["scope"] = self.scope

def start(self, rusage: "Optional[resource._RUsage]") -> None:
"""
Record that this logcontext is currently running.
@@ -542,13 +531,10 @@ class LoggingContext:
class LoggingContextFilter(logging.Filter):
"""Logging filter that adds values from the current logging context to each
record.
Args:
**defaults: Default values to avoid formatters complaining about
missing fields
"""

def __init__(self, **defaults) -> None:
self.defaults = defaults
def __init__(self, request: str = ""):
self._default_request = request

def filter(self, record) -> Literal[True]:
"""Add each fields from the logging contexts to the record.
@@ -556,14 +542,14 @@ class LoggingContextFilter(logging.Filter):
True to include the record in the log output.
"""
context = current_context()
for key, value in self.defaults.items():
setattr(record, key, value)
record.request = self._default_request

# context should never be None, but if it somehow ends up being, then
# we end up in a death spiral of infinite loops, so let's check, for
# robustness' sake.
if context is not None:
context.copy_to(record)
# Logging is interested in the request.
record.request = context.request

return True



+ 3
- 4
synapse/metrics/background_process_metrics.py View File

@@ -199,8 +199,7 @@ def run_as_background_process(desc: str, func, *args, bg_start_span=True, **kwar
_background_process_start_count.labels(desc).inc()
_background_process_in_flight_count.labels(desc).inc()

with BackgroundProcessLoggingContext(desc) as context:
context.request = "%s-%i" % (desc, count)
with BackgroundProcessLoggingContext(desc, "%s-%i" % (desc, count)) as context:
try:
ctx = noop_context_manager()
if bg_start_span:
@@ -244,8 +243,8 @@ class BackgroundProcessLoggingContext(LoggingContext):

__slots__ = ["_proc"]

def __init__(self, name: str):
super().__init__(name)
def __init__(self, name: str, request: Optional[str] = None):
super().__init__(name, request=request)

self._proc = _BackgroundProcess(name, self)



+ 1
- 2
synapse/replication/tcp/protocol.py View File

@@ -172,8 +172,7 @@ class BaseReplicationStreamProtocol(LineOnlyReceiver):
# a logcontext which we use for processing incoming commands. We declare it as a
# background process so that the CPU stats get reported to prometheus.
ctx_name = "replication-conn-%s" % self.conn_id
self._logging_context = BackgroundProcessLoggingContext(ctx_name)
self._logging_context.request = ctx_name
self._logging_context = BackgroundProcessLoggingContext(ctx_name, ctx_name)

def connectionMade(self):
logger.info("[%s] Connection established", self.id())


+ 3
- 3
tests/handlers/test_federation.py View File

@@ -126,7 +126,7 @@ class FederationTestCase(unittest.HomeserverTestCase):
room_version,
)

with LoggingContext(request="send_rejected"):
with LoggingContext("send_rejected"):
d = run_in_background(self.handler.on_receive_pdu, OTHER_SERVER, ev)
self.get_success(d)

@@ -178,7 +178,7 @@ class FederationTestCase(unittest.HomeserverTestCase):
room_version,
)

with LoggingContext(request="send_rejected"):
with LoggingContext("send_rejected"):
d = run_in_background(self.handler.on_receive_pdu, OTHER_SERVER, ev)
self.get_success(d)

@@ -198,7 +198,7 @@ class FederationTestCase(unittest.HomeserverTestCase):
# the auth code requires that a signature exists, but doesn't check that
# signature... go figure.
join_event.signatures[other_server] = {"x": "y"}
with LoggingContext(request="send_join"):
with LoggingContext("send_join"):
d = run_in_background(
self.handler.on_send_join_request, other_server, join_event
)


+ 2
- 5
tests/logging/test_terse_json.py View File

@@ -117,11 +117,10 @@ class TerseJsonTestCase(LoggerCleanupMixin, TestCase):
"""
handler = logging.StreamHandler(self.output)
handler.setFormatter(JsonFormatter())
handler.addFilter(LoggingContextFilter(request=""))
handler.addFilter(LoggingContextFilter())
logger = self.get_logger(handler)

with LoggingContext() as context_one:
context_one.request = "test"
with LoggingContext(request="test"):
logger.info("Hello there, %s!", "wally")

log = self.get_log_line()
@@ -132,9 +131,7 @@ class TerseJsonTestCase(LoggerCleanupMixin, TestCase):
"level",
"namespace",
"request",
"scope",
]
self.assertCountEqual(log.keys(), expected_log_keys)
self.assertEqual(log["log"], "Hello there, wally!")
self.assertEqual(log["request"], "test")
self.assertIsNone(log["scope"])

+ 1
- 1
tests/test_federation.py View File

@@ -134,7 +134,7 @@ class MessageAcceptTests(unittest.HomeserverTestCase):
}
)

with LoggingContext(request="lying_event"):
with LoggingContext():
failure = self.get_failure(
self.handler.on_receive_pdu(
"test.serv", lying_event, sent_to_us_directly=True


+ 1
- 1
tests/test_utils/logging_setup.py View File

@@ -48,7 +48,7 @@ def setup_logging():
handler = ToTwistedHandler()
formatter = logging.Formatter(log_format)
handler.setFormatter(formatter)
handler.addFilter(LoggingContextFilter(request=""))
handler.addFilter(LoggingContextFilter())
root_logger.addHandler(handler)

log_level = os.environ.get("SYNAPSE_TEST_LOG_LEVEL", "ERROR")


Loading…
Cancel
Save