-
Notifications
You must be signed in to change notification settings - Fork 65
Log lines merged to single message #107
Comments
Hi @chris-vest , Thanks for reporting this issue. |
MULTILINE_START_REGEXP = /^\w{3} \d{1,2}, \d{4}/ Example log: https://pastebin.com/SPr2rxFZ Just to clarify, this only seems to happen after we enable a new connector - this causes Connect to log a huge amount of lines very quickly, which seems to overload fluentd somehow - causing the massive, squashed log message to end up in SumoLogic. Restarting the fluentd pods then fixes this issue. |
I think there are still something not very clear. |
So Kafka Connect is logging very frequently after a new Connector is configured, which ends up overloading fluentd. The resulting log in SumoLogic looks like this - https://pastebin.com/SPr2rxFZ The initial logs from Kafka Connect are no different to the logs which it normally creates, the only difference is the frequency of those logs. |
I am still seeing this issue, this is what I'm getting from the sumologic-fluentd pods:
|
@bin3377 I've upgraded to 2.3.0 to see if this is mitigated at all. Looking at the buffer settings, I'm a little confused as to where to put them when configuring fluentd? I know I need to set the |
@bin3377 I've tried adding this:
But I'm still seeing log lines merged into single, large log messages in Sumologic - once this happens, fluentd stops sending messages to Sumologic. |
I think one of your original comments was correct, in that when I cause the event in my cluster, and the application logs busily, the logs are no longer picked up by the regex. However, shouldn't setting the |
In certain occasions the logs coming from our services are not correctly parsed. As a consequence, thousands of log lines are crammed together in one big message sent to Sumo Logic. Because the message is so large, it overflows Sumo Logic limit for an individual message and the content is split into multiple messages. As a result, it loses its JSON formatting and it becomes "invisible" for many queries that work under the assumption that the log entries are structured JSON.
This seems to be occurring only with Kafka Connect; when certain operations are performed it logs very noisily.
I have noticed that
sumologic-fluentd
is being OOM killed. I'm wondering if these log lines are being stuffed into a single message in order to 'catch up' aftersumologic-fluentd
is restarted?The text was updated successfully, but these errors were encountered: