-
Notifications
You must be signed in to change notification settings - Fork 103
Out of memory in case of Splunk indexer slowness/failure #423
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
@ludovic-boutros thanks for this. |
@ludovic-boutros can you attach the entire kafka connect logs? |
@VihasMakwana we will open a case on Splunk side in order to send you the complete logs in a more secure way. |
I still don't understand how the |
I have put some classes in debug log level.
I don't think this negative number is normal ;) |
This incorrect outstanding event count makes the outstanding event limit ineffective and this leads to an out of memory in case of Splunk slowness/failure. This is my understanding. |
@VihasMakwana I did not manage to understand how the outstanding event count could be lower than zero. |
@VihasMakwana I have patched the connector to prevent negative event count and we can see the effect on the number of events kept in memory. It does not fix the real issue but at least the "symptoms". |
Resolved by #431 |
Hello,
We are using the Splunk Sink Connector with these main parameters:
In my understanding, we should never have more than 50000 events per task kept in memory.
But that is not the case if Splunk indexers encounter slowness or failures.
We can observe in the Kafka Connect logs such errors and messages:
[2024-02-27 06:39:24,527] INFO [SplunkHECSinkConnector|task-5] handled 394 failed batches with 193452 events (com.splunk.kafka.connect.SplunkSinkTask:154)
I have attached the Kafka Connect metrics during a Splunk indexer stress test.
You can observe the out of memory and the number of active records.
The text was updated successfully, but these errors were encountered: