You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When a new Splunk indexer pod is created (or an existing one is re-created after restart/reschedule), startup can become extremely slow if the attached PVC already contains a very large number of files (hundreds of thousands).
From what we observe, OpenShift performs a recursive ownership/permission pass (chown/chmod) on the PVC during mount, and this step becomes a big bottleneck with high file counts.
In practice, the pod sits a long time in the “volume config / mount” phase before the container properly starts, and we often see timeout-like symptoms around that stage, then the pod become not Ready due to the ReadinessProbe.
Expected behavior
Indexer pod creation/restart should complete in a reasonable timeframe, even when the PVC is already populated with a lot of files.
Splunk setup on K8S
Splunk Operator managed deployment.
Indexer pods using PVCs for indexer data (high file count / large data volumes).
This affects normal ops actions too (rolling upgrade, node maintenance, probes, etc.), not only initial deployment.
Reproduction/Testing steps
Use an indexer PVC already populated with a very high number of files (typical indexer data layout).
Trigger indexer pod re-creation (restart/reschedule/rolling upgrade).
Observe pod stuck or very slow during volume mount / container volume configuration.
Check events/logs: we frequently see errors like CRI-O / kubelet “context deadline exceeded” and “name is reserved” during startup.
In worst cases, time-to-ready can reach ~30 mins per pod.
K8s environment
OpenShift (OCP) cluster.
Issue happens during pod start, around PVC mount / volume preparation.
Frequency: indexer pods can be restarted multiple times per month, so the operational impact adds up quickly.
Proposed changes(optional)
None for now — filing as an issue because the current behavior causes major delays for indexer restart/recovery scenarios.
Please select the type of request
Enhancement
Tell us more
Describe the request
chown/chmod) on the PVC during mount, and this step becomes a big bottleneck with high file counts.Expected behavior
Splunk setup on K8S
Reproduction/Testing steps
K8s environment
Proposed changes(optional)
K8s collector data(optional)
Additional context(optional)