The core dump handler is designed to quickly move the cores "off-box" to an object storage environment with as much additional runtime information as possible. In order to provide the following benefits:
-
The developer who needs to debug the core doesn't require access to the kubernetes host environment and can access the crash as soon as it has happened.
-
Cores will not use disk space by residing on the host longer than required.
-
As Object Storage APIs have migrated to S3 as a defacto standard post processing services for scrubbers and indexing the data are easier to implement.
It's strongly recommened that you maintain the upload pattern of moving the cores off the machine but you may wish to move them to a none S3 compabible host.
This scenario is possible but the following aspects need consideration:
-
The upload functionality needs to be disabled by commenting out the
schedule
andinterval
properties and setting theuseINotify
to befalse
. See main.rs for the details.e.g. values.yaml
# interval: 60000 # schedule: "* * * * * *" useINotify: false
N.B. The interval and schedule MUST be removed as the JSON schema validation makes them mutually exclusive.
-
File Locks need to be honoured.
The core dump composer uses libc flock in the advisory-lock-rs project and the agent then trys to obtain a shared lock to ensure that the file isn't still being written to when an upload is attempted.
Your own uploader should respect these semantics.
-
Disable the S3 credentials validation by setting the
manageStoreSecret
in the values file to false.manageStoreSecret: true
-
Your custom uploader should probably be configured and deployed outside of the core-dump-handler in order to facilitate upgrading the project in your environment. But if there is a hard requirement to integrate into this tool then please add a note to the operator project.
-
To assist in the configuration of the you custom container consider using the pvc and pv configuration that is provided as part of this project but ensure that you change the names.
In some scenarios the core file can be truncated and/or the JSON info files may not be captured.
This is usually due to the default grace period during for pod termination being exceeded.
If this is a potential issue then set the terminationGracePeriodSeconds
option in the Pod YAML to help mitigate this.
e.g. To change it to 120 seconds
apiVersion: v1
kind: Pod
metadata:
name: my-pod
spec:
containers
- name: my-container
image: busybox
terminationGracePeriodSeconds: 120
Also see Kubernetes best practices: terminating with grace
This appears to be a bug in some kubernetes services.
You should also notice that the command kubectl logs --tail=500 YOUR_POD_ID
exhibits the same behaviour.
Current workaround is to double the line count on the configuration.
Some users run the agent in schedule mode to ensure that bandwidth during peak times isn't impacted. In this scenario you may wish to force an upload if there is a critical core that you need to get access to.
This can be achieved by logging into the agent container and executing the sweep
command.
kubectl exec -it -n observe core-dump-handler-gcvtc -- /bin/bash
./core-dump-agent sweep
By default the upload to S3 compatible storage is configured using the storage parameters outlined in the install documents. However you may wish to integrate an external secrets management system to lay out your secrets outside of this helm chart.
In that case disable this chart from requiring the secrets by setting manageStoreSecret to false in the values.yaml
.
manageStoreSecret: false
Or by passing the following option when you deploy the chart:
--set manageStoreSecret=false
Ensure that your secrets manager has layed out your secret as defined in the secrets.yaml template.
apiVersion: v1
kind: Secret
metadata:
name: s3config
type: Opaque
stringData:
s3Secret: {{ .Values.daemonset.s3Secret }}
s3AccessKey: {{ .Values.daemonset.s3AccessKey }}
s3BucketName: {{ .Values.daemonset.s3BucketName }}
s3Region: {{ .Values.daemonset.s3Region }}
The custom endpoint feature is available to connect to certain object stores that require it. The agent checks for the S3_ENDPOINT environment variable and if it's present it creates a Custom Region Struct with the value along with that of the s3Region.
The S3_ENDPOINT
environment variable is set using the extraEnvVars
property in values.yaml
e.g.
extraEnvVars: |
- name: S3_ENDPOINT
value: https://the-endpoint