Greetings,
Is this expected to not show any status?
❯ k get pxc mysql
NAME ENDPOINT STATUS PXC PROXYSQL HAPROXY AGE
mysql 3h18m
Cluster seems fine otherwise
❯ k get pods --selector app.kubernetes.io/instance=mysql
NAME READY STATUS RESTARTS AGE
mysql-haproxy-0 2/2 Running 0 3h19m
mysql-haproxy-1 2/2 Running 0 3h16m
mysql-haproxy-2 2/2 Running 0 3h16m
mysql-pxc-0 3/3 Running 0 3h19m
mysql-pxc-1 3/3 Running 0 3h17m
mysql-pxc-2 3/3 Running 0 3h13m
❯ k get pxc mysql -o yaml
apiVersion: pxc.percona.com/v1
kind: PerconaXtraDBCluster
metadata:
annotations:
kubectl.kubernetes.io/last-applied-configuration: |
{"apiVersion":"pxc.percona.com/v1-9-0","kind":"PerconaXtraDBCluster"}
meta.helm.sh/release-name: app
meta.helm.sh/release-namespace: men
creationTimestamp: "2021-10-27T11:44:40Z"
finalizers:
- delete-pxc-pods-in-order
- delete-proxysql-pvc
- delete-pxc-pvc
generation: 2
labels:
app.kubernetes.io/instance: app
app.kubernetes.io/managed-by: Helm
app.kubernetes.io/name: mysql
app.kubernetes.io/version: 1.9.0
helm.sh/chart: pxc-db-1.9.1
name: mysql
namespace: men
resourceVersion: "290700413"
uid: fedd4665-5166-4e8d-8fbd-27ef6a3fc457
spec:
allowUnsafeConfigurations: true
backup:
image: percona/percona-xtradb-cluster-operator:1.9.0-pxc5.7-backup
imagePullPolicy: Always
pitr:
enabled: true
storageName: scw-poc-percona-backups
timeBetweenUploads: 60
schedule:
- keep: 5
name: daily-backup
schedule: 0 0 * * *
storageName: scw-poc-percona-backups
storages:
fs-pvc:
podSecurityContext:
fsGroup: 1001
supplementalGroups:
- 1001
s3:
bucket: ""
credentialsSecret: ""
type: filesystem
volume:
persistentVolumeClaim:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 6Gi
scw-poc-percona-backups:
podSecurityContext:
fsGroup: 1001
supplementalGroups:
- 1001
s3:
bucket: XXXXXXXXXXXXXXXXXXXXXx
credentialsSecret: mysql-s3-XXXXXXXXXXXXXXXXXXXXXx
endpointUrl: https://s3.fr-par.scw.cloud
region: fr-par
type: s3
crVersion: 1.9.0
enableCRValidationWebhook: false
haproxy:
affinity:
antiAffinityTopologyKey: kubernetes.io/hostname
annotations:
percona.com/configuration-hash: d41d8cd98f00b204e9800998ecf8427e
enabled: true
envVarsSecret: mysql-env-vars-haproxy
gracePeriod: 30
image: percona/percona-xtradb-cluster-operator:1.9.0-haproxy
imagePullPolicy: Always
livenessProbes:
exec:
command:
- /usr/local/bin/liveness-check.sh
failureThreshold: 4
initialDelaySeconds: 60
periodSeconds: 30
successThreshold: 1
timeoutSeconds: 5
nodeSelector:
andjaro.io/system: "true"
podDisruptionBudget:
maxUnavailable: 1
readinessProbes:
exec:
command:
- /usr/local/bin/readiness-check.sh
failureThreshold: 3
initialDelaySeconds: 15
periodSeconds: 5
successThreshold: 1
timeoutSeconds: 1
resources:
limits: {}
requests:
cpu: 600m
memory: 1G
serviceAccountName: default
sidecarResources:
limits: {}
requests: {}
size: 3
tolerations:
- effect: NoSchedule
key: k8s.scaleway.com/dedicated
operator: Equal
value: system
volumeSpec:
emptyDir: {}
logCollectorSecretName: mysql-log-collector
logcollector:
enabled: true
image: percona/percona-xtradb-cluster-operator:1.9.0-logcollector
imagePullPolicy: Always
resources:
limits: {}
requests: {}
platform: kubernetes
pmm:
resources:
limits: {}
requests:
cpu: 600m
memory: 1G
proxysql:
livenessProbes: {}
podSecurityContext:
fsGroup: 1001
supplementalGroups:
- 1001
readinessProbes: {}
pxc:
affinity:
antiAffinityTopologyKey: kubernetes.io/hostname
annotations:
percona.com/configuration-hash: d41d8cd98f00b204e9800998ecf8427e
autoRecovery: true
envVarsSecret: mysql-env-vars-pxc
expose: {}
gracePeriod: 600
image: percona/percona-xtradb-cluster:5.7.34-31.51
imagePullPolicy: Always
livenessDelaySec: 300
livenessProbes:
exec:
command:
- /var/lib/mysql/liveness-check.sh
failureThreshold: 3
initialDelaySeconds: 300
periodSeconds: 10
successThreshold: 1
timeoutSeconds: 5
nodeSelector:
andjaro.io/system: "true"
podDisruptionBudget:
maxUnavailable: 1
podSecurityContext:
fsGroup: 1001
supplementalGroups:
- 1001
readinessDelaySec: 15
readinessProbes:
exec:
command:
- /var/lib/mysql/readiness-check.sh
failureThreshold: 5
initialDelaySeconds: 15
periodSeconds: 30
successThreshold: 1
timeoutSeconds: 15
resources:
limits: {}
requests:
cpu: 600m
memory: 1G
serviceAccountName: default
sidecarResources:
limits: {}
requests: {}
size: 3
sslInternalSecretName: mysql-ssl-internal
sslSecretName: mysql-ssl
tolerations:
- effect: NoSchedule
key: k8s.scaleway.com/dedicated
operator: Equal
value: system
vaultSecretName: mysql-vault
volumeSpec:
persistentVolumeClaim:
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 10Gi
storageClassName: scw-bssd
secretsName: mysql
updateStrategy: SmartUpdate
upgradeOptions:
apply: Never
schedule: 0 4 * * *
versionServiceEndpoint: https://check.percona.com
vaultSecretName: mysql-vault
Also this is affecting backups - error from operator
{"level":"error","ts":1635347117.5467057,"logger":"controller","msg":"Reconciler error","controller":"perconaxtradbclusterbackup-controller","name":"cron-mysql-scw-poc-percona-20211027000-372f8","namespace":"men","error":"failed to run backup: there are no ready PXC nodes","errorVerbose":"there are no ready PXC nodes\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/apis/pxc/v1.(*PerconaXtraDBCluster).CanBackup\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/apis/pxc/v1/pxc_types.go:1086\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup.(*ReconcilePerconaXtraDBClusterBackup).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup/controller.go:177\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:235\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:209\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:188\nk8s.io/apimachinery/pkg/util/wait.BackoffUntil.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:155\nk8s.io/apimachinery/pkg/util/wait.BackoffUntil\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:156\nk8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\nk8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:90\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1371\nfailed to run backup\ngithub.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup.(*ReconcilePerconaXtraDBClusterBackup).Reconcile\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/pkg/controller/pxcbackup/controller.go:178\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:235\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:209\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:188\nk8s.io/apimachinery/pkg/util/wait.BackoffUntil.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:155\nk8s.io/apimachinery/pkg/util/wait.BackoffUntil\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:156\nk8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\nk8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:90\nruntime.goexit\n\t/usr/local/go/src/runtime/asm_amd64.s:1371","stacktrace":"sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:237\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:209\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).worker\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:188\nk8s.io/apimachinery/pkg/util/wait.BackoffUntil.func1\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:155\nk8s.io/apimachinery/pkg/util/wait.BackoffUntil\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:156\nk8s.io/apimachinery/pkg/util/wait.JitterUntil\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:133\nk8s.io/apimachinery/pkg/util/wait.Until\n\t/go/src/github.com/percona/percona-xtradb-cluster-operator/vendor/k8s.io/apimachinery/pkg/util/wait/wait.go:90"}
Why status is not populated? Am I missing something?#
Thanks