Skip to content

Replicas are never catch up after pod is recreated #588

Open
@ynnt

Description

@ynnt

Hello!
I have an issue with mysql cluster not catching up after failover happens.

Cluster spec:

  spec:
    image: percona:5.7.29-centos
    podSpec:
      metricsExporterResources: {}
      mysqlOperatorSidecarResources: {}
      resources: {}
    replicas: 2
    secretName: my-cluster-secret
    volumeSpec:
      persistentVolumeClaim:
        accessModes:
        - ReadWriteOnce
        resources:
          requests:
            storage: 1Gi

After I delete mysql-0 pod, mysql-1 starts complaining it can't sync with it:

//my-cluster-mysql-0.mysql.default:3306' - retry-time: 1  retries: 162, Error_code: 2005
2020-08-13T16:37:48.525821Z 939 [ERROR] Slave I/O for channel '': error connecting to master 'sys_replication@//my-cluster-mysql-0.mysql.default:3306' - retry-time: 1  retries: 163, Error_code: 2005
2020-08-13T16:37:49.526148Z 939 [ERROR] Slave I/O for channel '': error connecting to master 'sys_replication@//my-cluster-mysql-0.mysql.default:3306' - retry-time: 1  retries: 164, Error_code: 2005

Even when mysql-0 is recreated and successfully starts replication connection to mysql-1.

I wonder if using a pod name like my-cluster-mysql-0.mysql.default is an issue? Its IP address changes after pod is recreated.
Is it possible to experiment using services per each mysql pod for replication connection?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions