-
Notifications
You must be signed in to change notification settings - Fork 16.8k
[stable/redis] Recovered master pod fail to add back the original replication group when sentinel enabled #17244
Comments
Hi, I was unable to reproduce the issue. I scale down the master
Then one of the slaves got elected
Then I restored the master
And the master is a slave
Could you provide more details of your environment? |
I tried again and the problem still exists. Please find the details below and let me know if more information is needed. Thanks! Reproduce stepsWhen the master is just started.
I down scale the master to 0
One of the slaves becomes master as expected.
I scale up the master back to 1
The master won't join back to the original replication group
My deployment details
|
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Any further update will cause the issue/pull request to no longer be considered stale. Thank you for your contributions. |
This issue is being automatically closed due to inactivity. |
@carmenlau have you been able to fix this, we face the same issue :( |
Glad that I found you guys are facing the same issue. Especially since statefulsets can't set the restartpolicy to 'Never'. So everytime you even scale the nodegroup that the master is running on, you will face a new replication group. I'll be following this one, as this has been bugging me for quite sometime. |
I have not tried the new version chart yet, may try it later. So I still have the problem. Which version are you guys using? |
redis-10.6.13 @ AWS EKS. Latest Bitnami chart. We fetch the chart, modify production-values.yml to enable sentinel to true and network policy enabled. Rollout goes fine, master and slaves are connected properly. then:
Both master and slaves have this in their sentinel.conf: $ cat mounted-etc/sentinel.conf but they are all connecting to each other over IP (seems like it, no hostnames are mentioned in the logging), which I think is the root cause. |
I'm facing this issue. I delete the master redis node/pod and it doesn't connect to the new master node. Any new suggestions/workarounds? |
hey, sadly still facing same issue if some one has any input or suggestions will be greatly appreciated. fyi using v8.0.12 |
Describe the bug
When both cluster and sentinel are enabled. The recovered master pod fail to add back to the original cluster, it forms another replication group with new
master_replid
instead.Which chart:
stable/redis version 9.1.5
What happened:
stable/redis
with both cluster and sentinel enabled.master_replid
. There will be 2 replication groups in the cluster, the first one is the original replication group with new master (pod of slave statefulset), the second one is a new replication group with 1 master (pod of master statefulset) and no slave.What you expected to happen:
How to reproduce it (as minimally and precisely as possible):
info replication
of redis nodes.The text was updated successfully, but these errors were encountered: