[sheepdog] [PATCH v3] sheep: remove master node
Kai Zhang
kyle at zelin.io
Wed Jul 24 11:20:51 CEST 2013
On Jul 24, 2013, at 5:13 PM, Kai Zhang <kyle at zelin.io> wrote:
>
> On Jul 24, 2013, at 2:53 PM, MORITA Kazutaka <morita.kazutaka at lab.ntt.co.jp> wrote:
>
>> At Tue, 23 Jul 2013 17:30:03 +0800,
>> Kai Zhang wrote:
>>>
>>> On Jul 23, 2013, at 4:44 PM, MORITA Kazutaka <morita.kazutaka at lab.ntt.co.jp> wrote:
>>>
>>>> Ah, sorry. The node A doesn't start until the nodes B, C, and D come
>>>> back. It is because the latest epoch in the node A includes B, C, and
>>>> D. senerio
>>>
>>> Well, it seems I didn't fully understand the current implementation of cluster driver.
>>>
>>> A very silly question: if B, C come back but D does not, what is the status of
>>> the cluster? It can work or just wait for D?
>>
>> The cluster status will be SD_STATUS_WAIT. It will wait for the node
>> D to join Sheepdog if you don't run "collie cluster recover force".
>>
>
> Does this mean that sheepdog is not self-healing?
> Any persistent failure of sheep will be handled by administrator?
Sorry, my description is not correct.
What I mean is that sheepdog cluster cannot recover by themselves at this scenario.
And I'm a little disappointed with this.
Is there a possibility to solve this?
> If so, I think there is no need for auto-recover.
> Recover should happen when administrator call "collie cluster recover force".
>
> Thanks,
> Kyle
>
More information about the sheepdog
mailing list