Ok I was able to reproduce, I believe this happens when booting a VM
with a sheepdog volume fails (because the sheep daemon was down). Here's
the output:
<br>
<br>node 1:<br># sheep -f /data/sheep/<br>sheep: jrnl_recover(2221) Openning the directory /data/sheep//journal/00000009/.<br>sheep: set_addr(1595) addr = 172.16.1.1, port = 7000<br>sheep: main(144) Sheepdog daemon (version 0.2.3) started<br>
sheep: get_cluster_status(403) sheepdog is waiting with older epoch, 10 9 <a href="http://172.16.1.2:7000">172.16.1.2:7000</a><br><br>node 2:<br># sheep -f /data/sheep/<br>sheep: jrnl_recover(2221) Openning the directory /data/sheep//journal/00000010/.<br>
sheep: set_addr(1595) addr = 172.16.1.2, port = 7000<br>sheep: main(144) Sheepdog daemon (version 0.2.3) started<br>sheep: send_join_request(1048) 33624236 22579<br>sheep: update_cluster_info(568) failed to join sheepdog, 65<br>
<br># collie cluster info -a 172.16.1.1
<br>Waiting for other nodes joining
<br>
<br>Ctime Epoch Nodes
<br>2011-06-15 11:50:16 9 [<a href="http://172.16.1.1:7000">172.16.1.1:7000</a>, <a href="http://172.16.1.2:7000">172.16.1.2:7000</a>]
<br># collie cluster info -a 172.16.1.2
<br>The node had failed to join sheepdog
<br>
<br>Ctime Epoch Nodes
<br>2011-06-15 11:50:16 10 [<a href="http://172.16.1.2:7000">172.16.1.2:7000</a>]
<br>