[Sheepdog] Configuring simple cluster on CentOS 5.5 x86_64

Yuriy Kohut ykohut at onapp.com
Tue Oct 19 10:52:28 CEST 2010


Hi,

Sure. I'll let you know results.

Thank you.
---
Yura

On Oct 19, 2010, at 11:46 AM, MORITA Kazutaka wrote:

> Hi,
> 
> At Fri, 15 Oct 2010 17:33:18 +0300,
> Yuriy Kohut wrote:
>> 
>> One more new issue with TGTd.
>> 
>> Initially we have one sheepdog vdi (on which we would like to create iscsi unit) and no tgt targets/units:
>> [root at centos ~]# tgtadm --op show --mode target
>> [root at centos ~]# collie vdi list
>>  name        id    size    used  shared    creation time   vdi id
>> ------------------------------------------------------------------
>>  test0        1  4.0 GB  4.0 GB  0.0 MB 2010-10-15 17:55   fd34af
>> [root at centos ~]#  
>> 
>> 
>> Creating new target:
>> [root at centos ~]# tgtadm --op new --mode target --tid 1 -T some.vps:disk0
>> [root at centos ~]# tgtadm --op show --mode target
>> Target 1: some.vps:disk0
>>    System information:
>>        Driver: iscsi
>>        State: ready
>>    I_T nexus information:
>>    LUN information:
>>        LUN: 0
>>            Type: controller
>>            SCSI ID: IET     00010000
>>            SCSI SN: beaf10
>>            Size: 0 MB
>>            Online: Yes
>>            Removable media: No
>>            Readonly: No
>>            Backing store type: null
>>            Backing store path: None
>>            Backing store flags: 
>>    Account information:
>>    ACL information:
>> [root at centos ~]#
>> 
>> 
>> Try to create new logicalunit on existing tgt target and sheepdog vdi:
>> [root at centos ~]# tgtadm --op new --mode logicalunit --tid 1 --lun 1 -b test0 --bstype sheepdog
>> 
>> 
>> But the process never ends.
>> Please advise ...
> 
> Thanks for your report.
> 
> Can you try the following patch I sent minutes ago?
>  http://lists.wpkg.org/pipermail/sheepdog/2010-October/000741.html
> 
> Thanks,
> 
> Kazutaka
> 
>> ---
>> Yura
>> 
>> On Oct 15, 2010, at 4:55 PM, Yuriy Kohut wrote:
>> 
>>> Cool, that works.
>>> 
>>> Thanks
>>> ---
>>> Yura
>>> 
>>> On Oct 15, 2010, at 3:52 PM, MORITA Kazutaka wrote:
>>> 
>>>> At Fri, 15 Oct 2010 13:38:16 +0300,
>>>> Yuriy Kohut wrote:
>>>>> 
>>>>> Hi,
>>>>> 
>>>>> I'm using the following 'Getting Started' guide to configure simple cluster:
>>>>> http://www.osrg.net/sheepdog/usage.html
>>>>> 
>>>>> I have configured cluster on 1 node/box, so the first questions are:
>>>>> Can I configure cluster on single node (1 box) under CentOS 5.5 x86_64 ? 
>>>>> Is it required at least 3 nodes/boxes ... ?
>>>>> 
>>>>> I have faced with the following issue on my single-node cluster. I have rebooted the box after my first image creation. The following done for that:
>>>>> - corosync is up and running
>>>>> udp        0      0 192.168.128.195:5404        0.0.0.0:*                               3541/corosync       
>>>>> udp        0      0 192.168.128.195:5405        0.0.0.0:*                               3541/corosync       
>>>>> udp        0      0 226.94.1.1:5405             0.0.0.0:*                               3541/corosync 
>>>>> 
>>>>> - sheep is up and running
>>>>> tcp        0      0 0.0.0.0:7000                0.0.0.0:*                   LISTEN      3561/sheep
>>>>> 
>>>>> - cluster is formatted with 1 copy only
>>>>> #collie cluster format --copies=1
>>>>> 
>>>>> - the image with prelocated data is created
>>>>> # qemu-img create sheepdog:test0 -o preallocation=data 4G
>>>>> 
>>>>> 
>>>>> So after such siple steps I got:
>>>>> # collie vdi list
>>>>> name        id    size    used  shared    creation time   vdi id
>>>>> ------------------------------------------------------------------
>>>>> test0        1  4.0 GB  4.0 GB  0.0 MB 2010-10-15 12:42   fd34af
>>>>> 
>>>>> 
>>>>> Then I rebooted the box, and no image(s) are available for me after box came back. The vdi list just show nothing:
>>>>> # collie vdi list
>>>>> name        id    size    used  shared    creation time   vdi id
>>>>> ------------------------------------------------------------------
>>>>> 
>>>>> and 'collie vdi list' never ends ...
>>>>> corosync and sheep are still running.
>>>>> 
>>>>> 
>>>>> Could somebody assist me with that.
>>>> 
>>>> Sorry, the following patch will fixes the problem.
>>>> 
>>>> =
>>>> From: MORITA Kazutaka <morita.kazutaka at lab.ntt.co.jp>
>>>> Subject: [PATCH] sheep: call start_recovery when cluster restarts with one node
>>>> 
>>>> Sheepdog recovers objects before starting a storage service, and the
>>>> routine is called when nodes are joined.  However If sheepdog consists
>>>> of only one node, no node doesn't send join messages, so
>>>> start_recovery doesn't called.  This patch fixes the problem.
>>>> 
>>>> Signed-off-by: MORITA Kazutaka <morita.kazutaka at lab.ntt.co.jp>
>>>> ---
>>>> sheep/group.c |    3 +++
>>>> 1 files changed, 3 insertions(+), 0 deletions(-)
>>>> 
>>>> diff --git a/sheep/group.c b/sheep/group.c
>>>> index ba8cdfb..86cbdb8 100644
>>>> --- a/sheep/group.c
>>>> +++ b/sheep/group.c
>>>> @@ -1226,6 +1226,9 @@ static void __sd_confchg_done(struct cpg_event *cevent)
>>>> 
>>>> 		update_cluster_info(&msg);
>>>> 
>>>> +		if (sys->status == SD_STATUS_OK) /* sheepdog starts with one node */
>>>> +			start_recovery(sys->epoch, NULL, 0);
>>>> +
>>>> 		return;
>>>> 	}
>>>> 
>>>> -- 
>>>> 1.5.6.5
>>>> 
>>> 
>>> -- 
>>> sheepdog mailing list
>>> sheepdog at lists.wpkg.org
>>> http://lists.wpkg.org/mailman/listinfo/sheepdog
>> 
>> -- 
>> sheepdog mailing list
>> sheepdog at lists.wpkg.org
>> http://lists.wpkg.org/mailman/listinfo/sheepdog




More information about the sheepdog mailing list