On 10/21/2010 07:47 AM, Yuriy Kohut wrote: > As I know, fuse is ready to use filesystem. I would like to have block device mapped into existing VDI. Yes, the fuse fs exports all VDIs into a block devices. Not sure what the status of this work is though. Regards -steve > --- > Yura > > On Oct 21, 2010, at 5:33 PM, Steven Dake wrote: > >> Someone on the list was working on a fuse sheepdog daemon that exported >> the vdi's via the fs. >> >> Regards >> -steve >> >> On 10/21/2010 07:12 AM, Yuriy Kohut wrote: >>> Hi, >>> >>> Is there a way translate Sheepdog VDIs into block devices on the boxes which belong to the cluster without using iSCSI ? >>> --- >>> Yura >>> >>> On Oct 20, 2010, at 2:23 PM, Yuriy Kohut wrote: >>> >>>> One quick question. >>>> Is there any way to save/store created targets and logicalunits with 'tgtadm' ? >>>> They all are "lost" after machine reboot. >>>> >>>> --- >>>> Yura >>>> >>>> On Oct 19, 2010, at 6:40 PM, Yuriy Kohut wrote: >>>> >>>>> Got it working. >>>>> >>>>> The next step is to try all that on real hardware 3 node cluster. >>>>> >>>>> >>>>> Thank you for help. >>>>> --- >>>>> Yura >>>>> >>>>> On Oct 19, 2010, at 3:02 PM, MORITA Kazutaka wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> Your sheep.log says >>>>>> >>>>>> Oct 19 05:59:06 send_message(169) failed to send message, 2 >>>>>> >>>>>> This means that the sheep daemon failed to communicate with corosync. >>>>>> Unfortunatelly, I've never seen such an error... >>>>>> >>>>>> Try following things: >>>>>> - restart corosync daemon >>>>>> - disable iptable and restart corosync >>>>>> - disable selinux and restart corosync >>>>>> >>>>>> Did sheepdog work fine when you tested it on debian? >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Kazutaka >>>>>> >>>>>> On 2010/10/19 19:56, Yuriy Kohut wrote: >>>>>>> Attached. >>>>>>> >>>>>>> >>>>>>> >>>>>>> ------------------------------------------------------------------------ >>>>>>> >>>>>>> >>>>>>> >>>>>>> Please feel free to kick me if anything else required. >>>>>>> >>>>>>> --- >>>>>>> Yura >>>>>>> >>>>>>> On Oct 19, 2010, at 1:45 PM, MORITA Kazutaka wrote: >>>>>>> >>>>>>>> Could you send me a sheep.log in the store directory? >>>>>>>> It would be helpful for debugging. >>>>>>>> >>>>>>>> Kazutaka >>>>>>>> >>>>>>>> On 2010/10/19 19:16, Yuriy Kohut wrote: >>>>>>>>> The patch doesn't help. >>>>>>>>> >>>>>>>>> Probably I'm doing something wrong. but the following operation won't finish: >>>>>>>>> # tgtadm --op new --mode logicalunit --tid 1 --lun 1 -b test0 --bstype sheepdog >>>>>>>>> >>>>>>>>> >>>>>>>>> Attached please find the the operation/command strace log archived: >>>>>>>>> strace.log.tar.gz >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> ------------------------------------------------------------------------ >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Please advise. >>>>>>>>> >>>>>>>>> Thank you >>>>>>>>> --- >>>>>>>>> Yura >>>>>>>>> >>>>>>>>> On Oct 19, 2010, at 11:52 AM, Yuriy Kohut wrote: >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> Sure. I'll let you know results. >>>>>>>>>> >>>>>>>>>> Thank you. >>>>>>>>>> --- >>>>>>>>>> Yura >>>>>>>>>> >>>>>>>>>> On Oct 19, 2010, at 11:46 AM, MORITA Kazutaka wrote: >>>>>>>>>> >>>>>>>>>>> Hi, >>>>>>>>>>> >>>>>>>>>>> At Fri, 15 Oct 2010 17:33:18 +0300, >>>>>>>>>>> Yuriy Kohut wrote: >>>>>>>>>>>> One more new issue with TGTd. >>>>>>>>>>>> >>>>>>>>>>>> Initially we have one sheepdog vdi (on which we would like to create iscsi unit) and no tgt targets/units: >>>>>>>>>>>> [root at centos ~]# tgtadm --op show --mode target >>>>>>>>>>>> [root at centos ~]# collie vdi list >>>>>>>>>>>> name id size used shared creation time vdi id >>>>>>>>>>>> ------------------------------------------------------------------ >>>>>>>>>>>> test0 1 4.0 GB 4.0 GB 0.0 MB 2010-10-15 17:55 fd34af >>>>>>>>>>>> [root at centos ~]# >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Creating new target: >>>>>>>>>>>> [root at centos ~]# tgtadm --op new --mode target --tid 1 -T some.vps:disk0 >>>>>>>>>>>> [root at centos ~]# tgtadm --op show --mode target >>>>>>>>>>>> Target 1: some.vps:disk0 >>>>>>>>>>>> System information: >>>>>>>>>>>> Driver: iscsi >>>>>>>>>>>> State: ready >>>>>>>>>>>> I_T nexus information: >>>>>>>>>>>> LUN information: >>>>>>>>>>>> LUN: 0 >>>>>>>>>>>> Type: controller >>>>>>>>>>>> SCSI ID: IET 00010000 >>>>>>>>>>>> SCSI SN: beaf10 >>>>>>>>>>>> Size: 0 MB >>>>>>>>>>>> Online: Yes >>>>>>>>>>>> Removable media: No >>>>>>>>>>>> Readonly: No >>>>>>>>>>>> Backing store type: null >>>>>>>>>>>> Backing store path: None >>>>>>>>>>>> Backing store flags: >>>>>>>>>>>> Account information: >>>>>>>>>>>> ACL information: >>>>>>>>>>>> [root at centos ~]# >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Try to create new logicalunit on existing tgt target and sheepdog vdi: >>>>>>>>>>>> [root at centos ~]# tgtadm --op new --mode logicalunit --tid 1 --lun 1 -b test0 --bstype sheepdog >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> But the process never ends. >>>>>>>>>>>> Please advise ... >>>>>>>>>>> Thanks for your report. >>>>>>>>>>> >>>>>>>>>>> Can you try the following patch I sent minutes ago? >>>>>>>>>>> http://lists.wpkg.org/pipermail/sheepdog/2010-October/000741.html >>>>>>>>>>> >>>>>>>>>>> Thanks, >>>>>>>>>>> >>>>>>>>>>> Kazutaka >>>>>>>>>>> >>>>>>>>>>>> --- >>>>>>>>>>>> Yura >>>>>>>>>>>> >>>>>>>>>>>> On Oct 15, 2010, at 4:55 PM, Yuriy Kohut wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Cool, that works. >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks >>>>>>>>>>>>> --- >>>>>>>>>>>>> Yura >>>>>>>>>>>>> >>>>>>>>>>>>> On Oct 15, 2010, at 3:52 PM, MORITA Kazutaka wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> At Fri, 15 Oct 2010 13:38:16 +0300, >>>>>>>>>>>>>> Yuriy Kohut wrote: >>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I'm using the following 'Getting Started' guide to configure simple cluster: >>>>>>>>>>>>>>> http://www.osrg.net/sheepdog/usage.html >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I have configured cluster on 1 node/box, so the first questions are: >>>>>>>>>>>>>>> Can I configure cluster on single node (1 box) under CentOS 5.5 x86_64 ? >>>>>>>>>>>>>>> Is it required at least 3 nodes/boxes ... ? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I have faced with the following issue on my single-node cluster. I have rebooted the box after my first image creation. The following done for that: >>>>>>>>>>>>>>> - corosync is up and running >>>>>>>>>>>>>>> udp 0 0 192.168.128.195:5404 0.0.0.0:* 3541/corosync >>>>>>>>>>>>>>> udp 0 0 192.168.128.195:5405 0.0.0.0:* 3541/corosync >>>>>>>>>>>>>>> udp 0 0 226.94.1.1:5405 0.0.0.0:* 3541/corosync >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> - sheep is up and running >>>>>>>>>>>>>>> tcp 0 0 0.0.0.0:7000 0.0.0.0:* LISTEN 3561/sheep >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> - cluster is formatted with 1 copy only >>>>>>>>>>>>>>> #collie cluster format --copies=1 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> - the image with prelocated data is created >>>>>>>>>>>>>>> # qemu-img create sheepdog:test0 -o preallocation=data 4G >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> So after such siple steps I got: >>>>>>>>>>>>>>> # collie vdi list >>>>>>>>>>>>>>> name id size used shared creation time vdi id >>>>>>>>>>>>>>> ------------------------------------------------------------------ >>>>>>>>>>>>>>> test0 1 4.0 GB 4.0 GB 0.0 MB 2010-10-15 12:42 fd34af >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Then I rebooted the box, and no image(s) are available for me after box came back. The vdi list just show nothing: >>>>>>>>>>>>>>> # collie vdi list >>>>>>>>>>>>>>> name id size used shared creation time vdi id >>>>>>>>>>>>>>> ------------------------------------------------------------------ >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> and 'collie vdi list' never ends ... >>>>>>>>>>>>>>> corosync and sheep are still running. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Could somebody assist me with that. >>>>>>>>>>>>>> Sorry, the following patch will fixes the problem. >>>>>>>>>>>>>> >>>>>>>>>>>>>> = >>>>>>>>>>>>>> From: MORITA Kazutaka<morita.kazutaka at lab.ntt.co.jp> >>>>>>>>>>>>>> Subject: [PATCH] sheep: call start_recovery when cluster restarts with one node >>>>>>>>>>>>>> >>>>>>>>>>>>>> Sheepdog recovers objects before starting a storage service, and the >>>>>>>>>>>>>> routine is called when nodes are joined. However If sheepdog consists >>>>>>>>>>>>>> of only one node, no node doesn't send join messages, so >>>>>>>>>>>>>> start_recovery doesn't called. This patch fixes the problem. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Signed-off-by: MORITA Kazutaka<morita.kazutaka at lab.ntt.co.jp> >>>>>>>>>>>>>> --- >>>>>>>>>>>>>> sheep/group.c | 3 +++ >>>>>>>>>>>>>> 1 files changed, 3 insertions(+), 0 deletions(-) >>>>>>>>>>>>>> >>>>>>>>>>>>>> diff --git a/sheep/group.c b/sheep/group.c >>>>>>>>>>>>>> index ba8cdfb..86cbdb8 100644 >>>>>>>>>>>>>> --- a/sheep/group.c >>>>>>>>>>>>>> +++ b/sheep/group.c >>>>>>>>>>>>>> @@ -1226,6 +1226,9 @@ static void __sd_confchg_done(struct cpg_event *cevent) >>>>>>>>>>>>>> >>>>>>>>>>>>>> update_cluster_info(&msg); >>>>>>>>>>>>>> >>>>>>>>>>>>>> + if (sys->status == SD_STATUS_OK) /* sheepdog starts with one node */ >>>>>>>>>>>>>> + start_recovery(sys->epoch, NULL, 0); >>>>>>>>>>>>>> + >>>>>>>>>>>>>> return; >>>>>>>>>>>>>> } >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> 1.5.6.5 >>>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> sheepdog mailing list >>>>>>>>>>>>> sheepdog at lists.wpkg.org >>>>>>>>>>>>> http://lists.wpkg.org/mailman/listinfo/sheepdog >>>>>>>>>>>> -- >>>>>>>>>>>> sheepdog mailing list >>>>>>>>>>>> sheepdog at lists.wpkg.org >>>>>>>>>>>> http://lists.wpkg.org/mailman/listinfo/sheepdog >>>>>>>>>> -- >>>>>>>>>> sheepdog mailing list >>>>>>>>>> sheepdog at lists.wpkg.org >>>>>>>>>> http://lists.wpkg.org/mailman/listinfo/sheepdog >>>>>>> >>>>> >>>> >>>> -- >>>> sheepdog mailing list >>>> sheepdog at lists.wpkg.org >>>> http://lists.wpkg.org/mailman/listinfo/sheepdog >>> >> > |