As I know, fuse is ready to use filesystem. I would like to have block device mapped into existing VDI. --- Yura On Oct 21, 2010, at 5:33 PM, Steven Dake wrote: > Someone on the list was working on a fuse sheepdog daemon that exported > the vdi's via the fs. > > Regards > -steve > > On 10/21/2010 07:12 AM, Yuriy Kohut wrote: >> Hi, >> >> Is there a way translate Sheepdog VDIs into block devices on the boxes which belong to the cluster without using iSCSI ? >> --- >> Yura >> >> On Oct 20, 2010, at 2:23 PM, Yuriy Kohut wrote: >> >>> One quick question. >>> Is there any way to save/store created targets and logicalunits with 'tgtadm' ? >>> They all are "lost" after machine reboot. >>> >>> --- >>> Yura >>> >>> On Oct 19, 2010, at 6:40 PM, Yuriy Kohut wrote: >>> >>>> Got it working. >>>> >>>> The next step is to try all that on real hardware 3 node cluster. >>>> >>>> >>>> Thank you for help. >>>> --- >>>> Yura >>>> >>>> On Oct 19, 2010, at 3:02 PM, MORITA Kazutaka wrote: >>>> >>>>> Hi, >>>>> >>>>> Your sheep.log says >>>>> >>>>> Oct 19 05:59:06 send_message(169) failed to send message, 2 >>>>> >>>>> This means that the sheep daemon failed to communicate with corosync. >>>>> Unfortunatelly, I've never seen such an error... >>>>> >>>>> Try following things: >>>>> - restart corosync daemon >>>>> - disable iptable and restart corosync >>>>> - disable selinux and restart corosync >>>>> >>>>> Did sheepdog work fine when you tested it on debian? >>>>> >>>>> Thanks, >>>>> >>>>> Kazutaka >>>>> >>>>> On 2010/10/19 19:56, Yuriy Kohut wrote: >>>>>> Attached. >>>>>> >>>>>> >>>>>> >>>>>> ------------------------------------------------------------------------ >>>>>> >>>>>> >>>>>> >>>>>> Please feel free to kick me if anything else required. >>>>>> >>>>>> --- >>>>>> Yura >>>>>> >>>>>> On Oct 19, 2010, at 1:45 PM, MORITA Kazutaka wrote: >>>>>> >>>>>>> Could you send me a sheep.log in the store directory? >>>>>>> It would be helpful for debugging. >>>>>>> >>>>>>> Kazutaka >>>>>>> >>>>>>> On 2010/10/19 19:16, Yuriy Kohut wrote: >>>>>>>> The patch doesn't help. >>>>>>>> >>>>>>>> Probably I'm doing something wrong. but the following operation won't finish: >>>>>>>> # tgtadm --op new --mode logicalunit --tid 1 --lun 1 -b test0 --bstype sheepdog >>>>>>>> >>>>>>>> >>>>>>>> Attached please find the the operation/command strace log archived: >>>>>>>> strace.log.tar.gz >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> ------------------------------------------------------------------------ >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Please advise. >>>>>>>> >>>>>>>> Thank you >>>>>>>> --- >>>>>>>> Yura >>>>>>>> >>>>>>>> On Oct 19, 2010, at 11:52 AM, Yuriy Kohut wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> Sure. I'll let you know results. >>>>>>>>> >>>>>>>>> Thank you. >>>>>>>>> --- >>>>>>>>> Yura >>>>>>>>> >>>>>>>>> On Oct 19, 2010, at 11:46 AM, MORITA Kazutaka wrote: >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> At Fri, 15 Oct 2010 17:33:18 +0300, >>>>>>>>>> Yuriy Kohut wrote: >>>>>>>>>>> One more new issue with TGTd. >>>>>>>>>>> >>>>>>>>>>> Initially we have one sheepdog vdi (on which we would like to create iscsi unit) and no tgt targets/units: >>>>>>>>>>> [root at centos ~]# tgtadm --op show --mode target >>>>>>>>>>> [root at centos ~]# collie vdi list >>>>>>>>>>> name id size used shared creation time vdi id >>>>>>>>>>> ------------------------------------------------------------------ >>>>>>>>>>> test0 1 4.0 GB 4.0 GB 0.0 MB 2010-10-15 17:55 fd34af >>>>>>>>>>> [root at centos ~]# >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Creating new target: >>>>>>>>>>> [root at centos ~]# tgtadm --op new --mode target --tid 1 -T some.vps:disk0 >>>>>>>>>>> [root at centos ~]# tgtadm --op show --mode target >>>>>>>>>>> Target 1: some.vps:disk0 >>>>>>>>>>> System information: >>>>>>>>>>> Driver: iscsi >>>>>>>>>>> State: ready >>>>>>>>>>> I_T nexus information: >>>>>>>>>>> LUN information: >>>>>>>>>>> LUN: 0 >>>>>>>>>>> Type: controller >>>>>>>>>>> SCSI ID: IET 00010000 >>>>>>>>>>> SCSI SN: beaf10 >>>>>>>>>>> Size: 0 MB >>>>>>>>>>> Online: Yes >>>>>>>>>>> Removable media: No >>>>>>>>>>> Readonly: No >>>>>>>>>>> Backing store type: null >>>>>>>>>>> Backing store path: None >>>>>>>>>>> Backing store flags: >>>>>>>>>>> Account information: >>>>>>>>>>> ACL information: >>>>>>>>>>> [root at centos ~]# >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Try to create new logicalunit on existing tgt target and sheepdog vdi: >>>>>>>>>>> [root at centos ~]# tgtadm --op new --mode logicalunit --tid 1 --lun 1 -b test0 --bstype sheepdog >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> But the process never ends. >>>>>>>>>>> Please advise ... >>>>>>>>>> Thanks for your report. >>>>>>>>>> >>>>>>>>>> Can you try the following patch I sent minutes ago? >>>>>>>>>> http://lists.wpkg.org/pipermail/sheepdog/2010-October/000741.html >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> >>>>>>>>>> Kazutaka >>>>>>>>>> >>>>>>>>>>> --- >>>>>>>>>>> Yura >>>>>>>>>>> >>>>>>>>>>> On Oct 15, 2010, at 4:55 PM, Yuriy Kohut wrote: >>>>>>>>>>> >>>>>>>>>>>> Cool, that works. >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> --- >>>>>>>>>>>> Yura >>>>>>>>>>>> >>>>>>>>>>>> On Oct 15, 2010, at 3:52 PM, MORITA Kazutaka wrote: >>>>>>>>>>>> >>>>>>>>>>>>> At Fri, 15 Oct 2010 13:38:16 +0300, >>>>>>>>>>>>> Yuriy Kohut wrote: >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I'm using the following 'Getting Started' guide to configure simple cluster: >>>>>>>>>>>>>> http://www.osrg.net/sheepdog/usage.html >>>>>>>>>>>>>> >>>>>>>>>>>>>> I have configured cluster on 1 node/box, so the first questions are: >>>>>>>>>>>>>> Can I configure cluster on single node (1 box) under CentOS 5.5 x86_64 ? >>>>>>>>>>>>>> Is it required at least 3 nodes/boxes ... ? >>>>>>>>>>>>>> >>>>>>>>>>>>>> I have faced with the following issue on my single-node cluster. I have rebooted the box after my first image creation. The following done for that: >>>>>>>>>>>>>> - corosync is up and running >>>>>>>>>>>>>> udp 0 0 192.168.128.195:5404 0.0.0.0:* 3541/corosync >>>>>>>>>>>>>> udp 0 0 192.168.128.195:5405 0.0.0.0:* 3541/corosync >>>>>>>>>>>>>> udp 0 0 226.94.1.1:5405 0.0.0.0:* 3541/corosync >>>>>>>>>>>>>> >>>>>>>>>>>>>> - sheep is up and running >>>>>>>>>>>>>> tcp 0 0 0.0.0.0:7000 0.0.0.0:* LISTEN 3561/sheep >>>>>>>>>>>>>> >>>>>>>>>>>>>> - cluster is formatted with 1 copy only >>>>>>>>>>>>>> #collie cluster format --copies=1 >>>>>>>>>>>>>> >>>>>>>>>>>>>> - the image with prelocated data is created >>>>>>>>>>>>>> # qemu-img create sheepdog:test0 -o preallocation=data 4G >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> So after such siple steps I got: >>>>>>>>>>>>>> # collie vdi list >>>>>>>>>>>>>> name id size used shared creation time vdi id >>>>>>>>>>>>>> ------------------------------------------------------------------ >>>>>>>>>>>>>> test0 1 4.0 GB 4.0 GB 0.0 MB 2010-10-15 12:42 fd34af >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Then I rebooted the box, and no image(s) are available for me after box came back. The vdi list just show nothing: >>>>>>>>>>>>>> # collie vdi list >>>>>>>>>>>>>> name id size used shared creation time vdi id >>>>>>>>>>>>>> ------------------------------------------------------------------ >>>>>>>>>>>>>> >>>>>>>>>>>>>> and 'collie vdi list' never ends ... >>>>>>>>>>>>>> corosync and sheep are still running. >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Could somebody assist me with that. >>>>>>>>>>>>> Sorry, the following patch will fixes the problem. >>>>>>>>>>>>> >>>>>>>>>>>>> = >>>>>>>>>>>>> From: MORITA Kazutaka<morita.kazutaka at lab.ntt.co.jp> >>>>>>>>>>>>> Subject: [PATCH] sheep: call start_recovery when cluster restarts with one node >>>>>>>>>>>>> >>>>>>>>>>>>> Sheepdog recovers objects before starting a storage service, and the >>>>>>>>>>>>> routine is called when nodes are joined. However If sheepdog consists >>>>>>>>>>>>> of only one node, no node doesn't send join messages, so >>>>>>>>>>>>> start_recovery doesn't called. This patch fixes the problem. >>>>>>>>>>>>> >>>>>>>>>>>>> Signed-off-by: MORITA Kazutaka<morita.kazutaka at lab.ntt.co.jp> >>>>>>>>>>>>> --- >>>>>>>>>>>>> sheep/group.c | 3 +++ >>>>>>>>>>>>> 1 files changed, 3 insertions(+), 0 deletions(-) >>>>>>>>>>>>> >>>>>>>>>>>>> diff --git a/sheep/group.c b/sheep/group.c >>>>>>>>>>>>> index ba8cdfb..86cbdb8 100644 >>>>>>>>>>>>> --- a/sheep/group.c >>>>>>>>>>>>> +++ b/sheep/group.c >>>>>>>>>>>>> @@ -1226,6 +1226,9 @@ static void __sd_confchg_done(struct cpg_event *cevent) >>>>>>>>>>>>> >>>>>>>>>>>>> update_cluster_info(&msg); >>>>>>>>>>>>> >>>>>>>>>>>>> + if (sys->status == SD_STATUS_OK) /* sheepdog starts with one node */ >>>>>>>>>>>>> + start_recovery(sys->epoch, NULL, 0); >>>>>>>>>>>>> + >>>>>>>>>>>>> return; >>>>>>>>>>>>> } >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> 1.5.6.5 >>>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> sheepdog mailing list >>>>>>>>>>>> sheepdog at lists.wpkg.org >>>>>>>>>>>> http://lists.wpkg.org/mailman/listinfo/sheepdog >>>>>>>>>>> -- >>>>>>>>>>> sheepdog mailing list >>>>>>>>>>> sheepdog at lists.wpkg.org >>>>>>>>>>> http://lists.wpkg.org/mailman/listinfo/sheepdog >>>>>>>>> -- >>>>>>>>> sheepdog mailing list >>>>>>>>> sheepdog at lists.wpkg.org >>>>>>>>> http://lists.wpkg.org/mailman/listinfo/sheepdog >>>>>> >>>> >>> >>> -- >>> sheepdog mailing list >>> sheepdog at lists.wpkg.org >>> http://lists.wpkg.org/mailman/listinfo/sheepdog >> > |