[sheepdog-users] hyper volume issue

Hitoshi Mitake mitake.hitoshi at gmail.com
Mon Sep 21 16:32:38 CEST 2015


On Mon, Sep 21, 2015 at 10:57 PM, Glen Aidukas
<GAidukas at behaviormatrix.com> wrote:
> Hitoshi,
>
> In a  few days I will be setting up a new test cluster with three nodes, 1 journal and 3 data drives each and also 10gbit networking.  I will test with the new branch then.

Please never use journaling, it doesn't contribute to performance and
degrades stability. It will be removed in the future.

>
> What is the effect of using  '-z 23' ?

-z is the block size shift. -z 23 means a newly created VDI will have
1 << 23 B (8MB) object. It will solve the 4TB limitation.

Thanks,
Hitoshi

>
> Regards,
>
> Glen
>
>
> -----Original Message-----
> From: Hitoshi Mitake [mailto:mitake.hitoshi at gmail.com]
> Sent: Monday, September 21, 2015 4:10 AM
> To: Glen Aidukas
> Cc: sheepdog-users (sheepdog-users at lists.wpkg.org)
> Subject: Re: [sheepdog-users] hyper volume issue
>
> Hi Glen,
>
> hypervolume isn't a feature for VDIs, it is only used for http object store.
>
> Could you try below command line with master branch:
> $ dog vdi create <vdi name> 5T -z 23
>
> The master branch supports VDIs larger than 4TB. The feature isn't backported to 0.9.x yet.
>
> Thanks,
> Hitoshi
>
>
> On Fri, Sep 18, 2015 at 10:29 PM, Glen Aidukas <GAidukas at behaviormatrix.com> wrote:
>> Hello,
>>
>>
>>
>> I’m seeing an issue with hyper volumes on my two node test cluster
>> using v0.9.2.  While performing some tests, I tried to create a 5TB
>> vdi and saw that normal volumes are limited to 4TB and that I needed
>> to use the –y switch to create a Hyper Volume.  This worked when
>> creating it but then was not able to start my vm while connecting to
>> it.  I then tried to delete the hyper volume and the cluster then
>> crashed.  When I restarted the cluster I was still having issues.  I
>> was able to reformat the cluster, reimport my test vms and all was working again.
>>
>>
>>
>> To test and see if  trying to use the hv vdi is what caused the issue
>> I then created a 5TB hv-vdi and then delete it and I got the same issue.
>>
>>
>>
>> -Glen
>>
>>
>>
>> Here is a recreation of the issue:
>>
>>
>>
>> Take special note to the log file showing:
>>
>> Sep 18 09:08:25  EMERG [deletion] traverse_btree(190) PANIC: This
>> B-tree not support depth 0?
>>
>>
>>
>>
>>
>> root at pmsd-a01:~# sheep -v
>>
>> Sheepdog daemon version 0.9.2
>>
>>
>>
>> root at pmsd-a01:~# dog node md info --all
>>
>> Id      Size    Used    Avail   Use%    Path
>>
>> Node 0:
>>
>> 0      1.8 TB  10 GB   1.8 TB    0%    /var/lib/sheepdog//disc1
>>
>> 1      1.8 TB  9.7 GB  1.8 TB    0%    /var/lib/sheepdog//disc2
>>
>> 2      1.8 TB  10 GB   1.8 TB    0%    /var/lib/sheepdog//disc3
>>
>> Node 1:
>>
>> 0      1.8 TB  10 GB   1.8 TB    0%    /var/lib/sheepdog//disc1
>>
>> 1      1.8 TB  9.7 GB  1.8 TB    0%    /var/lib/sheepdog//disc2
>>
>> 2      1.8 TB  10 GB   1.8 TB    0%    /var/lib/sheepdog//disc3
>>
>>
>>
>>
>>
>> root at pmsd-a01:~# dog node info
>>
>> Id      Size    Used    Avail   Use%
>>
>> 0      5.4 TB  30 GB   5.3 TB    0%
>>
>> 1      5.4 TB  30 GB   5.3 TB    0%
>>
>> Total   11 TB   60 GB   11 TB     0%
>>
>>
>>
>> Total virtual image size        512 GB
>>
>>
>>
>> root at pmsd-a01:~# dog vdi list
>>
>>   Name        Id    Size    Used  Shared    Creation time   VDI id  Copies
>> Tag
>>
>>   vm-60153-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 11:13   2e24f0
>> 2
>>
>>   vm-60154-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 11:24   57e86b
>> 2
>>
>>   vm-60152-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 11:00   794d35
>> 2
>>
>>   vm-60151-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 10:35   c286fb
>> 2
>>
>>
>>
>> root at pmsd-a01:~# dog vdi create test-5tb-hv 5T -y
>>
>> root at pmsd-a01:~# dog vdi list
>>
>>   Name        Id    Size    Used  Shared    Creation time   VDI id  Copies
>> Tag
>>
>>   vm-60153-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 11:13   2e24f0
>> 2
>>
>>   test-5tb-hv         0  5.0 TB  0.0 MB  0.0 MB 2015-09-18 09:07   3af976
>> 2
>>
>>   vm-60154-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 11:24   57e86b
>> 2
>>
>>   vm-60152-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 11:00   794d35
>> 2
>>
>>   vm-60151-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 10:35   c286fb
>> 2
>>
>>
>>
>>
>>
>> root at pmsd-a01:~# dog vdi delete test-5tb-hv
>>
>> failed to read a response
>>
>>
>>
>>
>>
>> root at pmsd-a01:~# dog node info
>>
>> failed to connect to 127.0.0.1:7000: Connection refused
>>
>> failed to connect to 127.0.0.1:7000: Connection refused
>>
>> Failed to get node list
>>
>>
>>
>> root at pmsd-a01:~# service sheepdog restart
>>
>> Restarting Sheepdog Server: sheepdog.
>>
>>
>>
>> root at pmsd-a02:~# service sheepdog restart              (on second node)
>>
>> Restarting Sheepdog Server: sheepdog.
>>
>>
>>
>>
>>
>> root at pmsd-a01:~# dog vdi list
>>
>>
>>
>> PANIC: Depth of B-tree is out of range(depth: 0)
>>
>>   Name        Id    Size    Used  Shared    Creation time   VDI id  Copies
>> Tag
>>
>>   vm-60153-disk-1     0  128 GB  7.5 GB  0.0 MB 2015-09-17 11:13   2e24f0
>> 2
>>
>> dog exits unexpectedly (Aborted).
>>
>> sh: 1: addr2line: not found
>>
>> :
>>
>> sh: 1: addr2line: not found
>>
>> :
>>
>> sh: 1: addr2line: not found
>>
>> :
>>
>> sh: 1: addr2line: not found
>>
>>
>>
>>
>>
>> root at pmsd-a01:~# less sheedog.log
>>
>> Sep 18 09:07:52   INFO [main] rx_main(830) req=0x7f1d64032760, fd=189,
>> client=127.0.0.1:59938, op=NEW_VDI, data=(not string)
>>
>> Sep 18 09:07:52   INFO [main] tx_main(882) req=0x7f1d64032760, fd=189,
>> client=127.0.0.1:59938, op=NEW_VDI, result=00
>>
>> Sep 18 09:08:25   INFO [main] rx_main(830) req=0x7f1d6406aa40, fd=189,
>> client=127.0.0.1:59970, op=DEL_VDI, data=(not string)
>>
>> Sep 18 09:08:25  EMERG [deletion] traverse_btree(190) PANIC: This
>> B-tree not support depth 0
>>
>> Sep 18 09:08:25  EMERG [deletion] crash_handler(268) sheep exits
>> unexpectedly (Aborted).
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /usr/sbin/sheep(+0xb19b) [0x7f1d7c55319b]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /lib/x86_64-linux-gnu/libpthread.so.0(+0xf09f) [0x7f1d7c10f09f]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /lib/x86_64-linux-gnu/libc.so.6(gsignal+0x34) [0x7f1d7b2e2164]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /lib/x86_64-linux-gnu/libc.so.6(abort+0x17f) [0x7f1d7b2e53df]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /usr/sbin/sheep(+0x41566) [0x7f1d7c589566]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /usr/sbin/sheep(+0x16415) [0x7f1d7c55e415]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /usr/sbin/sheep(+0x3daba) [0x7f1d7c585aba]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /lib/x86_64-linux-gnu/libpthread.so.0(+0x6b4f) [0x7f1d7c106b4f]
>>
>> Sep 18 09:08:25  EMERG [deletion] sd_backtrace(840)
>> /lib/x86_64-linux-gnu/libc.so.6(clone+0x6c) [0x7f1d7b38b95c]
>>
>> Sep 18 09:08:54   INFO [main] md_add_disk(343) /var/lib/sheepdog//disc1,
>> vdisk nr 1834, total disk 1
>>
>> Sep 18 09:08:54   INFO [main] md_add_disk(343) /var/lib/sheepdog//disc2,
>> vdisk nr 1834, total disk 2
>>
>> Sep 18 09:08:54   INFO [main] md_add_disk(343) /var/lib/sheepdog//disc3,
>> vdisk nr 1834, total disk 3
>>
>> Sep 18 09:08:55 NOTICE [main] get_local_addr(522) found IPv4 address
>>
>> Sep 18 09:08:55   INFO [main] send_join_request(1032) IPv4 ip:10.30.20.81
>> port:7000 going to join the cluster
>>
>> Sep 18 09:08:55  ERROR [main] zk_join(975) Previous zookeeper session
>> exist, shoot myself. Please wait for 30 seconds to join me again.
>>
>> Sep 18 09:13:00   INFO [main] md_add_disk(343) /var/lib/sheepdog//disc1,
>> vdisk nr 1834, total disk 1
>>
>> Sep 18 09:13:00   INFO [main] md_add_disk(343) /var/lib/sheepdog//disc2,
>> vdisk nr 1834, total disk 2
>>
>> Sep 18 09:13:00   INFO [main] md_add_disk(343) /var/lib/sheepdog//disc3,
>> vdisk nr 1834, total disk 3
>>
>> Sep 18 09:13:01 NOTICE [main] get_local_addr(522) found IPv4 address
>>
>> Sep 18 09:13:01   INFO [main] send_join_request(1032) IPv4 ip:10.30.20.81
>> port:7000 going to join the cluster
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 59904, off 380928, 0
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 130560, off 448000, 0
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 53248, off 584192, 0
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 60416, off 642048, 0
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 45568, off 706560, 0
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 65536, off 756736, 0
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 29184, off 888832, 0
>>
>> Sep 18 09:13:01   INFO [main] replay_journal_entry(161)
>> /var/lib/sheepdog//disc2/0057e86b00007c02, size 84480, off 923136, 0
>>
>>
>>
>>
>> --
>> sheepdog-users mailing lists
>> sheepdog-users at lists.wpkg.org
>> https://lists.wpkg.org/mailman/listinfo/sheepdog-users
>>


More information about the sheepdog-users mailing list