[sheepdog] collie snapshot rollback cause the image unusable/undeletable

Wenhao Xu xuwenhao2008 at gmail.com
Wed Mar 6 15:01:07 CET 2013


I am running branch 0.5.x.

On Wed, Mar 6, 2013 at 9:57 PM, Wenhao Xu <xuwenhao2008 at gmail.com> wrote:

> Hi all,
>
> I did a snapshot rollback experiment by executing the following steps.
> After all these steps, segment fault occurred when executing 'collie tree'.
>
> $ collie vdi create  img1 2G
>
> $ collie vdi list
>   Name        Id    Size    Used  Shared    Creation time   VDI id  Copies
>  Tag
>   img1         1  2.0 GB  0.0 MB  0.0 MB 2013-03-06 21:39   9e028f     1
>
> $ collie vdi snapshot -s snap1 img1
>
> $ collie vdi snapshot -s snap2 img1
>
> $ collie vdi tree
> img1---[2013-03-06 21:39]---[2013-03-06 21:39]---(you are here)
>
> $ collie vdi rollback -s snap1 img1
>
> $ collie vdi tree
> img1---[2013-03-06 21:39]-+-[2013-03-06 21:39]
>                           `-(you are here)
>
> $ collie vdi list
>   Name        Id    Size    Used  Shared    Creation time   VDI id  Copies
>  Tag
> s img1         1  2.0 GB  0.0 MB  0.0 MB 2013-03-06 21:39   9e028f     1
>       snap1
> s img1         2  2.0 GB  0.0 MB  0.0 MB 2013-03-06 21:39   9e0290     1
>       snap2
>   img1         3  2.0 GB  0.0 MB  0.0 MB 2013-03-06 21:40   9e0291     1
>
>
> $ collie vdi rollback -s 3 img1     #This step cause the problem, Am I
> doing something wrong? If this is not allowed, collie should probably
> prevent the command executing.
>
> $ collie vdi tree
> Segmentation fault (core dumped)
>
> The worse thing is the volume cannot be deleted afterwards.
>
>  collie vdi delete -s 1 img1
>  collie vdi delete -s 2 img1
>
> works fine.
>
> But collie vdi delete  img1 just hung forever.
>
> The log shows sheep falls into a dead loop:
>
> <endless>
>  Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170902] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170903] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170904] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170905] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170906] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170907] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170908] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170909] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170910] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170911] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170912] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170913] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170914] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170915] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170916] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170917] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170918] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170919] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170920] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170921] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170922] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170923] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170924] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170925] do_process_work(1258) 2, 809e029100000000, 1
> Mar 06 21:55:29 [main] queue_request(355) READ_OBJ, 1
> Mar 06 21:55:29 [gway 170926] do_process_work(1258) 2, 809e029100000000, 1
> </endless>
>
> This seems a severe bug. Hopefully it is not hard to solve. ;) Thanks.
>
> Regards,
> Wenhao
> --
> ~_~
>



-- 
~_~
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.wpkg.org/pipermail/sheepdog/attachments/20130306/83a9748c/attachment-0004.html>


More information about the sheepdog mailing list