[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-discuss] Unable to start/deploy VMs after Qemu/Gluster upgrade
From: |
Martin Toth |
Subject: |
Re: [Qemu-discuss] Unable to start/deploy VMs after Qemu/Gluster upgrade to 2.0.0+dfsg-2ubuntu1.28glusterfs3.7.17trusty1 |
Date: |
Fri, 25 Nov 2016 17:54:39 +0100 |
Hi,
thanks for suggestion, but my VM disk file (qcow2 image) is 512 bytes aligned
(afaik).
address@hidden:/mnt/datastore0/64# stat disk.1
File: ‘disk.1.backup’
Size: 5228527616 Blocks: 10211968 IO Block: 131072 regular file
Device: 21h/33d Inode: 12589621850231693320 Links: 1
Access: (0644/-rw-r--r--) Uid: ( 0/ root) Gid: ( 0/ root)
Access: 2016-11-25 17:34:28.992251577 +0100
Modify: 2016-11-25 17:34:28.596250240 +0100
Change: 2016-11-25 17:34:28.924251347 +0100
Birth: -
address@hidden:/mnt/datastore0/64# qemu-img info disk.1
image: disk.1.backup
file format: qcow2
virtual size: 10G (10747904000 bytes)
disk size: 4.9G
cluster_size: 65536
Format specific information:
compat: 1.1
lazy refcounts: false
More over this Qemu-img create causes SegmentationFault. Log below.
address@hidden:/# qemu-img create -f qcow2 gluster://node1/vmvol/64/foo.qcow2 1G
Formatting 'gluster://node1/vmvol/64/foo.qcow2', fmt=qcow2 size=1073741824
encryption=off cluster_size=65536 lazy_refcounts=off
[2016-11-25 16:52:22.536760] I [MSGID: 104045] [glfs-master.c:96:notify]
0-gfapi: New graph 6e6f6465-312d-3331-3934-352d32303136 (0) coming up
[2016-11-25 16:52:22.536792] I [MSGID: 114020] [client.c:2113:notify]
0-vmvol-client-0: parent translators are ready, attempting connect on transport
[2016-11-25 16:52:22.537022] I [MSGID: 114020] [client.c:2113:notify]
0-vmvol-client-1: parent translators are ready, attempting connect on transport
[2016-11-25 16:52:22.537110] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
0-vmvol-client-0: changing port to 49152 (from 0)
[2016-11-25 16:52:22.537198] I [MSGID: 114020] [client.c:2113:notify]
0-vmvol-client-2: parent translators are ready, attempting connect on transport
[2016-11-25 16:52:22.537511] I [MSGID: 114057]
[client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-0:
Using Program GlusterFS 3.3, Num (1298437), Version (330)
[2016-11-25 16:52:22.537677] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
0-vmvol-client-1: changing port to 49152 (from 0)
[2016-11-25 16:52:22.537682] I [MSGID: 114046]
[client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-0: Connected to
vmvol-client-0, attached to remote volume '/gluster/vmvol/brick1'.
[2016-11-25 16:52:22.537705] I [MSGID: 114047]
[client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-0: Server and
Client lk-version numbers are not same, reopening the fds
[2016-11-25 16:52:22.537744] I [MSGID: 108005] [afr-common.c:4299:afr_notify]
0-vmvol-replicate-0: Subvolume 'vmvol-client-0' came back up; going online.
[2016-11-25 16:52:22.537769] I [MSGID: 114035]
[client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-0: Server lk
version = 1
[2016-11-25 16:52:22.537950] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
0-vmvol-client-2: changing port to 49152 (from 0)
[2016-11-25 16:52:22.538096] I [MSGID: 114057]
[client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-1:
Using Program GlusterFS 3.3, Num (1298437), Version (330)
[2016-11-25 16:52:22.538316] I [MSGID: 114046]
[client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-1: Connected to
vmvol-client-1, attached to remote volume '/gluster/vmvol/brick1'.
[2016-11-25 16:52:22.538332] I [MSGID: 114047]
[client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-1: Server and
Client lk-version numbers are not same, reopening the fds
[2016-11-25 16:52:22.538439] I [MSGID: 114035]
[client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-1: Server lk
version = 1
[2016-11-25 16:52:22.538548] I [MSGID: 114057]
[client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-2:
Using Program GlusterFS 3.3, Num (1298437), Version (330)
[2016-11-25 16:52:22.538984] I [MSGID: 114046]
[client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-2: Connected to
vmvol-client-2, attached to remote volume '/gluster/vmvol/brick1'.
[2016-11-25 16:52:22.538993] I [MSGID: 114047]
[client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-2: Server and
Client lk-version numbers are not same, reopening the fds
[2016-11-25 16:52:22.543005] I [MSGID: 114035]
[client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-2: Server lk
version = 1
[2016-11-25 16:52:22.543584] I [MSGID: 108031]
[afr-common.c:2071:afr_local_discovery_cbk] 0-vmvol-replicate-0: selecting
local read_child vmvol-client-0
[2016-11-25 16:52:22.544396] I [MSGID: 104041]
[glfs-resolve.c:890:__glfs_active_subvol] 0-vmvol: switched to graph
6e6f6465-312d-3331-3934-352d32303136 (0)
[2016-11-25 16:52:22.657150] I [MSGID: 114021] [client.c:2122:notify]
0-vmvol-client-0: current graph is no longer active, destroying rpc_client
[2016-11-25 16:52:22.657177] I [MSGID: 114021] [client.c:2122:notify]
0-vmvol-client-1: current graph is no longer active, destroying rpc_client
[2016-11-25 16:52:22.657195] I [MSGID: 114018]
[client.c:2037:client_rpc_notify] 0-vmvol-client-0: disconnected from
vmvol-client-0. Client process will keep trying to connect to glusterd until
brick's port is available
[2016-11-25 16:52:22.657199] I [MSGID: 114021] [client.c:2122:notify]
0-vmvol-client-2: current graph is no longer active, destroying rpc_client
[2016-11-25 16:52:22.657217] I [MSGID: 114018]
[client.c:2037:client_rpc_notify] 0-vmvol-client-1: disconnected from
vmvol-client-1. Client process will keep trying to connect to glusterd until
brick's port is available
[2016-11-25 16:52:22.657246] W [MSGID: 108001] [afr-common.c:4379:afr_notify]
0-vmvol-replicate-0: Client-quorum is not met
[2016-11-25 16:52:22.657246] I [MSGID: 114018]
[client.c:2037:client_rpc_notify] 0-vmvol-client-2: disconnected from
vmvol-client-2. Client process will keep trying to connect to glusterd until
brick's port is available
[2016-11-25 16:52:22.657270] E [MSGID: 108006] [afr-common.c:4321:afr_notify]
0-vmvol-replicate-0: All subvolumes are down. Going offline until atleast one
of them comes back up.
[2016-11-25 16:52:22.657350] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-gfapi: size=84 max=2 total=2
[2016-11-25 16:52:22.657488] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-gfapi: size=156 max=3 total=3
[2016-11-25 16:52:22.657599] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-gfapi: size=108 max=1 total=1
[2016-11-25 16:52:22.657611] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-vmvol-client-0: size=1300 max=2 total=19
[2016-11-25 16:52:22.657620] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-vmvol-client-1: size=1300 max=2 total=19
[2016-11-25 16:52:22.657628] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-vmvol-client-2: size=1300 max=3 total=19
[2016-11-25 16:52:22.657636] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-vmvol-replicate-0: size=10556 max=5 total=14
[2016-11-25 16:52:22.657705] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-vmvol-dht: size=1148 max=0 total=0
[2016-11-25 16:52:22.657740] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-vmvol-dht: size=2316 max=2 total=8
[2016-11-25 16:52:22.657799] I [MSGID: 101053]
[mem-pool.c:636:mem_pool_destroy] 0-vmvol-readdir-ahead: size=60 max=0 total=0
[2016-11-25 16:52:22.657808] I [io-stats.c:2951:fini] 0-vmvol: io-stats
translator unloaded
[2016-11-25 16:52:22.657843] I [MSGID: 101191]
[event-epoll.c:663:event_dispatch_epoll_worker] 0-epoll: Exited thread with
index 1
[2016-11-25 16:52:22.657843] I [MSGID: 101191]
[event-epoll.c:663:event_dispatch_epoll_worker] 0-epoll: Exited thread with
index 2
[2016-11-25 16:52:23.536847] I [MSGID: 104045] [glfs-master.c:96:notify]
0-gfapi: New graph 6e6f6465-312d-3331-3934-352d32303136 (0) coming up
[2016-11-25 16:52:23.536877] I [MSGID: 114020] [client.c:2113:notify]
0-vmvol-client-0: parent translators are ready, attempting connect on transport
[2016-11-25 16:52:23.537166] I [MSGID: 114020] [client.c:2113:notify]
0-vmvol-client-1: parent translators are ready, attempting connect on transport
[2016-11-25 16:52:23.537259] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
0-vmvol-client-0: changing port to 49152 (from 0)
[2016-11-25 16:52:23.537350] I [MSGID: 114020] [client.c:2113:notify]
0-vmvol-client-2: parent translators are ready, attempting connect on transport
[2016-11-25 16:52:23.537594] I [MSGID: 114057]
[client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-0:
Using Program GlusterFS 3.3, Num (1298437), Version (330)
[2016-11-25 16:52:23.537758] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
0-vmvol-client-1: changing port to 49152 (from 0)
[2016-11-25 16:52:23.537761] I [MSGID: 114046]
[client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-0: Connected to
vmvol-client-0, attached to remote volume '/gluster/vmvol/brick1'.
[2016-11-25 16:52:23.537784] I [MSGID: 114047]
[client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-0: Server and
Client lk-version numbers are not same, reopening the fds
[2016-11-25 16:52:23.537826] I [MSGID: 108005] [afr-common.c:4299:afr_notify]
0-vmvol-replicate-0: Subvolume 'vmvol-client-0' came back up; going online.
[2016-11-25 16:52:23.537852] I [MSGID: 114035]
[client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-0: Server lk
version = 1
[2016-11-25 16:52:23.538040] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
0-vmvol-client-2: changing port to 49152 (from 0)
[2016-11-25 16:52:23.538179] I [MSGID: 114057]
[client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-1:
Using Program GlusterFS 3.3, Num (1298437), Version (330)
[2016-11-25 16:52:23.538384] I [MSGID: 114046]
[client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-1: Connected to
vmvol-client-1, attached to remote volume '/gluster/vmvol/brick1'.
[2016-11-25 16:52:23.538399] I [MSGID: 114047]
[client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-1: Server and
Client lk-version numbers are not same, reopening the fds
[2016-11-25 16:52:23.538508] I [MSGID: 114035]
[client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-1: Server lk
version = 1
[2016-11-25 16:52:23.538612] I [MSGID: 114057]
[client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-2:
Using Program GlusterFS 3.3, Num (1298437), Version (330)
[2016-11-25 16:52:23.538994] I [MSGID: 114046]
[client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-2: Connected to
vmvol-client-2, attached to remote volume '/gluster/vmvol/brick1'.
[2016-11-25 16:52:23.539004] I [MSGID: 114047]
[client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-2: Server and
Client lk-version numbers are not same, reopening the fds
[2016-11-25 16:52:23.543058] I [MSGID: 114035]
[client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-2: Server lk
version = 1
[2016-11-25 16:52:23.543606] I [MSGID: 108031]
[afr-common.c:2071:afr_local_discovery_cbk] 0-vmvol-replicate-0: selecting
local read_child vmvol-client-0
[2016-11-25 16:52:23.544280] I [MSGID: 104041]
[glfs-resolve.c:890:__glfs_active_subvol] 0-vmvol: switched to graph
6e6f6465-312d-3331-3934-352d32303136 (0)
Segmentation fault (core dumped)
BR,
Martin
> On 25 Nov 2016, at 16:29, Thomas Lamprecht <address@hidden> wrote:
>
> Hi,
>
>
> On 11/25/2016 04:13 PM, Martin Toth wrote:
>> Hello all,
>>
>> we are using your qemu packages to deploy qemu VMs on our gluster via gfsapi.
>> Recent upgrade broken our qemu and we are not able to deploy / start VMs
>> anymore.
>>
>> Gluster is running OK, mounted with FUSE, everything looks ok, there is
>> probably some problem with qemu while accessing gluster with gfsapi.
>>
>> These are our current versions (Qemu is from
>> https://launchpad.net/~monotek/+archive/ubuntu/qemu-glusterfs-3.7 ):
>> ii glusterfs-client 3.7.17-ubuntu1~trusty2
>> amd64 clustered file-system (client
>> package)
>> ii glusterfs-common 3.7.17-ubuntu1~trusty2
>> amd64 GlusterFS common libraries and
>> translator modules
>> ii glusterfs-server 3.7.17-ubuntu1~trusty2
>> amd64 clustered file-system (server package)
>> ii qemu-keymaps
>> 2.0.0+dfsg-2ubuntu1.28glusterfs3.7.17trusty1 all QEMU keyboard maps
>> ii qemu-kvm
>> 2.0.0+dfsg-2ubuntu1.28glusterfs3.7.17trusty1 amd64 QEMU Full
>> virtualization
>> ii qemu-system-common
>> 2.0.0+dfsg-2ubuntu1.28glusterfs3.7.17trusty1 amd64 QEMU full system
>> emulation binaries (common files)
>> ii qemu-system-x86
>> 2.0.0+dfsg-2ubuntu1.28glusterfs3.7.17trusty1 amd64 QEMU full system
>> emulation binaries (x86)
>> ii qemu-utils
>> 2.0.0+dfsg-2ubuntu1.28glusterfs3.7.17trusty1 amd64 QEMU utilities
>>
>> We see error attached lower in mail. Do you have any suggestions what can
>> cause this problem ?
>>
>> Thanks in advance for your help.
>>
>> Regards,
>> Martin
>>
>> Volume Name: vmvol
>> Type: Replicate
>> Volume ID: a72b5c9e-b8ff-488e-b10f-5ba4b71e62b8
>> Status: Started
>> Number of Bricks: 1 x 3 = 3
>> Transport-type: tcp
>> Bricks:
>> Brick1: node1.storage.internal:/gluster/vmvol/brick1
>> Brick2: node2.storage.internal:/gluster/vmvol/brick1
>> Brick3: node3.storage.internal:/gluster/vmvol/brick1
>> Options Reconfigured:
>> cluster.self-heal-daemon: on
>> nfs.disable: on
>> cluster.server-quorum-type: server
>> cluster.quorum-type: auto
>> network.remote-dio: enable
>> performance.stat-prefetch: on
>> performance.io-cache: off
>> performance.read-ahead: off
>> performance.quick-read: off
>> cluster.eager-lock: enable
>> storage.owner-gid: 9869
>> storage.owner-uid: 9869
>> server.allow-insecure: on
>> performance.readdir-ahead: on
>>
>> 2016-11-25 12:54:06.121+0000: starting up
>> LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/bin:/usr/sbin:/sbin:/bin
>> QEMU_AUDIO_DRV=none /usr/bin/qemu-system-x86_64 -name one-67 -S -machine
>> pc-i440fx-trusty,accel=kvm,usb=off -m 1024 -realtime mlock=off -smp
>> 1,sockets=1,cores=1,threads=1 -uuid 3c703b26-3f57-44d0-8d76-bb281fd8902c
>> -no-user-config -nodefaults -chardev
>> socket,id=charmonitor,path=/var/lib/libvirt/qemu/one-67.monitor,server,nowait
>> -mon chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
>> -boot strict=on -device piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -drive
>> file=gluster://node1:24007/vmvol/67/disk.1,if=none,id=drive-ide0-0-0,format=qcow2,cache=none
>> -device
>> ide-hd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0,bootindex=2 -drive
>> file=/var/lib/one//datastores/112/67/disk.0,if=none,id=drive-ide0-0-1,readonly=on,format=raw,cache=none
>> -device
>> ide-cd,bus=ide.0,unit=1,drive=drive-ide0-0-1,id=ide0-0-1,bootindex=3 -drive
>> file=/var/lib/one//datastores/112/67/disk.2,if=none,id=drive-ide0-1-0,readonly=on,format=raw
>> -device ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -netdev
>> tap,fd=24,id=hostnet0 -device
>> rtl8139,netdev=hostnet0,id=net0,mac=02:00:0a:c8:64:1f,bus=pci.0,addr=0x3,bootindex=1
>> -vnc 0.0.0.0:67 -device cirrus-vga,id=video0,bus=pci.0,addr=0x2 -device
>> virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x4
>> [2016-11-25 12:54:06.254217] I [MSGID: 104045] [glfs-master.c:96:notify]
>> 0-gfapi: New graph 6e6f6465-322d-3231-3330-392d32303136 (0) coming up
>> [2016-11-25 12:54:06.254246] I [MSGID: 114020] [client.c:2113:notify]
>> 0-vmvol-client-0: parent translators are ready, attempting connect on
>> transport
>> [2016-11-25 12:54:06.254473] I [MSGID: 114020] [client.c:2113:notify]
>> 0-vmvol-client-1: parent translators are ready, attempting connect on
>> transport
>> [2016-11-25 12:54:06.254672] I [MSGID: 114020] [client.c:2113:notify]
>> 0-vmvol-client-2: parent translators are ready, attempting connect on
>> transport
>> [2016-11-25 12:54:06.254844] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
>> 0-vmvol-client-0: changing port to 49152 (from 0)
>> [2016-11-25 12:54:06.255303] I [MSGID: 114057]
>> [client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-0:
>> Using Program GlusterFS 3.3, Num (1298437), Version (330)
>> [2016-11-25 12:54:06.255391] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
>> 0-vmvol-client-2: changing port to 49152 (from 0)
>> [2016-11-25 12:54:06.255844] I [MSGID: 114057]
>> [client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-2:
>> Using Program GlusterFS 3.3, Num (1298437), Version (330)
>> [2016-11-25 12:54:06.256259] I [MSGID: 114046]
>> [client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-2: Connected
>> to vmvol-client-2, attached to remote volume '/gluster/vmvol/brick1'.
>> [2016-11-25 12:54:06.256268] I [MSGID: 114047]
>> [client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-2: Server and
>> Client lk-version numbers are not same, reopening the fds
>> [2016-11-25 12:54:06.256288] I [MSGID: 108005]
>> [afr-common.c:4299:afr_notify] 0-vmvol-replicate-0: Subvolume
>> 'vmvol-client-2' came back up; going online.
>> [2016-11-25 12:54:06.256464] I [MSGID: 114035]
>> [client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-2: Server
>> lk version = 1
>> [2016-11-25 12:54:06.262716] I [MSGID: 114046]
>> [client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-0: Connected
>> to vmvol-client-0, attached to remote volume '/gluster/vmvol/brick1'.
>> [2016-11-25 12:54:06.262729] I [MSGID: 114047]
>> [client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-0: Server and
>> Client lk-version numbers are not same, reopening the fds
>> [2016-11-25 12:54:06.262863] I [MSGID: 114035]
>> [client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-0: Server
>> lk version = 1
>> [2016-11-25 12:54:06.267906] I [rpc-clnt.c:1960:rpc_clnt_reconfig]
>> 0-vmvol-client-1: changing port to 49152 (from 0)
>> [2016-11-25 12:54:06.268148] I [MSGID: 114057]
>> [client-handshake.c:1437:select_server_supported_programs] 0-vmvol-client-1:
>> Using Program GlusterFS 3.3, Num (1298437), Version (330)
>> [2016-11-25 12:54:06.287724] I [MSGID: 114046]
>> [client-handshake.c:1213:client_setvolume_cbk] 0-vmvol-client-1: Connected
>> to vmvol-client-1, attached to remote volume '/gluster/vmvol/brick1'.
>> [2016-11-25 12:54:06.287734] I [MSGID: 114047]
>> [client-handshake.c:1224:client_setvolume_cbk] 0-vmvol-client-1: Server and
>> Client lk-version numbers are not same, reopening the fds
>> [2016-11-25 12:54:06.313439] I [MSGID: 114035]
>> [client-handshake.c:193:client_set_lk_version_cbk] 0-vmvol-client-1: Server
>> lk version = 1
>> [2016-11-25 12:54:06.314113] I [MSGID: 108031]
>> [afr-common.c:2071:afr_local_discovery_cbk] 0-vmvol-replicate-0: selecting
>> local read_child vmvol-client-1
>> [2016-11-25 12:54:06.314928] I [MSGID: 104041]
>> [glfs-resolve.c:890:__glfs_active_subvol] 0-vmvol: switched to graph
>> 6e6f6465-322d-3231-3330-392d32303136 (0)
>> [2016-11-25 12:54:16.331479] I [MSGID: 114021] [client.c:2122:notify]
>> 0-vmvol-client-0: current graph is no longer active, destroying rpc_client
>> [2016-11-25 12:54:16.331507] I [MSGID: 114021] [client.c:2122:notify]
>> 0-vmvol-client-1: current graph is no longer active, destroying rpc_client
>> [2016-11-25 12:54:16.331517] I [MSGID: 114018]
>> [client.c:2037:client_rpc_notify] 0-vmvol-client-0: disconnected from
>> vmvol-client-0. Client process will keep trying to connect to glusterd until
>> brick's port is available
>> [2016-11-25 12:54:16.331531] I [MSGID: 114018]
>> [client.c:2037:client_rpc_notify] 0-vmvol-client-1: disconnected from
>> vmvol-client-1. Client process will keep trying to connect to glusterd until
>> brick's port is available
>> [2016-11-25 12:54:16.331534] I [MSGID: 114021] [client.c:2122:notify]
>> 0-vmvol-client-2: current graph is no longer active, destroying rpc_client
>> [2016-11-25 12:54:16.331543] W [MSGID: 108001]
>> [afr-common.c:4379:afr_notify] 0-vmvol-replicate-0: Client-quorum is not met
>> [2016-11-25 12:54:16.331667] E [rpc-clnt.c:370:saved_frames_unwind] (-->
>> /usr/lib/x86_64-linux-gnu/libglusterfs.so.0(_gf_log_callingfn+0x192)[0x7f4dcee94502]
>> (-->
>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_unwind+0x1de)[0x7f4dcec642be]
>> (-->
>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(saved_frames_destroy+0xe)[0x7f4dcec643ce]
>> (-->
>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_connection_cleanup+0x84)[0x7f4dcec65aa4]
>> (-->
>> /usr/lib/x86_64-linux-gnu/libgfrpc.so.0(rpc_clnt_notify+0x120)[0x7f4dcec66360]
>> ))))) 0-vmvol-client-2: forced unwinding frame type(GlusterFS 3.3)
>> op(WRITE(13)) called at 2016-11-25 12:54:16.331128 (xid=0x2f62)
>> [2016-11-25 12:54:16.331684] W [MSGID: 114031]
>> [client-rpc-fops.c:907:client3_3_writev_cbk] 0-vmvol-client-2: remote
>> operation failed [Transport endpoint is not connected]
>> [2016-11-25 12:54:16.331701] I [MSGID: 114018]
>> [client.c:2037:client_rpc_notify] 0-vmvol-client-2: disconnected from
>> vmvol-client-2. Client process will keep trying to connect to glusterd until
>> brick's port is available
>> [2016-11-25 12:54:16.331710] E [MSGID: 108006]
>> [afr-common.c:4321:afr_notify] 0-vmvol-replicate-0: All subvolumes are down.
>> Going offline until atleast one of them comes back up.
>> [2016-11-25 12:54:16.331712] E [MSGID: 114031]
>> [client-rpc-fops.c:1624:client3_3_inodelk_cbk] 0-vmvol-client-0: remote
>> operation failed [Transport endpoint is not connected]
>> [2016-11-25 12:54:16.331727] E [MSGID: 114031]
>> [client-rpc-fops.c:1624:client3_3_inodelk_cbk] 0-vmvol-client-1: remote
>> operation failed [Transport endpoint is not connected]
>> [2016-11-25 12:54:16.331735] E [MSGID: 114031]
>> [client-rpc-fops.c:1624:client3_3_inodelk_cbk] 0-vmvol-client-2: remote
>> operation failed [Transport endpoint is not connected]
>> [2016-11-25 12:54:16.331749] E [MSGID: 114031]
>> [client-rpc-fops.c:1624:client3_3_inodelk_cbk] 0-vmvol-client-0: remote
>> operation failed [Transport endpoint is not connected]
>> [2016-11-25 12:54:16.331756] E [MSGID: 114031]
>> [client-rpc-fops.c:1624:client3_3_inodelk_cbk] 0-vmvol-client-1: remote
>> operation failed [Transport endpoint is not connected]
>> [2016-11-25 12:54:16.331762] E [MSGID: 114031]
>> [client-rpc-fops.c:1624:client3_3_inodelk_cbk] 0-vmvol-client-2: remote
>> operation failed [Transport endpoint is not connected]
>> [2016-11-25 12:54:16.331768] W [inode.c:1814:inode_table_destroy]
>> (-->/usr/lib/x86_64-linux-gnu/libgfapi.so.0(glfs_fini+0x3cf)
>> [0x7f4dd57ef49f]
>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(inode_table_destroy_all+0x51)
>> [0x7f4dceebe4d1]
>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(inode_table_destroy+0xd4)
>> [0x7f4dceebe3d4] ) 0-gfapi: Active inode(0x7f4db83101a4) with refcount(1)
>> found during cleanup
>> [2016-11-25 12:54:16.331834] E [inode.c:468:__inode_unref]
>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.7.17/xlator/cluster/replicate.so(afr_local_cleanup+0x128)
>> [0x7f4dbb5dafd8]
>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(inode_unref+0x21)
>> [0x7f4dceebc7d1] -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(+0x4a0b6)
>> [0x7f4dceebc0b6] ) 0-: Assertion failed: inode->ref
>> [2016-11-25 12:54:16.331876] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-gfapi: size=84 max=2 total=2
>> [2016-11-25 12:54:16.331924] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-gfapi: size=156 max=3 total=3
>> [2016-11-25 12:54:16.332037] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-gfapi: size=108 max=2 total=2
>> [2016-11-25 12:54:16.332046] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-vmvol-client-0: size=1300 max=2
>> total=3038
>> [2016-11-25 12:54:16.332054] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-vmvol-client-1: size=1300 max=2
>> total=3039
>> [2016-11-25 12:54:16.332060] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-vmvol-client-2: size=1300 max=3
>> total=3038
>> [2016-11-25 12:54:16.332068] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-vmvol-replicate-0: size=10556 max=4
>> total=3038
>> [2016-11-25 12:54:16.332107] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-vmvol-dht: size=1148 max=0 total=0
>> [2016-11-25 12:54:16.332141] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-vmvol-dht: size=2316 max=2 total=6
>> [2016-11-25 12:54:16.332200] I [MSGID: 101053]
>> [mem-pool.c:636:mem_pool_destroy] 0-vmvol-readdir-ahead: size=60 max=0
>> total=0
>> [2016-11-25 12:54:16.332212] I [io-stats.c:2951:fini] 0-vmvol: io-stats
>> translator unloaded
>> [2016-11-25 12:54:16.332245] I [MSGID: 101191]
>> [event-epoll.c:663:event_dispatch_epoll_worker] 0-epoll: Exited thread with
>> index 1
>> [2016-11-25 12:54:16.332259] I [MSGID: 101191]
>> [event-epoll.c:663:event_dispatch_epoll_worker] 0-epoll: Exited thread with
>> index 2
>> qemu-system-x86_64: -drive
>> file=gluster://node1:24007/vmvol/67/disk.1,if=none,id=drive-ide0-0-0,format=qcow2,cache=none:
>> could not open disk image gluster://node1:24007/vmvol/67/disk.1: Could not
>> read L1 table: Bad file descriptor
>> 2016-11-25 12:54:17.199+0000: shutting down
>>
>
>
> It looks like something we ran into:
> https://bugs.launchpad.net/qemu/+bug/1644754
> If you created the disk not directly with qemus glusterfs backend but access
> it with it (like your error says) it could be a file size alignment problem.
> Try using `truncate -s ...` to increase the size to the next 512 byte
> boundary, if that fixes your problem it should be that bug.
>
> cheers,
> Thomas
>
>
>