Re: [vdsm] [Users] How to configure sharedFS ?
by deepakcs@linux.vnet.ibm.com
On 02/29/2012 08:06 PM, Dan Kenigsberg wrote:
> On Wed, Feb 29, 2012 at 07:42:15PM +0530, Deepak C Shetty wrote:
>> On 02/27/2012 04:55 AM, Ayal Baron wrote:
>>> Any help on documenting this so people would not have to dig into the code would be greatly appreciated.
>>>
>>>> Great, let me start and get back if i land up into issues.
>>>> thanks for the pointer.
>>> Let us know if you need further assistance..
>>
> If you have that python script working for LOCALFS, I'd suggest you try
> making it work for SHAREDFS too.
>
>
I was able to use the VDSM_Standalone example, and modify it to use SHAREDFS
and export glusterfs as a DATA_DOMAIN and invoke VM backed by gluster
storage.
I edited http://www.ovirt.org/wiki/Vdsm_Standalone (scroll below) &
added it as
a SHAREDFS example.
11 years, 11 months
Re: [vdsm] [oVirt Jenkins] vdsm_check_code_violations - Build # 33 - Still Failing!
by abaron@redhat.com
----- Original Message -----
> From: "Jenkins oVirt Server" <jenkins(a)ovirt.org>
> To: abaron(a)redhat.com, vdsm-patches(a)lists.fedorahosted.org, danken(a)redhat.com, fsimonce(a)redhat.com
> Sent: Tuesday, March 20, 2012 6:38:17 PM
> Subject: [oVirt Jenkins] vdsm_check_code_violations - Build # 33 - Still Failing!
>
> vdsm_check_code_violations - Build # 33 - Still Failing:
>
> Check console output at
> http://jenkins.ovirt.org/job/vdsm_check_code_violations/33/console
> to view the results.
>
I reviewed the errors and the code, I couldn't find even one thing that is real (went through the first ~100 errors).
We need a conf file to filter these error types:
Instance of 'VmStatsThread' has no 'addStatsFunction' member
Passing unexpected keyword argument '__securityOverride' in function call
12 years, 2 months
PEP8 in VDSM code
by smizrahi@redhat.com
I suggest having pep8 a must for patch submission in VDSM.
http://www.python.org/dev/peps/pep-0008/
Currently there are a few people policing these rules in reviews but I suggest we make it automatic.
Unless someone objects I will put a gerrit hook that complains about pep8 violations.
It will not mark -1s until all (or at least most) source code has been converted because people might get complains about code they did not modify in this patch.
If you happy and you know it +1!
12 years, 2 months
Re: [vdsm] [node-devel] Not able to migrate (back) to node
by David Jaša
Michel van Horssen píše v Pá 09. 03. 2012 v 10:30 +0100:
> Hi,
>
> Not sure if it's an engine or node problem but seeing that the eningine is functioning fine I'm putting my bets on the node.
>
> I have a test install of ovirt on 3 servers.
>
> 1. FC 16 Engine/VDSM
>
> 2. FC 16 VDSM
>
> 3. Node version 2.2.3-1.1
>
> I have 2 networks I need access to on all servers.
>
> A. Default network for access to the servers
>
> B. Data network for sharing iSCSI from an OpenFiler server.
>
> I have an ISO nfs share on the engine server (1) and share an iSCSI disk on network B.
>
> I needed to do a "hand job" on the node so it would get my 2nd network correctly because I couldn't do it from the TUI (is a known) and not from the engine interface.
>
> Now for the problem:
>
> I can create virtuel guests on all three servers. Running just fine. I can migrate away from all 3 servers while the guests are running. The thing is I can only migrate towards the VDSM servers (1 and 2) not to the Node (3).
>
> The guest created on the node gets migrated away just fine but getting back on the node I get a:
>
> "Migration failed due to Error: Fatal error during migration (VM: guestname, Source Host: VDSM-Host)"
>
> Wich logs are needed to take a look at? Because I tried looking at the "/var/log/vdsm/vdsm.log" but that was to mixed up for me to find anything that pointed me in the right direction.
vdsm logs are most helpful. For starters, look up the VM names in the
log and note its UUID. Then grep the logs for that UUID and you should
see what happened to the VM.
qemu logs may also be of help, you can find them
at /var/log/libvirt/qemu/<VM Name>
David
>
> Any help is appreciated.
>
> Thnx
> Michel
> _______________________________________________
> node-devel mailing list
> node-devel(a)ovirt.org
> http://lists.ovirt.org/mailman/listinfo/node-devel
--
David Jaša, RHCE
SPICE QE based in Brno
GPG Key: 22C33E24
Fingerprint: 513A 060B D1B4 2A72 7F0D 0278 B125 CD00 22C3 3E24
12 years, 2 months
Following unix pipes
by smizrahi@redhat.com
I recently had several instances where I had to try and figure out who holds what end of a unix pipe.
To make this operation a bit more streamlined I created a small script to follow a pipe. I think it will be useful for other people debugging VDSM especially bugs related to "out of process" helpers not closing FDs properly.
To see all the exists of a pipe just input a known end of the pipe: stahlband <PID> <FD>
$ stahlband 5758 5
PID: 5758 FD: 5 KIND: r
PID: 5758 FD: 6 KIND: w
PID: 5770 FD: 5 KIND: r
PID: 5770 FD: 6 KIND: w
The code is available on github:
https://github.com/ficoos/stahlband
12 years, 2 months
Re: [vdsm] [node-devel] Not able to migrate (back) to node
by Perry Myers
On 03/09/2012 04:30 AM, Michel van Horssen wrote:
> Hi,
>
> Not sure if it's an engine or node problem but seeing that the eningine is functioning fine I'm putting my bets on the node.
>
> I have a test install of ovirt on 3 servers.
>
> 1. FC 16 Engine/VDSM
>
> 2. FC 16 VDSM
>
> 3. Node version 2.2.3-1.1
>
> I have 2 networks I need access to on all servers.
>
> A. Default network for access to the servers
>
> B. Data network for sharing iSCSI from an OpenFiler server.
>
> I have an ISO nfs share on the engine server (1) and share an iSCSI disk on network B.
>
> I needed to do a "hand job" on the node so it would get my 2nd network correctly because I couldn't do it from the TUI (is a known) and not from the engine interface.
>
> Now for the problem:
>
> I can create virtuel guests on all three servers. Running just fine. I can migrate away from all 3 servers while the guests are running. The thing is I can only migrate towards the VDSM servers (1 and 2) not to the Node (3).
>
> The guest created on the node gets migrated away just fine but getting back on the node I get a:
>
> "Migration failed due to Error: Fatal error during migration (VM: guestname, Source Host: VDSM-Host)"
>
> Wich logs are needed to take a look at? Because I tried looking at the "/var/log/vdsm/vdsm.log" but that was to mixed up for me to find anything that pointed me in the right direction.
>
> Any help is appreciated.
I think this would be something for the vdsm folks to take a look at...
If they determine there is something oVirt Node specific that is
causing this behavior, we can address it but they might need to help
diagnose.
Adding vdsm list.
Perry
12 years, 2 months
Installing VDSM on an empty Fedora 16 machine?
by tomic80@gmail.com
Hi to all,
Does anybody have step-by-step instructions how to install vdsm on an
empty Fedora 16 machine?
I need a git version of vdsm (to compile from source), so the binary
rpms will not do the work.
What I have so far is:
# for resolving the sanlock bug
sudo yum update --enablerepo=updates-testing
# installing additional packages for devel
sudo yum install @development-tools
sudo yum install fedora-packager
sudo yum install git autoconf automake gcc gcc-c++ pyflakes rpm-build
git clone http://gerrit.ovirt.org/p/vdsm.git
cd vdsm
./autogen.sh --system && make rpm
sudo yum -y install /home/stomic/rpmbuild/RPMS/x86_64/vdsm-*.rpm
/home/stomic/rpmbuild/RPMS/noarch/vdsm-cli*.rpm
vdsClient -s 0 getVdsCaps
#This should show you Vdsm's response about its capabilities, but
returns:
# Connection to 0:54321 refused
sudo vdsClient -s 0 getVdsStats
# If I try as root, nothing changes:
# Connection to root@localhost.localdomain:54321 refused
Thank you in advance!
Sasha
12 years, 2 months
safelease source location ?
by deepakcs@linux.vnet.ibm.com
Hi,
I could not find the source for `/usr/libexec/vdsm/safelease` in
the vdsm git.. which git / repo
do i need to look for ? thanks in advance.
12 years, 3 months
Re: [vdsm] Kernel crash on FC16
by Roi Dayan
Tried it now with sudo to run the commands.
everything is ok.
I tried with vdsm again and here is what I see in vdsm.log
I did the cut after multiple call to lvcreate for outbox,leases,.. etc
last one is master, then there is a call to chown and mkfs.
from here mkfs did fail and kernel crashed.
--
Thread-154::DEBUG::2012-03-07
11:00:14,560::lvm::956::Storage.Misc.excCmd::(createLV) '/usr/bin/sudo -n
/bin/chown vdsm:qemu /dev/18706999-9fab-45ab-9d29-75f2eaa225aa/master' (cwd
None)
Thread-154::DEBUG::2012-03-07
11:00:14,577::lvm::956::Storage.Misc.excCmd::(createLV) SUCCESS: <err> =
''; <rc> = 0
Thread-154::DEBUG::2012-03-07
11:00:14,578::blockSD::1036::Storage.Misc.excCmd::(_createVMSfs)
'/usr/bin/sudo -n /sbin/mkfs -q -j -E nodiscard
/dev/18706999-9fab-45ab-9d29-75f2eaa225aa/master' (cwd None)
Thread-154::DEBUG::2012-03-07
11:00:15,112::blockSD::1036::Storage.Misc.excCmd::(_createVMSfs) FAILED:
<err> = ''; <rc> = 139
Thread-154::ERROR::2012-03-07
11:00:15,113::task::853::TaskManager.Task::(_setError)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::Unexpected error
Traceback (most recent call last):
File "/usr/share/vdsm/storage/task.py", line 861, in _run
return fn(*args, **kargs)
File "/usr/share/vdsm/logUtils.py", line 38, in wrapper
res = f(*args, **kwargs)
File "/usr/share/vdsm/storage/hsm.py", line 2072, in createStorageDomain
domClass, typeSpecificArg, storageType, domVersion)
File "/usr/share/vdsm/storage/blockSD.py", line 357, in create
_createVMSfs(os.path.join("/dev", vgName, MASTERLV))
File "/usr/share/vdsm/storage/blockSD.py", line 1038, in _createVMSfs
raise se.MkfsError(dev)
MkfsError: Cannot create filesystem on device:
('/dev/18706999-9fab-45ab-9d29-75f2eaa225aa/master',)
Thread-154::DEBUG::2012-03-07
11:00:15,120::task::872::TaskManager.Task::(_run)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::Task._run:
349abfb0-1838-4a6f-9b95-6521ad880cb2 (3,
'18706999-9fab-45ab-9d29-75f2eaa225aa', 'xx',
'hkDKBq-XAjX-qE1A-WCi7-Sttt-Lmos-uzVqR7', 1, '2') {} failed - stopping task
Thread-154::DEBUG::2012-03-07
11:00:15,120::task::1199::TaskManager.Task::(stop)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::stopping in state preparing
(force False)
Thread-154::DEBUG::2012-03-07
11:00:15,120::task::978::TaskManager.Task::(_decref)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::ref 1 aborting True
Thread-154::INFO::2012-03-07
11:00:15,121::task::1157::TaskManager.Task::(prepare)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::aborting: Task is aborted:
'Cannot create filesystem on device' - code 608
Thread-154::DEBUG::2012-03-07
11:00:15,121::task::1162::TaskManager.Task::(prepare)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::Prepare: aborted: Cannot
create filesystem on device
Thread-154::DEBUG::2012-03-07
11:00:15,122::task::978::TaskManager.Task::(_decref)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::ref 0 aborting True
Thread-154::DEBUG::2012-03-07
11:00:15,122::task::913::TaskManager.Task::(_doAbort)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::Task._doAbort: force False
Thread-154::DEBUG::2012-03-07
11:00:15,123::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-154::DEBUG::2012-03-07
11:00:15,123::task::588::TaskManager.Task::(_updateState)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::moving from state preparing ->
state aborting
Thread-154::DEBUG::2012-03-07
11:00:15,123::task::537::TaskManager.Task::(__state_aborting)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::_aborting: recover policy none
Thread-154::DEBUG::2012-03-07
11:00:15,124::task::588::TaskManager.Task::(_updateState)
Task=`349abfb0-1838-4a6f-9b95-6521ad880cb2`::moving from state aborting ->
state failed
Thread-154::DEBUG::2012-03-07
11:00:15,124::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-154::DEBUG::2012-03-07
11:00:15,125::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-154::ERROR::2012-03-07
11:00:15,125::dispatcher::90::Storage.Dispatcher.Protect::(run) {'status':
{'message': "Cannot create filesystem on device:
('/dev/18706999-9fab-45ab-9d29-75f2eaa225aa/master',)", 'code': 608}}
Thread-157::DEBUG::2012-03-07
11:00:18,916::task::588::TaskManager.Task::(_updateState)
Task=`55a87a6e-6d63-4b40-94a5-1244cd7eb2aa`::moving from state init ->
state preparing
Thread-157::INFO::2012-03-07
11:00:18,917::logUtils::37::dispatcher::(wrapper) Run and protect:
repoStats(options=None)
Thread-157::INFO::2012-03-07
11:00:18,918::logUtils::39::dispatcher::(wrapper) Run and protect:
repoStats, Return response: {}
Thread-157::DEBUG::2012-03-07
11:00:18,918::task::1172::TaskManager.Task::(prepare)
Task=`55a87a6e-6d63-4b40-94a5-1244cd7eb2aa`::finished: {}
Thread-157::DEBUG::2012-03-07
11:00:18,919::task::588::TaskManager.Task::(_updateState)
Task=`55a87a6e-6d63-4b40-94a5-1244cd7eb2aa`::moving from state preparing ->
state finished
Thread-157::DEBUG::2012-03-07
11:00:18,919::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-157::DEBUG::2012-03-07
11:00:18,920::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-157::DEBUG::2012-03-07
11:00:18,920::task::978::TaskManager.Task::(_decref)
Task=`55a87a6e-6d63-4b40-94a5-1244cd7eb2aa`::ref 0 aborting False
--
12 years, 3 months
Kernel crash on FC16
by Roi Dayan
Hi,
After discovery I select a LUN and give it a name, I click OK and I get a
kernel crash.
The kernel version is 3.1.0-7
I pasted the log below.
Any idea?
Thanks,
Roi
Mar 6 18:00:55 xena018 kernel: [ 339.762969] ------------[ cut here
]------------
Mar 6 18:00:55 xena018 kernel: [ 339.763233] kernel BUG at
block/blk-flush.c:323!
Mar 6 18:00:55 xena018 kernel: [ 339.763496] invalid opcode: 0000 [#1] SMP
Mar 6 18:00:55 xena018 kernel: [ 339.763839] CPU 6
Mar 6 18:00:55 xena018 kernel: [ 339.763885] Modules linked in:
ebtable_nat ebtables 8021q garp bonding lockd bridge stp llc be2iscsi
iscsi_boot_sysfs bnx2i cnic uio cxgb4i cxgb4 cxgb3i libcxgbi cxgb3 mdio
iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi ib_ipoib rdma_ucm
xt_multiport ib_ucm ib_uverbs ib_umad rdma_cm ib_cm ip6t_REJECT iw_cm
nf_conntrack_ipv6 nf_defrag_ipv6 xt_state ib_addr nf_conntrack
ip6table_filter ib_sa ip6_tables dm_round_robin dm_multipath mlx4_ib ib_mad
ib_core sunrpc vhost_net igb i7core_edac edac_core i2c_i801 ioatdma macvtap
macvlan tun virtio_net iTCO_wdt i2c_core kvm_intel iTCO_vendor_support dca
kvm joydev mlx4_core serio_raw microcode uinput [last unloaded: ib_iser]
Mar 6 18:00:55 xena018 kernel: [ 339.769133]
Mar 6 18:00:55 xena018 kernel: [ 339.769387] Pid: 3217, comm: mkfs.ext2
Not tainted 3.1.0-7.fc16.x86_64 #1 Supermicro X8DTU/X8DTU
Mar 6 18:00:55 xena018 kernel: [ 339.770000] RIP:
0010:[<ffffffff8120e16f>] [<ffffffff8120e16f>] blk_insert_flush+0x84/0x101
Mar 6 18:00:55 xena018 kernel: [ 339.770529] RSP: 0018:ffff880629ab3c28
EFLAGS: 00010046
Mar 6 18:00:55 xena018 kernel: [ 339.770785] RAX: 0000000000000000 RBX:
ffff880319410e30 RCX: 0000000008018215
Mar 6 18:00:55 xena018 kernel: [ 339.771050] RDX: 0000000000000000 RSI:
0000000000000001 RDI: ffff8806291cd008
Mar 6 18:00:55 xena018 kernel: [ 339.771314] RBP: ffff880629ab3c38 R08:
ffff8806291cd008 R09: 0000000000000001
Mar 6 18:00:55 xena018 kernel: [ 339.771571] R10: 0000000000000001 R11:
ffff880622982378 R12: ffff880319410e30
Mar 6 18:00:55 xena018 kernel: [ 339.771827] R13: 0000000000000046 R14:
0000000000000005 R15: ffff88061f0c4c00
Mar 6 18:00:55 xena018 kernel: [ 339.772085] FS: 00007fedacb44780(0000)
GS:ffff88063fc00000(0000) knlGS:0000000000000000
Mar 6 18:00:55 xena018 kernel: [ 339.772542] CS: 0010 DS: 0000 ES: 0000
CR0: 0000000080050033
Mar 6 18:00:55 xena018 kernel: [ 339.772794] CR2: 00000000017f4e70 CR3:
0000000628c25000 CR4: 00000000000006e0
Mar 6 18:00:55 xena018 kernel: [ 339.773056] DR0: 0000000000000000 DR1:
0000000000000000 DR2: 0000000000000000
Mar 6 18:00:55 xena018 kernel: [ 339.773316] DR3: 0000000000000000 DR6:
00000000ffff0ff0 DR7: 0000000000000400
Mar 6 18:00:55 xena018 kernel: [ 339.773573] Process mkfs.ext2 (pid:
3217, threadinfo ffff880629ab2000, task ffff88062569dcc0)
Mar 6 18:00:55 xena018 kernel: [ 339.774033] Stack:
Mar 6 18:00:55 xena018 kernel: [ 339.774284] ffff8806291cd008
ffff8806291cd008 ffff880629ab3c68 ffffffff812069f0
Mar 6 18:00:55 xena018 kernel: [ 339.774909] ffff8806291cd008
ffff880319410e30 0000000000000046 0000000000000005
Mar 6 18:00:55 xena018 kernel: [ 339.775530] ffff880629ab3c98
ffffffff8120b234 ffff8806291cd008 ffff880622982378
Mar 6 18:00:55 xena018 kernel: [ 339.776153] Call Trace:
Mar 6 18:00:55 xena018 kernel: [ 339.776405] [<ffffffff812069f0>]
__elv_add_request+0x1a5/0x1df
Mar 6 18:00:55 xena018 kernel: [ 339.776659] [<ffffffff8120b234>]
blk_insert_cloned_request+0x67/0x89
Mar 6 18:00:55 xena018 kernel: [ 339.776917] [<ffffffff8138df81>]
dm_dispatch_request+0x3b/0x61
Mar 6 18:00:55 xena018 kernel: [ 339.777172] [<ffffffff8138f76a>]
dm_request_fn+0x151/0x1f4
Mar 6 18:00:55 xena018 kernel: [ 339.777429] [<ffffffff8120731d>]
__blk_run_queue+0x1b/0x1d
Mar 6 18:00:55 xena018 kernel: [ 339.777685] [<ffffffff8120cc25>]
__make_request+0x276/0x2a9
Mar 6 18:00:55 xena018 kernel: [ 339.777944] [<ffffffff8138ea4b>]
dm_request+0x30/0x172
Mar 6 18:00:55 xena018 kernel: [ 339.778201] [<ffffffff8120b5a1>]
generic_make_request+0x27d/0x312
Mar 6 18:00:55 xena018 kernel: [ 339.778464] [<ffffffff810e0425>] ?
mempool_alloc_slab+0x15/0x17
Mar 6 18:00:55 xena018 kernel: [ 339.778723] [<ffffffff8120b714>]
submit_bio+0xde/0xfd
Mar 6 18:00:55 xena018 kernel: [ 339.778981] [<ffffffff81150dfb>] ?
bio_alloc_bioset+0x4c/0xc3
Mar 6 18:00:55 xena018 kernel: [ 339.779240] [<ffffffff8120dd33>]
blkdev_issue_flush+0x8f/0xcf
Mar 6 18:00:55 xena018 kernel: [ 339.779501] [<ffffffff811522f1>]
blkdev_fsync+0x35/0x42
Mar 6 18:00:55 xena018 kernel: [ 339.779759] [<ffffffff8114b3d7>]
vfs_fsync_range+0x23/0x25
Mar 6 18:00:55 xena018 kernel: [ 339.780012] [<ffffffff8114b3f5>]
vfs_fsync+0x1c/0x1e
Mar 6 18:00:55 xena018 kernel: [ 339.780260] [<ffffffff8114b6b4>]
sys_fsync+0x29/0x3c
Mar 6 18:00:55 xena018 kernel: [ 339.780512] [<ffffffff814bc482>]
system_call_fastpath+0x16/0x1b
Mar 6 18:00:55 xena018 kernel: [ 339.780761] Code: 80 e4 f3 41 89 40 40
85 f6 75 11 31 c9 31 d2 4c 89 c7 e8 df e2 ff ff e9 87 00 00 00 49 8b 40 60
48 85 c0 74 06 49 3b 40 68 74 02 <0f> 0b 83 fe 02 75 19 48 8b 73 08 4c 89
c7 48 89 da e8 40 15 02
Mar 6 18:00:55 xena018 kernel: [ 339.784140] RIP [<ffffffff8120e16f>]
blk_insert_flush+0x84/0x101
Mar 6 18:00:55 xena018 kernel: [ 339.784450] RSP <ffff880629ab3c28>
Mar 6 18:00:55 xena018 kernel: [ 339.784737] ---[ end trace
91d8c37eac182b74 ]---
--
*Roi *
12 years, 3 months