OK - I think I'm getting closer - now that I'm on the correct box.
here is the output of the vdsClient command - which is the device id - the
first line?
[***@hv-02 etc]# vdsClient -s 0 list
41703d5c-6cdb-42b4-93df-d78be2776e2b
Status = Up
acpiEnable = true
emulatedMachine = rhel6.5.0
afterMigrationStatus =
pid = 27304
memGuaranteedSize = 2048
transparentHugePages = true
displaySecurePort = 5902
spiceSslCipherSuite = DEFAULT
cpuType = SandyBridge
smp = 2
numaTune = {'nodeset': '0,1', 'mode': 'interleave'}
custom =
{'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72':
'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
deviceId=ebd4c73d-12c4-435e-8cc5-f180d8f20a72, device=unix, type=CHANNEL,
bootOrder=0, specParams={}, address={bus=0, controller=0,
type=virtio-serial, port=2}, managed=false, plugged=true, readOnly=false,
deviceAlias=channel1, customProperties={}, snapshotId=null}',
'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0e':
'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
deviceId=ffd2796f-7644-4008-b920-5f0970b0ef0e, device=unix, type=CHANNEL,
bootOrder=0, specParams={}, address={bus=0, controller=0,
type=virtio-serial, port=1}, managed=false, plugged=true, readOnly=false,
deviceAlias=channel0, customProperties={}, snapshotId=null}',
'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6': 'VmDevice
{vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
deviceId=86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6, device=ide, type=CONTROLLER,
bootOrder=0, specParams={}, address={slot=0x01, bus=0x00, domain=0x0000,
type=pci, function=0x1}, managed=false, plugged=true, readOnly=false,
deviceAlias=ide0, customProperties={}, snapshotId=null}',
'device_86f1aa5a-aa3f-4e47-b546-aafcc86fcbb6device_ebd4c73d-12c4-435e-8cc5-f180d8f20a72device_ffd2796f-7644-4008-b920-5f0970b0ef0edevice_6693d023-9c1f-433c-870e-e9771be8474b':
'VmDevice {vmId=41703d5c-6cdb-42b4-93df-d78be2776e2b,
deviceId=6693d023-9c1f-433c-870e-e9771be8474b, device=spicevmc,
type=CHANNEL, bootOrder=0, specParams={}, address={bus=0, controller=0,
type=virtio-serial, port=3}, managed=false, plugged=true, readOnly=false,
deviceAlias=channel2, customProperties={}, snapshotId=null}'}
vmType = kvm
memSize = 2048
smpCoresPerSocket = 1
vmName = connect-turbo-stage-03
nice = 0
bootMenuEnable = false
copyPasteEnable = true
displayIp = 10.1.90.161
displayPort = -1
smartcardEnable = false
clientIp =
fileTransferEnable = true
nicModel = rtl8139,pv
keyboardLayout = en-us
kvmEnable = true
pitReinjection = false
displayNetwork = ovirtmgmt
devices = [{'target': 2097152, 'specParams': {'model': 'none'}, 'alias':
'balloon0', 'deviceType': 'balloon', 'device': 'memballoon', 'type':
'balloon'}, {'device': 'unix', 'alias': 'channel0', 'address': {'bus': '0',
'controller': '0', 'type': 'virtio-serial', 'port': '1'}, 'deviceType':
'channel', 'type': 'channel'}, {'device': 'unix', 'alias': 'channel1',
'address': {'bus': '0', 'controller': '0', 'type': 'virtio-serial', 'port':
'2'}, 'deviceType': 'channel', 'type': 'channel'}, {'device': 'spicevmc',
'alias': 'channel2', 'address': {'bus': '0', 'controller': '0', 'type':
'virtio-serial', 'port': '3'}, 'deviceType': 'channel', 'type': 'channel'},
{'index': '0', 'alias': 'scsi0', 'specParams': {}, 'deviceType':
'controller', 'deviceId': '88db8cb9-0960-4797-bd41-1694bf14b8a9',
'address': {'slot': '0x04', 'bus': '0x00', 'domain': '0x0000', 'type':
'pci', 'function': '0x0'}, 'device': 'scsi', 'model': 'virtio-scsi',
'type': 'controller'}, {'alias': 'virtio-serial0', 'specParams': {},
'deviceType': 'controller', 'deviceId':
'4bb9c112-e027-4e7d-8b1c-32f99c7040ee', 'address': {'slot': '0x05', 'bus':
'0x00', 'domain': '0x0000', 'type': 'pci', 'function': '0x0'}, 'device':
'virtio-serial', 'type': 'controller'}, {'device': 'usb', 'alias': 'usb0',
'address': {'slot': '0x01', 'bus': '0x00', 'domain': '0x0000', 'type':
'pci', 'function': '0x2'}, 'deviceType': 'controller', 'type':
'controller'}, {'device': 'ide', 'alias': 'ide0', 'address': {'slot':
'0x01', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'function':
'0x1'}, 'deviceType': 'controller', 'type': 'controller'}, {'alias':
'video0', 'specParams': {'vram': '32768', 'ram': '65536', 'heads': '1'},
'deviceType': 'video', 'deviceId': '23634541-3b7e-460d-9580-39504392ba36',
'address': {'slot': '0x02', 'bus': '0x00', 'domain': '0x0000', 'type':
'pci', 'function': '0x0'}, 'device': 'qxl', 'type': 'video'}, {'device':
'spice', 'specParams': {'displayNetwork': 'ovirtmgmt',
'spiceSecureChannels':
'smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard',
'keyMap': 'en-us', 'displayIp': '10.1.90.161', 'copyPasteEnable': 'true'},
'deviceType': 'graphics', 'tlsPort': '5902', 'type': 'graphics'},
{'nicModel': 'pv', 'macAddr': '00:01:a4:a2:b4:30', 'linkActive': True,
'network': 'ovirtmgmt', 'alias': 'net0', 'filter': 'vdsm-no-mac-spoofing',
'specParams': {'inbound': {}, 'outbound': {}}, 'deviceType': 'interface',
'deviceId': '63651662-2ddf-4611-b988-1a58d05982f6', 'address': {'slot':
'0x03', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'function':
'0x0'}, 'device': 'bridge', 'type': 'interface', 'name': 'vnet5'},
{'nicModel': 'pv', 'macAddr': '00:01:a4:a2:b4:31', 'linkActive': True,
'network': 'storage', 'alias': 'net1', 'filter': 'vdsm-no-mac-spoofing',
'specParams': {'inbound': {}, 'outbound': {}}, 'deviceType': 'interface',
'deviceId': 'b112d9c6-5144-4b67-912b-dcc27aabdae9', 'address': {'slot':
'0x07', 'bus': '0x00', 'domain': '0x0000', 'type': 'pci', 'function':
'0x0'}, 'device': 'bridge', 'type': 'interface', 'name': 'vnet6'},
{'index': '3', 'iface': 'ide', 'name': 'hdd', 'alias': 'ide0-1-1',
'specParams': {'vmPayload': {'volId': 'config-2', 'file':
{'openstack/latest/meta_data.json':
'ewogICJsYXVuY2hfaW5kZXgiIDogIjAiLAogICJhdmFpbGFiaWxpdHlfem9uZSIgOiAibm92YSIs\nCiAgIm5hbWUiIDogImNvbm5lY3QtdHVyYm8tc3RhZ2UtMDMiLAogICJob3N0bmFtZSIgOiAiY29u\nbmVjdC10dXJiby1zdGFnZS0wMyIsCiAgInV1aWQiIDogImJiNmIwMzdhLTZkY2ItNGZmZS04MjUw\nLTMwYjlkOWE0ZTlmZCIsCiAgIm1ldGEiIDogewogICAgImVzc2VudGlhbCIgOiAiZmFsc2UiLAog\nICAgInJvbGUiIDogInNlcnZlciIsCiAgICAiZHNtb2RlIiA6ICJsb2NhbCIKICB9Cn0=\n',
'openstack/latest/user_data':
'I2Nsb3VkLWNvbmZpZwpzc2hfcHdhdXRoOiB0cnVlCmRpc2FibGVfcm9vdDogMApvdXRwdXQ6CiAg\nYWxsOiAnPj4gL3Zhci9sb2cvY2xvdWQtaW5pdC1vdXRwdXQubG9nJwpjaHBhc3N3ZDoKICBleHBp\ncmU6IGZhbHNlCnJ1bmNtZDoKLSAnc2VkIC1pICcnL15kYXRhc291cmNlX2xpc3Q6IC9kJycgL2V0\nYy9jbG91ZC9jbG91ZC5jZmc7IGVjaG8gJydkYXRhc291cmNlX2xpc3Q6CiAgWyJOb0Nsb3VkIiwg\nIkNvbmZpZ0RyaXZlIl0nJyA+PiAvZXRjL2Nsb3VkL2Nsb3VkLmNmZycK\n'}}},
'readonly': 'True', 'deviceType': 'disk', 'deviceId':
'6de890b2-6454-4377-9a71-bea2e46d50a8', 'address': {'bus': '1',
'controller': '0', 'type': 'drive', 'target': '0', 'unit': '1'}, 'device':
'cdrom', 'shared': 'false', 'path': '', 'type': 'disk'}, {'index': '2',
'iface': 'ide', 'name': 'hdd', 'alias': 'ide0-1-1', 'specParams': {'path':
''}, 'readonly': 'True', 'deviceType': 'disk', 'deviceId':
'2763a41b-6576-4135-b349-4fe402c31246', 'address': {'bus': '1',
'controller': '0', 'type': 'drive', 'target': '0', 'unit': '1'}, 'device':
'cdrom', 'shared': 'false', 'path': '', 'type': 'disk'}, {'device': 'file',
'alias': 'ide0-1-0', 'address': {'bus': '1', 'controller': '0', 'type':
'drive', 'target': '0', 'unit': '0'}, 'deviceType': 'disk', 'type': 'disk'}]
timeOffset = -891891
maxVCpus = 16
spiceSecureChannels =
smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard
display = qxl
[***@hv-02 etc]#
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
800.885.8886 x128 | ***@telvue.com | http://www.telvue.com
twitter: http://twitter.com/telvue | facebook:
https://www.facebook.com/telvue
Post by Mark SteeleI think I may have not given you all the information.
I am not logging into the host - I am logging into the ovirt management.
Let me try logging into the host and checking
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
https://www.facebook.com/telvue
Post by Mark SteeleThat command returns nothing - I don't think qemu is running?
Not sure how to start it on CentOS
root 23279 23130 0 08:51 pts/0 00:00:00 grep qemu
that mean you don't have vm running on that host. so you can restart vdsm
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
http://www.telvue.com
twitter: <http://twitter.com/telvue>http://twitter.com/telvue |
facebook: <https://www.facebook.com/telvue>
https://www.facebook.com/telvue
Post by Mark SteeleI run into the same issue - I am unable to completely go into
maintenance mode because this VM is still on it - it cannot be migrated
because it is not managed.
pgrep -an qemu-kvm | grep external
and kill the process
[image: Inline image 1]
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
<http://www.telvue.com>http://www.telvue.com
twitter: <http://twitter.com/telvue>http://twitter.com/telvue |
facebook: <https://www.facebook.com/telvue>
https://www.facebook.com/telvue
Post by Mark Steeletotal 48K
drwxr-xr-x. 10 root root 4.0K Nov 14 2014 .
drwxr-xr-x. 118 root root 12K Jul 12 03:35 ..
drwxr-xr-x. 6 root root 4.0K Nov 14 2014 CA
drwxr-xr-x. 4 root root 4.0K Nov 14 2014 ca-trust
drwxr-xr-x. 2 root root 4.0K Nov 14 2014 java
drwxr-xr-x. 2 root root 4.0K Jul 12 07:03 nssdb
drwxr-xr-x. 6 ovirt ovirt 4.0K Nov 19 2014 ovirt-engine
drwxr-xr-x. 2 root root 4.0K Nov 14 2014 rpm-gpg
drwx------. 2 root root 4.0K Nov 22 2013 rsyslog
drwxr-xr-x. 5 root root 4.0K Nov 14 2014 tls
There is no vsdm directory under /etc/pki
This is an ovirt node. Version of software is 3.5.0.1-1.el6 from the
ovirt management console.
I'd like to add that I am not the person who originally installed
this instance - and am not entirely familiar with how it is setup and
installed - so I may ask ignorant questions from time to time.
not urgent but at this point it looks like it would be good to
reinstall this host from the webadmin. if you have the capacity,
you can put the host to maintenance, that will migrate vms to other
hosts, and then choose "reinstall" once its in "maintenance'
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
<http://www.telvue.com>http://www.telvue.com
https://www.facebook.com/telvue
Post by Mark SteeleThank you Roy,
I installed the client but am getting a permissions error when I run it
File "/usr/share/vdsm/vdsClient.py", line 2678, in <module>
serv.do_connect(hostPort)
File "/usr/share/vdsm/vdsClient.py", line 136, in do_connect
self.s = vdscli.connect(hostPort, self.useSSL, self.truststore)
File "/usr/lib/python2.6/site-packages/vdsm/vdscli.py", line 110, in
connect
raise Exception("No permission to read file: %s" % f)
Exception: No permission to read file: /etc/pki/vdsm/keys/vdsmkey.pem
This should work. something isn't right with your setup
is your host an ovirt-node? could be that you hit [1] . let me know
what version are you running.
please try the same with user vdsm. it should have permissions to
/etc/pki/vdsm
[1] <https://gerrit.ovirt.org/#/c/27779/>
https://gerrit.ovirt.org/#/c/27779/
If I restart vdsm, will that cause any issues with running VM's on
this ovirt installation? This is our production environment.
Generatlly the answer is no but lets avoid it if we can for this is
a minor cosmetic issue I guess.
just as FYI - vdsm only reconnects to the socket exposed by libvirt to
control the vm lifecycle. VDSM doesn't manadate the lifecycle of a VM
unless the engine tells is so. Storage wise there could be some operations
but i'm almost sure they must not have effect on running VMs.
Thank you
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
<http://www.telvue.com>http://www.telvue.com
<https://www.facebook.com/telvue>https://www.facebook.com/telvue
Post by Mark SteeleYes,
[image: Inline image 1]
the vdsm on that host reports it back to the engine . since this vm
isn't in the engine DB it is concidered as EXTERNAL (thus the error 400
from the API)
do yo know if the qemu-kvm proccess isn't running anymore?
if the process isn't running then vdsm must clean its cache
yum install vdsm-cli
vdsClient -s 0 list
vdsClient -s 0 destroy {vmId}
alternativly a vdsm restart will work (if the qemu proccess isn't running)
The VM is not really running - the IP addresses that are being
reported are from another VM that was recently removed. All attempts to
control the VM have failed. It does not have any NICS or disk associated
with it - so this seems to be a ghost in the machine. I attempted to unlock
it using the unlock_entity.sh script - it reports successful, however I
still cannot do anything with the VM.
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
<http://www.telvue.com>http://www.telvue.com
<https://www.facebook.com/telvue>https://www.facebook.com/telvue
On Thu, Jul 9, 2015 at 11:11 AM, Artyom Lukianov <
Post by Artyom LukianovCan you sea via engine, on what host run VM?
Anyway if you have really run VM on host you can try to figure it
with 'ps aux | grep qemu', if it will return you some process, you can just
kill process via 'kill pid'.
I hope it will help you.
----- Original Message -----
Sent: Thursday, July 9, 2015 5:42:20 PM
Subject: Re: [ovirt-users] This VM is not managed by the engine
Artyom,
Thank you - I don't have vdsClient installed - can you point me to the
download?
***
*Mark Steele*
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
twitter: <http://twitter.com/telvue>http://twitter.com/telvue |
<https://www.facebook.com/telvue>https://www.facebook.com/telvue
On Thu, Jul 9, 2015 at 10:11 AM, Artyom Lukianov <
Post by Artyom LukianovPlease check host where VM run(vdsClient -s 0 list table), and you
can
Post by Artyom Lukianovdestroy it via vdsClient(vdsClient -s 0 destroy vm_id).
Thanks
----- Original Message -----
Sent: Thursday, July 9, 2015 4:38:32 PM
Subject: [ovirt-users] This VM is not managed by the engine
I have a VM that was not started and is now showing as running.
When I
Post by Artyom Lukianovattempt to suspend or stop it in the ovirt-shell, I get the
status: 400
reason: bad request
detail: Cannot hibernate VM. This VM is not managed by the engine.
Not sure how the VM was initially created on the ovirt manager.
This VM is
Post by Artyom Lukianovnot needed - how can I 'shutdown' and remove this VM?
Thanks
***
Mark Steele
CIO / VP Technical Operations | TelVue Corporation
TelVue - We Share Your Vision
twitter: <http://twitter.com/telvue>http://twitter.com/telvue |
<https://www.facebook.com/telvue>https://www.facebook.com/telvue
_______________________________________________
Users mailing list
<http://lists.ovirt.org/mailman/listinfo/users>
http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________