Puncte:0

openstack create instance stuck at scheduling and took long time to fail with ERROR

drapel us

anyone help please, i m unable to create an instance.

Build Environment

Based Hypevisory is ESXI (single Node)

Two VM Created with:

OS Centos 8 Steam with latest update
Openstack Yoga
Network = Flat Bridged
Hypervisor qemu on compute node

Compute Node: 8CPU ( Expose hardware assisted virtualization to the guest OS enabled) 50GBRAM 50GB for OS and 400GBvdisk for Cinder Volume ( Thick provisioned, eagerly zeroed )

Compute Node: 8CPU ( Expose hardware assisted virtualization to the guest OS enabled) 12GBRAM 50GBvdisk (Thick provisioned, eagerly zeroed)

both node DNS name is resolvable from DNS server and pingable.

Create instance using dashboard and command line is same result

ERROR

BlockquoteTraceback (most recent call last): File "/usr/lib/python3.6/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 441, in get return self._queues[msg_id].get(block=True, timeout=timeout) File "/usr/lib/python3.6/site-packages/eventlet/queue.py", line 322, in get return waiter.wait() File "/usr/lib/python3.6/site-packages/eventlet/queue.py", line 141, in wait return get_hub().switch() File "/usr/lib/python3.6/site-packages/eventlet/hubs/hub.py", line 313, in switch return self.greenlet.switch() queue.Empty During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.6/site-packages/nova/conductor/manager.py", line 1549, in schedule_and_build_instances instance_uuids, return_alternates=True) File "/usr/lib/python3.6/site-packages/nova/conductor/manager.py", line 910, in _schedule_instances return_alternates=return_alternates) File "/usr/lib/python3.6/site-packages/nova/scheduler/client/query.py", line 42, in select_destinations instance_uuids, return_objects, return_alternates) File "/usr/lib/python3.6/site-packages/nova/scheduler/rpcapi.py", line 160, in select_destinations return cctxt.call(ctxt, 'select_destinations', **msg_args) File "/usr/lib/python3.6/site-packages/oslo_messaging/rpc/client.py", line 192, in call retry=self.retry, transport_options=self.transport_options) File "/usr/lib/python3.6/site-packages/oslo_messaging/transport.py", line 128, in _send transport_options=transport_options) File "/usr/lib/python3.6/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 691, in send transport_options=transport_options) File "/usr/lib/python3.6/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 679, in _send call_monitor_timeout) File "/usr/lib/python3.6/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 567, in wait message = self.waiters.get(msg_id, timeout=timeout) File "/usr/lib/python3.6/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 445, in get 'to message ID %s' % msg_id) oslo_messaging.exceptions.MessagingTimeout: Timed out waiting for a reply to message ID c4216d4a20dd47b5860e15b52d5e99c2 Blockquote

from controller node:

journalctl -f SYSLOG_IDENTIFIER=nova-scheduler | grep -E "DEBUG|WARNING|ERROR"

instances from host 'nova-compute01'. sync_instance_info /usr/lib/python3.6/site-packages/nova/scheduler/host_manager.py:957 May 04 08:22:21 openstack.ad.local nova-scheduler[3142]: 2022-05-04 08:22:21.777 3142 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" "released" by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: held 0.001s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405 May 04 08:22:21 openstack.ad.local nova-scheduler[3144]: 2022-05-04 08:22:21.777 3144 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" "released" by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: held 0.001s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405 May 04 08:22:21 openstack.ad.local nova-scheduler[3139]: 2022-05-04 08:22:21.777 3139 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" "released" by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: held 0.001s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405 May 04 08:22:21 openstack.ad.local nova-scheduler[3137]: 2022-05-04 08:22:21.777 3137 DEBUG nova.scheduler.host_manager [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Successfully synced instances from host 'nova-compute01'. sync_instance_info /usr/lib/python3.6/site-packages/nova/scheduler/host_manager.py:957 May 04 08:22:21 openstack.ad.local nova-scheduler[3138]: 2022-05-04 08:22:21.777 3138 DEBUG nova.scheduler.host_manager [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Successfully synced instances from host 'nova-compute01'. sync_instance_info /usr/lib/python3.6/site-packages/nova/scheduler/host_manager.py:957 May 04 08:22:21 openstack.ad.local nova-scheduler[3136]: 2022-05-04 08:22:21.777 3136 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" acquired by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: waited 0.000s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:390 May 04 08:22:21 openstack.ad.local nova-scheduler[3143]: 2022-05-04 08:22:21.777 3143 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" "released" by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: held 0.001s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405 May 04 08:22:21 openstack.ad.local nova-scheduler[3137]: 2022-05-04 08:22:21.777 3137 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" "released" by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: held 0.001s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405 May 04 08:22:21 openstack.ad.local nova-scheduler[3138]: 2022-05-04 08:22:21.777 3138 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" "released" by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: held 0.001s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405 May 04 08:22:21 openstack.ad.local nova-scheduler[3136]: 2022-05-04 08:22:21.777 3136 DEBUG nova.scheduler.host_manager [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Successfully synced instances from host 'nova-compute01'. sync_instance_info /usr/lib/python3.6/site-packages/nova/scheduler/host_manager.py:957 May 04 08:22:21 openstack.ad.local nova-scheduler[3136]: 2022-05-04 08:22:21.777 3136 DEBUG oslo_concurrency.lockutils [req-2402ad44-ecfa-4b7a-a8c1-8aff6b4d8c20 - - - - -] Lock "host_instance" "released" by "nova.scheduler.host_manager.HostManager.sync_instance_info" :: held 0.001s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405

From Compute Node

journalctl -f SYSLOG_IDENTIFIER=nova-compute | grep -E "DEBUG|WARNING|ERROR"

May 04 08:26:21 nova-compute01 nova-compute[2137]: 2022-05-04 08:26:21.661 2137 DEBUG oslo_concurrency.lockutils [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 0.031s inner /usr/lib/python3.6/site-packages/oslo_concurrency/lockutils.py:405 May 04 08:26:26 nova-compute01 nova-compute[2137]: 2022-05-04 08:26:26.657 2137 DEBUG oslo_service.periodic_task [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Running periodic task ComputeManager._sync_scheduler_instance_info run_periodic_tasks /usr/lib/python3.6/site-packages/oslo_service/periodic_task.py:211 May 04 08:26:37 nova-compute01 nova-compute[2137]: 2022-05-04 08:26:37.567 2137 DEBUG oslo_service.periodic_task [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.6/site-packages/oslo_service/periodic_task.py:211 May 04 08:26:37 nova-compute01 nova-compute[2137]: 2022-05-04 08:26:37.567 2137 DEBUG nova.compute.manager [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.6/site-packages/nova/compute/manager.py:9444 May 04 08:26:37 nova-compute01 nova-compute[2137]: 2022-05-04 08:26:37.567 2137 DEBUG nova.compute.manager [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.6/site-packages/nova/compute/manager.py:9448 May 04 08:26:37 nova-compute01 nova-compute[2137]: 2022-05-04 08:26:37.573 2137 DEBUG nova.compute.manager [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Didn't find any instances for network info cache update. _heal_instance_info_cache /usr/lib/python3.6/site-packages/nova/compute/manager.py:9530 May 04 08:27:04 nova-compute01 nova-compute[2137]: 2022-05-04 08:27:04.566 2137 DEBUG oslo_service.periodic_task [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Running periodic task ComputeManager._instance_usage_audit run_periodic_tasks /usr/lib/python3.6/site-packages/oslo_service/periodic_task.py:211 May 04 08:27:07 nova-compute01 nova-compute[2137]: 2022-05-04 08:27:07.568 2137 DEBUG oslo_service.periodic_task [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Running periodic task ComputeManager._heal_instance_info_cache run_periodic_tasks /usr/lib/python3.6/site-packages/oslo_service/periodic_task.py:211 May 04 08:27:07 nova-compute01 nova-compute[2137]: 2022-05-04 08:27:07.568 2137 DEBUG nova.compute.manager [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Starting heal instance info cache _heal_instance_info_cache /usr/lib/python3.6/site-packages/nova/compute/manager.py:9444 May 04 08:27:07 nova-compute01 nova-compute[2137]: 2022-05-04 08:27:07.569 2137 DEBUG nova.compute.manager [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Rebuilding the list of instances to heal _heal_instance_info_cache /usr/lib/python3.6/site-packages/nova/compute/manager.py:9448 May 04 08:27:07 nova-compute01 nova-compute[2137]: 2022-05-04 08:27:07.574 2137 DEBUG nova.compute.manager [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Didn't find any instances for network info cache update. _heal_instance_info_cache /usr/lib/python3.6/site-packages/nova/compute/manager.py:9530 May 04 08:27:08 nova-compute01 nova-compute[2137]: 2022-05-04 08:27:08.568 2137 DEBUG oslo_service.periodic_task [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Running periodic task ComputeManager._poll_volume_usage run_periodic_tasks /usr/lib/python3.6/site-packages/oslo_service/periodic_task.py:211 May 04 08:27:09 nova-compute01 nova-compute[2137]: 2022-05-04 08:27:09.568 2137 DEBUG oslo_service.periodic_task [req-584789e4-2699-48bc-81f6-1fa6f3e2045e - - - - -] Running periodic task ComputeManager._poll_rebooting_instances run_periodic_tasks /usr/lib/python3.6/site-packages/oslo_service/periodic_task.py:211

drapel us
Este rabbitmq în funcțiune? Mesajul timeout ar putea fi legat de asta. Sunt serviciile de rețea în funcțiune? `Openstack Network Agent List`, `Openstack Compute Service List` ne pot oferi mai multe informații. Ce conține `nova-conductor.log`, orice indicii?

Postează un răspuns

Majoritatea oamenilor nu înțeleg că a pune multe întrebări deblochează învățarea și îmbunătățește legătura interpersonală. În studiile lui Alison, de exemplu, deși oamenii își puteau aminti cu exactitate câte întrebări au fost puse în conversațiile lor, ei nu au intuit legătura dintre întrebări și apreciere. În patru studii, în care participanții au fost implicați în conversații ei înșiși sau au citit transcrieri ale conversațiilor altora, oamenii au avut tendința să nu realizeze că întrebarea ar influența – sau ar fi influențat – nivelul de prietenie dintre conversatori.