Ceph: After OSD node was added to Ceph cluster Open Stack Nova instances shut down

Solution Verified - Updated -

Issue

  • An OSD node was added to an existing Ceph Storage cluster to expand space. After the OSD node was successfully added all Nova instances in Openstack region to shut down.

  • From Nova Compute log:

015-09-03 11:22:17.903 18735 ERROR nova.openstack.common.periodic_task [-] Error during ComputeManager.update_available_resource: error calling connect
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task Traceback (most recent call last):
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/openstack/common/periodic_task.py", line 198, in run_periodic_tasks
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    task(self, context)
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 6027, in update_available_resource
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    rt.update_available_resource(context)
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/compute/resource_tracker.py", line 370, in update_available_resource
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    resources = self.driver.get_available_resource(self.nodename)
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 5095, in get_available_resource
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    stats = self.get_host_stats(refresh=True)
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 6178, in get_host_stats
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    return self.host_state.get_host_stats(refresh=refresh)
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 6752, in get_host_stats
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    self.update_status()
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 6775, in update_status
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    disk_info_dict = self.driver._get_local_gb_info()
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/driver.py", line 4688, in _get_local_gb_info
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    info = LibvirtDriver._get_rbd_driver().get_pool_info()
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/rbd_utils.py", line 286, in get_pool_info
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    with RADOSClient(self) as client:
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/rbd_utils.py", line 86, in __init__
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    self.cluster, self.ioctx = driver._connect_to_rados(pool)
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/nova/virt/libvirt/rbd_utils.py", line 110, in _connect_to_rados
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    client.connect()
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task  File "/usr/lib/python2.7/site-packages/rados.py", line 419, in connect
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task    raise make_ex(ret, "error calling connect")
2015-09-03 11:22:17.903 18735 TRACE nova.openstack.common.periodic_task TimedOut: error calling connect
cant connect
  • Display from 'rbd list' command on Ceph cluster:
terminate called after throwing an instance of 'ceph::buffer::malformed_input'
what():  buffer::malformed_input: unsupported bucket algorithm: 5

Environment

  • Red Had Ceph Storage 1.3
  • Red Hat Open Stack 6.0

Subscriber exclusive content

A Red Hat subscription provides unlimited access to our knowledgebase, tools, and much more.

Current Customers and Partners

Log in for full access

Log In

New to Red Hat?

Learn more about Red Hat subscriptions

Using a Red Hat product through a public cloud?

How to access this content