nova service & hypervisors going up & down

Latest response

Hi Community,

I am deploying one PoC with Redhat OPenstack v10 in my LAB & facing problem in overcloud with the nova services & hypervisors which is continuously going up & down. Below is the environment details i have in my LAB.

1) HP C7000 Chassis with BL460 G9 Blades
2) Cloud Environment is with 3 Controllers & 4 Computes with No Ceph & Block Storage.
3) Using Single NIC networking for Control as well as Traffic.

network-environment.yaml

This file is an example of an environment file for defining the isolated

networks and related parameters.

resource_registry:
# Network Interface templates to use (these files must exist)
OS::TripleO::BlockStorage::Net::SoftwareConfig: /home/stack/templates/network/config/single-nic-vlans/cinder-storage.yaml
OS::TripleO::Compute::Net::SoftwareConfig: /home/stack/templates/network/config/single-nic-vlans/compute.yaml
OS::TripleO::Controller::Net::SoftwareConfig: /home/stack/templates/network/config/single-nic-vlans/controller.yaml
OS::TripleO::ObjectStorage::Net::SoftwareConfig: /home/stack/templates/network/config/single-nic-vlans/swift-storage.yaml
OS::TripleO::CephStorage::Net::SoftwareConfig: /home/stack/templates/network/config/single-nic-vlans/ceph-storage.yaml
OS::TripleO::NodeUserData: /home/stack/templates/first-boot.yaml
parameter_defaults:
# This section is where deployment-specific configuration is done
# CIDR subnet mask length for provisioning network
ControlPlaneSubnetCidr: '24'
# Gateway router for the provisioning network (or Undercloud IP)
ControlPlaneDefaultRoute: 192.168.20.1
EC2MetadataIp: 192.168.20.1 # Generally the IP of the Undercloud
# Customize the IP subnets to match the local environment
InternalApiNetCidr: 172.17.0.0/24
StorageNetCidr: 172.18.0.0/24
StorageMgmtNetCidr: 172.19.0.0/24
TenantNetCidr: 172.16.0.0/24
ExternalNetCidr: 10.50.166.0/27
# Customize the VLAN IDs to match the local environment
InternalApiNetworkVlanID: 201
StorageNetworkVlanID: 202
StorageMgmtNetworkVlanID: 203
TenantNetworkVlanID: 210
ExternalNetworkVlanID: 205
# Customize the IP ranges on each network to use for static IPs and VIPs
InternalApiAllocationPools: [{'start': '172.17.0.10', 'end': '172.17.0.200'}]
StorageAllocationPools: [{'start': '172.18.0.10', 'end': '172.18.0.200'}]
StorageMgmtAllocationPools: [{'start': '172.19.0.10', 'end': '172.19.0.200'}]
TenantAllocationPools: [{'start': '172.16.0.10', 'end': '172.16.0.200'}]
# Leave room if the external network is also used for floating IPs
ExternalAllocationPools: [{'start': '10.50.166.10', 'end': '10.50.166.25'}]
# Gateway router for the external network
ExternalInterfaceDefaultRoute: 10.50.166.1
# Uncomment if using the Management Network (see network-management.yaml)
# ManagementNetCidr: 10.0.1.0/24
# ManagementAllocationPools: [{'start': '10.0.1.10', 'end', '10.0.1.50'}]
# Use either this parameter or ControlPlaneDefaultRoute in the NIC templates
# ManagementInterfaceDefaultRoute: 10.0.1.1
# Define the DNS servers (maximum 2) for the overcloud nodes
DnsServers: ["8.8.8.8","8.8.4.4","192.168.20.1","147.128.148.150"]
# Set to empty string to enable multiple external networks or VLANs
NeutronExternalNetworkBridge: "''"
TimeZone: 'Asia/Calcutta'
NeutronTunnelTypes: "''"
NeutronNetworkType: 'vlan'
# The OVS logical->physical bridge mappings to use.

NeutronBridgeMappings: 'datacentre: br-ex'

NeutronNetworkVLANRanges: 'physnet2:1100:1500,datacentre:250:400'

Responses

Hi there,

Just to clarify, it's just the services that are going up and down, not the power states of the Compute nodes switching off then on?

Here are a couple of things to check for:

On a Compute node:

  • Log into a Compute node and check /var/log/nova/nova-compute.log for any errors.
  • Check if the openstack-nova-compute systemd service is continuously running on the Compute node or if it's continuously stopping and starting (or even just stopped dead). If continuously running, the Controller might be having trouble communicating with the Compute node. If stopping and starting, there might be a configuration issue with the Compute node.

On the Controller nodes:

  • Check /var/log/nova/nova-conductor.log on each Controller node for any errors.
  • Check /var/log/nova/nova-api.log on each Controller node to see if there are any odd API calls being made.
  • Check the openstack-nova-* and rabbitmq-server systemd services on each controller nodes.
Close

Welcome! Check out the Getting Started with Red Hat page for quick tours and guides for common tasks.