[victoria][octavia] Operating status OFFLINE

Malik Obaid malik.obaid at rapidcompute.com
Fri Feb 12 04:49:03 UTC 2021


Hi,

 

I have configured Octavia using Openstack Victoria release on Ubuntu 20.04.

 

Everything is working as expected like loadbalancing, healthmonitoring etc.

 

The issue I am facing is that the operating_status of Loadbalancer,
listener, pool and its members always stay OFFLINE.

 

I am new to Openstack  and would really appreciate your help in this regard.

 

 

-------------------------------------------------------------------------

 

loadbalancer status:

 

{

    "loadbalancer": {

        "id": "b33d7de7-4bf6-4815-a151-5ca7a7a3a40e",

        "name": "testlb",

        "operating_status": "OFFLINE",

        "provisioning_status": "ACTIVE",

        "listeners": [

            {

                "id": "03972160-dd11-4eac-855b-870ee9ee909b",

                "name": "testlistener",

                "operating_status": "OFFLINE",

                "provisioning_status": "ACTIVE",

                "pools": [

                    {

                        "id": "5ad57d34-a9e7-4aa1-982d-e528b38c84ed",

                        "name": "testpool",

                        "provisioning_status": "ACTIVE",

                        "operating_status": "OFFLINE",

                        "health_monitor": {

                            "id": "3524ea0f-ce2b-4957-9967-27d71033f964",

                            "name": "testhm",

                            "type": "HTTP",

                            "provisioning_status": "ACTIVE",

                            "operating_status": "ONLINE"

                        },

                       "members": [

                            {

                                "id":
"25b94432-6464-4245-a5db-6ecedb286721",

                                "name": "",

                                "operating_status": "OFFLINE",

                                "provisioning_status": "ACTIVE",

                                "address": "192.168.100.44",

                                "protocol_port": 80

                            },

                            {

                                "id":
"9157600c-280e-4eb9-a9aa-6b683da76420",

                                "name": "",

                                "operating_status": "OFFLINE",

                                "provisioning_status": "ACTIVE",

                                "address": "192.168.100.90",

                                "protocol_port": 80

                            }

                        ]

                    }

                ]

            }

        ]

    }

}

 

----------------------------------------------------------------------------
------------

 

octavia.conf

 

#create new

[DEFAULT]

# RabbitMQ connection info

transport_url = rabbit://openstack:password@172.16.30.46

 

[api_settings]

# IP address this host listens

bind_host = 172.16.30.46

bind_port = 9876

auth_strategy = keystone

api_base_uri = http://172.16.30.46:9876

 

# MariaDB connection info

[database]

connection = mysql+pymysql://octavia:password@172.16.30.45/octavia

 

[health_manager]

bind_ip = 172.16.30.46

bind_port = 5555

controller_ip_port_list = 127.0.0.1:5555, 172.16.30.46:5555

 

#healthcheck_enabled = True

 

[healthcheck]

backends = octavia_db_check

 

#debug = true

 

# Keystone auth info

[keystone_authtoken]

www_authenticate_uri = http://172.16.30.46:5000

auth_url = http://172.16.30.46:5000

memcached_servers = 172.16.30.46:11211

auth_type = password

project_domain_name = default

user_domain_name = default

project_name = service

username = octavia

password = servicepassword

 

# specify certificates created on [2]

[certificates]

ca_private_key = /etc/octavia/certs/private/server_ca.key.pem

ca_certificate = /etc/octavia/certs/server_ca.cert.pem

server_certs_key_passphrase = insecure-key-do-not-use-this-key

ca_private_key_passphrase = not-secure-passphrase

 

# specify certificates created on [2]

[haproxy_amphora]

server_ca = /etc/octavia/certs/server_ca-chain.cert.pem

client_cert = /etc/octavia/certs/private/client.cert-and-key.pem

 

# specify certificates created on [2]

[controller_worker]

client_ca = /etc/octavia/certs/client_ca.cert.pem

amp_image_tag = Amphora

# specify [flavor] ID for Amphora instance

amp_flavor_id = 200

# specify security group ID Amphora instance

amp_secgroup_list = 4fcb5a29-06b3-4a5d-8804-23c4670c200e

# specify network ID to boot Amphora instance (example below specifies
public network [public])

amp_boot_network_list = 7d6af354-206f-4c30-a0d6-dcf0f7f35f08

network_driver = allowed_address_pairs_driver

compute_driver = compute_nova_driver

amphora_driver = amphora_haproxy_rest_driver

 

[oslo_messaging]

topic = octavia_prov

 

# Keystone auth info

[service_auth]

auth_url = http://172.16.30.46:5000

memcached_servers = 172.16.30.46:11211

auth_type = password

project_domain_name = default

user_domain_name = default

project_name = service

username = octavia

password = servicepassword

 

-----------------------------------------------------------------------

 

Regards,

 

Malik Obaid

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-discuss/attachments/20210212/e545eec5/attachment-0001.html>


More information about the openstack-discuss mailing list