Hi,
I am having trouble configuring Openstack to use Ceph RGW as the Object store backend for Swift and S3.

My setup is an HCI, I have 3 controllers which are also my ceph mgrs, mons and rgws and 9 compte/storage servers (osds).
Xena is deployed with Ceph Pacific.

Ceph public network is a private network on vlan10 with 10.10.1.0/24 as a subnet.

Here is a snippet from my globals.yml :
---
kolla_base_distro: "centos"
kolla_install_type: "source"
openstack_release: "xena"
kolla_internal_vip_address: "10.10.3.1"
kolla_internal_fqdn: "dashint.cloud.example.com"
kolla_external_vip_address: "x.x.x.x"
kolla_external_fqdn: "dash.cloud.example.com "
docker_registry: 192.168.1.16:4000
network_interface: "bond0"
kolla_external_vip_interface: "bond1"
api_interface: "bond1.30"
storage_interface: "bond1.10"       <---------------- VLAN10 (public ceph network)
tunnel_interface: "bond1.40"
dns_interface: "bond1"
octavia_network_interface: "bond1.301"
neutron_external_interface: "bond2"
neutron_plugin_agent: "openvswitch"
keepalived_virtual_router_id: "51"
kolla_enable_tls_internal: "yes"
kolla_enable_tls_external: "yes"
kolla_certificates_dir: "{{ node_config }}/certificates"
kolla_external_fqdn_cert: "{{ kolla_certificates_dir }}/haproxy.pem"
kolla_internal_fqdn_cert: "{{ kolla_certificates_dir }}/haproxy-internal.pem"
kolla_admin_openrc_cacert: "{{ kolla_certificates_dir }}/ca.pem"
kolla_copy_ca_into_containers: "yes"
kolla_enable_tls_backend: "yes"
kolla_verify_tls_backend: "no"
kolla_tls_backend_cert: "{{ kolla_certificates_dir }}/backend-cert.pem"
kolla_tls_backend_key: "{{ kolla_certificates_dir }}/backend-key.pem"
enable_openstack_core: "yes"
enable_hacluster: "yes"
enable_haproxy: "yes"
enable_aodh: "yes"
enable_barbican: "yes"
enable_ceilometer: "yes"
enable_central_logging: "yes"
enable_ceph_rgw: "yes"
enable_ceph_rgw_loadbalancer: "{{ enable_ceph_rgw | bool }}"

enable_cinder: "yes"
enable_cinder_backup: "yes"
enable_collectd: "yes"
enable_designate: "yes"
enable_elasticsearch_curator: "yes"
enable_freezer: "no"
enable_gnocchi: "yes"
enable_gnocchi_statsd: "yes"
enable_magnum: "yes"
enable_manila: "yes"
enable_manila_backend_cephfs_native: "yes"
enable_mariabackup: "yes"
enable_masakari: "yes"
enable_neutron_vpnaas: "yes"
enable_neutron_qos: "yes"
enable_neutron_agent_ha: "yes"
enable_neutron_provider_networks: "yes"
enable_neutron_segments: "yes"
enable_octavia: "yes"
enable_trove: "yes"
external_ceph_cephx_enabled: "yes"
ceph_glance_keyring: "ceph.client.glance.keyring"
ceph_glance_user: "glance"
ceph_glance_pool_name: "images"
ceph_cinder_keyring: "ceph.client.cinder.keyring"
ceph_cinder_user: "cinder"
ceph_cinder_pool_name: "volumes"
ceph_cinder_backup_keyring: "ceph.client.cinder-backup.keyring"
ceph_cinder_backup_user: "cinder-backup"
ceph_cinder_backup_pool_name: "backups"
ceph_nova_keyring: "{{ ceph_cinder_keyring }}"
ceph_nova_user: "cinder"
ceph_nova_pool_name: "vms"
ceph_gnocchi_keyring: "ceph.client.gnocchi.keyring"
ceph_gnocchi_user: "gnocchi"
ceph_gnocchi_pool_name: "metrics"
ceph_manila_keyring: "ceph.client.manila.keyring"
ceph_manila_user: "manila"
glance_backend_ceph: "yes"
glance_backend_file: "no"
gnocchi_backend_storage: "ceph"
cinder_backend_ceph: "yes"
cinder_backup_driver: "ceph"
cloudkitty_collector_backend: "gnocchi"
designate_ns_record: "cloud.example.com "
nova_backend_ceph: "yes"
nova_compute_virt_type: "kvm"
octavia_auto_configure: yes
octavia_amp_flavor:
  name: "amphora"
  is_public: no
  vcpus: 1
  ram: 1024
  disk: 5
octavia_amp_network:
  name: lb-mgmt-net
  provider_network_type: vlan
  provider_segmentation_id: 301
  provider_physical_network: physnet1
  external: false
  shared: false
  subnet:
    name: lb-mgmt-subnet
    cidr: "10.7.0.0/16"
    allocation_pool_start: "10.7.0.50"
    allocation_pool_end: "10.7.255.200"
    no_gateway_ip: yes
    enable_dhcp: yes
    mtu: 9000
octavia_amp_network_cidr: 10.10.7.0/24
octavia_amp_image_tag: "amphora"
octavia_certs_country: XZ
octavia_certs_state: Gotham
octavia_certs_organization: WAYNE
octavia_certs_organizational_unit: IT
horizon_keystone_multidomain: true
elasticsearch_curator_dry_run: "no"
enable_cluster_user_trust: true
ceph_rgw_hosts:
        - host: controllera
          ip: 10.10.1.5
          port: 8080
        - host: controllerb
          ip: 10.10.1.9
          port: 8080
        - host: controllerc
          ip: 10.10.1.13
          port: 8080
ceph_rgw_swift_account_in_url: true
ceph_rgw_swift_compatibility: true


And Here is my ceph all.yml file
---
dummy:
ceph_release_num: 16
cluster: ceph
configure_firewall: False
monitor_interface: bond1.10
monitor_address_block: 10.10.1.0/24
is_hci: true
hci_safety_factor: 0.2
osd_memory_target: 4294967296
public_network: 10.10.1.0/24
cluster_network: 10.10.2.0/24
radosgw_interface: "{{ monitor_interface }}"
radosgw_address_block: 10.10.1.0/24
nfs_file_gw: true
nfs_obj_gw: true
ceph_docker_image: "ceph/daemon"
ceph_docker_image_tag: latest-pacific
ceph_docker_registry: 192.168.1.16:4000
containerized_deployment: True
openstack_config: true
openstack_glance_pool:
  name: "images"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 128
  pgp_num: 128
  target_size_ratio: 5.00
  rule_name: "SSD"
openstack_cinder_pool:
  name: "volumes"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 1024
  pgp_num: 1024
  target_size_ratio: 42.80
  rule_name: "SSD"
openstack_nova_pool:
  name: "vms"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 256
  pgp_num: 256
  target_size_ratio: 10.00
  rule_name: "SSD"
openstack_cinder_backup_pool:
  name: "backups"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 512
  pgp_num: 512
  target_size_ratio: 18.00
  rule_name: "SSD"
openstack_gnocchi_pool:
  name: "metrics"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 32
  pgp_num: 32
  target_size_ratio: 0.10
  rule_name: "SSD"
openstack_cephfs_data_pool:
  name: "cephfs_data"
  pg_autoscale_mode: False
  application: "cephfs"
  pg_num: 256
  pgp_num: 256
  target_size_ratio: 10.00
  rule_name: "SSD"
openstack_cephfs_metadata_pool:
  name: "cephfs_metadata"
  pg_autoscale_mode: False
  application: "cephfs"
  pg_num: 32
  pgp_num: 32
  target_size_ratio: 0.10
  rule_name: "SSD"
openstack_pools:
  - "{{ openstack_glance_pool }}"
  - "{{ openstack_cinder_pool }}"
  - "{{ openstack_nova_pool }}"
  - "{{ openstack_cinder_backup_pool }}"
  - "{{ openstack_gnocchi_pool }}"
  - "{{ openstack_cephfs_data_pool }}"
  - "{{ openstack_cephfs_metadata_pool }}"
openstack_keys:
  - { name: client.glance, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" }
  - { name: client.cinder, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" }
  - { name: client.cinder-backup, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" }
  - { name: client.gnocchi, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_gnocchi_pool.name }}"}, mode: "0600", }
  - { name: client.openstack, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_glance_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" }
dashboard_enabled: True
dashboard_protocol: https
dashboard_port: 8443
dashboard_network: "192.168.1.0/24"
dashboard_admin_user: admin
dashboard_admin_user_ro: true
dashboard_admin_password: ***********
dashboard_crt: '/home/deployer/work/site-central/chaininv.crt'
dashboard_key: '/home/deployer/work/site-central/cloud_example.com.priv'
dashboard_grafana_api_no_ssl_verify: true
dashboard_rgw_api_user_id: admin
dashboard_rgw_api_no_ssl_verify: true
dashboard_frontend_vip: '192.168.1.5'
node_exporter_container_image: "192.168.1.16:4000/prom/node-exporter:v0.17.0"
grafana_admin_user: admin
grafana_admin_password: *********
grafana_crt: '/home/deployer/work/site-central/chaininv.crt'
grafana_key: '/home/deployer/work/site-central/cloud_example.com.priv'
grafana_server_fqdn: 'grafanasrv.cloud.example.com'
grafana_container_image: "192.168.1.16:4000/grafana/grafana:6.7.4"
grafana_dashboard_version: pacific
prometheus_container_image: "192.168.1.16:4000/prom/prometheus:v2.7.2"
alertmanager_container_image: "192.168.1.16:4000/prom/alertmanager:v0.16.2"

And my rgws.yml
---
dummy:
copy_admin_key: true
rgw_create_pools:
  "{{ rgw_zone }}.rgw.buckets.data":
    pg_num: 256
    pgp_num: 256
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.buckets.index":
    pg_num: 64
    pgp_num: 64
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.meta":
    pg_num: 32
    pgp_num: 32
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.log":
    pg_num: 32
    pgp_num: 32
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.control":
    pg_num: 32
    pgp_num: 32
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1

The ceph_rgw user was created by kolla
(xenavenv) [deployer@rscdeployer ~]$ openstack user list | grep ceph
| 3262aa7e03ab49c8a5710dfe3b16a136 | ceph_rgw

This is my ceph.conf from one of my controllers :
[root@controllera ~]# cat /etc/ceph/ceph.conf
[client.rgw.controllera.rgw0]
host = controllera
rgw_keystone_url = https://dash.cloud.example.com:5000
##Authentication using username, password and tenant. Preferred.
rgw_keystone_verify_ssl = false
rgw_keystone_api_version = 3
rgw_keystone_admin_user = ceph_rgw
rgw_keystone_admin_password = cos2Jcnpnw9BhGwvPm**************************
rgw_keystone_admin_domain = Default
rgw_keystone_admin_project = service
rgw_s3_auth_use_keystone = true
rgw_keystone_accepted_roles = admin
rgw_keystone_implicit_tenants = true
rgw_swift_account_in_url = true
keyring = /var/lib/ceph/radosgw/ceph-rgw.controllera.rgw0/keyring
log file = /var/log/ceph/ceph-rgw-controllera.rgw0.log
rgw frontends = beast endpoint=10.10.1.5:8080
rgw thread pool size = 512
#For Debug
debug ms = 1
debug rgw = 20


# Please do not change this file directly since it is managed by Ansible and will be overwritten
[global]
cluster network = 10.10.2.0/24
fsid = da094354-6ade-415a-a424-************
mon host = [v2:10.10.1.5:3300,v1:10.10.1.5:6789],[v2:10.10.1.9:3300,v1:10.10.1.9:6789],[v2:10.10.1.13:3300,v1:10.10.1.13:6789]
mon initial members = controllera,controllerb,controllerc
osd pool default crush rule = 1
public network = 10.10.1.0/24


Here are my swift endpoints
(xenavenv) [deployer@rscdeployer ~]$ openstack endpoint list | grep swift
| 4082b4acf8bc4e4c9efc6e2d0e293724 | RegionOne | swift        | object-store    | True    | admin     | https://dashint.cloud.example.com:6780/v1/AUTH_%(project_id)s |
| b13a2f53e13e4650b4efdb8184eb0211 | RegionOne | swift        | object-store    | True    | internal  | https://dashint.cloud.example.com:6780/v1/AUTH_%(project_id)s |
| f85b36ff9a2b49bc9eaadf1aafdee28c | RegionOne | swift        | object-store    | True    | public    | https://dash.cloud.example.com:6780/v1/AUTH_%(project_id)s    |

When I connect to Horizon -> Project -> Object Store -> Containers I get theses errors :
I cannot create a new container from the WebUI, the Storage policy parameter is empty.
If I try to create a new container from the CLI, I get this :
 (xenavenv) [deployer@rscdeployer ~]$ source cephrgw-openrc.sh
(xenavenv) [deployer@rscdeployer ~]$ openstack container create demo -v
START with options: container create demo -v
command: container create -> openstackclient.object.v1.container.CreateContainer (auth=True)
Using auth plugin: password
Not Found (HTTP 404)
END return value: 1


This is the log from RGW service when I execute the above command :
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 CONTENT_LENGTH=0
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_ACCEPT=*/*
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_ACCEPT_ENCODING=gzip, deflate
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_HOST=dashint.cloud.example.com:6780
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_USER_AGENT=openstacksdk/0.59.0 keystoneauth1/4.4.0 python-requests/2.26.0 CPython/3.8.8
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_VERSION=1.1
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_AUTH_TOKEN=gAAAAABiXUrjDFNzXx03mt1lbpUiCqNND1HACspSfg6h_TMxKYND5Hb9BO3FxH0a7CYoBXgRJywGszlK8cl-7zbUNRjHmxgIzmyh-CrWyGv793ZLOAmT_XShcrIKThjIIH3gTxYoX1TXwOKbsvMuZnI5EKKsol2y2MhcqPLeLGc28_AwoOr_b80
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_FORWARDED_FOR=10.10.3.16
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_FORWARDED_PROTO=https
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REMOTE_ADDR=10.10.1.13
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REQUEST_METHOD=PUT
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REQUEST_URI=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 SCRIPT_URI=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 SERVER_PORT=8080
2022-04-18T12:26:27.995+0100 7f22e07a9700  1 ====== starting new request req=0x7f23221aa620 =====
2022-04-18T12:26:27.995+0100 7f22e07a9700  2 req 728157015944164764 0.000000000s initializing for trans_id = tx000000a1aeef2b40f759c-00625d4ae3-4b389-default
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s rgw api priority: s3=8 s3website=7
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s host=dashint.cloud.example.com
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s subdomain= domain= in_hosted_domain=0 in_hosted_domain_s3website=0
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s final domain/bucket subdomain= domain= in_hosted_domain=0 in_hosted_domain_s3website=0 s->info.domain= s->info.request_uri=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s get_handler handler=22RGWHandler_REST_Obj_S3
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s handler=22RGWHandler_REST_Obj_S3
2022-04-18T12:26:27.995+0100 7f22e07a9700  2 req 728157015944164764 0.000000000s getting op 1
2022-04-18T12:26:27.995+0100 7f22e07a9700  1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.7:6801/4815,v1:10.10.1.7:6803/4815] -- osd_op(unknown.0.0:1516 12.3 12:c14cb721:::script.prerequest.:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2c400 con 0x56055e53b000
2022-04-18T12:26:27.996+0100 7f230d002700  1 -- 10.10.1.13:0/2715436964 <== osd.23 v2:10.10.1.7:6801/4815 22 ==== osd_op_reply(1516 script.prerequest. [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 246+0+0 (crc 0 0 0) 0x56055ea18b40 con 0x56055e53b000
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj scheduling with throttler client=2 cost=1
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj op=21RGWPutObj_ObjStore_S3
2022-04-18T12:26:27.996+0100 7f22ddfa4700  2 req 728157015944164764 0.001000002s s3:put_obj verifying requester
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::StrategyRegistry::s3_main_strategy_t: trying rgw::auth::s3::AWSAuthStrategy
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::AWSAuthStrategy: trying rgw::auth::s3::S3AnonymousEngine
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::S3AnonymousEngine granted access
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::AWSAuthStrategy granted access
2022-04-18T12:26:27.996+0100 7f22ddfa4700  2 req 728157015944164764 0.001000002s s3:put_obj normalizing buckets and tenants
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s->object=AUTH_971efa4cb18f42f7a405342072c39c9d/demo s->bucket=v1
2022-04-18T12:26:27.996+0100 7f22ddfa4700  2 req 728157015944164764 0.001000002s s3:put_obj init permissions
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj get_system_obj_state: rctx=0x7f23221a9000 obj=default.rgw.meta:root:v1 state=0x56055ea8c520 s->prefetch_data=0
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj cache get: name=default.rgw.meta+root+v1 : miss
2022-04-18T12:26:27.996+0100 7f22ddfa4700  1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.3:6802/4933,v1:10.10.1.3:6806/4933] -- osd_op(unknown.0.0:1517 11.b 11:d05f7b30:root::v1:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2cc00 con 0x56055e585000
2022-04-18T12:26:27.997+0100 7f230c801700  1 -- 10.10.1.13:0/2715436964 <== osd.3 v2:10.10.1.3:6802/4933 9 ==== osd_op_reply(1517 v1 [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 230+0+0 (crc 0 0 0) 0x56055e39db00 con 0x56055e585000
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj cache put: name=default.rgw.meta+root+v1 info.flags=0x0
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj adding default.rgw.meta+root+v1 to cache LRU end
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj init_permissions on <NULL> failed, ret=-2002
2022-04-18T12:26:27.997+0100 7f22dd7a3700  1 req 728157015944164764 0.002000004s op->ERRORHANDLER: err_no=-2002 new_err_no=-2002
2022-04-18T12:26:27.997+0100 7f22dbfa0700  1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.8:6804/4817,v1:10.10.1.8:6805/4817] -- osd_op(unknown.0.0:1518 12.1f 12:fb11263f:::script.postrequest.:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2d000 con 0x56055e94c800
2022-04-18T12:26:27.998+0100 7f230d002700  1 -- 10.10.1.13:0/2715436964 <== osd.9 v2:10.10.1.8:6804/4817 10 ==== osd_op_reply(1518 script.postrequest. [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 247+0+0 (crc 0 0 0) 0x56055ea18b40 con 0x56055e94c800
2022-04-18T12:26:27.998+0100 7f22d8f9a700  2 req 728157015944164764 0.003000006s s3:put_obj op status=0
2022-04-18T12:26:27.998+0100 7f22d8f9a700  2 req 728157015944164764 0.003000006s s3:put_obj http status=404
2022-04-18T12:26:27.998+0100 7f22d8f9a700  1 ====== req done req=0x7f23221aa620 op status=0 http_status=404 latency=0.003000006s ======
2022-04-18T12:26:27.998+0100 7f22d8f9a700  1 beast: 0x7f23221aa620: 10.10.1.13 - anonymous [18/Apr/2022:12:26:27.995 +0100] "PUT /v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo HTTP/1.1" 404 214 - "openstacksdk/0.59.0 keystoneauth1/4.4.0 python-requests/2.26.0 CPython/3.8.8" - latency=0.003000006s

Could you help please.

Regards.