Dear Wodel,I think that default endpoint for swift when using cephrgw is /swift/v1 (unless you have changed it in ceph),so your endpoints should be| 4082b4acf8bc4e4c9efc6e2d0e293724 | RegionOne | swift | object-store | True | admin | https://dashint.cloud.example.com:6780/swift/v1/AUTH_%(project_id)s |
| b13a2f53e13e4650b4efdb8184eb0211 | RegionOne | swift | object-store | True | internal | https://dashint.cloud.example.com:6780/swift/v1/AUTH_%(project_id)s || f85b36ff9a2b49bc9eaadf1aafdee28c | RegionOne | swift | object-store | True | public | https://dash.cloud.example.com:6780/swift/v1/AUTH_%(project_id)s |On Mon, 18 Apr 2022 at 23:52, wodel youchi <wodel.youchi@gmail.com> wrote:Hi,I am having trouble configuring Openstack to use Ceph RGW as the Object store backend for Swift and S3.My setup is an HCI, I have 3 controllers which are also my ceph mgrs, mons and rgws and 9 compte/storage servers (osds).Xena is deployed with Ceph Pacific.Ceph public network is a private network on vlan10 with 10.10.1.0/24 as a subnet.Here is a snippet from my globals.yml :---
kolla_base_distro: "centos"
kolla_install_type: "source"
openstack_release: "xena"
kolla_internal_vip_address: "10.10.3.1"
kolla_internal_fqdn: "dashint.cloud.example.com"
kolla_external_vip_address: "x.x.x.x"
kolla_external_fqdn: "dash.cloud.example.com "
docker_registry: 192.168.1.16:4000
network_interface: "bond0"
kolla_external_vip_interface: "bond1"
api_interface: "bond1.30"
storage_interface: "bond1.10" <---------------- VLAN10 (public ceph network)
tunnel_interface: "bond1.40"
dns_interface: "bond1"
octavia_network_interface: "bond1.301"
neutron_external_interface: "bond2"
neutron_plugin_agent: "openvswitch"
keepalived_virtual_router_id: "51"
kolla_enable_tls_internal: "yes"
kolla_enable_tls_external: "yes"
kolla_certificates_dir: "{{ node_config }}/certificates"
kolla_external_fqdn_cert: "{{ kolla_certificates_dir }}/haproxy.pem"
kolla_internal_fqdn_cert: "{{ kolla_certificates_dir }}/haproxy-internal.pem"
kolla_admin_openrc_cacert: "{{ kolla_certificates_dir }}/ca.pem"
kolla_copy_ca_into_containers: "yes"
kolla_enable_tls_backend: "yes"
kolla_verify_tls_backend: "no"
kolla_tls_backend_cert: "{{ kolla_certificates_dir }}/backend-cert.pem"
kolla_tls_backend_key: "{{ kolla_certificates_dir }}/backend-key.pem"
enable_openstack_core: "yes"
enable_hacluster: "yes"
enable_haproxy: "yes"
enable_aodh: "yes"
enable_barbican: "yes"
enable_ceilometer: "yes"
enable_central_logging: "yes"
enable_ceph_rgw: "yes"
enable_ceph_rgw_loadbalancer: "{{ enable_ceph_rgw | bool }}"
enable_cinder: "yes"
enable_cinder_backup: "yes"
enable_collectd: "yes"
enable_designate: "yes"
enable_elasticsearch_curator: "yes"
enable_freezer: "no"
enable_gnocchi: "yes"
enable_gnocchi_statsd: "yes"
enable_magnum: "yes"
enable_manila: "yes"
enable_manila_backend_cephfs_native: "yes"
enable_mariabackup: "yes"
enable_masakari: "yes"
enable_neutron_vpnaas: "yes"
enable_neutron_qos: "yes"
enable_neutron_agent_ha: "yes"
enable_neutron_provider_networks: "yes"
enable_neutron_segments: "yes"
enable_octavia: "yes"
enable_trove: "yes"
external_ceph_cephx_enabled: "yes"
ceph_glance_keyring: "ceph.client.glance.keyring"
ceph_glance_user: "glance"
ceph_glance_pool_name: "images"
ceph_cinder_keyring: "ceph.client.cinder.keyring"
ceph_cinder_user: "cinder"
ceph_cinder_pool_name: "volumes"
ceph_cinder_backup_keyring: "ceph.client.cinder-backup.keyring"
ceph_cinder_backup_user: "cinder-backup"
ceph_cinder_backup_pool_name: "backups"
ceph_nova_keyring: "{{ ceph_cinder_keyring }}"
ceph_nova_user: "cinder"
ceph_nova_pool_name: "vms"
ceph_gnocchi_keyring: "ceph.client.gnocchi.keyring"
ceph_gnocchi_user: "gnocchi"
ceph_gnocchi_pool_name: "metrics"
ceph_manila_keyring: "ceph.client.manila.keyring"
ceph_manila_user: "manila"
glance_backend_ceph: "yes"
glance_backend_file: "no"
gnocchi_backend_storage: "ceph"
cinder_backend_ceph: "yes"
cinder_backup_driver: "ceph"
cloudkitty_collector_backend: "gnocchi"
designate_ns_record: "cloud.example.com "
nova_backend_ceph: "yes"
nova_compute_virt_type: "kvm"
octavia_auto_configure: yes
octavia_amp_flavor:
name: "amphora"
is_public: no
vcpus: 1
ram: 1024
disk: 5
octavia_amp_network:
name: lb-mgmt-net
provider_network_type: vlan
provider_segmentation_id: 301
provider_physical_network: physnet1
external: false
shared: false
subnet:
name: lb-mgmt-subnet
cidr: "10.7.0.0/16"
allocation_pool_start: "10.7.0.50"
allocation_pool_end: "10.7.255.200"
no_gateway_ip: yes
enable_dhcp: yes
mtu: 9000
octavia_amp_network_cidr: 10.10.7.0/24
octavia_amp_image_tag: "amphora"
octavia_certs_country: XZ
octavia_certs_state: Gotham
octavia_certs_organization: WAYNE
octavia_certs_organizational_unit: IT
horizon_keystone_multidomain: true
elasticsearch_curator_dry_run: "no"
enable_cluster_user_trust: true
ceph_rgw_hosts:
- host: controllera
ip: 10.10.1.5
port: 8080
- host: controllerb
ip: 10.10.1.9
port: 8080
- host: controllerc
ip: 10.10.1.13
port: 8080
ceph_rgw_swift_account_in_url: true
ceph_rgw_swift_compatibility: trueAnd Here is my ceph all.yml file---
dummy:
ceph_release_num: 16
cluster: ceph
configure_firewall: False
monitor_interface: bond1.10
monitor_address_block: 10.10.1.0/24
is_hci: true
hci_safety_factor: 0.2
osd_memory_target: 4294967296
public_network: 10.10.1.0/24
cluster_network: 10.10.2.0/24
radosgw_interface: "{{ monitor_interface }}"
radosgw_address_block: 10.10.1.0/24
nfs_file_gw: true
nfs_obj_gw: true
ceph_docker_image: "ceph/daemon"
ceph_docker_image_tag: latest-pacific
ceph_docker_registry: 192.168.1.16:4000
containerized_deployment: True
openstack_config: true
openstack_glance_pool:
name: "images"
pg_autoscale_mode: False
application: "rbd"
pg_num: 128
pgp_num: 128
target_size_ratio: 5.00
rule_name: "SSD"
openstack_cinder_pool:
name: "volumes"
pg_autoscale_mode: False
application: "rbd"
pg_num: 1024
pgp_num: 1024
target_size_ratio: 42.80
rule_name: "SSD"
openstack_nova_pool:
name: "vms"
pg_autoscale_mode: False
application: "rbd"
pg_num: 256
pgp_num: 256
target_size_ratio: 10.00
rule_name: "SSD"
openstack_cinder_backup_pool:
name: "backups"
pg_autoscale_mode: False
application: "rbd"
pg_num: 512
pgp_num: 512
target_size_ratio: 18.00
rule_name: "SSD"
openstack_gnocchi_pool:
name: "metrics"
pg_autoscale_mode: False
application: "rbd"
pg_num: 32
pgp_num: 32
target_size_ratio: 0.10
rule_name: "SSD"
openstack_cephfs_data_pool:
name: "cephfs_data"
pg_autoscale_mode: False
application: "cephfs"
pg_num: 256
pgp_num: 256
target_size_ratio: 10.00
rule_name: "SSD"
openstack_cephfs_metadata_pool:
name: "cephfs_metadata"
pg_autoscale_mode: False
application: "cephfs"
pg_num: 32
pgp_num: 32
target_size_ratio: 0.10
rule_name: "SSD"
openstack_pools:
- "{{ openstack_glance_pool }}"
- "{{ openstack_cinder_pool }}"
- "{{ openstack_nova_pool }}"
- "{{ openstack_cinder_backup_pool }}"
- "{{ openstack_gnocchi_pool }}"
- "{{ openstack_cephfs_data_pool }}"
- "{{ openstack_cephfs_metadata_pool }}"
openstack_keys:
- { name: client.glance, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" }
- { name: client.cinder, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" }
- { name: client.cinder-backup, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" }
- { name: client.gnocchi, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_gnocchi_pool.name }}"}, mode: "0600", }
- { name: client.openstack, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_glance_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" }
dashboard_enabled: True
dashboard_protocol: https
dashboard_port: 8443
dashboard_network: "192.168.1.0/24"
dashboard_admin_user: admin
dashboard_admin_user_ro: true
dashboard_admin_password: ***********
dashboard_crt: '/home/deployer/work/site-central/chaininv.crt'
dashboard_key: '/home/deployer/work/site-central/cloud_example.com.priv'
dashboard_grafana_api_no_ssl_verify: true
dashboard_rgw_api_user_id: admin
dashboard_rgw_api_no_ssl_verify: true
dashboard_frontend_vip: '192.168.1.5'
node_exporter_container_image: "192.168.1.16:4000/prom/node-exporter:v0.17.0"
grafana_admin_user: admin
grafana_admin_password: *********
grafana_crt: '/home/deployer/work/site-central/chaininv.crt'
grafana_key: '/home/deployer/work/site-central/cloud_example.com.priv'
grafana_server_fqdn: 'grafanasrv.cloud.example.com'
grafana_container_image: "192.168.1.16:4000/grafana/grafana:6.7.4"
grafana_dashboard_version: pacific
prometheus_container_image: "192.168.1.16:4000/prom/prometheus:v2.7.2"
alertmanager_container_image: "192.168.1.16:4000/prom/alertmanager:v0.16.2"And my rgws.yml---
dummy:
copy_admin_key: true
rgw_create_pools:
"{{ rgw_zone }}.rgw.buckets.data":
pg_num: 256
pgp_num: 256
size: 3
type: replicated
pg_autoscale_mode: False
rule_id: 1
"{{ rgw_zone }}.rgw.buckets.index":
pg_num: 64
pgp_num: 64
size: 3
type: replicated
pg_autoscale_mode: False
rule_id: 1
"{{ rgw_zone }}.rgw.meta":
pg_num: 32
pgp_num: 32
size: 3
type: replicated
pg_autoscale_mode: False
rule_id: 1
"{{ rgw_zone }}.rgw.log":
pg_num: 32
pgp_num: 32
size: 3
type: replicated
pg_autoscale_mode: False
rule_id: 1
"{{ rgw_zone }}.rgw.control":
pg_num: 32
pgp_num: 32
size: 3
type: replicated
pg_autoscale_mode: False
rule_id: 1The ceph_rgw user was created by kolla(xenavenv) [deployer@rscdeployer ~]$ openstack user list | grep ceph
| 3262aa7e03ab49c8a5710dfe3b16a136 | ceph_rgwThis is my ceph.conf from one of my controllers :[root@controllera ~]# cat /etc/ceph/ceph.conf
[client.rgw.controllera.rgw0]
host = controllera
rgw_keystone_url = https://dash.cloud.example.com:5000
##Authentication using username, password and tenant. Preferred.
rgw_keystone_verify_ssl = false
rgw_keystone_api_version = 3
rgw_keystone_admin_user = ceph_rgw
rgw_keystone_admin_password = cos2Jcnpnw9BhGwvPm**************************
rgw_keystone_admin_domain = Default
rgw_keystone_admin_project = service
rgw_s3_auth_use_keystone = true
rgw_keystone_accepted_roles = admin
rgw_keystone_implicit_tenants = true
rgw_swift_account_in_url = true
keyring = /var/lib/ceph/radosgw/ceph-rgw.controllera.rgw0/keyring
log file = /var/log/ceph/ceph-rgw-controllera.rgw0.log
rgw frontends = beast endpoint=10.10.1.5:8080
rgw thread pool size = 512
#For Debug
debug ms = 1
debug rgw = 20
# Please do not change this file directly since it is managed by Ansible and will be overwritten
[global]
cluster network = 10.10.2.0/24
fsid = da094354-6ade-415a-a424-************
mon host = [v2:10.10.1.5:3300,v1:10.10.1.5:6789],[v2:10.10.1.9:3300,v1:10.10.1.9:6789],[v2:10.10.1.13:3300,v1:10.10.1.13:6789]
mon initial members = controllera,controllerb,controllerc
osd pool default crush rule = 1
public network = 10.10.1.0/24Here are my swift endpoints(xenavenv) [deployer@rscdeployer ~]$ openstack endpoint list | grep swift
| 4082b4acf8bc4e4c9efc6e2d0e293724 | RegionOne | swift | object-store | True | admin | https://dashint.cloud.example.com:6780/v1/AUTH_%(project_id)s |
| b13a2f53e13e4650b4efdb8184eb0211 | RegionOne | swift | object-store | True | internal | https://dashint.cloud.example.com:6780/v1/AUTH_%(project_id)s |
| f85b36ff9a2b49bc9eaadf1aafdee28c | RegionOne | swift | object-store | True | public | https://dash.cloud.example.com:6780/v1/AUTH_%(project_id)s |When I connect to Horizon -> Project -> Object Store -> Containers I get theses errors :
- Unable to get the swift container listing
- Unable to fetch the policy details.
I cannot create a new container from the WebUI, the Storage policy parameter is empty.If I try to create a new container from the CLI, I get this :(xenavenv) [deployer@rscdeployer ~]$ source cephrgw-openrc.sh
(xenavenv) [deployer@rscdeployer ~]$ openstack container create demo -v
START with options: container create demo -v
command: container create -> openstackclient.object.v1.container.CreateContainer (auth=True)
Using auth plugin: password
Not Found (HTTP 404)
END return value: 1This is the log from RGW service when I execute the above command :2022-04-18T12:26:27.995+0100 7f22e07a9700 20 CONTENT_LENGTH=0
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_ACCEPT=*/*
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_ACCEPT_ENCODING=gzip, deflate
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_HOST=dashint.cloud.example.com:6780
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_USER_AGENT=openstacksdk/0.59.0 keystoneauth1/4.4.0 python-requests/2.26.0 CPython/3.8.8
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_VERSION=1.1
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_AUTH_TOKEN=gAAAAABiXUrjDFNzXx03mt1lbpUiCqNND1HACspSfg6h_TMxKYND5Hb9BO3FxH0a7CYoBXgRJywGszlK8cl-7zbUNRjHmxgIzmyh-CrWyGv793ZLOAmT_XShcrIKThjIIH3gTxYoX1TXwOKbsvMuZnI5EKKsol2y2MhcqPLeLGc28_AwoOr_b80
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_FORWARDED_FOR=10.10.3.16
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_FORWARDED_PROTO=https
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REMOTE_ADDR=10.10.1.13
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REQUEST_METHOD=PUT
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REQUEST_URI=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 SCRIPT_URI=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 SERVER_PORT=8080
2022-04-18T12:26:27.995+0100 7f22e07a9700 1 ====== starting new request req=0x7f23221aa620 =====
2022-04-18T12:26:27.995+0100 7f22e07a9700 2 req 728157015944164764 0.000000000s initializing for trans_id = tx000000a1aeef2b40f759c-00625d4ae3-4b389-default
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s rgw api priority: s3=8 s3website=7
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s host=dashint.cloud.example.com
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s subdomain= domain= in_hosted_domain=0 in_hosted_domain_s3website=0
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s final domain/bucket subdomain= domain= in_hosted_domain=0 in_hosted_domain_s3website=0 s->info.domain= s->info.request_uri=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s get_handler handler=22RGWHandler_REST_Obj_S3
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s handler=22RGWHandler_REST_Obj_S3
2022-04-18T12:26:27.995+0100 7f22e07a9700 2 req 728157015944164764 0.000000000s getting op 1
2022-04-18T12:26:27.995+0100 7f22e07a9700 1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.7:6801/4815,v1:10.10.1.7:6803/4815] -- osd_op(unknown.0.0:1516 12.3 12:c14cb721:::script.prerequest.:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2c400 con 0x56055e53b000
2022-04-18T12:26:27.996+0100 7f230d002700 1 -- 10.10.1.13:0/2715436964 <== osd.23 v2:10.10.1.7:6801/4815 22 ==== osd_op_reply(1516 script.prerequest. [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 246+0+0 (crc 0 0 0) 0x56055ea18b40 con 0x56055e53b000
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj scheduling with throttler client=2 cost=1
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj op=21RGWPutObj_ObjStore_S3
2022-04-18T12:26:27.996+0100 7f22ddfa4700 2 req 728157015944164764 0.001000002s s3:put_obj verifying requester
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::StrategyRegistry::s3_main_strategy_t: trying rgw::auth::s3::AWSAuthStrategy
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::AWSAuthStrategy: trying rgw::auth::s3::S3AnonymousEngine
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::S3AnonymousEngine granted access
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::AWSAuthStrategy granted access
2022-04-18T12:26:27.996+0100 7f22ddfa4700 2 req 728157015944164764 0.001000002s s3:put_obj normalizing buckets and tenants
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s->object=AUTH_971efa4cb18f42f7a405342072c39c9d/demo s->bucket=v1
2022-04-18T12:26:27.996+0100 7f22ddfa4700 2 req 728157015944164764 0.001000002s s3:put_obj init permissions
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj get_system_obj_state: rctx=0x7f23221a9000 obj=default.rgw.meta:root:v1 state=0x56055ea8c520 s->prefetch_data=0
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj cache get: name=default.rgw.meta+root+v1 : miss
2022-04-18T12:26:27.996+0100 7f22ddfa4700 1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.3:6802/4933,v1:10.10.1.3:6806/4933] -- osd_op(unknown.0.0:1517 11.b 11:d05f7b30:root::v1:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2cc00 con 0x56055e585000
2022-04-18T12:26:27.997+0100 7f230c801700 1 -- 10.10.1.13:0/2715436964 <== osd.3 v2:10.10.1.3:6802/4933 9 ==== osd_op_reply(1517 v1 [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 230+0+0 (crc 0 0 0) 0x56055e39db00 con 0x56055e585000
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj cache put: name=default.rgw.meta+root+v1 info.flags=0x0
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj adding default.rgw.meta+root+v1 to cache LRU end
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj init_permissions on <NULL> failed, ret=-2002
2022-04-18T12:26:27.997+0100 7f22dd7a3700 1 req 728157015944164764 0.002000004s op->ERRORHANDLER: err_no=-2002 new_err_no=-2002
2022-04-18T12:26:27.997+0100 7f22dbfa0700 1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.8:6804/4817,v1:10.10.1.8:6805/4817] -- osd_op(unknown.0.0:1518 12.1f 12:fb11263f:::script.postrequest.:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2d000 con 0x56055e94c800
2022-04-18T12:26:27.998+0100 7f230d002700 1 -- 10.10.1.13:0/2715436964 <== osd.9 v2:10.10.1.8:6804/4817 10 ==== osd_op_reply(1518 script.postrequest. [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 247+0+0 (crc 0 0 0) 0x56055ea18b40 con 0x56055e94c800
2022-04-18T12:26:27.998+0100 7f22d8f9a700 2 req 728157015944164764 0.003000006s s3:put_obj op status=0
2022-04-18T12:26:27.998+0100 7f22d8f9a700 2 req 728157015944164764 0.003000006s s3:put_obj http status=404
2022-04-18T12:26:27.998+0100 7f22d8f9a700 1 ====== req done req=0x7f23221aa620 op status=0 http_status=404 latency=0.003000006s ======
2022-04-18T12:26:27.998+0100 7f22d8f9a700 1 beast: 0x7f23221aa620: 10.10.1.13 - anonymous [18/Apr/2022:12:26:27.995 +0100] "PUT /v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo HTTP/1.1" 404 214 - "openstacksdk/0.59.0 keystoneauth1/4.4.0 python-requests/2.26.0 CPython/3.8.8" - latency=0.003000006sCould you help please.Regards.--බුද්ධික සංජීව ගොඩාකුරුBuddhika Sanjeewa GodakuruSystems Analyst/ProgrammerDeputy Webmaster / University of KelaniyaInformation and Communication Technology Centre (ICTC)University of Kelaniya, Sri Lanka,Kelaniya,Sri Lanka.Mobile : (+94) 071 5696981Office : (+94) 011 2903420 / 2903424
++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++University of Kelaniya Sri Lanka, accepts no liability for the content of this email, or for the consequences of any actions taken on the basis of the information provided, unless that information is subsequently confirmed in writing. If you are not the intended recipient, this email and/or any information it contains should not be copied, disclosed, retained or used by you or any other party and the email and all its contents should be promptly deleted fully from our system and the sender informed.
E-mail transmission cannot be guaranteed to be secure or error-free as information could be intercepted, corrupted, lost, destroyed, arrive late or incomplete.
+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++