Dear Wodel,
It seems you need to set 
ceph_rgw_swift_compatibility: false

Seems when you enable compatibility, it assumes ceph is working just like swift itself. (without /swift/ part)

Hope this helps.

On Tue, 19 Apr 2022 at 20:40, wodel youchi <wodel.youchi@gmail.com> wrote:
Hi,

Many thanks, after changing the endpoints it worked.
So the question is why kolla-ansible did not create the correct urls? Did I miss something?

Regards.

Le mar. 19 avr. 2022 à 11:09, wodel youchi <wodel.youchi@gmail.com> a écrit :
Hi,
Thanks.

The endpoints were created by Kolla-ansible upon deployment.

I did configure kolla-ansible to enable cross project tenant access by using :
ceph_rgw_swift_account_in_url: true

And I did add the rgw_swift_account_in_url = true in ceph.conf in the Rados servers. But the endpoints were created by kolla.

I will modify them and try again.

Regards.

Le mar. 19 avr. 2022 à 08:12, Buddhika S. Godakuru - University of Kelaniya <bsanjeewa@kln.ac.lk> a écrit :
Dear Wodel,
I think that default endpoint for swift when using cephrgw is /swift/v1 (unless you have changed it in ceph),
so your endpoints should be 
| 4082b4acf8bc4e4c9efc6e2d0e293724 | RegionOne | swift        | object-store    | True    | admin     | https://dashint.cloud.example.com:6780/swift/v1/AUTH_%(project_id)s |
| b13a2f53e13e4650b4efdb8184eb0211 | RegionOne | swift        | object-store    | True    | internal  | https://dashint.cloud.example.com:6780/swift/v1/AUTH_%(project_id)s |
| f85b36ff9a2b49bc9eaadf1aafdee28c | RegionOne | swift        | object-store    | True    | public    | https://dash.cloud.example.com:6780/swift/v1/AUTH_%(project_id)s    |



On Mon, 18 Apr 2022 at 23:52, wodel youchi <wodel.youchi@gmail.com> wrote:
Hi,
I am having trouble configuring Openstack to use Ceph RGW as the Object store backend for Swift and S3.

My setup is an HCI, I have 3 controllers which are also my ceph mgrs, mons and rgws and 9 compte/storage servers (osds).
Xena is deployed with Ceph Pacific.

Ceph public network is a private network on vlan10 with 10.10.1.0/24 as a subnet.

Here is a snippet from my globals.yml :
---
kolla_base_distro: "centos"
kolla_install_type: "source"
openstack_release: "xena"
kolla_internal_vip_address: "10.10.3.1"
kolla_internal_fqdn: "dashint.cloud.example.com"
kolla_external_vip_address: "x.x.x.x"
kolla_external_fqdn: "dash.cloud.example.com "
docker_registry: 192.168.1.16:4000
network_interface: "bond0"
kolla_external_vip_interface: "bond1"
api_interface: "bond1.30"
storage_interface: "bond1.10"       <---------------- VLAN10 (public ceph network)
tunnel_interface: "bond1.40"
dns_interface: "bond1"
octavia_network_interface: "bond1.301"
neutron_external_interface: "bond2"
neutron_plugin_agent: "openvswitch"
keepalived_virtual_router_id: "51"
kolla_enable_tls_internal: "yes"
kolla_enable_tls_external: "yes"
kolla_certificates_dir: "{{ node_config }}/certificates"
kolla_external_fqdn_cert: "{{ kolla_certificates_dir }}/haproxy.pem"
kolla_internal_fqdn_cert: "{{ kolla_certificates_dir }}/haproxy-internal.pem"
kolla_admin_openrc_cacert: "{{ kolla_certificates_dir }}/ca.pem"
kolla_copy_ca_into_containers: "yes"
kolla_enable_tls_backend: "yes"
kolla_verify_tls_backend: "no"
kolla_tls_backend_cert: "{{ kolla_certificates_dir }}/backend-cert.pem"
kolla_tls_backend_key: "{{ kolla_certificates_dir }}/backend-key.pem"
enable_openstack_core: "yes"
enable_hacluster: "yes"
enable_haproxy: "yes"
enable_aodh: "yes"
enable_barbican: "yes"
enable_ceilometer: "yes"
enable_central_logging: "yes"
enable_ceph_rgw: "yes"
enable_ceph_rgw_loadbalancer: "{{ enable_ceph_rgw | bool }}"

enable_cinder: "yes"
enable_cinder_backup: "yes"
enable_collectd: "yes"
enable_designate: "yes"
enable_elasticsearch_curator: "yes"
enable_freezer: "no"
enable_gnocchi: "yes"
enable_gnocchi_statsd: "yes"
enable_magnum: "yes"
enable_manila: "yes"
enable_manila_backend_cephfs_native: "yes"
enable_mariabackup: "yes"
enable_masakari: "yes"
enable_neutron_vpnaas: "yes"
enable_neutron_qos: "yes"
enable_neutron_agent_ha: "yes"
enable_neutron_provider_networks: "yes"
enable_neutron_segments: "yes"
enable_octavia: "yes"
enable_trove: "yes"
external_ceph_cephx_enabled: "yes"
ceph_glance_keyring: "ceph.client.glance.keyring"
ceph_glance_user: "glance"
ceph_glance_pool_name: "images"
ceph_cinder_keyring: "ceph.client.cinder.keyring"
ceph_cinder_user: "cinder"
ceph_cinder_pool_name: "volumes"
ceph_cinder_backup_keyring: "ceph.client.cinder-backup.keyring"
ceph_cinder_backup_user: "cinder-backup"
ceph_cinder_backup_pool_name: "backups"
ceph_nova_keyring: "{{ ceph_cinder_keyring }}"
ceph_nova_user: "cinder"
ceph_nova_pool_name: "vms"
ceph_gnocchi_keyring: "ceph.client.gnocchi.keyring"
ceph_gnocchi_user: "gnocchi"
ceph_gnocchi_pool_name: "metrics"
ceph_manila_keyring: "ceph.client.manila.keyring"
ceph_manila_user: "manila"
glance_backend_ceph: "yes"
glance_backend_file: "no"
gnocchi_backend_storage: "ceph"
cinder_backend_ceph: "yes"
cinder_backup_driver: "ceph"
cloudkitty_collector_backend: "gnocchi"
designate_ns_record: "cloud.example.com "
nova_backend_ceph: "yes"
nova_compute_virt_type: "kvm"
octavia_auto_configure: yes
octavia_amp_flavor:
  name: "amphora"
  is_public: no
  vcpus: 1
  ram: 1024
  disk: 5
octavia_amp_network:
  name: lb-mgmt-net
  provider_network_type: vlan
  provider_segmentation_id: 301
  provider_physical_network: physnet1
  external: false
  shared: false
  subnet:
    name: lb-mgmt-subnet
    cidr: "10.7.0.0/16"
    allocation_pool_start: "10.7.0.50"
    allocation_pool_end: "10.7.255.200"
    no_gateway_ip: yes
    enable_dhcp: yes
    mtu: 9000
octavia_amp_network_cidr: 10.10.7.0/24
octavia_amp_image_tag: "amphora"
octavia_certs_country: XZ
octavia_certs_state: Gotham
octavia_certs_organization: WAYNE
octavia_certs_organizational_unit: IT
horizon_keystone_multidomain: true
elasticsearch_curator_dry_run: "no"
enable_cluster_user_trust: true
ceph_rgw_hosts:
        - host: controllera
          ip: 10.10.1.5
          port: 8080
        - host: controllerb
          ip: 10.10.1.9
          port: 8080
        - host: controllerc
          ip: 10.10.1.13
          port: 8080
ceph_rgw_swift_account_in_url: true
ceph_rgw_swift_compatibility: true


And Here is my ceph all.yml file
---
dummy:
ceph_release_num: 16
cluster: ceph
configure_firewall: False
monitor_interface: bond1.10
monitor_address_block: 10.10.1.0/24
is_hci: true
hci_safety_factor: 0.2
osd_memory_target: 4294967296
public_network: 10.10.1.0/24
cluster_network: 10.10.2.0/24
radosgw_interface: "{{ monitor_interface }}"
radosgw_address_block: 10.10.1.0/24
nfs_file_gw: true
nfs_obj_gw: true
ceph_docker_image: "ceph/daemon"
ceph_docker_image_tag: latest-pacific
ceph_docker_registry: 192.168.1.16:4000
containerized_deployment: True
openstack_config: true
openstack_glance_pool:
  name: "images"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 128
  pgp_num: 128
  target_size_ratio: 5.00
  rule_name: "SSD"
openstack_cinder_pool:
  name: "volumes"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 1024
  pgp_num: 1024
  target_size_ratio: 42.80
  rule_name: "SSD"
openstack_nova_pool:
  name: "vms"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 256
  pgp_num: 256
  target_size_ratio: 10.00
  rule_name: "SSD"
openstack_cinder_backup_pool:
  name: "backups"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 512
  pgp_num: 512
  target_size_ratio: 18.00
  rule_name: "SSD"
openstack_gnocchi_pool:
  name: "metrics"
  pg_autoscale_mode: False
  application: "rbd"
  pg_num: 32
  pgp_num: 32
  target_size_ratio: 0.10
  rule_name: "SSD"
openstack_cephfs_data_pool:
  name: "cephfs_data"
  pg_autoscale_mode: False
  application: "cephfs"
  pg_num: 256
  pgp_num: 256
  target_size_ratio: 10.00
  rule_name: "SSD"
openstack_cephfs_metadata_pool:
  name: "cephfs_metadata"
  pg_autoscale_mode: False
  application: "cephfs"
  pg_num: 32
  pgp_num: 32
  target_size_ratio: 0.10
  rule_name: "SSD"
openstack_pools:
  - "{{ openstack_glance_pool }}"
  - "{{ openstack_cinder_pool }}"
  - "{{ openstack_nova_pool }}"
  - "{{ openstack_cinder_backup_pool }}"
  - "{{ openstack_gnocchi_pool }}"
  - "{{ openstack_cephfs_data_pool }}"
  - "{{ openstack_cephfs_metadata_pool }}"
openstack_keys:
  - { name: client.glance, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" }
  - { name: client.cinder, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_glance_pool.name }}"}, mode: "0600" }
  - { name: client.cinder-backup, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" }
  - { name: client.gnocchi, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_gnocchi_pool.name }}"}, mode: "0600", }
  - { name: client.openstack, caps: { mon: "profile rbd", osd: "profile rbd pool={{ openstack_glance_pool.name }}, profile rbd pool={{ openstack_nova_pool.name }}, profile rbd pool={{ openstack_cinder_pool.name }}, profile rbd pool={{ openstack_cinder_backup_pool.name }}"}, mode: "0600" }
dashboard_enabled: True
dashboard_protocol: https
dashboard_port: 8443
dashboard_network: "192.168.1.0/24"
dashboard_admin_user: admin
dashboard_admin_user_ro: true
dashboard_admin_password: ***********
dashboard_crt: '/home/deployer/work/site-central/chaininv.crt'
dashboard_key: '/home/deployer/work/site-central/cloud_example.com.priv'
dashboard_grafana_api_no_ssl_verify: true
dashboard_rgw_api_user_id: admin
dashboard_rgw_api_no_ssl_verify: true
dashboard_frontend_vip: '192.168.1.5'
node_exporter_container_image: "192.168.1.16:4000/prom/node-exporter:v0.17.0"
grafana_admin_user: admin
grafana_admin_password: *********
grafana_crt: '/home/deployer/work/site-central/chaininv.crt'
grafana_key: '/home/deployer/work/site-central/cloud_example.com.priv'
grafana_server_fqdn: 'grafanasrv.cloud.example.com'
grafana_container_image: "192.168.1.16:4000/grafana/grafana:6.7.4"
grafana_dashboard_version: pacific
prometheus_container_image: "192.168.1.16:4000/prom/prometheus:v2.7.2"
alertmanager_container_image: "192.168.1.16:4000/prom/alertmanager:v0.16.2"

And my rgws.yml
---
dummy:
copy_admin_key: true
rgw_create_pools:
  "{{ rgw_zone }}.rgw.buckets.data":
    pg_num: 256
    pgp_num: 256
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.buckets.index":
    pg_num: 64
    pgp_num: 64
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.meta":
    pg_num: 32
    pgp_num: 32
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.log":
    pg_num: 32
    pgp_num: 32
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1
  "{{ rgw_zone }}.rgw.control":
    pg_num: 32
    pgp_num: 32
    size: 3
    type: replicated
    pg_autoscale_mode: False
    rule_id: 1

The ceph_rgw user was created by kolla
(xenavenv) [deployer@rscdeployer ~]$ openstack user list | grep ceph
| 3262aa7e03ab49c8a5710dfe3b16a136 | ceph_rgw

This is my ceph.conf from one of my controllers :
[root@controllera ~]# cat /etc/ceph/ceph.conf
[client.rgw.controllera.rgw0]
host = controllera
rgw_keystone_url = https://dash.cloud.example.com:5000
##Authentication using username, password and tenant. Preferred.
rgw_keystone_verify_ssl = false
rgw_keystone_api_version = 3
rgw_keystone_admin_user = ceph_rgw
rgw_keystone_admin_password = cos2Jcnpnw9BhGwvPm**************************
rgw_keystone_admin_domain = Default
rgw_keystone_admin_project = service
rgw_s3_auth_use_keystone = true
rgw_keystone_accepted_roles = admin
rgw_keystone_implicit_tenants = true
rgw_swift_account_in_url = true
keyring = /var/lib/ceph/radosgw/ceph-rgw.controllera.rgw0/keyring
log file = /var/log/ceph/ceph-rgw-controllera.rgw0.log
rgw frontends = beast endpoint=10.10.1.5:8080
rgw thread pool size = 512
#For Debug
debug ms = 1
debug rgw = 20


# Please do not change this file directly since it is managed by Ansible and will be overwritten
[global]
cluster network = 10.10.2.0/24
fsid = da094354-6ade-415a-a424-************
mon host = [v2:10.10.1.5:3300,v1:10.10.1.5:6789],[v2:10.10.1.9:3300,v1:10.10.1.9:6789],[v2:10.10.1.13:3300,v1:10.10.1.13:6789]
mon initial members = controllera,controllerb,controllerc
osd pool default crush rule = 1
public network = 10.10.1.0/24


Here are my swift endpoints
(xenavenv) [deployer@rscdeployer ~]$ openstack endpoint list | grep swift
| 4082b4acf8bc4e4c9efc6e2d0e293724 | RegionOne | swift        | object-store    | True    | admin     | https://dashint.cloud.example.com:6780/v1/AUTH_%(project_id)s |
| b13a2f53e13e4650b4efdb8184eb0211 | RegionOne | swift        | object-store    | True    | internal  | https://dashint.cloud.example.com:6780/v1/AUTH_%(project_id)s |
| f85b36ff9a2b49bc9eaadf1aafdee28c | RegionOne | swift        | object-store    | True    | public    | https://dash.cloud.example.com:6780/v1/AUTH_%(project_id)s    |

When I connect to Horizon -> Project -> Object Store -> Containers I get theses errors :
  • Unable to get the swift container listing
  • Unable to fetch the policy details.
I cannot create a new container from the WebUI, the Storage policy parameter is empty.
If I try to create a new container from the CLI, I get this :
 (xenavenv) [deployer@rscdeployer ~]$ source cephrgw-openrc.sh
(xenavenv) [deployer@rscdeployer ~]$ openstack container create demo -v
START with options: container create demo -v
command: container create -> openstackclient.object.v1.container.CreateContainer (auth=True)
Using auth plugin: password
Not Found (HTTP 404)
END return value: 1


This is the log from RGW service when I execute the above command :
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 CONTENT_LENGTH=0
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_ACCEPT=*/*
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_ACCEPT_ENCODING=gzip, deflate
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_HOST=dashint.cloud.example.com:6780
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_USER_AGENT=openstacksdk/0.59.0 keystoneauth1/4.4.0 python-requests/2.26.0 CPython/3.8.8
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_VERSION=1.1
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_AUTH_TOKEN=gAAAAABiXUrjDFNzXx03mt1lbpUiCqNND1HACspSfg6h_TMxKYND5Hb9BO3FxH0a7CYoBXgRJywGszlK8cl-7zbUNRjHmxgIzmyh-CrWyGv793ZLOAmT_XShcrIKThjIIH3gTxYoX1TXwOKbsvMuZnI5EKKsol2y2MhcqPLeLGc28_AwoOr_b80
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_FORWARDED_FOR=10.10.3.16
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 HTTP_X_FORWARDED_PROTO=https
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REMOTE_ADDR=10.10.1.13
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REQUEST_METHOD=PUT
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 REQUEST_URI=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 SCRIPT_URI=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 SERVER_PORT=8080
2022-04-18T12:26:27.995+0100 7f22e07a9700  1 ====== starting new request req=0x7f23221aa620 =====
2022-04-18T12:26:27.995+0100 7f22e07a9700  2 req 728157015944164764 0.000000000s initializing for trans_id = tx000000a1aeef2b40f759c-00625d4ae3-4b389-default
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s rgw api priority: s3=8 s3website=7
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s host=dashint.cloud.example.com
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s subdomain= domain= in_hosted_domain=0 in_hosted_domain_s3website=0
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s final domain/bucket subdomain= domain= in_hosted_domain=0 in_hosted_domain_s3website=0 s->info.domain= s->info.request_uri=/v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo
2022-04-18T12:26:27.995+0100 7f22e07a9700 20 req 728157015944164764 0.000000000s get_handler handler=22RGWHandler_REST_Obj_S3
2022-04-18T12:26:27.995+0100 7f22e07a9700 10 req 728157015944164764 0.000000000s handler=22RGWHandler_REST_Obj_S3
2022-04-18T12:26:27.995+0100 7f22e07a9700  2 req 728157015944164764 0.000000000s getting op 1
2022-04-18T12:26:27.995+0100 7f22e07a9700  1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.7:6801/4815,v1:10.10.1.7:6803/4815] -- osd_op(unknown.0.0:1516 12.3 12:c14cb721:::script.prerequest.:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2c400 con 0x56055e53b000
2022-04-18T12:26:27.996+0100 7f230d002700  1 -- 10.10.1.13:0/2715436964 <== osd.23 v2:10.10.1.7:6801/4815 22 ==== osd_op_reply(1516 script.prerequest. [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 246+0+0 (crc 0 0 0) 0x56055ea18b40 con 0x56055e53b000
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj scheduling with throttler client=2 cost=1
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj op=21RGWPutObj_ObjStore_S3
2022-04-18T12:26:27.996+0100 7f22ddfa4700  2 req 728157015944164764 0.001000002s s3:put_obj verifying requester
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::StrategyRegistry::s3_main_strategy_t: trying rgw::auth::s3::AWSAuthStrategy
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::AWSAuthStrategy: trying rgw::auth::s3::S3AnonymousEngine
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::S3AnonymousEngine granted access
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj rgw::auth::s3::AWSAuthStrategy granted access
2022-04-18T12:26:27.996+0100 7f22ddfa4700  2 req 728157015944164764 0.001000002s s3:put_obj normalizing buckets and tenants
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s->object=AUTH_971efa4cb18f42f7a405342072c39c9d/demo s->bucket=v1
2022-04-18T12:26:27.996+0100 7f22ddfa4700  2 req 728157015944164764 0.001000002s s3:put_obj init permissions
2022-04-18T12:26:27.996+0100 7f22ddfa4700 20 req 728157015944164764 0.001000002s s3:put_obj get_system_obj_state: rctx=0x7f23221a9000 obj=default.rgw.meta:root:v1 state=0x56055ea8c520 s->prefetch_data=0
2022-04-18T12:26:27.996+0100 7f22ddfa4700 10 req 728157015944164764 0.001000002s s3:put_obj cache get: name=default.rgw.meta+root+v1 : miss
2022-04-18T12:26:27.996+0100 7f22ddfa4700  1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.3:6802/4933,v1:10.10.1.3:6806/4933] -- osd_op(unknown.0.0:1517 11.b 11:d05f7b30:root::v1:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2cc00 con 0x56055e585000
2022-04-18T12:26:27.997+0100 7f230c801700  1 -- 10.10.1.13:0/2715436964 <== osd.3 v2:10.10.1.3:6802/4933 9 ==== osd_op_reply(1517 v1 [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 230+0+0 (crc 0 0 0) 0x56055e39db00 con 0x56055e585000
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj cache put: name=default.rgw.meta+root+v1 info.flags=0x0
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj adding default.rgw.meta+root+v1 to cache LRU end
2022-04-18T12:26:27.997+0100 7f22dd7a3700 10 req 728157015944164764 0.002000004s s3:put_obj init_permissions on <NULL> failed, ret=-2002
2022-04-18T12:26:27.997+0100 7f22dd7a3700  1 req 728157015944164764 0.002000004s op->ERRORHANDLER: err_no=-2002 new_err_no=-2002
2022-04-18T12:26:27.997+0100 7f22dbfa0700  1 -- 10.10.1.13:0/2715436964 --> [v2:10.10.1.8:6804/4817,v1:10.10.1.8:6805/4817] -- osd_op(unknown.0.0:1518 12.1f 12:fb11263f:::script.postrequest.:head [call version.read in=11b,getxattrs,stat] snapc 0=[] ondisk+read+known_if_redirected e1182) v8 -- 0x56055eb2d000 con 0x56055e94c800
2022-04-18T12:26:27.998+0100 7f230d002700  1 -- 10.10.1.13:0/2715436964 <== osd.9 v2:10.10.1.8:6804/4817 10 ==== osd_op_reply(1518 script.postrequest. [call,getxattrs,stat] v0'0 uv0 ondisk = -2 ((2) No such file or directory)) v8 ==== 247+0+0 (crc 0 0 0) 0x56055ea18b40 con 0x56055e94c800
2022-04-18T12:26:27.998+0100 7f22d8f9a700  2 req 728157015944164764 0.003000006s s3:put_obj op status=0
2022-04-18T12:26:27.998+0100 7f22d8f9a700  2 req 728157015944164764 0.003000006s s3:put_obj http status=404
2022-04-18T12:26:27.998+0100 7f22d8f9a700  1 ====== req done req=0x7f23221aa620 op status=0 http_status=404 latency=0.003000006s ======
2022-04-18T12:26:27.998+0100 7f22d8f9a700  1 beast: 0x7f23221aa620: 10.10.1.13 - anonymous [18/Apr/2022:12:26:27.995 +0100] "PUT /v1/AUTH_971efa4cb18f42f7a405342072c39c9d/demo HTTP/1.1" 404 214 - "openstacksdk/0.59.0 keystoneauth1/4.4.0 python-requests/2.26.0 CPython/3.8.8" - latency=0.003000006s

Could you help please.

Regards.


--

බුද්ධික සංජීව ගොඩාකුරු 
Buddhika Sanjeewa Godakuru 

Systems Analyst/Programmer 
Deputy Webmaster / University of Kelaniya

Information and Communication Technology Centre (ICTC) 
University of Kelaniya, Sri Lanka, 
Kelaniya, 
Sri Lanka. 

Mobile : (+94) 071 5696981 
Office : (+94) 011 2903420 / 2903424 

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ 
University of Kelaniya Sri Lanka, accepts no liability for the content of this email, or for the consequences of any actions taken on the basis of the information provided, unless that information is subsequently confirmed in writing. If you are not the intended recipient, this email and/or any information it contains should not be copied, disclosed, retained or used by you or any other party and the email and all its contents should be promptly deleted fully from our system and the sender informed.

E-mail transmission cannot be guaranteed to be secure or error-free as information could be intercepted, corrupted, lost, destroyed, arrive late or incomplete.

+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Virus-free. www.avast.com


--

බුද්ධික සංජීව ගොඩාකුරු 
Buddhika Sanjeewa Godakuru 

Systems Analyst/Programmer 
Deputy Webmaster / University of Kelaniya

Information and Communication Technology Centre (ICTC) 
University of Kelaniya, Sri Lanka, 
Kelaniya, 
Sri Lanka. 

Mobile : (+94) 071 5696981 
Office : (+94) 011 2903420 / 2903424 

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ 
University of Kelaniya Sri Lanka, accepts no liability for the content of this email, or for the consequences of any actions taken on the basis of the information provided, unless that information is subsequently confirmed in writing. If you are not the intended recipient, this email and/or any information it contains should not be copied, disclosed, retained or used by you or any other party and the email and all its contents should be promptly deleted fully from our system and the sender informed.

E-mail transmission cannot be guaranteed to be secure or error-free as information could be intercepted, corrupted, lost, destroyed, arrive late or incomplete.

+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++