[Openstack-operators] Cinder api enpoint not found error while attach volume to instance

Abel Lopez alopgeek at gmail.com
Thu Jan 15 06:04:36 UTC 2015


Do your compute nodes have the required iSCSI packages installed?

On Wednesday, January 14, 2015, Geo Varghese <gvarghese at aqorn.com> wrote:

> Hi Jay,
>
> Thanks for the reply. Just pasting the details below
>
> keystone catalog
> ================================================
> Service: compute
>
> +-------------+------------------------------------------------------------+
> |   Property  |                           Value
> |
>
> +-------------+------------------------------------------------------------+
> |   adminURL  | http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1
> |
> |      id     |              02028b1f4c0849c68eb79f5887516299
> |
> | internalURL | http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1
> |
> |  publicURL  | http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1
> |
> |    region   |                         RegionOne
> |
>
> +-------------+------------------------------------------------------------+
> Service: network
> +-------------+----------------------------------+
> |   Property  |              Value               |
> +-------------+----------------------------------+
> |   adminURL  |      http://controller:9696      |
> |      id     | 32f687d4f7474769852d88932288b893 |
> | internalURL |      http://controller:9696      |
> |  publicURL  |      http://controller:9696      |
> |    region   |            RegionOne             |
> +-------------+----------------------------------+
> Service: volumev2
>
> +-------------+------------------------------------------------------------+
> |   Property  |                           Value
> |
>
> +-------------+------------------------------------------------------------+
> |   adminURL  | http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1
> |
> |      id     |              5bca493cdde2439887d54fb805c4d2d4
> |
> | internalURL | http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1
> |
> |  publicURL  | http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1
> |
> |    region   |                         RegionOne
> |
>
> +-------------+------------------------------------------------------------+
> Service: image
> +-------------+----------------------------------+
> |   Property  |              Value               |
> +-------------+----------------------------------+
> |   adminURL  |      http://controller:9292      |
> |      id     | 2e2294b9151e4fb9b6efccf33c62181b |
> | internalURL |      http://controller:9292      |
> |  publicURL  |      http://controller:9292      |
> |    region   |            RegionOne             |
> +-------------+----------------------------------+
> Service: volume
>
> +-------------+------------------------------------------------------------+
> |   Property  |                           Value
> |
>
> +-------------+------------------------------------------------------------+
> |   adminURL  | http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1
> |
> |      id     |              0e29cfaa785e4e148c57601b182a5e26
> |
> | internalURL | http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1
> |
> |  publicURL  | http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1
> |
> |    region   |                         RegionOne
> |
>
> +-------------+------------------------------------------------------------+
> Service: ec2
> +-------------+---------------------------------------+
> |   Property  |                 Value                 |
> +-------------+---------------------------------------+
> |   adminURL  | http://controller:8773/services/Admin |
> |      id     |    8f4957d98cd04130b055b8b80b051833   |
> | internalURL | http://controller:8773/services/Cloud |
> |  publicURL  | http://controller:8773/services/Cloud |
> |    region   |               RegionOne               |
> +-------------+---------------------------------------+
> Service: identity
> +-------------+----------------------------------+
> |   Property  |              Value               |
> +-------------+----------------------------------+
> |   adminURL  |   http://controller:35357/v2.0   |
> |      id     | 146f7bbb0ad54740b95f8499f04b2ee2 |
> | internalURL |   http://controller:5000/v2.0    |
> |  publicURL  |   http://controller:5000/v2.0    |
> |    region   |            RegionOne             |
> +-------------+----------------------------------+
> ==============================================
>
> Nova.conf
>
> ================================================
> # This file autogenerated by Chef
> # Do not edit, changes will be overwritten
>
>
> [DEFAULT]
>
> # LOGS/STATE
> debug=False
> verbose=False
> auth_strategy=keystone
> dhcpbridge_flagfile=/etc/nova/nova.conf
> dhcpbridge=/usr/bin/nova-dhcpbridge
> log_dir=/var/log/nova
> state_path=/var/lib/nova
> instances_path=/var/lib/nova/instances
> instance_name_template=instance-%08x
> network_allocate_retries=0
> lock_path=/var/lib/nova/lock
>
> ssl_only=false
> cert=self.pem
> key=
>
> # Command prefix to use for running commands as root (default: sudo)
> rootwrap_config=/etc/nova/rootwrap.conf
>
> # Should unused base images be removed? (default: false)
> remove_unused_base_images=true
>
> # Unused unresized base images younger than this will not be removed
> (default: 86400)
> remove_unused_original_minimum_age_seconds=3600
>
> # Options defined in nova.openstack.common.rpc
> rpc_thread_pool_size=64
> rpc_conn_pool_size=30
> rpc_response_timeout=60
> rpc_backend=nova.openstack.common.rpc.impl_kombu
> amqp_durable_queues=false
> amqp_auto_delete=false
>
> ##### RABBITMQ #####
> rabbit_userid=guest
> rabbit_password=guest
> rabbit_virtual_host=/
> rabbit_hosts=rabbit1:5672,rabbit2:5672
> rabbit_retry_interval=1
> rabbit_retry_backoff=2
> rabbit_max_retries=0
> rabbit_durable_queues=false
> rabbit_ha_queues=True
>
>
>
> ##### SCHEDULER #####
> scheduler_manager=nova.scheduler.manager.SchedulerManager
> scheduler_driver=nova.scheduler.filter_scheduler.FilterScheduler
> scheduler_available_filters=nova.scheduler.filters.all_filters
> # which filter class names to use for filtering hosts when not specified
> in the request.
>
> scheduler_default_filters=AvailabilityZoneFilter,RamFilter,ComputeFilter,CoreFilter,SameHostFilter,DifferentHostFilter
> default_availability_zone=nova
> default_schedule_zone=nova
>
> ##### NETWORK #####
>
>
>
> # N.B. due to https://bugs.launchpad.net/nova/+bug/1206330
> # we override the endpoint scheme below, ignore the port
> # and essentially force http
> neutron_url=http://controller:9696
> neutron_api_insecure=false
> network_api_class=nova.network.neutronv2.api.API
> neutron_auth_strategy=keystone
> neutron_admin_tenant_name=service
> neutron_admin_username=neutron
> neutron_admin_password=openstack-network
> neutron_admin_auth_url=http://controller:5000/v2.0
> neutron_url_timeout=30
> neutron_region_name=
> neutron_ovs_bridge=br-int
> neutron_extension_sync_interval=600
> neutron_ca_certificates_file=
> linuxnet_interface_driver=nova.network.linux_net.LinuxOVSInterfaceDriver
> firewall_driver = nova.virt.firewall.NoopFirewallDriver
> security_group_api=neutron
> service_neutron_metadata_proxy=true
> neutron_metadata_proxy_shared_secret=secret123
> default_floating_pool=public
> dns_server=8.8.8.8
>
> use_ipv6=false
>
> ##### GLANCE #####
> image_service=nova.image.glance.GlanceImageService
> glance_api_servers=http://controller:9292
> glance_api_insecure=false
>
> ##### Cinder #####
> # Location of ca certificates file to use for cinder client requests
> cinder_ca_certificates_file=
>
> # Allow to perform insecure SSL requests to cinder
> cinder_api_insecure=false
>
> # Info to match when looking for cinder in the service catalog
> cinder_catalog_info=volumev2:cinderv2:publicURL
>
> ##### COMPUTE #####
> compute_driver=libvirt.LibvirtDriver
> preallocate_images=none
> use_cow_images=true
> vif_plugging_is_fatal=false
> vif_plugging_timeout=0
> compute_manager=nova.compute.manager.ComputeManager
> sql_connection=mysql://nova:nova@loadbalancer:3306/nova?charset=utf8
> connection_type=libvirt
>
> ##### NOTIFICATIONS #####
> # Driver or drivers to handle sending notifications (multi valued)
>
> # AMQP topic used for OpenStack notifications. (list value)
> # Deprecated group/name - [rpc_notifier2]/topics
> notification_topics=notifications
>
> # Generate periodic compute.instance.exists notifications
> instance_usage_audit=False
>
> # Time period to generate instance usages for.  Time period
> # must be hour, day, month or year (string value)
> instance_usage_audit_period=month
>
>
> # The IP address on which the OpenStack API will listen. (string value)
> osapi_compute_listen=0.0.0.0
> # The port on which the OpenStack API will listen. (integer value)
> osapi_compute_listen_port=8774
>
> # The IP address on which the metadata will listen. (string value)
> metadata_listen=0.0.0.0
> # The port on which the metadata will listen. (integer value)
> metadata_listen_port=8775
>
> ##### VNCPROXY #####
> novncproxy_base_url=http://controller:6080/vnc_auto.html
> xvpvncproxy_base_url=http://controller:6081/console
>
> # This is only required on the server running xvpvncproxy
> xvpvncproxy_host=0.0.0.0
> xvpvncproxy_port=6081
>
> # This is only required on the server running novncproxy
> novncproxy_host=0.0.0.0
> novncproxy_port=6080
>
> vncserver_listen=0.0.0.0
> vncserver_proxyclient_address=0.0.0.0
>
> vnc_keymap=en-us
>
> # store consoleauth tokens in memcached
>
> ##### MISC #####
> # force backing images to raw format
> force_raw_images=false
> allow_same_net_traffic=true
> osapi_max_limit=1000
> # If you terminate SSL with a load balancer, the HTTP_HOST environ
> # variable that generates the request_uri in webob.Request will lack
> # the HTTPS scheme. Setting this overrides the default and allows
> # URIs returned in the various links collections to contain the proper
> # HTTPS endpoint.
> osapi_compute_link_prefix = http://controller:8774/v2/%(tenant_id)s
> start_guests_on_host_boot=false
> resume_guests_state_on_host_boot=true
> allow_resize_to_same_host=false
> resize_confirm_window=0
> live_migration_retry_count=30
>
> ##### QUOTAS #####
> # (StrOpt) default driver to use for quota checks (default:
> nova.quota.DbQuotaDriver)
> quota_driver=nova.quota.DbQuotaDriver
> # number of security groups per project (default: 10)
> quota_security_groups=50
> # number of security rules per security group (default: 20)
> quota_security_group_rules=20
> # number of instance cores allowed per project (default: 20)
> quota_cores=20
> # number of fixed ips allowed per project (this should be at least the
> number of instances allowed) (default: -1)
> quota_fixed_ips=-1
> # number of floating ips allowed per project (default: 10)
> quota_floating_ips=10
> # number of bytes allowed per injected file (default: 10240)
> quota_injected_file_content_bytes=10240
> # number of bytes allowed per injected file path (default: 255)
> quota_injected_file_path_length=255
> # number of injected files allowed (default: 5)
> quota_injected_files=5
> # number of instances allowed per project (defailt: 10)
> quota_instances=10
> # number of key pairs per user (default: 100)
> quota_key_pairs=100
> # number of metadata items allowed per instance (default: 128)
> quota_metadata_items=128
> # megabytes of instance ram allowed per project (default: 51200)
> quota_ram=51200
>
> # virtual CPU to Physical CPU allocation ratio (default: 16.0)
> cpu_allocation_ratio=16.0
> # virtual ram to physical ram allocation ratio (default: 1.5)
> ram_allocation_ratio=1.5
>
> mkisofs_cmd=genisoimage
> injected_network_template=$pybasedir/nova/virt/interfaces.template
> flat_injected=false
>
> # The IP address on which the EC2 API will listen. (string value)
> ec2_listen=0.0.0.0
> # The port on which the EC2 API will listen. (integer value)
> ec2_listen_port=8773
>
>
> ##### WORKERS ######
>
> ##### KEYSTONE #####
> keystone_ec2_url=http://controller:5000/v2.0/ec2tokens
>
> # a list of APIs to enable by default (list value)
> enabled_apis=ec2,osapi_compute,metadata
>
> ##### WORKERS ######
>
> ##### MONITORS ######
> # Monitor classes available to the compute which may be
> # specified more than once. (multi valued)
> compute_available_monitors=nova.compute.monitors.all_monitors
>
> # A list of monitors that can be used for getting compute
> # metrics. (list value)
> compute_monitors=
>
> ##### VOLUMES #####
> # iscsi target user-land tool to use
> iscsi_helper=tgtadm
> volume_api_class=nova.volume.cinder.API
> # Region name of this node (string value)
> os_region_name=RegionOne
>
> # Override the default dnsmasq settings with this file (String value)
> dnsmasq_config_file=
>
> ##### THIRD PARTY ADDITIONS #####
>
>
> [ssl]
>
> # CA certificate file to use to verify connecting clients
> ca_file=
>
> # Certificate file to use when starting the server securely
> cert_file=
>
> # Private key file to use when starting the server securely
> key_file=
>
> [conductor]
>
> use_local=False
>
>
> [libvirt]
>
> #
> # Options defined in nova.virt.libvirt.driver
> #
>
> # Rescue ami image (string value)
> #rescue_image_id=<None>
>
> # Rescue aki image (string value)
> #rescue_kernel_id=<None>
>
> # Rescue ari image (string value)
> #rescue_ramdisk_id=<None>
>
> # Libvirt domain type (valid options are: kvm, lxc, qemu, uml,
> # xen) (string value)
> # Deprecated group/name - [DEFAULT]/libvirt_type
> virt_type=kvm
>
> # Override the default libvirt URI (which is dependent on
> # virt_type) (string value)
> # Deprecated group/name - [DEFAULT]/libvirt_uri
> #connection_uri=
>
> # Inject the admin password at boot time, without an agent.
> # (boolean value)
> # Deprecated group/name - [DEFAULT]/libvirt_inject_password
> inject_password=false
>
> # Inject the ssh public key at boot time (boolean value)
> # Deprecated group/name - [DEFAULT]/libvirt_inject_key
> inject_key=true
>
> # The partition to inject to : -2 => disable, -1 => inspect
> # (libguestfs only), 0 => not partitioned, >0 => partition
> # number (integer value)
> # Deprecated group/name - [DEFAULT]/libvirt_inject_partition
> inject_partition=-2
>
> # Sync virtual and real mouse cursors in Windows VMs (boolean
> # value)
> #use_usb_tablet=true
>
> # Migration target URI (any included "%s" is replaced with the
> # migration target hostname) (string value)
> live_migration_uri=qemu+tcp://%s/system
>
> # Migration flags to be set for live migration (string value)
> live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER
>
> # Migration flags to be set for block migration (string value)
> block_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER,
> VIR_MIGRATE_NON_SHARED_INC
>
> # Maximum bandwidth to be used during migration, in Mbps
> # (integer value)
> live_migration_bandwidth=0
>
> # Snapshot image format (valid options are : raw, qcow2, vmdk,
> # vdi). Defaults to same as source image (string value)
> snapshot_image_format=qcow2
>
> # The libvirt VIF driver to configure the VIFs. (string value)
> # Deprecated group/name - [DEFAULT]/libvirt_vif_driver
> vif_driver=nova.virt.libvirt.vif.LibvirtGenericVIFDriver
>
> # Libvirt handlers for remote volumes. (list value)
> # Deprecated group/name - [DEFAULT]/libvirt_volume_drivers
>
> #volume_drivers=iscsi=nova.virt.libvirt.volume.LibvirtISCSIVolumeDriver,iser=nova.virt.libvirt.volume.LibvirtISERVolumeDriver,local=nova.virt.libvirt.volume.LibvirtVolumeDriver,fake=nova.virt.libvirt.volume.LibvirtFakeVolumeDriver,rbd=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,sheepdog=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,nfs=nova.virt.libvirt.volume.LibvirtNFSVolumeDriver,aoe=nova.virt.libvirt.volume.LibvirtAOEVolumeDriver,glusterfs=nova.virt.libvirt.volume.LibvirtGlusterfsVolumeDriver,fibre_channel=nova.virt.libvirt.volume.LibvirtFibreChannelVolumeDriver,scality=nova.virt.libvirt.volume.LibvirtScalityVolumeDriver
>
> # Override the default disk prefix for the devices attached to
> # a server, which is dependent on virt_type. (valid options
> # are: sd, xvd, uvd, vd) (string value)
> # Deprecated group/name - [DEFAULT]/libvirt_disk_prefix
> #disk_prefix=<None>
>
> # Number of seconds to wait for instance to shut down after
> # soft reboot request is made. We fall back to hard reboot if
> # instance does not shutdown within this window. (integer
> # value)
> # Deprecated group/name - [DEFAULT]/libvirt_wait_soft_reboot_seconds
> #wait_soft_reboot_seconds=120
>
> # Set to "host-model" to clone the host CPU feature flags; to
> # "host-passthrough" to use the host CPU model exactly; to
> # "custom" to use a named CPU model; to "none" to not set any
> # CPU model. If virt_type="kvm|qemu", it will default to
> # "host-model", otherwise it will default to "none" (string
> # value)
> # Deprecated group/name - [DEFAULT]/libvirt_cpu_mode
>
> # Set to a named libvirt CPU model (see names listed in
> # /usr/share/libvirt/cpu_map.xml). Only has effect if
> # cpu_mode="custom" and virt_type="kvm|qemu" (string value)
> # Deprecated group/name - [DEFAULT]/libvirt_cpu_model
> #cpu_model=<none>
>
> # Location where libvirt driver will store snapshots before
> # uploading them to image service (string value)
> # Deprecated group/name - [DEFAULT]/libvirt_snapshots_directory
> #snapshots_directory=$instances_path/snapshots
>
> # Location where the Xen hvmloader is kept (string value)
> #xen_hvmloader_path=/usr/lib/xen/boot/hvmloader
>
> # Specific cachemodes to use for different disk types e.g:
> # file=directsync,block=none (list value)
>
> # A path to a device that will be used as source of entropy on
> # the host. Permitted options are: /dev/random or /dev/hwrng
> # (string value)
>
> #
> # Options defined in nova.virt.libvirt.imagecache
> #
>
> # Unused resized base images younger than this will not be removed
> (default: 3600)
> remove_unused_resized_minimum_age_seconds=3600
>
> # Write a checksum for files in _base to disk (default: false)
> checksum_base_images=false
>
> #
> # Options defined in nova.virt.libvirt.vif
> #
>
> use_virtio_for_bridges=true
>
> #
> # Options defined in nova.virt.libvirt.imagebackend
> #
>
> # VM Images format. Acceptable values are: raw, qcow2, lvm, rbd, default.
> If default is specified,
> # then use_cow_images flag is used instead of this one.
> images_type=default
>
>
> [keystone_authtoken]
> auth_uri = http://controller:5000/v2.0
> auth_host = controller
> auth_port = 35357
> auth_protocol = http
> auth_version = v2.0
> admin_tenant_name = service
> admin_user = nova
> admin_password = openstack-compute
> signing_dir = /var/cache/nova/api
> hash_algorithms = md5
> insecure = false
> ========================================
>
>
> Please check it.
>
>
> On Wed, Jan 14, 2015 at 8:23 PM, Jay Pipes <jaypipes at gmail.com
> <javascript:_e(%7B%7D,'cvml','jaypipes at gmail.com');>> wrote:
>
>> Could you pastebin the output of:
>>
>>  keystone catalog
>>
>> and also pastebin your nova.conf for the node running the Nova API
>> service?
>>
>> Thanks!
>> -jay
>>
>>
>> On 01/14/2015 02:25 AM, Geo Varghese wrote:
>>
>>> Hi Team,
>>>
>>> I need a help with cinder volume attachment with an instance.
>>>
>>> I have succesfully created cinder volume and it is in available state.
>>> Please check attached screenshot.
>>>
>>> Later I tried to attach volume to an instance but attachment failed.
>>> While checking logs in /var/log/nova/nova-api-os-compute.log
>>>
>>> ========
>>>
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     res =
>>> method(self, ctx, volume_id, *args, **kwargs)
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 206, in
>>> get
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     item =
>>> cinderclient(context).volumes.get(volume_id)
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 91, in
>>> cinderclient
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack
>>> endpoint_type=endpoint_type)
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>> "/usr/lib/python2.7/dist-packages/cinderclient/service_catalog.py", line
>>> 80, in url_for
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     raise
>>> cinderclient.exceptions.EndpointNotFound()
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack EndpointNotFound
>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack
>>> =========================================
>>>
>>>
>>> I have already created endpints in v1 and v2.
>>> Please check endpints I have created for cinder below
>>>
>>> ============================================================
>>> =============
>>> root at controller:/home/geo# keystone endpoint-list | grep 8776
>>> | 5c7bcc79daa74532ac9ca19949e0d872 | regionOne |
>>> http://controller:8776/v1/%(tenant_id)s |
>>> http://controller:8776/v1/%(tenant_id)s |
>>> http://controller:8776/v1/%(tenant_id)s | 8ce0898aa7c84fec9b011823d34b55cb
>>> |
>>> | 5d71e0a1237c483990b84c36346602b4 | RegionOne |
>>> http://controller:8776/v2/%(tenant_id)s |
>>> http://controller:8776/v2/%(tenant_id)s |
>>> http://controller:8776/v2/%(tenant_id)s | 251eca5fdb6b4550a9f521c10fa9f2ca
>>> |
>>> ============================================================
>>> ===================
>>>
>>> Anyone please help me. Thanks for the support guys.
>>>
>>> --
>>> Regards,
>>> Geo Varghese
>>>
>>>
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators at lists.openstack.org
>>> <javascript:_e(%7B%7D,'cvml','OpenStack-operators at lists.openstack.org');>
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>
>>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators at lists.openstack.org
>> <javascript:_e(%7B%7D,'cvml','OpenStack-operators at lists.openstack.org');>
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>
>
>
>
> --
> --
> Regards,
> Geo Varghese
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20150114/7f3b8a9d/attachment.html>


More information about the OpenStack-operators mailing list