[Openstack-operators] Cinder api enpoint not found error while attach volume to instance

Geo Varghese gvarghese at aqorn.com
Thu Jan 15 06:08:41 UTC 2015


Hi Abel,

Thanks for the reply.

I have created volume and its in available state. Please check attached
screenshot.



On Thu, Jan 15, 2015 at 11:34 AM, Abel Lopez <alopgeek at gmail.com> wrote:

> Do your compute nodes have the required iSCSI packages installed?
>
>
> On Wednesday, January 14, 2015, Geo Varghese <gvarghese at aqorn.com> wrote:
>
>> Hi Jay,
>>
>> Thanks for the reply. Just pasting the details below
>>
>> keystone catalog
>> ================================================
>> Service: compute
>>
>> +-------------+------------------------------------------------------------+
>> |   Property  |
>> Value                            |
>>
>> +-------------+------------------------------------------------------------+
>> |   adminURL  |
>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 |
>> |      id     |
>> 02028b1f4c0849c68eb79f5887516299              |
>> | internalURL |
>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 |
>> |  publicURL  |
>> http://controller:8774/v2/e600ba9727924a3b97ede34aea8279c1 |
>> |    region   |
>> RegionOne                          |
>>
>> +-------------+------------------------------------------------------------+
>> Service: network
>> +-------------+----------------------------------+
>> |   Property  |              Value               |
>> +-------------+----------------------------------+
>> |   adminURL  |      http://controller:9696      |
>> |      id     | 32f687d4f7474769852d88932288b893 |
>> | internalURL |      http://controller:9696      |
>> |  publicURL  |      http://controller:9696      |
>> |    region   |            RegionOne             |
>> +-------------+----------------------------------+
>> Service: volumev2
>>
>> +-------------+------------------------------------------------------------+
>> |   Property  |
>> Value                            |
>>
>> +-------------+------------------------------------------------------------+
>> |   adminURL  |
>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 |
>> |      id     |
>> 5bca493cdde2439887d54fb805c4d2d4              |
>> | internalURL |
>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 |
>> |  publicURL  |
>> http://controller:8776/v2/e600ba9727924a3b97ede34aea8279c1 |
>> |    region   |
>> RegionOne                          |
>>
>> +-------------+------------------------------------------------------------+
>> Service: image
>> +-------------+----------------------------------+
>> |   Property  |              Value               |
>> +-------------+----------------------------------+
>> |   adminURL  |      http://controller:9292      |
>> |      id     | 2e2294b9151e4fb9b6efccf33c62181b |
>> | internalURL |      http://controller:9292      |
>> |  publicURL  |      http://controller:9292      |
>> |    region   |            RegionOne             |
>> +-------------+----------------------------------+
>> Service: volume
>>
>> +-------------+------------------------------------------------------------+
>> |   Property  |
>> Value                            |
>>
>> +-------------+------------------------------------------------------------+
>> |   adminURL  |
>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 |
>> |      id     |
>> 0e29cfaa785e4e148c57601b182a5e26              |
>> | internalURL |
>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 |
>> |  publicURL  |
>> http://controller:8776/v1/e600ba9727924a3b97ede34aea8279c1 |
>> |    region   |
>> RegionOne                          |
>>
>> +-------------+------------------------------------------------------------+
>> Service: ec2
>> +-------------+---------------------------------------+
>> |   Property  |                 Value                 |
>> +-------------+---------------------------------------+
>> |   adminURL  | http://controller:8773/services/Admin |
>> |      id     |    8f4957d98cd04130b055b8b80b051833   |
>> | internalURL | http://controller:8773/services/Cloud |
>> |  publicURL  | http://controller:8773/services/Cloud |
>> |    region   |               RegionOne               |
>> +-------------+---------------------------------------+
>> Service: identity
>> +-------------+----------------------------------+
>> |   Property  |              Value               |
>> +-------------+----------------------------------+
>> |   adminURL  |   http://controller:35357/v2.0   |
>> |      id     | 146f7bbb0ad54740b95f8499f04b2ee2 |
>> | internalURL |   http://controller:5000/v2.0    |
>> |  publicURL  |   http://controller:5000/v2.0    |
>> |    region   |            RegionOne             |
>> +-------------+----------------------------------+
>> ==============================================
>>
>> Nova.conf
>>
>> ================================================
>> # This file autogenerated by Chef
>> # Do not edit, changes will be overwritten
>>
>>
>> [DEFAULT]
>>
>> # LOGS/STATE
>> debug=False
>> verbose=False
>> auth_strategy=keystone
>> dhcpbridge_flagfile=/etc/nova/nova.conf
>> dhcpbridge=/usr/bin/nova-dhcpbridge
>> log_dir=/var/log/nova
>> state_path=/var/lib/nova
>> instances_path=/var/lib/nova/instances
>> instance_name_template=instance-%08x
>> network_allocate_retries=0
>> lock_path=/var/lib/nova/lock
>>
>> ssl_only=false
>> cert=self.pem
>> key=
>>
>> # Command prefix to use for running commands as root (default: sudo)
>> rootwrap_config=/etc/nova/rootwrap.conf
>>
>> # Should unused base images be removed? (default: false)
>> remove_unused_base_images=true
>>
>> # Unused unresized base images younger than this will not be removed
>> (default: 86400)
>> remove_unused_original_minimum_age_seconds=3600
>>
>> # Options defined in nova.openstack.common.rpc
>> rpc_thread_pool_size=64
>> rpc_conn_pool_size=30
>> rpc_response_timeout=60
>> rpc_backend=nova.openstack.common.rpc.impl_kombu
>> amqp_durable_queues=false
>> amqp_auto_delete=false
>>
>> ##### RABBITMQ #####
>> rabbit_userid=guest
>> rabbit_password=guest
>> rabbit_virtual_host=/
>> rabbit_hosts=rabbit1:5672,rabbit2:5672
>> rabbit_retry_interval=1
>> rabbit_retry_backoff=2
>> rabbit_max_retries=0
>> rabbit_durable_queues=false
>> rabbit_ha_queues=True
>>
>>
>>
>> ##### SCHEDULER #####
>> scheduler_manager=nova.scheduler.manager.SchedulerManager
>> scheduler_driver=nova.scheduler.filter_scheduler.FilterScheduler
>> scheduler_available_filters=nova.scheduler.filters.all_filters
>> # which filter class names to use for filtering hosts when not specified
>> in the request.
>>
>> scheduler_default_filters=AvailabilityZoneFilter,RamFilter,ComputeFilter,CoreFilter,SameHostFilter,DifferentHostFilter
>> default_availability_zone=nova
>> default_schedule_zone=nova
>>
>> ##### NETWORK #####
>>
>>
>>
>> # N.B. due to https://bugs.launchpad.net/nova/+bug/1206330
>> # we override the endpoint scheme below, ignore the port
>> # and essentially force http
>> neutron_url=http://controller:9696
>> neutron_api_insecure=false
>> network_api_class=nova.network.neutronv2.api.API
>> neutron_auth_strategy=keystone
>> neutron_admin_tenant_name=service
>> neutron_admin_username=neutron
>> neutron_admin_password=openstack-network
>> neutron_admin_auth_url=http://controller:5000/v2.0
>> neutron_url_timeout=30
>> neutron_region_name=
>> neutron_ovs_bridge=br-int
>> neutron_extension_sync_interval=600
>> neutron_ca_certificates_file=
>> linuxnet_interface_driver=nova.network.linux_net.LinuxOVSInterfaceDriver
>> firewall_driver = nova.virt.firewall.NoopFirewallDriver
>> security_group_api=neutron
>> service_neutron_metadata_proxy=true
>> neutron_metadata_proxy_shared_secret=secret123
>> default_floating_pool=public
>> dns_server=8.8.8.8
>>
>> use_ipv6=false
>>
>> ##### GLANCE #####
>> image_service=nova.image.glance.GlanceImageService
>> glance_api_servers=http://controller:9292
>> glance_api_insecure=false
>>
>> ##### Cinder #####
>> # Location of ca certificates file to use for cinder client requests
>> cinder_ca_certificates_file=
>>
>> # Allow to perform insecure SSL requests to cinder
>> cinder_api_insecure=false
>>
>> # Info to match when looking for cinder in the service catalog
>> cinder_catalog_info=volumev2:cinderv2:publicURL
>>
>> ##### COMPUTE #####
>> compute_driver=libvirt.LibvirtDriver
>> preallocate_images=none
>> use_cow_images=true
>> vif_plugging_is_fatal=false
>> vif_plugging_timeout=0
>> compute_manager=nova.compute.manager.ComputeManager
>> sql_connection=mysql://nova:nova@loadbalancer:3306/nova?charset=utf8
>> connection_type=libvirt
>>
>> ##### NOTIFICATIONS #####
>> # Driver or drivers to handle sending notifications (multi valued)
>>
>> # AMQP topic used for OpenStack notifications. (list value)
>> # Deprecated group/name - [rpc_notifier2]/topics
>> notification_topics=notifications
>>
>> # Generate periodic compute.instance.exists notifications
>> instance_usage_audit=False
>>
>> # Time period to generate instance usages for.  Time period
>> # must be hour, day, month or year (string value)
>> instance_usage_audit_period=month
>>
>>
>> # The IP address on which the OpenStack API will listen. (string value)
>> osapi_compute_listen=0.0.0.0
>> # The port on which the OpenStack API will listen. (integer value)
>> osapi_compute_listen_port=8774
>>
>> # The IP address on which the metadata will listen. (string value)
>> metadata_listen=0.0.0.0
>> # The port on which the metadata will listen. (integer value)
>> metadata_listen_port=8775
>>
>> ##### VNCPROXY #####
>> novncproxy_base_url=http://controller:6080/vnc_auto.html
>> xvpvncproxy_base_url=http://controller:6081/console
>>
>> # This is only required on the server running xvpvncproxy
>> xvpvncproxy_host=0.0.0.0
>> xvpvncproxy_port=6081
>>
>> # This is only required on the server running novncproxy
>> novncproxy_host=0.0.0.0
>> novncproxy_port=6080
>>
>> vncserver_listen=0.0.0.0
>> vncserver_proxyclient_address=0.0.0.0
>>
>> vnc_keymap=en-us
>>
>> # store consoleauth tokens in memcached
>>
>> ##### MISC #####
>> # force backing images to raw format
>> force_raw_images=false
>> allow_same_net_traffic=true
>> osapi_max_limit=1000
>> # If you terminate SSL with a load balancer, the HTTP_HOST environ
>> # variable that generates the request_uri in webob.Request will lack
>> # the HTTPS scheme. Setting this overrides the default and allows
>> # URIs returned in the various links collections to contain the proper
>> # HTTPS endpoint.
>> osapi_compute_link_prefix = http://controller:8774/v2/%(tenant_id)s
>> start_guests_on_host_boot=false
>> resume_guests_state_on_host_boot=true
>> allow_resize_to_same_host=false
>> resize_confirm_window=0
>> live_migration_retry_count=30
>>
>> ##### QUOTAS #####
>> # (StrOpt) default driver to use for quota checks (default:
>> nova.quota.DbQuotaDriver)
>> quota_driver=nova.quota.DbQuotaDriver
>> # number of security groups per project (default: 10)
>> quota_security_groups=50
>> # number of security rules per security group (default: 20)
>> quota_security_group_rules=20
>> # number of instance cores allowed per project (default: 20)
>> quota_cores=20
>> # number of fixed ips allowed per project (this should be at least the
>> number of instances allowed) (default: -1)
>> quota_fixed_ips=-1
>> # number of floating ips allowed per project (default: 10)
>> quota_floating_ips=10
>> # number of bytes allowed per injected file (default: 10240)
>> quota_injected_file_content_bytes=10240
>> # number of bytes allowed per injected file path (default: 255)
>> quota_injected_file_path_length=255
>> # number of injected files allowed (default: 5)
>> quota_injected_files=5
>> # number of instances allowed per project (defailt: 10)
>> quota_instances=10
>> # number of key pairs per user (default: 100)
>> quota_key_pairs=100
>> # number of metadata items allowed per instance (default: 128)
>> quota_metadata_items=128
>> # megabytes of instance ram allowed per project (default: 51200)
>> quota_ram=51200
>>
>> # virtual CPU to Physical CPU allocation ratio (default: 16.0)
>> cpu_allocation_ratio=16.0
>> # virtual ram to physical ram allocation ratio (default: 1.5)
>> ram_allocation_ratio=1.5
>>
>> mkisofs_cmd=genisoimage
>> injected_network_template=$pybasedir/nova/virt/interfaces.template
>> flat_injected=false
>>
>> # The IP address on which the EC2 API will listen. (string value)
>> ec2_listen=0.0.0.0
>> # The port on which the EC2 API will listen. (integer value)
>> ec2_listen_port=8773
>>
>>
>> ##### WORKERS ######
>>
>> ##### KEYSTONE #####
>> keystone_ec2_url=http://controller:5000/v2.0/ec2tokens
>>
>> # a list of APIs to enable by default (list value)
>> enabled_apis=ec2,osapi_compute,metadata
>>
>> ##### WORKERS ######
>>
>> ##### MONITORS ######
>> # Monitor classes available to the compute which may be
>> # specified more than once. (multi valued)
>> compute_available_monitors=nova.compute.monitors.all_monitors
>>
>> # A list of monitors that can be used for getting compute
>> # metrics. (list value)
>> compute_monitors=
>>
>> ##### VOLUMES #####
>> # iscsi target user-land tool to use
>> iscsi_helper=tgtadm
>> volume_api_class=nova.volume.cinder.API
>> # Region name of this node (string value)
>> os_region_name=RegionOne
>>
>> # Override the default dnsmasq settings with this file (String value)
>> dnsmasq_config_file=
>>
>> ##### THIRD PARTY ADDITIONS #####
>>
>>
>> [ssl]
>>
>> # CA certificate file to use to verify connecting clients
>> ca_file=
>>
>> # Certificate file to use when starting the server securely
>> cert_file=
>>
>> # Private key file to use when starting the server securely
>> key_file=
>>
>> [conductor]
>>
>> use_local=False
>>
>>
>> [libvirt]
>>
>> #
>> # Options defined in nova.virt.libvirt.driver
>> #
>>
>> # Rescue ami image (string value)
>> #rescue_image_id=<None>
>>
>> # Rescue aki image (string value)
>> #rescue_kernel_id=<None>
>>
>> # Rescue ari image (string value)
>> #rescue_ramdisk_id=<None>
>>
>> # Libvirt domain type (valid options are: kvm, lxc, qemu, uml,
>> # xen) (string value)
>> # Deprecated group/name - [DEFAULT]/libvirt_type
>> virt_type=kvm
>>
>> # Override the default libvirt URI (which is dependent on
>> # virt_type) (string value)
>> # Deprecated group/name - [DEFAULT]/libvirt_uri
>> #connection_uri=
>>
>> # Inject the admin password at boot time, without an agent.
>> # (boolean value)
>> # Deprecated group/name - [DEFAULT]/libvirt_inject_password
>> inject_password=false
>>
>> # Inject the ssh public key at boot time (boolean value)
>> # Deprecated group/name - [DEFAULT]/libvirt_inject_key
>> inject_key=true
>>
>> # The partition to inject to : -2 => disable, -1 => inspect
>> # (libguestfs only), 0 => not partitioned, >0 => partition
>> # number (integer value)
>> # Deprecated group/name - [DEFAULT]/libvirt_inject_partition
>> inject_partition=-2
>>
>> # Sync virtual and real mouse cursors in Windows VMs (boolean
>> # value)
>> #use_usb_tablet=true
>>
>> # Migration target URI (any included "%s" is replaced with the
>> # migration target hostname) (string value)
>> live_migration_uri=qemu+tcp://%s/system
>>
>> # Migration flags to be set for live migration (string value)
>> live_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER
>>
>> # Migration flags to be set for block migration (string value)
>> block_migration_flag=VIR_MIGRATE_UNDEFINE_SOURCE, VIR_MIGRATE_PEER2PEER,
>> VIR_MIGRATE_NON_SHARED_INC
>>
>> # Maximum bandwidth to be used during migration, in Mbps
>> # (integer value)
>> live_migration_bandwidth=0
>>
>> # Snapshot image format (valid options are : raw, qcow2, vmdk,
>> # vdi). Defaults to same as source image (string value)
>> snapshot_image_format=qcow2
>>
>> # The libvirt VIF driver to configure the VIFs. (string value)
>> # Deprecated group/name - [DEFAULT]/libvirt_vif_driver
>> vif_driver=nova.virt.libvirt.vif.LibvirtGenericVIFDriver
>>
>> # Libvirt handlers for remote volumes. (list value)
>> # Deprecated group/name - [DEFAULT]/libvirt_volume_drivers
>>
>> #volume_drivers=iscsi=nova.virt.libvirt.volume.LibvirtISCSIVolumeDriver,iser=nova.virt.libvirt.volume.LibvirtISERVolumeDriver,local=nova.virt.libvirt.volume.LibvirtVolumeDriver,fake=nova.virt.libvirt.volume.LibvirtFakeVolumeDriver,rbd=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,sheepdog=nova.virt.libvirt.volume.LibvirtNetVolumeDriver,nfs=nova.virt.libvirt.volume.LibvirtNFSVolumeDriver,aoe=nova.virt.libvirt.volume.LibvirtAOEVolumeDriver,glusterfs=nova.virt.libvirt.volume.LibvirtGlusterfsVolumeDriver,fibre_channel=nova.virt.libvirt.volume.LibvirtFibreChannelVolumeDriver,scality=nova.virt.libvirt.volume.LibvirtScalityVolumeDriver
>>
>> # Override the default disk prefix for the devices attached to
>> # a server, which is dependent on virt_type. (valid options
>> # are: sd, xvd, uvd, vd) (string value)
>> # Deprecated group/name - [DEFAULT]/libvirt_disk_prefix
>> #disk_prefix=<None>
>>
>> # Number of seconds to wait for instance to shut down after
>> # soft reboot request is made. We fall back to hard reboot if
>> # instance does not shutdown within this window. (integer
>> # value)
>> # Deprecated group/name - [DEFAULT]/libvirt_wait_soft_reboot_seconds
>> #wait_soft_reboot_seconds=120
>>
>> # Set to "host-model" to clone the host CPU feature flags; to
>> # "host-passthrough" to use the host CPU model exactly; to
>> # "custom" to use a named CPU model; to "none" to not set any
>> # CPU model. If virt_type="kvm|qemu", it will default to
>> # "host-model", otherwise it will default to "none" (string
>> # value)
>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_mode
>>
>> # Set to a named libvirt CPU model (see names listed in
>> # /usr/share/libvirt/cpu_map.xml). Only has effect if
>> # cpu_mode="custom" and virt_type="kvm|qemu" (string value)
>> # Deprecated group/name - [DEFAULT]/libvirt_cpu_model
>> #cpu_model=<none>
>>
>> # Location where libvirt driver will store snapshots before
>> # uploading them to image service (string value)
>> # Deprecated group/name - [DEFAULT]/libvirt_snapshots_directory
>> #snapshots_directory=$instances_path/snapshots
>>
>> # Location where the Xen hvmloader is kept (string value)
>> #xen_hvmloader_path=/usr/lib/xen/boot/hvmloader
>>
>> # Specific cachemodes to use for different disk types e.g:
>> # file=directsync,block=none (list value)
>>
>> # A path to a device that will be used as source of entropy on
>> # the host. Permitted options are: /dev/random or /dev/hwrng
>> # (string value)
>>
>> #
>> # Options defined in nova.virt.libvirt.imagecache
>> #
>>
>> # Unused resized base images younger than this will not be removed
>> (default: 3600)
>> remove_unused_resized_minimum_age_seconds=3600
>>
>> # Write a checksum for files in _base to disk (default: false)
>> checksum_base_images=false
>>
>> #
>> # Options defined in nova.virt.libvirt.vif
>> #
>>
>> use_virtio_for_bridges=true
>>
>> #
>> # Options defined in nova.virt.libvirt.imagebackend
>> #
>>
>> # VM Images format. Acceptable values are: raw, qcow2, lvm, rbd, default.
>> If default is specified,
>> # then use_cow_images flag is used instead of this one.
>> images_type=default
>>
>>
>> [keystone_authtoken]
>> auth_uri = http://controller:5000/v2.0
>> auth_host = controller
>> auth_port = 35357
>> auth_protocol = http
>> auth_version = v2.0
>> admin_tenant_name = service
>> admin_user = nova
>> admin_password = openstack-compute
>> signing_dir = /var/cache/nova/api
>> hash_algorithms = md5
>> insecure = false
>> ========================================
>>
>>
>> Please check it.
>>
>>
>> On Wed, Jan 14, 2015 at 8:23 PM, Jay Pipes <jaypipes at gmail.com> wrote:
>>
>>> Could you pastebin the output of:
>>>
>>>  keystone catalog
>>>
>>> and also pastebin your nova.conf for the node running the Nova API
>>> service?
>>>
>>> Thanks!
>>> -jay
>>>
>>>
>>> On 01/14/2015 02:25 AM, Geo Varghese wrote:
>>>
>>>> Hi Team,
>>>>
>>>> I need a help with cinder volume attachment with an instance.
>>>>
>>>> I have succesfully created cinder volume and it is in available state.
>>>> Please check attached screenshot.
>>>>
>>>> Later I tried to attach volume to an instance but attachment failed.
>>>> While checking logs in /var/log/nova/nova-api-os-compute.log
>>>>
>>>> ========
>>>>
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     res =
>>>> method(self, ctx, volume_id, *args, **kwargs)
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 206, in
>>>> get
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     item =
>>>> cinderclient(context).volumes.get(volume_id)
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>>> "/usr/lib/python2.7/dist-packages/nova/volume/cinder.py", line 91, in
>>>> cinderclient
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack
>>>> endpoint_type=endpoint_type)
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack   File
>>>> "/usr/lib/python2.7/dist-packages/cinderclient/service_catalog.py",
>>>> line
>>>> 80, in url_for
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack     raise
>>>> cinderclient.exceptions.EndpointNotFound()
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack EndpointNotFound
>>>> 2015-01-13 19:14:46.563 1736 TRACE nova.api.openstack
>>>> =========================================
>>>>
>>>>
>>>> I have already created endpints in v1 and v2.
>>>> Please check endpints I have created for cinder below
>>>>
>>>> ============================================================
>>>> =============
>>>> root at controller:/home/geo# keystone endpoint-list | grep 8776
>>>> | 5c7bcc79daa74532ac9ca19949e0d872 | regionOne |
>>>> http://controller:8776/v1/%(tenant_id)s |
>>>> http://controller:8776/v1/%(tenant_id)s |
>>>> http://controller:8776/v1/%(tenant_id)s |
>>>> 8ce0898aa7c84fec9b011823d34b55cb |
>>>> | 5d71e0a1237c483990b84c36346602b4 | RegionOne |
>>>> http://controller:8776/v2/%(tenant_id)s |
>>>> http://controller:8776/v2/%(tenant_id)s |
>>>> http://controller:8776/v2/%(tenant_id)s |
>>>> 251eca5fdb6b4550a9f521c10fa9f2ca |
>>>> ============================================================
>>>> ===================
>>>>
>>>> Anyone please help me. Thanks for the support guys.
>>>>
>>>> --
>>>> Regards,
>>>> Geo Varghese
>>>>
>>>>
>>>> _______________________________________________
>>>> OpenStack-operators mailing list
>>>> OpenStack-operators at lists.openstack.org
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>
>>>>
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators at lists.openstack.org
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>
>>
>>
>>
>> --
>> --
>> Regards,
>> Geo Varghese
>>
>


-- 
--
Regards,
Geo Varghese
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20150115/095c7aa3/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: Screenshot-Volumes & Snapshots - OpenStack Dashboard - Mozilla Firefox.png
Type: image/png
Size: 101748 bytes
Desc: not available
URL: <http://lists.openstack.org/pipermail/openstack-operators/attachments/20150115/095c7aa3/attachment-0001.png>


More information about the OpenStack-operators mailing list