Thread Previous • Date Previous • Date Next • Thread Next |
On 05/30/2013 02:18 PM, Martin Mailand wrote:
Hi Josh, that's working. I have to more things. 1. The volume_driver=cinder.volume.driver.RBDDriver is deprecated, update your configuration to the new path. What is the new path?
cinder.volume.drivers.rbd.RBDDriver
2. I have in the glance-api.conf show_image_direct_url=True, but the volumes are not clones of the original which are in the images pool.
Set glance_api_version=2 in cinder.conf. The default was changed in Grizzly.
That's what I did. root@controller:~/vm_images# !1228 glance add name="Precise Server" is_public=true container_format=ovf disk_format=raw < ./precise-server-cloudimg-amd64-disk1.raw Added new image with ID: 6fbf4dfd-adce-470b-87fe-9b6ddb3993c8 root@controller:~/vm_images# rbd -p images -l ls NAME SIZE PARENT FMT PROT LOCK 6fbf4dfd-adce-470b-87fe-9b6ddb3993c8 2048M 2 6fbf4dfd-adce-470b-87fe-9b6ddb3993c8@snap 2048M 2 yes root@controller:~/vm_images# cinder create --image-id 6fbf4dfd-adce-470b-87fe-9b6ddb3993c8 --display-name volcli1 10+---------------------+--------------------------------------+ | Property | Value | +---------------------+--------------------------------------+ | attachments | [] | | availability_zone | nova | | bootable | false | | created_at | 2013-05-30T21:08:16.506094 | | display_description | None | | display_name | volcli1 | | id | 34838911-6613-4140-93e0-e1565054a2d3 | | image_id | 6fbf4dfd-adce-470b-87fe-9b6ddb3993c8 | | metadata | {} | | size | 10 | | snapshot_id | None | | source_volid | None | | status | creating | | volume_type | None | +---------------------+--------------------------------------+ root@controller:~/vm_images# cinder list +--------------------------------------+-------------+--------------+------+-------------+----------+-------------+ | ID | Status | Display Name | Size | Volume Type | Bootable | Attached to | +--------------------------------------+-------------+--------------+------+-------------+----------+-------------+ | 34838911-6613-4140-93e0-e1565054a2d3 | downloading | volcli1 | 10 | None | false | | +--------------------------------------+-------------+--------------+------+-------------+----------+-------------+ root@controller:~/vm_images# rbd -p volumes -l ls NAME SIZE PARENT FMT PROT LOCK volume-34838911-6613-4140-93e0-e1565054a2d3 10240M 2 root@controller:~/vm_images# -martin On 30.05.2013 22:56, Josh Durgin wrote:On 05/30/2013 01:50 PM, Martin Mailand wrote:Hi Josh, I found the problem, nova-compute tries to connect to the publicurl (xxx.xxx.240.10) of the keytone endpoints, this ip is not reachable from the management network. I thought the internalurl is the one, which is used for the internal communication of the openstack components and the publicurl is the ip for "customer" of the cluster? Am I wrong here?I'd expect that too, but it's determined in nova by the cinder_catalog_info option, which defaults to volume:cinder:publicURL. You can also override it explicitly with cinder_endpoint_template=http://192.168.192.2:8776/v1/$(tenant_id)s in your nova.conf. Josh-martin On 30.05.2013 22:22, Martin Mailand wrote:Hi Josh, On 30.05.2013 21:17, Josh Durgin wrote:It's trying to talk to the cinder api, and failing to connect at all. Perhaps there's a firewall preventing that on the compute host, or it's trying to use the wrong endpoint for cinder (check the keystone service and endpoint tables for the volume service).the keystone endpoint looks like this: | dd21ed74a9ac4744b2ea498609f0a86e | RegionOne | http://xxx.xxx.240.10:8776/v1/$(tenant_id)s | http://192.168.192.2:8776/v1/$(tenant_id)s | http://192.168.192.2:8776/v1/$(tenant_id)s | 5ad684c5a0154c13b54283b01744181b where 192.168.192.2 is the IP from the controller node. And from the compute node a telnet 192.168.192.2 8776 is working. -martin _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
Thread Previous • Date Previous • Date Next • Thread Next |