openstack team mailing list archive
-
openstack team
-
Mailing list archive
-
Message #19297
Re: CRITICAL nova [-] [Errno 98] Address already in use
Hey,
Thanks, That seems to have done it!
Now I have some whole new errors to work on :)
2012-12-10 20:08:44 ERROR keystone.middleware.auth_token [-] HTTP connection exception: [Errno 1] _ssl.c:490: error:140770FC:SSL routines:SSL23_GET_SERVER_HELLO:unknown protocol
2012-12-10 20:08:44 WARNING keystone.middleware.auth_token [-] Authorization failed for token fbb8f2fe3fcb4155b9428b862b5bb943
2012-12-10 20:08:44 INFO keystone.middleware.auth_token [-] Invalid user token - rejecting request
2012-12-10 20:08:44 INFO nova.osapi_compute.wsgi.server [-] 127.0.0.1 - - [10/Dec/2012 20:08:44] "GET /v2/3efa0ffe282c4a0c8b0b3a2812b1b4d0/servers/detail HTTP/1.1" 401 462 0.012388
2012-12-10 20:08:44 ERROR keystone.middleware.auth_token [-] HTTP connection exception: [Errno 1] _ssl.c:490: error:140770FC:SSL routines:SSL23_GET_SERVER_HELLO:unknown protocol
2012-12-10 20:08:44 WARNING keystone.middleware.auth_token [-] Authorization failed for token e6f30024fc1142638452854f03a7735c
2012-12-10 20:08:44 INFO keystone.middleware.auth_token [-] Invalid user token - rejecting request
2012-12-10 20:08:44 INFO nova.osapi_compute.wsgi.server [-] 127.0.0.1 - - [10/Dec/2012 20:08:44] "GET /v2/3efa0ffe282c4a0c8b0b3a2812b1b4d0/servers/detail HTTP/1.1" 401 462 0.002427
Take care,
Andrew
On Dec 10, 2012, at 8:01 PM, Vishvananda Ishaya wrote:
> I just realized the problem. Your issue is actually the metadata api since you have something listening on 8775. If you are running nova-api-metadata separately then you can remove it from your list of enabled apis:
>
> enabled_apis=ec2,osapi_compute
>
> Alternatively just kill nova-api-metadata and allow it to run as wone of the nova-api components.
>
> Just for your reference, nova-api is an easy way to run all of the apis as one service. In this case it uses the enabled_apis config option. You can also run all of the apis separately by using the individual binaries:
>
> nova-api-ec2
> nova-api-metadata
> nova-api-os-compute
> nova-api-os-volume
>
> Vish
>
> On Dec 10, 2012, at 10:42 AM, Andrew Holway <a.holway@xxxxxxxxxxxx> wrote:
>
>> Hi,
>>
>> I have actually no idea how do do that. But the service opts look vaguely relevant:
>>
>> Does anyone have a working installation on Centos 6.3?
>>
>> Thanks,
>>
>> Andrew
>>
>>
>>
>> service_opts = [
>> cfg.IntOpt('report_interval',
>> default=10,
>> help='seconds between nodes reporting state to datastore'),
>> cfg.IntOpt('periodic_interval',
>> default=60,
>> help='seconds between running periodic tasks'),
>> cfg.IntOpt('periodic_fuzzy_delay',
>> default=60,
>> help='range of seconds to randomly delay when starting the'
>> ' periodic task scheduler to reduce stampeding.'
>> ' (Disable by setting to 0)'),
>> cfg.StrOpt('ec2_listen',
>> default="0.0.0.0",
>> help='IP address for EC2 API to listen'),
>> cfg.IntOpt('ec2_listen_port',
>> default=8773,
>> help='port for ec2 api to listen'),
>> cfg.IntOpt('ec2_workers',
>> default=None,
>> help='Number of workers for EC2 API service'),
>> cfg.StrOpt('osapi_compute_listen',
>> default="0.0.0.0",
>> help='IP address for OpenStack API to listen'),
>> cfg.IntOpt('osapi_compute_listen_port',
>> default=8774,
>> help='list port for osapi compute'),
>> cfg.IntOpt('osapi_compute_workers',
>> default=None,
>> help='Number of workers for OpenStack API service'),
>> cfg.StrOpt('metadata_manager',
>> default='nova.api.manager.MetadataManager',
>> help='OpenStack metadata service manager'),
>> cfg.StrOpt('metadata_listen',
>> default="0.0.0.0",
>> help='IP address for metadata api to listen'),
>> cfg.IntOpt('metadata_listen_port',
>> default=8775,
>> help='port for metadata api to listen'),
>> cfg.IntOpt('metadata_workers',
>> default=None,
>> help='Number of workers for metadata service'),
>> cfg.StrOpt('osapi_volume_listen',
>> default="0.0.0.0",
>> help='IP address for OpenStack Volume API to listen'),
>> cfg.IntOpt('osapi_volume_listen_port',
>> default=8776,
>> help='port for os volume api to listen'),
>> cfg.IntOpt('osapi_volume_workers',
>> default=None,
>> help='Number of workers for OpenStack Volume API service'),
>> ]
>>
>> On Dec 10, 2012, at 7:29 PM, Vishvananda Ishaya wrote:
>>
>>> Nope. Best i can think of is to throw some log statements into nova/service.py right before the exception gets thrown. See which api it is trying to start and what it thinks the value of enabled_apis is. Etc.
>>>
>>> Vish
>>>
>>> On Dec 10, 2012, at 10:24 AM, Andrew Holway <a.holway@xxxxxxxxxxxx> wrote:
>>>
>>>> Hi,
>>>>
>>>> maybe this will shed some light on it..?
>>>>
>>>> Thanks,
>>>>
>>>> Andrew
>>>>
>>>> [root@blade02 init.d]# cat /etc/nova/api-paste.ini
>>>> ############
>>>> # Metadata #
>>>> ############
>>>> [composite:metadata]
>>>> use = egg:Paste#urlmap
>>>> /: meta
>>>>
>>>> [pipeline:meta]
>>>> pipeline = ec2faultwrap logrequest metaapp
>>>>
>>>> [app:metaapp]
>>>> paste.app_factory = nova.api.metadata.handler:MetadataRequestHandler.factory
>>>>
>>>> #######
>>>> # EC2 #
>>>> #######
>>>>
>>>> [composite:ec2]
>>>> use = egg:Paste#urlmap
>>>> /services/Cloud: ec2cloud
>>>>
>>>> [composite:ec2cloud]
>>>> use = call:nova.api.auth:pipeline_factory
>>>> noauth = ec2faultwrap logrequest ec2noauth cloudrequest validator ec2executor
>>>> keystone = ec2faultwrap logrequest ec2keystoneauth cloudrequest validator ec2executor
>>>>
>>>> [filter:ec2faultwrap]
>>>> paste.filter_factory = nova.api.ec2:FaultWrapper.factory
>>>>
>>>> [filter:logrequest]
>>>> paste.filter_factory = nova.api.ec2:RequestLogging.factory
>>>>
>>>> [filter:ec2lockout]
>>>> paste.filter_factory = nova.api.ec2:Lockout.factory
>>>>
>>>> [filter:ec2keystoneauth]
>>>> paste.filter_factory = nova.api.ec2:EC2KeystoneAuth.factory
>>>>
>>>> [filter:ec2noauth]
>>>> paste.filter_factory = nova.api.ec2:NoAuth.factory
>>>>
>>>> [filter:cloudrequest]
>>>> controller = nova.api.ec2.cloud.CloudController
>>>> paste.filter_factory = nova.api.ec2:Requestify.factory
>>>>
>>>> [filter:authorizer]
>>>> paste.filter_factory = nova.api.ec2:Authorizer.factory
>>>>
>>>> [filter:validator]
>>>> paste.filter_factory = nova.api.ec2:Validator.factory
>>>>
>>>> [app:ec2executor]
>>>> paste.app_factory = nova.api.ec2:Executor.factory
>>>>
>>>> #############
>>>> # Openstack #
>>>> #############
>>>>
>>>> [composite:osapi_compute]
>>>> use = call:nova.api.openstack.urlmap:urlmap_factory
>>>> /: oscomputeversions
>>>> /v1.1: openstack_compute_api_v2
>>>> /v2: openstack_compute_api_v2
>>>>
>>>> [composite:osapi_volume]
>>>> use = call:nova.api.openstack.urlmap:urlmap_factory
>>>> /: osvolumeversions
>>>> /v1: openstack_volume_api_v1
>>>>
>>>> [composite:openstack_compute_api_v2]
>>>> use = call:nova.api.auth:pipeline_factory
>>>> noauth = faultwrap sizelimit noauth ratelimit osapi_compute_app_v2
>>>> keystone = faultwrap sizelimit authtoken keystonecontext ratelimit osapi_compute_app_v2
>>>> keystone_nolimit = faultwrap sizelimit authtoken keystonecontext osapi_compute_app_v2
>>>>
>>>> [composite:openstack_volume_api_v1]
>>>> use = call:nova.api.auth:pipeline_factory
>>>> noauth = faultwrap sizelimit noauth ratelimit osapi_volume_app_v1
>>>> keystone = faultwrap sizelimit authtoken keystonecontext ratelimit osapi_volume_app_v1
>>>> keystone_nolimit = faultwrap sizelimit authtoken keystonecontext osapi_volume_app_v1
>>>>
>>>> [filter:faultwrap]
>>>> paste.filter_factory = nova.api.openstack:FaultWrapper.factory
>>>>
>>>> [filter:noauth]
>>>> paste.filter_factory = nova.api.openstack.auth:NoAuthMiddleware.factory
>>>>
>>>> [filter:ratelimit]
>>>> paste.filter_factory = nova.api.openstack.compute.limits:RateLimitingMiddleware.factory
>>>>
>>>> [filter:sizelimit]
>>>> paste.filter_factory = nova.api.sizelimit:RequestBodySizeLimiter.factory
>>>>
>>>> [app:osapi_compute_app_v2]
>>>> paste.app_factory = nova.api.openstack.compute:APIRouter.factory
>>>>
>>>> [pipeline:oscomputeversions]
>>>> pipeline = faultwrap oscomputeversionapp
>>>>
>>>> [app:osapi_volume_app_v1]
>>>> paste.app_factory = nova.api.openstack.volume:APIRouter.factory
>>>>
>>>> [app:oscomputeversionapp]
>>>> paste.app_factory = nova.api.openstack.compute.versions:Versions.factory
>>>>
>>>> [pipeline:osvolumeversions]
>>>> pipeline = faultwrap osvolumeversionapp
>>>>
>>>> [app:osvolumeversionapp]
>>>> paste.app_factory = nova.api.openstack.volume.versions:Versions.factory
>>>>
>>>> ##########
>>>> # Shared #
>>>> ##########
>>>>
>>>> [filter:keystonecontext]
>>>> paste.filter_factory = nova.api.auth:NovaKeystoneContext.factory
>>>>
>>>> [filter:authtoken]
>>>> paste.filter_factory = keystone.middleware.auth_token:filter_factory
>>>> admin_tenant_name = service
>>>> admin_user = nova
>>>> admin_password = x7deix7dei
>>>> auth_uri = http://controller:5000/
>>>> On Dec 10, 2012, at 7:10 PM, Vishvananda Ishaya wrote:
>>>>
>>>>> Odd. This looks remarkably like it is trying to start osapi_volume even though you don't have it specified in enabled apis. Your enabled_apis setting looks correct to me.
>>>>>
>>>>> Vish
>>>>>
>>>>>
>>>>> On Dec 10, 2012, at 9:24 AM, Andrew Holway <a.holway@xxxxxxxxxxxx> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I cannot start the nova-api service.
>>>>>>
>>>>>> [root@blade02 07-openstack-controller]# nova list
>>>>>> ERROR: ConnectionRefused: '[Errno 111] Connection refused'
>>>>>>
>>>>>> I followed this guide very carefully:
>>>>>>
>>>>>> https://github.com/beloglazov/openstack-centos-kvm-glusterfs/#07-openstack-controller-controller
>>>>>>
>>>>>> Here is api.log
>>>>>>
>>>>>> 2012-12-10 17:51:31 DEBUG nova.wsgi [-] Loading app metadata from /etc/nova/api-paste.ini from (pid=2536) load_app /usr/lib/python2.6/site-packages/nova/wsgi.py:371
>>>>>> 2012-12-10 17:51:31 CRITICAL nova [-] [Errno 98] Address already in use
>>>>>> 2012-12-10 17:51:31 TRACE nova Traceback (most recent call last):
>>>>>> 2012-12-10 17:51:31 TRACE nova File "/usr/bin/nova-api", line 50, in <module>
>>>>>> 2012-12-10 17:51:31 TRACE nova server = service.WSGIService(api)
>>>>>> 2012-12-10 17:51:31 TRACE nova File "/usr/lib/python2.6/site-packages/nova/service.py", line 584, in __init__
>>>>>> 2012-12-10 17:51:31 TRACE nova port=self.port)
>>>>>> 2012-12-10 17:51:31 TRACE nova File "/usr/lib/python2.6/site-packages/nova/wsgi.py", line 72, in __init__
>>>>>> 2012-12-10 17:51:31 TRACE nova self._socket = eventlet.listen((host, port), backlog=backlog)
>>>>>> 2012-12-10 17:51:31 TRACE nova File "/usr/lib/python2.6/site-packages/eventlet/convenience.py", line 38, in listen
>>>>>> 2012-12-10 17:51:31 TRACE nova sock.bind(addr)
>>>>>> 2012-12-10 17:51:31 TRACE nova File "<string>", line 1, in bind
>>>>>> 2012-12-10 17:51:31 TRACE nova error: [Errno 98] Address already in use
>>>>>> 2012-12-10 17:51:31 TRACE nova
>>>>>> 2012-12-10 17:51:31 INFO nova.service [-] Parent process has died unexpectedly, exiting
>>>>>> 2012-12-10 17:51:31 INFO nova.service [-] Parent process has died unexpectedly, exiting
>>>>>> 2012-12-10 17:51:31 INFO nova.wsgi [-] Stopping WSGI server.
>>>>>> 2012-12-10 17:51:31 INFO nova.wsgi [-] Stopping WSGI server.
>>>>>>
>>>>>> [root@blade02 07-openstack-controller]# cat /etc/nova/nova.conf
>>>>>> [DEFAULT]
>>>>>> logdir = /var/log/nova
>>>>>> state_path = /var/lib/nova
>>>>>> lock_path = /var/lib/nova/tmp
>>>>>> volumes_dir = /etc/nova/volumes
>>>>>> dhcpbridge = /usr/bin/nova-dhcpbridge
>>>>>> dhcpbridge_flagfile = /etc/nova/nova.conf
>>>>>> force_dhcp_release = False
>>>>>> injected_network_template = /usr/share/nova/interfaces.template
>>>>>> libvirt_nonblocking = True
>>>>>> libvirt_inject_partition = -1
>>>>>> network_manager = nova.network.manager.FlatDHCPManager
>>>>>> iscsi_helper = tgtadm
>>>>>> sql_connection = mysql://nova:x7deix7dei@controller/nova
>>>>>> compute_driver = libvirt.LibvirtDriver
>>>>>> firewall_driver = nova.virt.libvirt.firewall.IptablesFirewallDriver
>>>>>> rpc_backend = nova.openstack.common.rpc.impl_qpid
>>>>>> rootwrap_config = /etc/nova/rootwrap.conf
>>>>>> verbose = True
>>>>>> auth_strategy = keystone
>>>>>> qpid_hostname = controller
>>>>>> network_host = compute1
>>>>>> fixed_range = 10.0.0.0/24
>>>>>> flat_interface = eth1
>>>>>> flat_network_bridge = br100
>>>>>> public_interface = eth1
>>>>>> glance_host = controller
>>>>>> vncserver_listen = 0.0.0.0
>>>>>> vncserver_proxyclient_address = controller
>>>>>> novncproxy_base_url = http://37.123.104.3:6080/vnc_auto.html
>>>>>> xvpvncproxy_base_url = http://37.123.104.3:6081/console
>>>>>> metadata_host = 10.141.6.2
>>>>>> enabled_apis=ec2,osapi_compute,metadata
>>>>>>
>>>>>> #[keystone_authtoken]
>>>>>> admin_tenant_name = %SERVICE_TENANT_NAME%
>>>>>> admin_user = %SERVICE_USER%
>>>>>> admin_password = %SERVICE_PASSWORD%
>>>>>> auth_host = 127.0.0.1
>>>>>> auth_port = 35357
>>>>>> auth_protocol = http
>>>>>> signing_dirname = /tmp/keystone-signing-nova
>>>>>>
>>>>>> There is no process using port 8774.
>>>>>>
>>>>>> [root@blade02 07-openstack-controller]# netstat -tunlp | grep 877
>>>>>> tcp 0 0 0.0.0.0:8775 0.0.0.0:* LISTEN 2157/python
>>>>>>
>>>>>> Maybe it is something similar to:
>>>>>>
>>>>>> https://bugzilla.redhat.com/show_bug.cgi?id=877606#c3
>>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>> Andrew
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Mailing list: https://launchpad.net/~openstack
>>>>>> Post to : openstack@xxxxxxxxxxxxxxxxxxx
>>>>>> Unsubscribe : https://launchpad.net/~openstack
>>>>>> More help : https://help.launchpad.net/ListHelp
>>>>>
>>>>
>>>>
>>>
>>
>>
>
References