No. Doing all stuffs manaully.
On Sunday, July 27, 2014, Remo Mattei <r...@italy1.com> wrote: > Did u use packstack? > > Inviato da iPhone () > Il giorno Jul 26, 2014, alle ore 20:34, john decot <johnde...@gmail.com> ha scritto: > > CentOS > > On Sunday, July 27, 2014, Remo Mattei <r...@italy1.com> wrote: >> Is this centos/ rh or Ubuntu? >> >> Inviato da iPhone () >> Il giorno Jul 26, 2014, alle ore 19:52, john decot <johnde...@gmail.com> ha scritto: >> >> Neutron.conf and ML2_conf.ini files are as follows: >> neutron.conf: >> [DEFAULT] >> # Print more verbose output (set logging level to INFO instead of default WARNING level). >> # verbose = True >> # Print debugging output (set logging level to DEBUG instead of default WARNING level). >> # debug = False >> # Where to store Neutron state files. This directory must be writable by the >> # user executing the agent. >> # state_path = /var/lib/neutron >> # Where to store lock files >> # lock_path = $state_path/lock >> # log_format = %(asctime)s %(levelname)8s [%(name)s] %(message)s >> # log_date_format = %Y-%m-%d %H:%M:%S >> # use_syslog -> syslog >> # log_file and log_dir -> log_dir/log_file >> # (not log_file) and log_dir -> log_dir/{binary_name}.log >> # use_stderr -> stderr >> # (not user_stderr) and (not log_file) -> stdout >> # publish_errors -> notification system >> # use_syslog = False >> # syslog_log_facility = LOG_USER >> # use_stderr = False >> # log_file = >> # log_dir = >> # publish_errors = False >> # Address to bind the API server to >> # bind_host = 0.0.0.0 >> # Port the bind the API server to >> # bind_port = 9696 >> # Path to the extensions. Note that this can be a colon-separated list of >> # paths. For example: >> # api_extensions_path = extensions:/path/to/more/extensions:/even/more/extensions >> # The __path__ of neutron.extensions is appended to this, so if your >> # extensions are in there you don't need to specify them here >> # api_extensions_path = >> # (StrOpt) Neutron core plugin entrypoint to be loaded from the >> # neutron.core_plugins namespace. See setup.cfg for the entrypoint names of the >> # plugins included in the neutron source distribution. For compatibility with >> # previous versions, the class name of a plugin can be specified instead of its >> # entrypoint name. >> # >> # core_plugin = >> # Example: core_plugin = ml2 >> core_plugin = ml2 >> # (ListOpt) List of service plugin entrypoints to be loaded from the >> # neutron.service_plugins namespace. See setup.cfg for the entrypoint names of >> # the plugins included in the neutron source distribution. For compatibility >> # with previous versions, the class name of a plugin can be specified instead >> # of its entrypoint name. >> # >> # service_plugins = >> # Example: service_plugins = router,firewall,lbaas,vpnaas,metering >> service_plugins = router >> # Paste configuration file >> # api_paste_config = /usr/share/neutron/api-paste.ini >> # The strategy to be used for auth. >> # Supported values are 'keystone'(default), 'noauth'. >> # auth_strategy = noauth >> auth_strategy = keystone >> # Base MAC address. The first 3 octets will remain unchanged. If the >> # 4h octet is not 00, it will also be used. The others will be >> # randomly generated. >> # 3 octet >> # base_mac = fa:16:3e:00:00:00 >> # 4 octet >> # base_mac = fa:16:3e:4f:00:00 >> # Maximum amount of retries to generate a unique MAC address >> # mac_generation_retries = 16 >> # DHCP Lease duration (in seconds) >> # dhcp_lease_duration = 86400 >> # Allow sending resource operation notification to DHCP agent >> dhcp_agent_notification = True >> # Enable or disable bulk create/update/delete operations >> # allow_bulk = True >> # Enable or disable pagination >> # allow_pagination = False >> # Enable or disable sorting >> # allow_sorting = False >> # Enable or disable overlapping IPs for subnets >> # Attention: the following parameter MUST be set to False if Neutron is >> # being used in conjunction with nova security groups >> allow_overlapping_ips = True >> # Ensure that configured gateway is on subnet >> # force_gateway_on_subnet = False >> >> # RPC configuration options. Defined in rpc __init__ >> # The messaging module to use, defaults to kombu. >> rpc_backend = neutron.openstack.common.rpc.impl_kombu >> # Size of RPC thread pool >> # rpc_thread_pool_size = 64 >> # Size of RPC connection pool >> # rpc_conn_pool_size = 30 >> # Seconds to wait for a response from call or multicall >> # rpc_response_timeout = 60 >> # Seconds to wait before a cast expires (TTL). Only supported by impl_zmq. >> # rpc_cast_timeout = 30 >> # Modules of exceptions that are permitted to be recreated >> # upon receiving exception data from an rpc call. >> # allowed_rpc_exception_modules = neutron.openstack.common.exception, nova.exception >> # AMQP exchange to connect to if using RabbitMQ or QPID >> control_exchange = neutron >> # If passed, use a fake RabbitMQ provider >> # fake_rabbit = False >> # Configuration options if sending notifications via kombu rpc (these are >> # the defaults) >> # SSL version to use (valid only if SSL enabled) >> # kombu_ssl_version = >> # SSL key file (valid only if SSL enabled) >> # kombu_ssl_keyfile = >> # SSL cert file (valid only if SSL enabled) >> # kombu_ssl_certfile = >> # SSL certification authority file (valid only if SSL enabled) >> # kombu_ssl_ca_certs = >> # IP address of the RabbitMQ installation >> rabbit_host = localhost >> # Password of the RabbitMQ server >> rabbit_password = xxxxxx >> # Port where RabbitMQ server is running/listening >> rabbit_port = 5672 >> # RabbitMQ single or HA cluster (host:port pairs i.e: host1:5672, host2:5672) >> # rabbit_hosts is defaulted to '$rabbit_host:$rabbit_port' >> # rabbit_hosts = localhost:5672 >> # User ID used for RabbitMQ connections >> rabbit_userid = guest >> # Location of a virtual RabbitMQ installation. >> # rabbit_virtual_host = / >> # Maximum retries with trying to connect to RabbitMQ >> # (the default of 0 implies an infinite retry count) >> # rabbit_max_retries = 0 >> # RabbitMQ connection retry interval >> # rabbit_retry_interval = 1 >> # Use HA queues in RabbitMQ (x-ha-policy: all). You need to >> # wipe RabbitMQ database when changing this option. (boolean value) >> # rabbit_ha_queues = false >> # QPID >> # rpc_backend=neutron.openstack.common.rpc.impl_qpid >> # Qpid broker hostname >> # qpid_hostname = localhost >> # Qpid broker port >> # qpid_port = 5672 >> # Qpid single or HA cluster (host:port pairs i.e: host1:5672, host2:5672) >> # qpid_hosts is defaulted to '$qpid_hostname:$qpid_port' >> # qpid_hosts = localhost:5672 >> # Username for qpid connection >> # qpid_username = '' >> # Password for qpid connection >> # qpid_password = '' >> # Space separated list of SASL mechanisms to use for auth >> # qpid_sasl_mechanisms = '' >> # Seconds between connection keepalive heartbeats >> # qpid_heartbeat = 60 >> # Transport to use, either 'tcp' or 'ssl' >> # qpid_protocol = tcp >> # Disable Nagle algorithm >> # qpid_tcp_nodelay = True >> # ZMQ >> # rpc_backend=neutron.openstack.common.rpc.impl_zmq >> # ZeroMQ bind address. Should be a wildcard (*), an ethernet interface, or IP. >> # The "host" option should point or resolve to this address. >> # rpc_zmq_bind_address = * >> # ============ Notification System Options ===================== >> # Notifications can be sent when network/subnet/port are created, updated or deleted. >> # There are three methods of sending notifications: logging (via the >> # log_file directive), rpc (via a message queue) and >> # noop (no notifications sent, the default) >> # Notification_driver can be defined multiple times >> # Do nothing driver >> # notification_driver = neutron.openstack.common.notifier.no_op_notifier >> # Logging driver >> # notification_driver = neutron.openstack.common.notifier.log_notifier >> # RPC driver. >> # notification_driver = neutron.openstack.common.notifier.rpc_notifier >> # default_notification_level is used to form actual topic name(s) or to set logging level >> # default_notification_level = INFO >> # default_publisher_id is a part of the notification payload >> # host = myhost.com >> # default_publisher_id = $host >> # Defined in rpc_notifier, can be comma separated values. >> # The actual topic names will be %s.%(default_notification_level)s >> # notification_topics = notifications >> # Default maximum number of items returned in a single response, >> # value == infinite and value < 0 means no max limit, and value must >> # be greater than 0. If the number of items requested is greater than >> # pagination_max_limit, server will just return pagination_max_limit >> # of number of items. >> # pagination_max_limit = -1 >> # Maximum number of DNS nameservers per subnet >> # max_dns_nameservers = 5 >> # Maximum number of host routes per subnet >> # max_subnet_host_routes = 20 >> # Maximum number of fixed ips per port >> # max_fixed_ips_per_port = 5 >> # =========== items for agent management extension ============= >> # Seconds to regard the agent as down; should be at least twice >> # report_interval, to be sure the agent is down for good >> # agent_down_time = 75 >> # =========== end of items for agent management extension ===== >> # =========== items for agent scheduler extension ============= >> # Driver to use for scheduling network to DHCP agent >> # network_scheduler_driver = neutron.scheduler.dhcp_agent_scheduler.ChanceScheduler >> # Driver to use for scheduling router to a default L3 agent >> # router_scheduler_driver = neutron.scheduler.l3_agent_scheduler.ChanceScheduler >> # Driver to use for scheduling a loadbalancer pool to an lbaas agent >> # loadbalancer_pool_scheduler_driver = neutron.services.loadbalancer.agent_scheduler.ChanceScheduler >> # Allow auto scheduling networks to DHCP agent. It will schedule non-hosted >> # networks to first DHCP agent which sends get_active_networks message to >> # neutron server >> # network_auto_schedule = True >> # Allow auto scheduling routers to L3 agent. It will schedule non-hosted >> # routers to first L3 agent which sends sync_routers message to neutron server >> # router_auto_schedule = True >> # Number of DHCP agents scheduled to host a network. This enables redundant >> # DHCP agents for configured networks. >> # dhcp_agents_per_network = 1 >> # =========== end of items for agent scheduler extension ===== >> # =========== WSGI parameters related to the API server ============== >> # Number of separate worker processes to spawn. The default, 0, runs the >> # worker thread in the current process. Greater than 0 launches that number of >> # child processes as workers. The parent process manages them. >> # api_workers = 0 >> # Number of separate RPC worker processes to spawn. The default, 0, runs the >> # worker thread in the current process. Greater than 0 launches that number of >> # child processes as RPC workers. The parent process manages them. >> # This feature is experimental until issues are addressed and testing has been >> # enabled for various plugins for compatibility. >> # rpc_workers = 0 >> # Sets the value of TCP_KEEPIDLE in seconds to use for each server socket when >> # starting API server. Not supported on OS X. >> # tcp_keepidle = 600 >> # Number of seconds to keep retrying to listen >> # retry_until_window = 30 >> # Number of backlog requests to configure the socket with. >> # backlog = 4096 >> # Max header line to accommodate large tokens >> # max_header_line = 16384 >> # Enable SSL on the API server >> # use_ssl = False >> # Certificate file to use when starting API server securely >> # ssl_cert_file = /path/to/certfile >> # Private key file to use when starting API server securely >> # ssl_key_file = /path/to/keyfile >> # CA certificate file to use when starting API server securely to >> # verify connecting clients. This is an optional parameter only required if >> # API clients need to authenticate to the API server using SSL certificates >> # signed by a trusted CA >> # ssl_ca_file = /path/to/cafile >> # ======== end of WSGI parameters related to the API server ========== >> >> # ======== neutron nova interactions ========== >> # Send notification to nova when port status is active. >> notify_nova_on_port_status_changes = True >> # Send notifications to nova when port data (fixed_ips/floatingips) change >> # so nova can update it's cache. >> notify_nova_on_port_data_changes = True >> # URL for connection to nova (Only supports one nova region currently). >> nova_url = http://127.0.0.1:8774/v2 >> # Name of nova region to use. Useful if keystone manages more than one region >> # nova_region_name = >> nova_region_name = RegionOne >> # Username for connection to nova in admin context >> nova_admin_username = nova >> # The uuid of the admin nova tenant >> # nova_admin_tenant_id = >> nova_admin_tenant_id = 8b83c2b6d22f4dfd82e15b105ea10ae3 >> # Password for connection to nova in admin context. >> nova_admin_password = xxxxxx >> # Authorization URL for connection to nova in admin context. >> # nova_admin_auth_url = >> nova_admin_auth_url = http://localhost:35357/v2.0 >> # Number of seconds between sending events to nova if there are any events to send >> # send_events_interval = 2 >> # ======== end of neutron nova interactions ========== >> [quotas] >> # Default driver to use for quota checks >> # quota_driver = neutron.db.quota_db.DbQuotaDriver >> # Resource name(s) that are supported in quota features >> # quota_items = network,subnet,port >> # Default number of resource allowed per tenant. A negative value means >> # unlimited. >> # default_quota = -1 >> # Number of networks allowed per tenant. A negative value means unlimited. >> # quota_network = 10 >> # Number of subnets allowed per tenant. A negative value means unlimited. >> # quota_subnet = 10 >> # Number of ports allowed per tenant. A negative value means unlimited. >> # quota_port = 50 >> # Number of security groups allowed per tenant. A negative value means >> # unlimited. >> # quota_security_group = 10 >> # Number of security group rules allowed per tenant. A negative value means >> # unlimited. >> # quota_security_group_rule = 100 >> # Number of vips allowed per tenant. A negative value means unlimited. >> # quota_vip = 10 >> # Number of pools allowed per tenant. A negative value means unlimited. >> # quota_pool = 10 >> # Number of pool members allowed per tenant. A negative value means unlimited. >> # The default is unlimited because a member is not a real resource consumer >> # on Openstack. However, on back-end, a member is a resource consumer >> # and that is the reason why quota is possible. >> # quota_member = -1 >> # Number of health monitors allowed per tenant. A negative value means >> # unlimited. >> # The default is unlimited because a health monitor is not a real resource >> # consumer on Openstack. However, on back-end, a member is a resource consumer >> # and that is the reason why quota is possible. >> # quota_health_monitors = -1 >> # Number of routers allowed per tenant. A negative value means unlimited. >> # quota_router = 10 >> # Number of floating IPs allowed per tenant. A negative value means unlimited. >> # quota_floatingip = 50 >> [agent] >> # Use "sudo neutron-rootwrap /etc/neutron/rootwrap.conf" to use the real >> # root filter facility. >> # Change to "sudo" to skip the filtering and just run the comand directly >> root_helper = sudo neutron-rootwrap /etc/neutron/rootwrap.conf >> # =========== items for agent management extension ============= >> # seconds between nodes reporting state to server; should be less than >> # agent_down_time, best if it is half or less than agent_down_time >> # report_interval = 30 >> # =========== end of items for agent management extension ===== >> [keystone_authtoken] >> auth_uri = http://192.168.168.100:5000 >> auth_host = 127.0.0.1 >> auth_port = 35357 >> auth_protocol = http >> admin_tenant_name = service >> admin_user = neutron >> admin_password = xxxxxx >> signing_dir = $state_path/keystone-signing >> [database] >> # This line MUST be changed to actually run the plugin. >> # Example: >> connection = mysql://neutron:xxxxxx@127.0.0.1/neutron_ml2 >> # Replace 127.0.0.1 above with the IP address of the database used by the >> # main neutron server. (Leave it as is if the database runs on this host.) >> # connection = sqlite:// >> # The SQLAlchemy connection string used to connect to the slave database >> # slave_connection = >> # Database reconnection retry times - in event connectivity is lost >> # set to -1 implies an infinite retry count >> # max_retries = 10 >> # Database reconnection interval in seconds - if the initial connection to the >> # database fails >> # retry_interval = 10 >> # Minimum number of SQL connections to keep open in a pool >> # min_pool_size = 1 >> # Maximum number of SQL connections to keep open in a pool >> # max_pool_size = 10 >> # Timeout in seconds before idle sql connections are reaped >> # idle_timeout = 3600 >> # If set, use this value for max_overflow with sqlalchemy >> # max_overflow = 20 >> # Verbosity of SQL debugging information. 0=None, 100=Everything >> # connection_debug = 0 >> # Add python stack traces to SQL as comment strings >> # connection_trace = False >> # If set, use this value for pool_timeout with sqlalchemy >> # pool_timeout = 10 >> [service_providers] >> # Specify service providers (drivers) for advanced services like loadbalancer, VPN, Firewall. >> # Must be in form: >> # service_provider=<service_type>:<name>:<driver>[:default] >> # List of allowed service types includes LOADBALANCER, FIREWALL, VPN >> # Combination of <service type> and <name> must be unique; <driver> must also be unique >> # This is multiline option, example for default provider: >> # service_provider=LOADBALANCER:name:lbaas_plugin_driver_path:default >> # example of non-default provider: >> # service_provider=FIREWALL:name2:firewall_driver_path >> # --- Reference implementations --- >> # service_provider = LOADBALANCER:Haproxy:neutron.services.loadbalancer.drivers.haproxy.plugin_driver.HaproxyOnHostPluginDriver:default >> service_provider=VPN:openswan:neutron.services.vpn.service_drivers.ipsec.IPsecVPNDriver:default >> # In order to activate Radware's lbaas driver you need to uncomment the next line. >> # If you want to keep the HA Proxy as the default lbaas driver, remove the attribute default from the line below. >> # Otherwise comment the HA Proxy line >> # service_provider = LOADBALANCER:Radware:neutron.services.loadbalancer.drivers.radware.driver.LoadBalancerDriver:default >> # uncomment the following line to make the 'netscaler' LBaaS provider available. >> # service_provider=LOADBALANCER:NetScaler:neutron.services.loadbalancer.drivers.netscaler.netscaler_driver.NetScalerPluginDriver >> # Uncomment the following line (and comment out the OpenSwan VPN line) to enable Cisco's VPN driver. >> # service_provider=VPN:cisco:neutron.services.vpn.service_drivers.cisco_ipsec.CiscoCsrIPsecVPNDriver:default >> # Uncomment the line below to use Embrane heleos as Load Balancer service provider. >> # service_provider=LOADBALANCER:Embrane:neutron.services.loadbalancer.drivers.embrane.driver.EmbraneLbaas:default >> >> >> and ml2_conf.ini: >> [ml2] >> # (ListOpt) List of network type driver entrypoints to be loaded from >> # the neutron.ml2.type_drivers namespace. >> # >> # type_drivers = local,flat,vlan,gre,vxlan >> # Example: type_drivers = flat,vlan,gre,vxlan >> type_drivers = flat,vlan,gre >> tenant_network_types = vlan,gre >> mechanism_drivers = openvswitch >> # (ListOpt) Ordered list of network_types to allocate as tenant >> # networks. The default value 'local' is useful for single-box testing >> # but provides no connectivity between hosts. >> # >> # tenant_network_types = local >> # Example: tenant_network_types = vlan,gre,vxlan >> # (ListOpt) Ordered list of networking mechanism driver entrypoints >> # to be loaded from the neutron.ml2.mechanism_drivers namespace. >> # mechanism_drivers = >> # Example: mechanism drivers = openvswitch,mlnx >> # Example: mechanism_drivers = arista >> # Example: mechanism_drivers = cisco,logger >> # Example: mechanism_drivers = openvswitch,brocade >> # Example: mechanism_drivers = linuxbridge,brocade >> [ml2_type_flat] >> # (ListOpt) List of physical_network names with which flat networks >> # can be created. Use * to allow flat networks with arbitrary >> # physical_network names. >> # >> # flat_networks = >> # Example:flat_networks = physnet1,physnet2 >> # Example:flat_networks = * >> [ml2_type_vlan] >> # (ListOpt) List of <physical_network>[:<vlan_min>:<vlan_max>] tuples >> # specifying physical_network names usable for VLAN provider and >> # tenant networks, as well as ranges of VLAN tags on each >> # physical_network available for allocation as tenant networks. >> # >> # network_vlan_ranges = >> # Example: network_vlan_ranges = physnet1:1000:2999,physnet2 >> [ml2_type_gre] >> # (ListOpt) Comma-separated list of <tun_min>:<tun_max> tuples enumerating ranges of GRE tunnel IDs that are available for tenant network allocation >> # tunnel_id_ranges = >> tunnel_id_ranges = 1:1000 >> [ml2_type_vxlan] >> # (ListOpt) Comma-separated list of <vni_min>:<vni_max> tuples enumerating >> # ranges of VXLAN VNI IDs that are available for tenant network allocation. >> # >> # vni_ranges = >> # (StrOpt) Multicast group for the VXLAN interface. When configured, will >> # enable sending all broadcast traffic to this multicast group. When left >> # unconfigured, will disable multicast VXLAN mode. >> # >> # vxlan_group = >> # Example: vxlan_group = 239.1.1.1 >> [securitygroup] >> # Controls if neutron security group is enabled or not. >> # It should be false when you use nova security group. >> firewall_driver = neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver >> enable_security_group = True >> >> Rgds, >> John >> >> On Sun, Jul 27, 2014 at 8:22 AM, john decot <johnde...@gmail.com> wrote: >>> >>> Hi, >>> >>> the is no upstart file in log directory. >>> >>> Regards, >>> John. >>> >>> On Sun, Jul 27, 2014 at 8:14 AM, Gangur, Hrushikesh (R & D HP Cloud) < hrushikesh.gan...@hp.com> wrote: >>>> >>>> Do you see any neutron logs getting created in /var/log/upstart >>>> >>>> >>>> >>>> From: john decot [mailto:johnde...@gmail.com] >>>> Sent: Saturday, July 26, 2014 6:14 AM >>>> To: openstack@lists.openstack.org >>>> Subject: [Openstack] neutron-server cannot start >>>> >>>> >>>> >>>> Hi, >>>> >>>> I am new to openstack. I am trying to bring neutron-server up but try to fail. >>>> >>>> >>>> >>>> output of /etc/init.d/neutron-server status >>>> >>>> >>>> >>>> is >>>> >>>> neutron dead but pid file exists >>>> >>>> >>>> >>>> tail -f /var/log/messages is >>>> >>>> >>>> >>>> detected unhandled Python exception in '/usr/bin/neutron-server' >>>> >>>> Not saving repeating crash in '/usr/bin/neutron-server' >>>> >>>> >>>> >>>> >>>> >>>> direcotry /var/log/neutron is empty >>>> >>>> >>>> >>>> Any help is appreciated. >>>> >>>> >>>> >>>> Thanking You, >>>> >>>> >>>> >>>> Regards, >>>> >>>> John >> >> >> >> _______________________________________________ >> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >> Post to : openstack@lists.openstack.org >> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >> >> >> !DSPAM:1,53d46cbd143596451224348! >> !DSPAM:1,53d47657156749327115755!
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack