cfg.StrOpt('share_mount_path', default='/shares', help="Parent path in service instance where shares " "will be mounted."), cfg.IntOpt('max_time_to_create_volume', default=180, help="Maximum time to wait for creating cinder volume."), cfg.IntOpt('max_time_to_attach', default=120, help="Maximum time to wait for attaching cinder volume."), cfg.StrOpt('service_instance_smb_config_path', default='$share_mount_path/smb.conf', help="Path to SMB config in service instance."), cfg.ListOpt('share_helpers', default=[ 'CIFS=manila.share.drivers.generic.CIFSHelper', 'NFS=manila.share.drivers.generic.NFSHelper', ], help='Specify list of share export helpers.'), cfg.StrOpt('share_volume_fstype', default='ext4', choices=['ext4', 'ext3'], help='Filesystem type of the share volume.'), cfg.StrOpt('cinder_volume_type', default=None, help='Name or id of cinder volume type which will be used ' 'for all volumes created by driver.'), ] CONF = cfg.CONF CONF.register_opts(share_opts)
from nova.scheduler import weights host_manager_opts = [ cfg.MultiStrOpt('scheduler_available_filters', default=['nova.scheduler.filters.all_filters'], help='Filter classes available to the scheduler which may ' 'be specified more than once. An entry of ' '"nova.scheduler.filters.standard_filters" ' 'maps to all filters included with nova.'), cfg.ListOpt('scheduler_default_filters', default=[ 'RetryFilter', 'AvailabilityZoneFilter', 'RamFilter', 'ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', ], help='Which filter class names to use for filtering hosts ' 'when not specified in the request.'), cfg.ListOpt('scheduler_weight_classes', default=['nova.scheduler.weights.all_weighers'], help='Which weight class names to use for weighing hosts'), ] CONF = cfg.CONF CONF.register_opts(host_manager_opts) LOG = logging.getLogger(__name__)
default=paths.state_path_def('mnt'), help='Directory where the glusterfs volume is mounted on the ' 'compute node'), cfg.BoolOpt('iscsi_use_multipath', default=False, help='Use multipath connection of the iSCSI volume'), cfg.BoolOpt('iser_use_multipath', default=False, help='Use multipath connection of the iSER volume'), cfg.StrOpt('scality_sofs_config', help='Path or URL to Scality SOFS configuration file'), cfg.StrOpt('scality_sofs_mount_point', default='$state_path/scality', help='Base dir where Scality SOFS shall be mounted'), cfg.ListOpt('qemu_allowed_storage_drivers', default=[], help='Protocols listed here will be accessed directly ' 'from QEMU. Currently supported protocols: [gluster]') ] CONF = cfg.CONF CONF.register_opts(volume_opts, 'libvirt') class LibvirtBaseVolumeDriver(object): """Base class for volume drivers.""" def __init__(self, connection, is_block_dev): self.connection = connection self.is_block_dev = is_block_dev def get_config(self, connection_info, disk_info): """Returns xml for libvirt."""
from quark import exceptions LOG = logging.getLogger(__name__) CONF = cfg.CONF nvp_opts = [ cfg.IntOpt('max_ports_per_switch', default=0, help=_('Maximum amount of NVP ports on an NVP lswitch')), cfg.StrOpt('default_tz_type', help=_('The type of connector to use for the default tz'), default="stt"), cfg.StrOpt('default_tz', help=_('The default transport zone UUID')), cfg.ListOpt('controller_connection', default=[], help=_('NVP Controller connection string')), cfg.IntOpt('max_rules_per_group', default=30, help=_('Maxiumum size of NVP SecurityRule list per group')), cfg.IntOpt('max_rules_per_port', default=30, help=_('Maximum rules per NVP lport across all groups')), cfg.IntOpt('backoff', default=0, help=_('Base seconds for exponential backoff')), ] physical_net_type_map = { "stt": "stt", "gre": "gre",
from nova.api.openstack import wsgi from nova import exception from nova import notifications from nova.openstack.common import gettextutils from nova.openstack.common.gettextutils import _ from nova.openstack.common import log as logging from nova import utils from nova import wsgi as base_wsgi api_opts = [ cfg.BoolOpt('enabled', default=False, help='Whether the V3 API is enabled or not'), cfg.ListOpt('extensions_blacklist', default=[], help='A list of v3 API extensions to never load. ' 'Specify the extension aliases here.'), cfg.ListOpt('extensions_whitelist', default=[], help='If the list is not empty then a v3 API extension ' 'will only be loaded if it exists in this list. Specify ' 'the extension aliases here.') ] api_opts_group = cfg.OptGroup(name='osapi_v3', title='API v3 Options') LOG = logging.getLogger(__name__) CONF = cfg.CONF CONF.register_group(api_opts_group) CONF.register_opts(api_opts, api_opts_group) # List of v3 API extensions which are considered to form
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo.config import cfg from savanna.openstack.common import context as req_context from savanna.openstack.common.gettextutils import _ # noqa from savanna.openstack.common import log as logging from savanna.openstack.common import rpc LOG = logging.getLogger(__name__) notification_topic_opt = cfg.ListOpt( 'notification_topics', default=[ 'notifications', ], help='AMQP topic used for openstack notifications') CONF = cfg.CONF CONF.register_opt(notification_topic_opt) def notify(context, message): """Sends a notification via RPC.""" if not context: context = req_context.get_admin_context() priority = message.get('priority', CONF.default_notification_level) priority = priority.lower() for topic in CONF.notification_topics: topic = '%s.%s' % (topic, priority)
from oslo.config import cfg from cloudbaseinit.openstack.common import log as logging from cloudbaseinit.plugins.common import base from cloudbaseinit.utils.windows import vds ole32 = ctypes.windll.ole32 ole32.CoTaskMemFree.restype = None ole32.CoTaskMemFree.argtypes = [ctypes.c_void_p] opts = [ cfg.ListOpt('volumes_to_extend', default=None, help='List of volumes that need to be extended ' 'if contiguous space is available on the disk. By default ' 'all the available volumes can be extended. Volumes must ' 'be specified using a comma separated list of volume indexes, ' 'e.g.: "1,2"'), ] CONF = cfg.CONF CONF.register_opts(opts) LOG = logging.getLogger(__name__) class ExtendVolumesPlugin(base.BasePlugin): def _extend_volumes(self, pack, volume_idxs=None): enum = pack.QueryVolumes()
default=True, help='delegation and impersonation features can be ' 'optionally disabled.'), cfg.StrOpt('driver', default='keystone.trust.backends.sql.Trust', help='Keystone Trust backend driver.') ], 'os_inherit': [ cfg.BoolOpt('enabled', default=False, help='role-assignment inheritance to projects from ' 'owning domain can be optionally enabled.') ], 'token': [ cfg.ListOpt('bind', default=[], help='External auth mechanisms that should add bind ' 'information to token e.g. kerberos, x509.'), cfg.StrOpt('enforce_token_bind', default='permissive', help='Enforcement policy on tokens presented to keystone ' 'with bind information. One of disabled, permissive, ' 'strict, required or a specifically required bind ' 'mode e.g. kerberos or x509 to require binding to ' 'that authentication.'), cfg.IntOpt('expiration', default=3600, help='Amount of time a token should remain valid ' '(in seconds).'), cfg.StrOpt('provider', default=None, help='Controls the token construction, validation, and '
secret=True, help=_('Password for NSX controllers in this cluster')), cfg.IntOpt('req_timeout', default=30, help=_('Total time limit for a cluster request')), cfg.IntOpt('http_timeout', default=10, help=_('Time before aborting a request')), cfg.IntOpt('retries', default=2, help=_('Number of time a request should be retried')), cfg.IntOpt('redirects', default=2, help=_('Number of times a redirect should be followed')), cfg.ListOpt('nsx_controllers', deprecated_name='nvp_controllers', help=_("Lists the NSX controllers in this cluster")), ] cluster_opts = [ cfg.StrOpt('default_tz_uuid', help=_("This is uuid of the default NSX Transport zone that " "will be used for creating tunneled isolated " "\"Neutron\" networks. It needs to be created in NSX " "before starting Neutron with the nsx plugin.")), cfg.StrOpt('default_l3_gw_service_uuid', help=_("Unique identifier of the NSX L3 Gateway service " "which will be used for implementing routers and " "floating IPs")), cfg.StrOpt('default_l2_gw_service_uuid', help=_("Unique identifier of the NSX L2 Gateway service "
'%(instance)s%(message)s', help='format string to use for log messages with context'), cfg.StrOpt('logging_default_format_string', default='%(asctime)s.%(msecs)03d %(process)d %(levelname)s ' '%(name)s [-] %(instance)s%(message)s', help='format string to use for log messages without context'), cfg.StrOpt('logging_debug_format_suffix', default='%(funcName)s %(pathname)s:%(lineno)d', help='data to append to log format when level is DEBUG'), cfg.StrOpt('logging_exception_prefix', default='%(asctime)s.%(msecs)03d %(process)d TRACE %(name)s ' '%(instance)s', help='prefix each line of exception output with this format'), cfg.ListOpt('default_log_levels', default=[ 'amqplib=WARN', 'sqlalchemy=WARN', 'boto=WARN', 'suds=INFO', 'keystone=INFO', 'eventlet.wsgi.server=WARN' ], help='list of logger=LEVEL pairs'), cfg.BoolOpt('publish_errors', default=False, help='publish error events'), cfg.BoolOpt('fatal_deprecations', default=False, help='make deprecations fatal'), # NOTE(mikal): there are two options here because sometimes we are handed # a full instance (and could include more information), and other times we # are just handed a UUID for the instance. cfg.StrOpt('instance_format', default='[instance: %(uuid)s] ', help='If an instance is passed with the log message, format ' 'it like this'), cfg.StrOpt('instance_uuid_format',
from marconi import common from marconi.common import decorators from marconi.common import utils from marconi.openstack.common.gettextutils import _ from marconi.openstack.common import log as logging from marconi.queues.storage import base LOG = logging.getLogger(__name__) _PIPELINE_RESOURCES = ('queue', 'message', 'claim') _PIPELINE_CONFIGS = [ cfg.ListOpt(resource + '_pipeline', default=[], help=_('Pipeline to use for processing {0} operations. ' 'This pipeline will be consumed before calling ' 'the storage driver\'s controller methods, ' 'which will always be appended to this ' 'pipeline.').format(resource)) for resource in _PIPELINE_RESOURCES ] _PIPELINE_GROUP = 'storage' def _config_options(): return utils.options_iter(_PIPELINE_CONFIGS, _PIPELINE_GROUP) def _get_storage_pipeline(resource_name, conf): """Constructs and returns a storage resource pipeline.
help='SSL key file (valid only if SSL enabled)'), cfg.StrOpt('kombu_ssl_certfile', default='', help='SSL cert file (valid only if SSL enabled)'), cfg.StrOpt('kombu_ssl_ca_certs', default='', help=('SSL certification authority file ' '(valid only if SSL enabled)')), cfg.StrOpt('rabbit_host', default='localhost', help='The RabbitMQ broker address where a single node is used'), cfg.IntOpt('rabbit_port', default=5672, help='The RabbitMQ broker port where a single node is used'), cfg.ListOpt('rabbit_hosts', default=['$rabbit_host:$rabbit_port'], help='RabbitMQ HA cluster host:port pairs'), cfg.BoolOpt('rabbit_use_ssl', default=False, help='connect over SSL for RabbitMQ'), cfg.StrOpt('rabbit_userid', default='guest', help='the RabbitMQ userid'), cfg.StrOpt('rabbit_password', default='guest', help='the RabbitMQ password', secret=True), cfg.StrOpt('rabbit_virtual_host', default='/', help='the RabbitMQ virtual host'), cfg.IntOpt('rabbit_retry_interval', default=1, help='how frequently to retry connecting with RabbitMQ'),
""" from oslo.config import cfg from oslo.serialization import jsonutils from oslo.utils import timeutils from nova.openstack.common import log as logging from nova.scheduler import host_manager host_manager_opts = [ cfg.ListOpt('baremetal_scheduler_default_filters', default=[ 'RetryFilter', 'AvailabilityZoneFilter', 'ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ExactRamFilter', 'ExactDiskFilter', 'ExactCoreFilter', ], help='Which filter class names to use for filtering ' 'baremetal hosts when not specified in the request.'), cfg.BoolOpt('scheduler_use_baremetal_filters', default=False, help='Flag to decide whether to use ' 'baremetal_scheduler_default_filters or not.'), ] CONF = cfg.CONF CONF.register_opts(host_manager_opts)
coordinator_db_options = [ cfg.BoolOpt('active', default=True, help="""Determines whether the handler for this store is registered for use """ ), cfg.StrOpt('adapter_name', default='mongodb', help="""Sets the name of the handler to load for datasource interactions. e.g. mongodb """ ), cfg.ListOpt('servers', default='localhost:27017', help="""hostanme:port for db servers """ ), cfg.StrOpt('database', default='test', help="""database name """ ), cfg.StrOpt('index', default=None, help="""datasource index """ ), cfg.StrOpt('username', default='test', help="""db username
from nova.scheduler import client as scheduler_client from nova import utils from nova.virt import hardware resource_tracker_opts = [ cfg.IntOpt('reserved_host_disk_mb', default=0, help='Amount of disk in MB to reserve for the host'), cfg.IntOpt('reserved_host_memory_mb', default=512, help='Amount of memory in MB to reserve for the host'), cfg.StrOpt('compute_stats_class', default='nova.compute.stats.Stats', help='Class that will manage stats for the local compute host'), cfg.ListOpt('compute_resources', default=['vcpu'], help='The names of the extra resources to track.'), ] CONF = cfg.CONF CONF.register_opts(resource_tracker_opts) LOG = logging.getLogger(__name__) COMPUTE_RESOURCE_SEMAPHORE = "compute_resources" CONF.import_opt('my_ip', 'nova.netconf') class ResourceTracker(object): """Compute helper class for keeping track of resource usage as instances are built and destroyed.
from nova.openstack.common import importutils from nova.openstack.common import lockutils from nova.openstack.common import log as logging from nova.openstack.common import processutils from nova.openstack.common.rpc import common as rpc_common from nova.openstack.common import timeutils notify_decorator = 'nova.openstack.common.notifier.api.notify_decorator' monkey_patch_opts = [ cfg.BoolOpt('monkey_patch', default=False, help='Whether to log monkey patching'), cfg.ListOpt('monkey_patch_modules', default=[ 'nova.api.ec2.cloud:%s' % (notify_decorator), 'nova.compute.api:%s' % (notify_decorator) ], help='List of modules/decorators to monkey patch'), ] utils_opts = [ cfg.IntOpt('password_length', default=12, help='Length of generated instance admin passwords'), cfg.StrOpt('instance_usage_audit_period', default='month', help='time period to generate instance usages for. ' 'Time period must be hour, day, month or year'), cfg.StrOpt('rootwrap_config', default="/etc/nova/rootwrap.conf", help='Path to the rootwrap configuration file to use for ' 'running commands as root'),
import logging import sys from ryu import log log.early_init_log(logging.DEBUG) from ryu import flags from ryu import version from ryu.app import wsgi from ryu.base.app_manager import AppManager from ryu.controller import controller from ryu.topology import switches CONF = cfg.CONF CONF.register_cli_opts([ cfg.ListOpt('app-lists', default=[], help='application module name to run'), cfg.MultiStrOpt('app', positional=True, default=[], help='application module name to run') ]) def main(): try: CONF(project='ryu', version='ryu-manager %s' % version, default_config_files=['/usr/local/etc/ryu/ryu.conf']) except cfg.ConfigFilesNotFoundError: CONF(project='ryu', version='ryu-manager %s' % version)
import webob from heat.api.aws import exception from heat.openstack.common import log as logging logger = logging.getLogger(__name__) opts = [ cfg.StrOpt('auth_uri', default=None, help=_("Authentication Endpoint URI")), cfg.BoolOpt('multi_cloud', default=False, help=_('Allow orchestration of multiple clouds')), cfg.ListOpt('allowed_auth_uris', default=[], help=_('Allowed keystone endpoints for auth_uri when ' 'multi_cloud is enabled. At least one endpoint needs ' 'to be specified.')) ] cfg.CONF.register_opts(opts, group='ec2authtoken') class EC2Token(wsgi.Middleware): """Authenticate an EC2 request with keystone and convert to token.""" def __init__(self, app, conf): self.conf = conf self.application = app def _conf_get(self, name): # try config from paste-deploy first if name in self.conf:
# under the License. from oslo.config import cfg from swiftclient import utils as swift_utils from ironic.common import exception as exc from ironic.common.glance_service import base_image_service from ironic.common.glance_service import service from ironic.common.glance_service import service_utils from ironic.common.i18n import _ from ironic.common import utils glance_opts = [ cfg.ListOpt('allowed_direct_url_schemes', default=[], help='A list of URL schemes that can be downloaded directly ' 'via the direct_url. Currently supported schemes: ' '[file].'), # To upload this key to Swift: # swift post -m Temp-Url-Key:correcthorsebatterystaple cfg.StrOpt('swift_temp_url_key', help='The secret token given to Swift to allow temporary URL ' 'downloads. Required for temporary URLs.', secret=True), cfg.IntOpt('swift_temp_url_duration', default=1200, help='The length of time in seconds that the temporary URL ' 'will be valid for. Defaults to 20 minutes. If some ' 'deploys get a 401 response code when trying to download ' 'from the temporary URL, try raising this duration.'), cfg.StrOpt('swift_endpoint_url',
from oslo.config import cfg from designate.openstack.common import log as logging from designate.objects import Record from designate.notification_handler.base import NotificationHandler LOG = logging.getLogger(__name__) # Setup a config group cfg.CONF.register_group( cfg.OptGroup(name='handler:sample', title="Configuration for Sample Notification Handler")) # Setup the config options cfg.CONF.register_opts([ cfg.StrOpt('control-exchange', default='nova'), cfg.ListOpt('notification-topics', default=['designate']), cfg.StrOpt('domain-name', default='example.org.'), cfg.StrOpt('domain-id', default='12345'), ], group='handler:sample') class SampleHandler(NotificationHandler): """ Sample Handler """ __plugin_name__ = 'sample' def get_exchange_topics(self): """ Return a tuple of (exchange, [topics]) this handler wants to receive events from. """
cfg.IntOpt('ssh_channel_timeout', default=20, help="Timeout in seconds to wait for output from ssh " "channel."), cfg.IntOpt('ip_version_for_ssh', default=4, help="IP version used for SSH connections."), cfg.StrOpt('catalog_type', default='compute', help="Catalog type of the Compute service."), cfg.StrOpt('path_to_private_key', default='/root/.ssh/id_rsa', help="Path to a private key file for SSH access to remote " "hosts"), cfg.ListOpt('controller_nodes', default=[], help="IP addresses of controller nodes"), cfg.ListOpt('controller_names', default=[], help="FQDNs of controller nodes"), cfg.ListOpt('online_controllers', default=[], help="ips of online controller nodes"), cfg.ListOpt('online_controller_names', default=[], help="FQDNs of online controller nodes"), cfg.ListOpt('compute_nodes', default=[], help="IP addresses of compute nodes"), cfg.ListOpt('online_computes', default=[],
help='Keystone account password'), cfg.StrOpt('admin_tenant_name', default='admin', help='Keystone service account tenant name to validate' ' user tokens'), cfg.StrOpt('cache', default=None, help='Env key for the swift cache'), cfg.StrOpt('certfile', help='Required if Keystone server requires client certificate'), cfg.StrOpt('keyfile', help='Required if Keystone server requires client certificate'), cfg.StrOpt('signing_dir', help='Directory used to cache files related to PKI tokens'), cfg.ListOpt('memcached_servers', deprecated_name='memcache_servers', help='If defined, the memcache server(s) to use for' ' caching'), cfg.IntOpt('token_cache_time', default=300, help='In order to prevent excessive requests and validations,' ' the middleware uses an in-memory cache for the tokens the' ' Keystone API returns. This is only valid if memcache_servers' ' is defined. Set to -1 to disable caching completely.'), cfg.IntOpt('revocation_cache_time', default=1, help='Value only used for unit testing'), cfg.StrOpt('memcache_security_strategy', default=None, help='(optional) if defined, indicate whether token data' ' should be authenticated or authenticated and encrypted.' ' Acceptable values are MAC or ENCRYPT. If MAC, token data is'
cfg.BoolOpt('inject_password', default=True, help='Whether baremetal compute injects password or not'), cfg.StrOpt('injected_network_template', default=paths.basedir_def('nova/virt/' 'baremetal/interfaces.template'), help='Template file for injected network'), cfg.StrOpt('vif_driver', default='nova.virt.baremetal.vif_driver.BareMetalVIFDriver', help='Baremetal VIF driver.'), cfg.StrOpt('volume_driver', default='nova.virt.baremetal.volume_driver.LibvirtVolumeDriver', help='Baremetal volume driver.'), cfg.ListOpt('instance_type_extra_specs', default=[], help='a list of additional capabilities corresponding to ' 'instance_type_extra_specs for this compute ' 'host to advertise. Valid entries are name=value, pairs ' 'For example, "key1:val1, key2:val2"'), cfg.StrOpt('driver', default='nova.virt.baremetal.pxe.PXE', help='Baremetal driver back-end (pxe or tilera)'), cfg.StrOpt('power_manager', default='nova.virt.baremetal.ipmi.IPMI', help='Baremetal power management method'), cfg.StrOpt('tftp_root', default='/tftpboot', help='Baremetal compute node\'s tftp root path'), ] LOG = logging.getLogger(__name__)
from nova.compute import vm_states from nova import conductor from nova.db import base from nova import exception from nova.objects import base as obj_base from nova.objects import instance as instance_obj from nova.openstack.common.gettextutils import _ from nova.openstack.common import log as logging from nova.scheduler import rpcapi as scheduler_rpcapi from nova.scheduler import utils as scheduler_utils from nova import utils cell_scheduler_opts = [ cfg.ListOpt('scheduler_filter_classes', default=['nova.cells.filters.all_filters'], help='Filter classes the cells scheduler should use. ' 'An entry of "nova.cells.filters.all_filters" ' 'maps to all cells filters included with nova.'), cfg.ListOpt('scheduler_weight_classes', default=['nova.cells.weights.all_weighers'], help='Weigher classes the cells scheduler should use. ' 'An entry of "nova.cells.weights.all_weighers" ' 'maps to all cell weighers included with nova.'), cfg.IntOpt('scheduler_retries', default=10, help='How many retries when no cells are available.'), cfg.IntOpt('scheduler_retry_delay', default=2, help='How often to retry in seconds when no cells are ' 'available.') ]
cfg.IntOpt('swift_store_large_object_chunk_size', default=DEFAULT_LARGE_OBJECT_CHUNK_SIZE, help=_('The amount of data written to a temporary disk buffer ' 'during the process of chunking the image file.')), cfg.BoolOpt('swift_store_create_container_on_put', default=False, help=_('A boolean value that determines if we create the ' 'container if it does not exist.')), cfg.BoolOpt('swift_store_multi_tenant', default=False, help=_('If set to True, enables multi-tenant storage ' 'mode which causes Glance images to be stored in ' 'tenant specific Swift accounts.')), cfg.ListOpt('swift_store_admin_tenants', default=[], help=_('A list of tenants that will be granted read/write ' 'access on all Swift containers created by Glance in ' 'multi-tenant mode.')), cfg.BoolOpt('swift_store_ssl_compression', default=True, help=_('If set to False, disables SSL layer compression of ' 'https swift requests. Setting to False may improve ' 'performance for images which are already in a ' 'compressed format, eg qcow2.')), cfg.IntOpt('swift_store_retry_get_count', default=0, help=_('The number of times a Swift download will be retried ' 'before the request fails.')) ] CONF = cfg.CONF
cfg.IntOpt('bind_port', default=9696, help=_("The port to bind to")), cfg.StrOpt('api_paste_config', default="api-paste.ini", help=_("The API paste config file to use")), cfg.StrOpt('api_extensions_path', default="", help=_("The path for API extensions")), cfg.StrOpt('policy_file', default="policy.json", help=_("The policy file to use")), cfg.StrOpt('auth_strategy', default='keystone', help=_("The type of authentication to use")), cfg.StrOpt('core_plugin', help=_("The core plugin Neutron will use")), cfg.ListOpt('service_plugins', default=[], help=_("The service plugins Neutron will use")), cfg.StrOpt('base_mac', default="fa:16:3e:00:00:00", help=_("The base MAC address Neutron will use for VIFs")), cfg.IntOpt('mac_generation_retries', default=16, help=_("How many times Neutron will retry MAC generation")), cfg.BoolOpt('allow_bulk', default=True, help=_("Allow the usage of the bulk API")), cfg.BoolOpt('allow_pagination', default=False, help=_("Allow the usage of the pagination")), cfg.BoolOpt('allow_sorting', default=False,
'%(name)s [%(request_id)s %(user_identity)s] ' '%(instance)s%(message)s', help='Format string to use for log messages with context.'), cfg.StrOpt('logging_default_format_string', default='%(asctime)s.%(msecs)03d %(process)d %(levelname)s ' '%(name)s [-] %(instance)s%(message)s', help='Format string to use for log messages without context.'), cfg.StrOpt('logging_debug_format_suffix', default='%(funcName)s %(pathname)s:%(lineno)d', help='Data to append to log format when level is DEBUG.'), cfg.StrOpt('logging_exception_prefix', default='%(asctime)s.%(msecs)03d %(process)d TRACE %(name)s ' '%(instance)s', help='Prefix each line of exception output with this format.'), cfg.ListOpt('default_log_levels', default=DEFAULT_LOG_LEVELS, help='List of logger=LEVEL pairs.'), cfg.BoolOpt('publish_errors', default=False, help='Enables or disables publication of error events.'), cfg.BoolOpt('fatal_deprecations', default=False, help='Enables or disables fatal status of deprecations.'), # NOTE(mikal): there are two options here because sometimes we are handed # a full instance (and could include more information), and other times we # are just handed a UUID for the instance. cfg.StrOpt('instance_format', default='[instance: %(uuid)s] ', help='The format for an instance that is passed with the log ' 'message.'),
'%(name)s [-] %(instance)s%(message)s', help='Format string to use for log messages without context'), cfg.StrOpt('logging_debug_format_suffix', default='%(funcName)s %(pathname)s:%(lineno)d', help='Data to append to log format when level is DEBUG'), cfg.StrOpt('logging_exception_prefix', default='%(asctime)s.%(msecs)03d %(process)d TRACE %(name)s ' '%(instance)s', help='Prefix each line of exception output with this format'), cfg.ListOpt('default_log_levels', default=[ 'amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN' ], help='List of logger=LEVEL pairs'), cfg.BoolOpt('publish_errors', default=False, help='Publish error events'), cfg.BoolOpt('fatal_deprecations', default=False, help='Make deprecations fatal'), # NOTE(mikal): there are two options here because sometimes we are handed # a full instance (and could include more information), and other times we
from oslo.config import cfg from neutron.agent.common import config DEFAULT_VLAN_RANGES = [] DEFAULT_INTERFACE_MAPPINGS = [] DEFAULT_VXLAN_GROUP = '224.0.0.1' vlan_opts = [ cfg.StrOpt('tenant_network_type', default='local', help=_("Network type for tenant networks " "(local, vlan, or none)")), cfg.ListOpt('network_vlan_ranges', default=DEFAULT_VLAN_RANGES, help=_("List of <physical_network>:<vlan_min>:<vlan_max> " "or <physical_network>")), ] vxlan_opts = [ cfg.BoolOpt('enable_vxlan', default=False, help=_("Enable VXLAN on the agent. Can be enabled when " "agent is managed by ml2 plugin using linuxbridge " "mechanism driver")), cfg.IntOpt('ttl', help=_("TTL for vxlan interface protocol packets.")), cfg.IntOpt('tos', help=_("TOS for vxlan interface protocol packets.")), cfg.StrOpt('vxlan_group', default=DEFAULT_VXLAN_GROUP, help=_("Multicast group for vxlan interface.")), cfg.StrOpt('local_ip',
import fnmatch import functools import glob import os import os.path import re import shutil import sys from oslo.config import cfg _OBSOLETE_LIST = None opts = [ cfg.ListOpt('modules', default=[], help='The list of modules to copy from oslo-incubator ' '(deprecated in favor of --module).'), cfg.MultiStrOpt('module', default=[], help='The list of modules to copy from oslo-incubator.'), cfg.MultiStrOpt( 'script', default=[], help='The list of stand-alone scripts to copy from oslo-incubator.'), cfg.StrOpt('base', help='The base module to hold the copy of openstack.common.'), cfg.StrOpt('dest-dir', help='Destination project directory.'), cfg.StrOpt('configfile_or_destdir', help='A configuration file or destination project directory.', positional=True), cfg.BoolOpt('nodeps',