class runner(SuprocBenchmarks): """Some rudimentary tests to see if there is no major slowdowns from Runner """ def setup(self): from datalad.cmd import Runner self.runner = Runner() # older versions might not have it try: from datalad.cmd import GitRunner self.git_runner = GitRunner() except ImportError: pass def time_echo(self): self.runner.run("echo") def time_echo_gitrunner(self): self.git_runner.run("echo")
class RunnerSuite(SuprocBenchmarks): """Some rudimentary tests to see if there is no major slowdowns from Runner """ def setup(self): from datalad.cmd import Runner self.runner = Runner() # older versions might not have it try: from datalad.cmd import GitRunner self.git_runner = GitRunner() except ImportError: pass def time_echo(self): self.runner.run("echo") def time_echo_gitrunner(self): self.git_runner.run("echo")
def _describe_annex(): from datalad.cmd import GitRunner runner = GitRunner() try: out, err = runner.run(['git', 'annex', 'version']) except CommandError as e: return dict( version='not available', message=exc_str(e), ) info = {} for line in out.split(os.linesep): key = line.split(':')[0] if not key: continue value = line[len(key) + 2:].strip() key = key.replace('git-annex ', '') if key.endswith('s'): value = value.split() info[key] = value return info
class ConfigManager(object): """Thin wrapper around `git-config` with support for a dataset configuration. The general idea is to have an object that is primarily used to read/query configuration option. Upon creation, current configuration is read via one (or max two, in the case of the presence of dataset-specific configuration) calls to `git config`. If this class is initialized with a Dataset instance, it supports reading and writing configuration from ``.datalad/config`` inside a dataset too. This file is committed to Git and hence useful to ship certain configuration items with a dataset. The API aims to provide the most significant read-access API of a dictionary, the Python ConfigParser, and GitPython's config parser implementations. This class is presently not capable of efficiently writing multiple configurations items at once. Instead, each modification results in a dedicated call to `git config`. This author thinks this is OK, as he cannot think of a situation where a large number of items need to be written during normal operation. If such need arises, various solutions are possible (via GitPython, or an independent writer). Each instance carries a public `overrides` attribute. This dictionary contains variables that override any setting read from a file. The overrides are persistent across reloads, and are not modified by any of the manipulation methods, such as `set` or `unset`. Any DATALAD_* environment variable is also presented as a configuration item. Settings read from environment variables are not stored in any of the configuration file, but are read dynamically from the environment at each `reload()` call. Their values take precedence over any specification in configuration files, and even overrides. Parameters ---------- dataset : Dataset, optional If provided, all `git config` calls are executed in this dataset's directory. Moreover, any modifications are, by default, directed to this dataset's configuration file (which will be created on demand) dataset_only : bool If True, configuration items are only read from a datasets persistent configuration file, if any present (the one in ``.datalad/config``, not ``.git/config``). overrides : dict, optional Variable overrides, see general class documentation for details. """ def __init__(self, dataset=None, dataset_only=False, overrides=None): # store in a simple dict # no subclassing, because we want to be largely read-only, and implement # config writing separately self._store = {} self._cfgfiles = set() self._cfgmtimes = None # public dict to store variables that always override any setting # read from a file # `hasattr()` is needed because `datalad.cfg` is generated upon first module # import, hence when this code runs first, there cannot be any config manager # to inherit from self.overrides = datalad.cfg.overrides.copy() if hasattr( datalad, 'cfg') else {} if overrides is not None: self.overrides.update(overrides) if dataset is None: self._dataset_path = None self._dataset_cfgfname = None self._repo_cfgfname = None else: self._dataset_path = dataset.path self._dataset_cfgfname = opj(self._dataset_path, '.datalad', 'config') if not dataset_only: self._repo_cfgfname = opj(self._dataset_path, '.git', 'config') self._dataset_only = dataset_only # Since configs could contain sensitive information, to prevent # any "facilitated" leakage -- just disable logging of outputs for # this runner run_kwargs = dict(log_outputs=False) if dataset is not None: # make sure we run the git config calls in the dataset # to pick up the right config files run_kwargs['cwd'] = dataset.path self._runner = GitRunner(**run_kwargs) try: self._gitconfig_has_showorgin = \ LooseVersion(get_git_version(self._runner)) >= '2.8.0' except: # no git something else broken, assume git is present anyway # to not delay this, but assume it is old self._gitconfig_has_showorgin = False self.reload(force=True) def reload(self, force=False): """Reload all configuration items from the configured sources If `force` is False, all files configuration was previously read from are checked for differences in the modification times. If no difference is found for any file no reload is performed. This mechanism will not detect newly created global configuration files, use `force` in this case. """ if not force and self._cfgmtimes: # we aren't forcing and we have read files before # check if any file we read from has changed current_time = time() curmtimes = {c: getmtime(c) for c in self._cfgfiles if exists(c)} if all(curmtimes[c] == self._cfgmtimes.get(c) and # protect against low-res mtimes (FAT32 has 2s, EXT3 has 1s!) # if mtime age is less than worst resolution assume modified (current_time - curmtimes[c]) > 2.0 for c in curmtimes): # all the same, nothing to do, except for # superimpose overrides, could have changed in the meantime self._store.update(self.overrides) # reread env, is quick self._store = _parse_env(self._store) return self._store = {} # 2-step strategy: # - load datalad dataset config from dataset # - load git config from all supported by git sources # in doing so we always stay compatible with where Git gets its # config from, but also allow to override persistent information # from dataset locally or globally run_args = ['-z', '-l'] if self._gitconfig_has_showorgin: run_args.append('--show-origin') if self._dataset_cfgfname: if exists(self._dataset_cfgfname): stdout, stderr = self._run(run_args + ['--file', self._dataset_cfgfname], log_stderr=True) # overwrite existing value, do not amend to get multi-line # values self._store, self._cfgfiles = _parse_gitconfig_dump( stdout, self._store, self._cfgfiles, replace=False) if self._dataset_only: # superimpose overrides self._store.update(self.overrides) return stdout, stderr = self._run(run_args, log_stderr=True) self._store, self._cfgfiles = _parse_gitconfig_dump(stdout, self._store, self._cfgfiles, replace=True) # always monitor the dataset cfg location, we know where it is in all cases if self._dataset_cfgfname: self._cfgfiles.add(self._dataset_cfgfname) self._cfgfiles.add(self._repo_cfgfname) self._cfgmtimes = {c: getmtime(c) for c in self._cfgfiles if exists(c)} # superimpose overrides self._store.update(self.overrides) # override with environment variables self._store = _parse_env(self._store) @_where_reload def obtain(self, var, default=None, dialog_type=None, valtype=None, store=False, where=None, reload=True, **kwargs): """ Convenience method to obtain settings interactively, if needed A UI will be used to ask for user input in interactive sessions. Questions to ask, and additional explanations can be passed directly as arguments, or retrieved from a list of pre-configured items. Additionally, this method allows for type conversion and storage of obtained settings. Both aspects can also be pre-configured. Parameters ---------- var : str Variable name including any section like `git config` expects them, e.g. 'core.editor' default : any type In interactive sessions and if `store` is True, this default value will be presented to the user for confirmation (or modification). In all other cases, this value will be silently assigned unless there is an existing configuration setting. dialog_type : {'question', 'yesno', None} Which dialog type to use in interactive sessions. If `None`, pre-configured UI options are used. store : bool Whether to store the obtained value (or default) %s `**kwargs` Additional arguments for the UI function call, such as a question `text`. """ # do local import, as this module is import prominently and the # could theroetically import all kind of weired things for type # conversion from datalad.interface.common_cfg import definitions as cfg_defs # fetch what we know about this variable cdef = cfg_defs.get(var, {}) # type conversion setup if valtype is None and 'type' in cdef: valtype = cdef['type'] if valtype is None: valtype = lambda x: x # any default? if default is None and 'default' in cdef: default = cdef['default'] _value = None if var in self: # nothing needs to be obtained, it is all here already _value = self[var] elif store is False and default is not None: # nothing will be stored, and we have a default -> no user confirmation # we cannot use logging, because we want to use the config to confiugre # the logging #lgr.debug('using default {} for config setting {}'.format(default, var)) _value = default if _value is not None: # we got everything we need and can exit early try: return valtype(_value) except Exception as e: raise ValueError( "value '{}' of existing configuration for '{}' cannot be " "converted to the desired type '{}' ({})".format( _value, var, valtype, exc_str(e))) # now we need to try to obtain something from the user from datalad.ui import ui # configure UI dialog_opts = kwargs if dialog_type is None: # no override # check for common knowledge on how to obtain a value if 'ui' in cdef: dialog_type = cdef['ui'][0] # pull standard dialog settings dialog_opts = cdef['ui'][1] # update with input dialog_opts.update(kwargs) if (not ui.is_interactive or dialog_type is None) and default is None: raise RuntimeError( "cannot obtain value for configuration item '{}', " "not preconfigured, no default, no UI available".format(var)) if not hasattr(ui, dialog_type): raise ValueError( "UI '{}' does not support dialog type '{}'".format( ui, dialog_type)) # configure storage destination, if needed if store: if where is None and 'destination' in cdef: where = cdef['destination'] if where is None: raise ValueError( "request to store configuration item '{}', but no " "storage destination specified".format(var)) # obtain via UI dialog = getattr(ui, dialog_type) _value = dialog(default=default, **dialog_opts) if _value is None: # we got nothing if default is None: raise RuntimeError( "could not obtain value for configuration item '{}', " "not preconfigured, no default".format(var)) # XXX maybe we should return default here, even it was returned # from the UI -- if that is even possible # execute type conversion before storing to check that we got # something that looks like what we want try: value = valtype(_value) except Exception as e: raise ValueError( "cannot convert user input `{}` to desired type ({})".format( _value, exc_str(e))) # XXX we could consider "looping" until we have a value of proper # type in case of a user typo... if store: # store value as it was before any conversion, needs to be str # anyway # needs string conversion nevertheless, because default could come # in as something else self.add(var, '{}'.format(_value), where=where, reload=reload) return value # # Compatibility with dict API # def __len__(self): return len(self._store) def __getitem__(self, key): return self._store.__getitem__(key) def __contains__(self, key): return self._store.__contains__(key) def keys(self): """Returns list of configuration item names""" return self._store.keys() # XXX should this be *args? def get(self, key, default=None): """D.get(k[,d]) -> D[k] if k in D, else d. d defaults to None.""" return self._store.get(key, default) # # Compatibility with ConfigParser API # def sections(self): """Returns a list of the sections available""" return list( set([cfg_section_regex.match(k).group(1) for k in self._store])) def options(self, section): """Returns a list of options available in the specified section.""" opts = [] for k in self._store: sec, opt = cfg_sectionoption_regex.match(k).groups() if sec == section: opts.append(opt) return opts def has_section(self, section): """Indicates whether a section is present in the configuration""" for k in self._store: if k.startswith(section): return True return False def has_option(self, section, option): """If the given section exists, and contains the given option""" for k in self._store: sec, opt = cfg_sectionoption_regex.match(k).groups() if sec == section and opt == option: return True return False def getint(self, section, option): """A convenience method which coerces the option value to an integer""" return int(self.get_value(section, option)) def getbool(self, section, option, default=None): """A convenience method which coerces the option value to a bool Values "on", "yes", "true" and any int!=0 are considered True Values which evaluate to bool False, "off", "no", "false" are considered False TypeError is raised for other values. """ val = self.get_value(section, option, default=default) return anything2bool(val) def getfloat(self, section, option): """A convenience method which coerces the option value to a float""" return float(self.get_value(section, option)) # this is a hybrid of ConfigParser and dict API def items(self, section=None): """Return a list of (name, value) pairs for each option Optionally limited to a given section. """ if section is None: return self._store.items() return [(k, v) for k, v in self._store.items() if cfg_section_regex.match(k).group(1) == section] # # Compatibility with GitPython's ConfigParser # def get_value(self, section, option, default=None): """Like `get()`, but with an optional default value If the default is not None, the given default value will be returned in case the option did not exist. This behavior imitates GitPython's config parser. """ try: return self['.'.join((section, option))] except KeyError as e: # this strange dance is needed because gitpython does it this way if default is not None: return default else: raise e # # Modify configuration (proxy respective git-config call) # @_where_reload def _run(self, args, where=None, reload=False, **kwargs): """Centralized helper to run "git config" calls Parameters ---------- args : list Arguments to pass for git config %s **kwargs Keywords arguments for Runner's call """ if where: args = self._get_location_args(where) + args out = self._runner.run(['git', 'config'] + args, **kwargs) if reload: self.reload() return out def _get_location_args(self, where, args=None): if args is None: args = [] cfg_labels = ('dataset', 'local', 'global') if where not in cfg_labels: raise ValueError( "unknown configuration label '{}' (not in {})".format( where, cfg_labels)) if where == 'dataset': if not self._dataset_cfgfname: raise ValueError( 'ConfigManager cannot store to configuration to dataset, ' 'none specified') # create an empty config file if none exists, `git config` will # fail otherwise dscfg_dirname = opj(self._dataset_path, '.datalad') if not exists(dscfg_dirname): os.makedirs(dscfg_dirname) if not exists(self._dataset_cfgfname): open(self._dataset_cfgfname, 'w').close() args.extend(['--file', self._dataset_cfgfname]) elif where == 'global': args.append('--global') elif where == 'local': args.append('--local') return args @_where_reload def add(self, var, value, where='dataset', reload=True): """Add a configuration variable and value Parameters ---------- var : str Variable name including any section like `git config` expects them, e.g. 'core.editor' value : str Variable value %s""" self._run(['--add', var, value], where=where, reload=reload, log_stderr=True) @_where_reload def set(self, var, value, where='dataset', reload=True, force=False): """Set a variable to a value. In opposition to `add`, this replaces the value of `var` if there is one already. Parameters ---------- var : str Variable name including any section like `git config` expects them, e.g. 'core.editor' value : str Variable value force: bool if set, replaces all occurrences of `var` by a single one with the given `value`. Otherwise raise if multiple entries for `var` exist already %s""" from datalad.support.gitrepo import to_options self._run(to_options(replace_all=force) + [var, value], where=where, reload=reload, log_stderr=True) @_where_reload def rename_section(self, old, new, where='dataset', reload=True): """Rename a configuration section Parameters ---------- old : str Name of the section to rename. new : str Name of the section to rename to. %s""" self._run(['--rename-section', old, new], where=where, reload=reload) @_where_reload def remove_section(self, sec, where='dataset', reload=True): """Rename a configuration section Parameters ---------- sec : str Name of the section to remove. %s""" self._run(['--remove-section', sec], where=where, reload=reload) @_where_reload def unset(self, var, where='dataset', reload=True): """Remove all occurrences of a variable Parameters ---------- var : str Name of the variable to remove %s""" # use unset all as it is simpler for now self._run(['--unset-all', var], where=where, reload=reload)
class runner(SuprocBenchmarks): """Some rudimentary tests to see if there is no major slowdowns from Runner """ def setup(self): self.runner = Runner() # older versions might not have it try: from datalad.cmd import GitRunner self.git_runner = GitRunner() except ImportError: pass def time_echo(self): self.runner.run("echo") def time_echo_gitrunner(self): self.git_runner.run("echo") # Following "track" measures computing overhead comparing to the simplest # os.system call on the same command without carrying for in/out unit = "% overhead" def _get_overhead(self, cmd, nrepeats=3, **run_kwargs): """Estimate overhead over running command via the simplest os.system and to not care about any output """ # asv does not repeat tracking ones I think, so nrepeats overheads = [] for _ in range(nrepeats): t0 = time() os.system(cmd + " >/dev/null 2>&1") t1 = time() self.runner.run(cmd, **run_kwargs) t2 = time() overhead = 100 * ((t2 - t1) / (t1 - t0) - 1.0) # print("O :", t1 - t0, t2 - t0, overhead) overheads.append(overhead) overhead = round(sum(overheads) / len(overheads), 2) #overhead = round(min(overheads), 2) return overhead def track_overhead_echo(self): return self._get_overhead("echo") # 100ms chosen below as providing some sensible stability for me. # at 10ms -- too much variability def track_overhead_100ms(self): return self._get_overhead("sleep 0.1") def track_overhead_heavyout(self): # run busyloop for 100ms outputing as much as it could return self._get_overhead(heavyout_cmd) def track_overhead_heavyout_online_through(self): return self._get_overhead( heavyout_cmd, log_stderr='offline', # needed to would get stuck log_online=True) def track_overhead_heavyout_online_process(self): return self._get_overhead( heavyout_cmd, log_stdout=lambda s: '', log_stderr='offline', # needed to would get stuck log_online=True)
class runner(SuprocBenchmarks): """Some rudimentary tests to see if there is no major slowdowns from Runner """ def setup(self): self.runner = Runner() # older versions might not have it try: from datalad.cmd import GitRunner self.git_runner = GitRunner() except ImportError: pass def time_echo(self): self.runner.run("echo") def time_echo_gitrunner(self): self.git_runner.run("echo") # Following "track" measures computing overhead comparing to the simplest # os.system call on the same command without carrying for in/out unit = "% overhead" def _get_overhead(self, cmd, nrepeats=3, **run_kwargs): """Estimate overhead over running command via the simplest os.system and to not care about any output """ # asv does not repeat tracking ones I think, so nrepeats overheads = [] for _ in range(nrepeats): t0 = time() os.system(cmd + " >/dev/null 2>&1") t1 = time() self.runner.run(cmd, **run_kwargs) t2 = time() overhead = 100 * ((t2 - t1) / (t1 - t0) - 1.0) # print("O :", t1 - t0, t2 - t0, overhead) overheads.append(overhead) overhead = round(sum(overheads) / len(overheads), 2) #overhead = round(min(overheads), 2) return overhead def track_overhead_echo(self): return self._get_overhead("echo") # 100ms chosen below as providing some sensible stability for me. # at 10ms -- too much variability def track_overhead_100ms(self): return self._get_overhead("sleep 0.1") def track_overhead_heavyout(self): # run busyloop for 100ms outputing as much as it could return self._get_overhead(heavyout_cmd) def track_overhead_heavyout_online_through(self): return self._get_overhead(heavyout_cmd, log_stderr='offline', # needed to would get stuck log_online=True) def track_overhead_heavyout_online_process(self): return self._get_overhead(heavyout_cmd, log_stdout=lambda s: '', log_stderr='offline', # needed to would get stuck log_online=True) # # Probably not really interesting, and good lord wobbles around 0 # def track_overhead_heavyout_offline(self): # return self._get_overhead(heavyout_cmd, # log_stdout='offline', # log_stderr='offline') # TODO: track the one with in/out, i.e. for those BatchedProcesses
class ConfigManager(object): """Thin wrapper around `git-config` with support for a dataset configuration. The general idea is to have an object that is primarily used to read/query configuration option. Upon creation, current configuration is read via one (or max two, in the case of the presence of dataset-specific configuration) calls to `git config`. If this class is initialized with a Dataset instance, it supports reading and writing configuration from ``.datalad/config`` inside a dataset too. This file is committed to Git and hence useful to ship certain configuration items with a dataset. The API aims to provide the most significant read-access API of a dictionary, the Python ConfigParser, and GitPython's config parser implementations. This class is presently not capable of efficiently writing multiple configurations items at once. Instead, each modification results in a dedicated call to `git config`. This author thinks this is OK, as he cannot think of a situation where a large number of items need to be written during normal operation. If such need arises, various solutions are possible (via GitPython, or an independent writer). Each instance carries a public `overrides` attribute. This dictionary contains variables that override any setting read from a file. The overrides are persistent across reloads, and are not modified by any of the manipulation methods, such as `set` or `unset`. Any DATALAD_* environment variable is also presented as a configuration item. Settings read from environment variables are not stored in any of the configuration file, but are read dynamically from the environment at each `reload()` call. Their values take precedence over any specification in configuration files, and even overrides. Parameters ---------- dataset : Dataset, optional If provided, all `git config` calls are executed in this dataset's directory. Moreover, any modifications are, by default, directed to this dataset's configuration file (which will be created on demand) dataset_only : bool If True, configuration items are only read from a datasets persistent configuration file, if any present (the one in ``.datalad/config``, not ``.git/config``). overrides : dict, optional Variable overrides, see general class documentation for details. """ def __init__(self, dataset=None, dataset_only=False, overrides=None): # store in a simple dict # no subclassing, because we want to be largely read-only, and implement # config writing separately self._store = {} self._cfgfiles = set() self._cfgmtimes = None # public dict to store variables that always override any setting # read from a file # `hasattr()` is needed because `datalad.cfg` is generated upon first module # import, hence when this code runs first, there cannot be any config manager # to inherit from self.overrides = datalad.cfg.overrides.copy() if hasattr(datalad, 'cfg') else {} if overrides is not None: self.overrides.update(overrides) if dataset is None: self._dataset_path = None self._dataset_cfgfname = None self._repo_cfgfname = None else: self._dataset_path = dataset.path self._dataset_cfgfname = opj(self._dataset_path, DATASET_CONFIG_FILE) if not dataset_only: self._repo_cfgfname = opj(self._dataset_path, '.git', 'config') self._dataset_only = dataset_only # Since configs could contain sensitive information, to prevent # any "facilitated" leakage -- just disable logging of outputs for # this runner run_kwargs = dict(log_outputs=False) if dataset is not None: # make sure we run the git config calls in the dataset # to pick up the right config files run_kwargs['cwd'] = dataset.path self._runner = GitRunner(**run_kwargs) try: self._gitconfig_has_showorgin = \ LooseVersion(get_git_version(self._runner)) >= '2.8.0' except: # no git something else broken, assume git is present anyway # to not delay this, but assume it is old self._gitconfig_has_showorgin = False self.reload(force=True) def reload(self, force=False): """Reload all configuration items from the configured sources If `force` is False, all files configuration was previously read from are checked for differences in the modification times. If no difference is found for any file no reload is performed. This mechanism will not detect newly created global configuration files, use `force` in this case. """ if not force and self._cfgmtimes: # we aren't forcing and we have read files before # check if any file we read from has changed current_time = time() curmtimes = {c: getmtime(c) for c in self._cfgfiles if exists(c)} if all(curmtimes[c] == self._cfgmtimes.get(c) and # protect against low-res mtimes (FAT32 has 2s, EXT3 has 1s!) # if mtime age is less than worst resolution assume modified (current_time - curmtimes[c]) > 2.0 for c in curmtimes): # all the same, nothing to do, except for # superimpose overrides, could have changed in the meantime self._store.update(self.overrides) # reread env, is quick self._store = _parse_env(self._store) return self._store = {} # 2-step strategy: # - load datalad dataset config from dataset # - load git config from all supported by git sources # in doing so we always stay compatible with where Git gets its # config from, but also allow to override persistent information # from dataset locally or globally run_args = ['-z', '-l'] if self._gitconfig_has_showorgin: run_args.append('--show-origin') if self._dataset_cfgfname: if exists(self._dataset_cfgfname): stdout, stderr = self._run( run_args + ['--file', self._dataset_cfgfname], log_stderr=True ) # overwrite existing value, do not amend to get multi-line # values self._store, self._cfgfiles = _parse_gitconfig_dump( stdout, self._store, self._cfgfiles, replace=False) if self._dataset_only: # superimpose overrides self._store.update(self.overrides) return stdout, stderr = self._run(run_args, log_stderr=True) self._store, self._cfgfiles = _parse_gitconfig_dump( stdout, self._store, self._cfgfiles, replace=True) # always monitor the dataset cfg location, we know where it is in all cases if self._dataset_cfgfname: self._cfgfiles.add(self._dataset_cfgfname) self._cfgfiles.add(self._repo_cfgfname) self._cfgmtimes = {c: getmtime(c) for c in self._cfgfiles if exists(c)} # superimpose overrides self._store.update(self.overrides) # override with environment variables self._store = _parse_env(self._store) @_where_reload def obtain(self, var, default=None, dialog_type=None, valtype=None, store=False, where=None, reload=True, **kwargs): """ Convenience method to obtain settings interactively, if needed A UI will be used to ask for user input in interactive sessions. Questions to ask, and additional explanations can be passed directly as arguments, or retrieved from a list of pre-configured items. Additionally, this method allows for type conversion and storage of obtained settings. Both aspects can also be pre-configured. Parameters ---------- var : str Variable name including any section like `git config` expects them, e.g. 'core.editor' default : any type In interactive sessions and if `store` is True, this default value will be presented to the user for confirmation (or modification). In all other cases, this value will be silently assigned unless there is an existing configuration setting. dialog_type : {'question', 'yesno', None} Which dialog type to use in interactive sessions. If `None`, pre-configured UI options are used. store : bool Whether to store the obtained value (or default) %s `**kwargs` Additional arguments for the UI function call, such as a question `text`. """ # do local import, as this module is import prominently and the # could theroetically import all kind of weired things for type # conversion from datalad.interface.common_cfg import definitions as cfg_defs # fetch what we know about this variable cdef = cfg_defs.get(var, {}) # type conversion setup if valtype is None and 'type' in cdef: valtype = cdef['type'] if valtype is None: valtype = lambda x: x # any default? if default is None and 'default' in cdef: default = cdef['default'] _value = None if var in self: # nothing needs to be obtained, it is all here already _value = self[var] elif store is False and default is not None: # nothing will be stored, and we have a default -> no user confirmation # we cannot use logging, because we want to use the config to confiugre # the logging #lgr.debug('using default {} for config setting {}'.format(default, var)) _value = default if _value is not None: # we got everything we need and can exit early try: return valtype(_value) except Exception as e: raise ValueError( "value '{}' of existing configuration for '{}' cannot be " "converted to the desired type '{}' ({})".format( _value, var, valtype, exc_str(e))) # now we need to try to obtain something from the user from datalad.ui import ui # configure UI dialog_opts = kwargs if dialog_type is None: # no override # check for common knowledge on how to obtain a value if 'ui' in cdef: dialog_type = cdef['ui'][0] # pull standard dialog settings dialog_opts = cdef['ui'][1] # update with input dialog_opts.update(kwargs) if (not ui.is_interactive or dialog_type is None) and default is None: raise RuntimeError( "cannot obtain value for configuration item '{}', " "not preconfigured, no default, no UI available".format(var)) if not hasattr(ui, dialog_type): raise ValueError("UI '{}' does not support dialog type '{}'".format( ui, dialog_type)) # configure storage destination, if needed if store: if where is None and 'destination' in cdef: where = cdef['destination'] if where is None: raise ValueError( "request to store configuration item '{}', but no " "storage destination specified".format(var)) # obtain via UI dialog = getattr(ui, dialog_type) _value = dialog(default=default, **dialog_opts) if _value is None: # we got nothing if default is None: raise RuntimeError( "could not obtain value for configuration item '{}', " "not preconfigured, no default".format(var)) # XXX maybe we should return default here, even it was returned # from the UI -- if that is even possible # execute type conversion before storing to check that we got # something that looks like what we want try: value = valtype(_value) except Exception as e: raise ValueError( "cannot convert user input `{}` to desired type ({})".format( _value, exc_str(e))) # XXX we could consider "looping" until we have a value of proper # type in case of a user typo... if store: # store value as it was before any conversion, needs to be str # anyway # needs string conversion nevertheless, because default could come # in as something else self.add(var, '{}'.format(_value), where=where, reload=reload) return value # # Compatibility with dict API # def __len__(self): return len(self._store) def __getitem__(self, key): return self._store.__getitem__(key) def __contains__(self, key): return self._store.__contains__(key) def keys(self): """Returns list of configuration item names""" return self._store.keys() # XXX should this be *args? def get(self, key, default=None): """D.get(k[,d]) -> D[k] if k in D, else d. d defaults to None.""" return self._store.get(key, default) # # Compatibility with ConfigParser API # def sections(self): """Returns a list of the sections available""" return list(set([cfg_section_regex.match(k).group(1) for k in self._store])) def options(self, section): """Returns a list of options available in the specified section.""" opts = [] for k in self._store: sec, opt = cfg_sectionoption_regex.match(k).groups() if sec == section: opts.append(opt) return opts def has_section(self, section): """Indicates whether a section is present in the configuration""" for k in self._store: if k.startswith(section): return True return False def has_option(self, section, option): """If the given section exists, and contains the given option""" for k in self._store: sec, opt = cfg_sectionoption_regex.match(k).groups() if sec == section and opt == option: return True return False def getint(self, section, option): """A convenience method which coerces the option value to an integer""" return int(self.get_value(section, option)) def getbool(self, section, option, default=None): """A convenience method which coerces the option value to a bool Values "on", "yes", "true" and any int!=0 are considered True Values which evaluate to bool False, "off", "no", "false" are considered False TypeError is raised for other values. """ val = self.get_value(section, option, default=default) return anything2bool(val) def getfloat(self, section, option): """A convenience method which coerces the option value to a float""" return float(self.get_value(section, option)) # this is a hybrid of ConfigParser and dict API def items(self, section=None): """Return a list of (name, value) pairs for each option Optionally limited to a given section. """ if section is None: return self._store.items() return [(k, v) for k, v in self._store.items() if cfg_section_regex.match(k).group(1) == section] # # Compatibility with GitPython's ConfigParser # def get_value(self, section, option, default=None): """Like `get()`, but with an optional default value If the default is not None, the given default value will be returned in case the option did not exist. This behavior imitates GitPython's config parser. """ try: return self['.'.join((section, option))] except KeyError as e: # this strange dance is needed because gitpython does it this way if default is not None: return default else: raise e # # Modify configuration (proxy respective git-config call) # @_where_reload def _run(self, args, where=None, reload=False, **kwargs): """Centralized helper to run "git config" calls Parameters ---------- args : list Arguments to pass for git config %s **kwargs Keywords arguments for Runner's call """ if where: args = self._get_location_args(where) + args out = self._runner.run(['git', 'config'] + args, **kwargs) if reload: self.reload() return out def _get_location_args(self, where, args=None): if args is None: args = [] cfg_labels = ('dataset', 'local', 'global') if where not in cfg_labels: raise ValueError( "unknown configuration label '{}' (not in {})".format( where, cfg_labels)) if where == 'dataset': if not self._dataset_cfgfname: raise ValueError( 'ConfigManager cannot store to configuration to dataset, ' 'none specified') # create an empty config file if none exists, `git config` will # fail otherwise dscfg_dirname = opj(self._dataset_path, DATALAD_DOTDIR) if not exists(dscfg_dirname): os.makedirs(dscfg_dirname) if not exists(self._dataset_cfgfname): open(self._dataset_cfgfname, 'w').close() args.extend(['--file', self._dataset_cfgfname]) elif where == 'global': args.append('--global') elif where == 'local': args.append('--local') return args @_where_reload def add(self, var, value, where='dataset', reload=True): """Add a configuration variable and value Parameters ---------- var : str Variable name including any section like `git config` expects them, e.g. 'core.editor' value : str Variable value %s""" self._run(['--add', var, value], where=where, reload=reload, log_stderr=True) @_where_reload def set(self, var, value, where='dataset', reload=True, force=False): """Set a variable to a value. In opposition to `add`, this replaces the value of `var` if there is one already. Parameters ---------- var : str Variable name including any section like `git config` expects them, e.g. 'core.editor' value : str Variable value force: bool if set, replaces all occurrences of `var` by a single one with the given `value`. Otherwise raise if multiple entries for `var` exist already %s""" from datalad.support.gitrepo import to_options self._run(to_options(replace_all=force) + [var, value], where=where, reload=reload, log_stderr=True) @_where_reload def rename_section(self, old, new, where='dataset', reload=True): """Rename a configuration section Parameters ---------- old : str Name of the section to rename. new : str Name of the section to rename to. %s""" self._run(['--rename-section', old, new], where=where, reload=reload) @_where_reload def remove_section(self, sec, where='dataset', reload=True): """Rename a configuration section Parameters ---------- sec : str Name of the section to remove. %s""" self._run(['--remove-section', sec], where=where, reload=reload) @_where_reload def unset(self, var, where='dataset', reload=True): """Remove all occurrences of a variable Parameters ---------- var : str Name of the variable to remove %s""" # use unset all as it is simpler for now self._run(['--unset-all', var], where=where, reload=reload)