def prepare_process_inputs(process, inputs): """ Prepare the inputs for submission for the given process, according to its spec. That is to say that when an input is found in the inputs that corresponds to an input port in the spec of the process that expects a Dict, yet the value in the inputs is a plain dictionary, the value will be wrapped in by the Dict class to create a valid input. :param process: sub class of Process for which to prepare the inputs dictionary :param inputs: a dictionary of inputs intended for submission of the process :return: a dictionary with all bare dictionaries wrapped in Dict if dictated by process spec """ prepared_inputs = AttributeDict() try: process_spec = process.spec() except AttributeError as exc: raise ValueError('Process {} does not have a spec') from exc for key, value in six.iteritems(inputs): if key not in process_spec.inputs: continue if process_spec.inputs[key].valid_type == Dict and isinstance( value, dict): prepared_inputs[key] = Dict(dict=value) else: prepared_inputs[key] = value return prepared_inputs
def update_mapping(original, source): """ Update a nested dictionary with another optionally nested dictionary. The dictionaries may be plain Mapping objects or Dict nodes. If the original dictionary is an instance of Dict the returned dictionary will also be wrapped in Dict. :param original: Mapping object or Dict instance :param source: Mapping object or Dict instance :return: the original dictionary updated with the source dictionary """ return_parameter_data = False if isinstance(original, Dict): return_parameter_data = True original = original.get_dict() if isinstance(source, Dict): source = source.get_dict() for key, value in six.iteritems(source): if key in original and isinstance(value, (Dict, Mapping)) and isinstance( original[key], (Dict, Mapping)): original[key] = update_mapping(original[key], value) else: original[key] = value if return_parameter_data: original = Dict(dict=original) return original
def get_structure(self, store=False, **kwargs): """ Creates :py:class:`aiida.orm.nodes.data.structure.StructureData`. .. versionadded:: 1.0 Renamed from _get_aiida_structure :param converter: specify the converter. Default 'ase'. :param store: If True, intermediate calculation gets stored in the AiiDA database for record. Default False. :return: :py:class:`aiida.orm.nodes.data.structure.StructureData` node. """ from aiida.orm.nodes.data.dict import Dict from aiida.tools.data.array.trajectory import _get_aiida_structure_inline param = Dict(dict=kwargs) ret_dict = _get_aiida_structure_inline(trajectory=self, parameters=param, metadata={'store_provenance': store}) # pylint: disable=unexpected-keyword-arg return ret_dict['structure']
def create_input_nodes(self, open_transport, input_file_name=None, output_file_name=None, remote_workdir=None): """Create calculation input nodes based on the job's files. :param open_transport: An open instance of the transport class of the calculation's computer. See the tutorial for more information. :type open_transport: aiida.transport.plugins.local.LocalTransport or aiida.transport.plugins.ssh.SshTransport This method parses the files in the job's remote working directory to create the input nodes that would exist if the calculation were submitted using AiiDa. These nodes are * a ``'parameters'`` Dict node, based on the namelists and their variable-value pairs; * a ``'kpoints'`` KpointsData node, based on the *K_POINTS* card; * a ``'structure'`` StructureData node, based on the *ATOMIC_POSITIONS* and *CELL_PARAMETERS* cards; * one ``'pseudo_X'`` UpfData node for the pseudopotential used for the atomic species with name ``X``, as specified in the *ATOMIC_SPECIES* card; * a ``'settings'`` Dict node, if there are any fixed coordinates, or if the gamma kpoint is used; and can be retrieved as a dictionary using the ``get_incoming()`` method. *These input links are cached-links; nothing is stored by this method (including the calculation node itself).* .. note:: QE stores the calculation's pseudopotential files in the ``<outdir>/<prefix>.save/`` subfolder of the job's working directory, where ``outdir`` and ``prefix`` are QE *CONTROL* variables (see `pw input file description <http://www.quantum-espresso.org/wp-content/uploads/Doc/INPUT_PW.html>`_). This method uses these files to either get--if the a node already exists for the pseudo--or create a UpfData node for each pseudopotential. **Keyword arguments** .. note:: These keyword arguments can also be set when instantiating the class or using the ``set_`` methods (e.g. ``set_remote_workdir``). Offering to set them here simply offers the user an additional place to set their values. *Only the values that have not yet been set need to be specified.* :param input_file_name: The file name of the job's input file. :type input_file_name: str :param output_file_name: The file name of the job's output file (i.e. the file containing the stdout of QE). :type output_file_name: str :param remote_workdir: Absolute path to the directory where the job was run. The transport of the computer you link ask input to the calculation is the transport that will be used to retrieve the calculation's files. Therefore, ``remote_workdir`` should be the absolute path to the job's directory on that computer. :type remote_workdir: str :raises aiida.common.exceptions.InputValidationError: if ``open_transport`` is a different type of transport than the computer's. :raises aiida.common.exceptions.InvalidOperation: if ``open_transport`` is not open. :raises aiida.common.exceptions.InputValidationError: if ``remote_workdir``, ``input_file_name``, and/or ``output_file_name`` are not set prior to or during the call of this method. :raises aiida.common.exceptions.FeatureNotAvailable: if the input file uses anything other than ``ibrav = 0``, which is not currently implimented in aiida. :raises aiida.common.exceptions.ParsingError: if there are issues parsing the input file. :raises IOError: if there are issues reading the input file. """ import re # Make sure the remote workdir and input + output file names were # provided either before or during the call to this method. If they # were just provided during this method call, store the values. if remote_workdir is not None: self.set_remote_workdir(remote_workdir) elif self.get_attr('remote_workdir', None) is None: raise InputValidationError( 'The remote working directory has not been specified.\n' 'Please specify it using one of the following...\n ' '(a) pass as a keyword argument to create_input_nodes\n' ' [create_input_nodes(remote_workdir=your_remote_workdir)]\n' '(b) pass as a keyword argument when instantiating\n ' ' [calc = PwCalculationImport(remote_workdir=' 'your_remote_workdir)]\n' '(c) use the set_remote_workdir method\n' ' [calc.set_remote_workdir(your_remote_workdir)]') if input_file_name is not None: self._INPUT_FILE_NAME = input_file_name elif self._INPUT_FILE_NAME is None: raise InputValidationError( 'The input file_name has not been specified.\n' 'Please specify it using one of the following...\n ' '(a) pass as a keyword argument to create_input_nodes\n' ' [create_input_nodes(input_file_name=your_file_name)]\n' '(b) pass as a keyword argument when instantiating\n ' ' [calc = PwCalculationImport(input_file_name=' 'your_file_name)]\n' '(c) use the set_input_file_name method\n' ' [calc.set_input_file_name(your_file_name)]') if output_file_name is not None: self._OUTPUT_FILE_NAME = output_file_name elif self._OUTPUT_FILE_NAME is None: raise InputValidationError( 'The input file_name has not been specified.\n' 'Please specify it using one of the following...\n ' '(a) pass as a keyword argument to create_input_nodes\n' ' [create_input_nodes(output_file_name=your_file_name)]\n' '(b) pass as a keyword argument when instantiating\n ' ' [calc = PwCalculationImport(output_file_name=' 'your_file_name)]\n' '(c) use the set_output_file_name method\n' ' [calc.set_output_file_name(your_file_name)]') # Check that open_transport is the correct transport type. if type(open_transport) is not self.get_computer().get_transport_class( ): raise InputValidationError( 'The transport passed as the `open_transport` parameter is ' 'not the same transport type linked to the computer. Please ' 'obtain the correct transport class using the ' "`get_transport_class` method of the calculation's computer. " 'See the tutorial for more information.') # Check that open_transport is actually open. if not open_transport._is_open: raise InvalidOperation( 'The transport passed as the `open_transport` parameter is ' "not open. Please execute the open the transport using it's " '`open` method, or execute the call to this method within a ' '`with` statement context guard. See the tutorial for more ' 'information.') # Copy the input file and psuedo files to a temp folder for parsing. with SandboxFolder() as folder: # Copy the input file to the temp folder. remote_path = os.path.join(self._get_remote_workdir(), self._INPUT_FILE_NAME) open_transport.get(remote_path, folder.abspath) # Parse the input file. local_path = os.path.join(folder.abspath, self._INPUT_FILE_NAME) with open(local_path) as fin: pwinputfile = pwinputparser.PwInputFile(fin) # Determine PREFIX, if it hasn't already been set by the user. if self._PREFIX is None: control_dict = pwinputfile.namelists['CONTROL'] # If prefix is not set in input file, use the default, # 'pwscf'. self._PREFIX = control_dict.get('prefix', 'pwscf') # Determine _OUTPUT_SUBFOLDER, if it hasn't already been set by # the user. # TODO: Prompt user before using the environment variable??? if self._OUTPUT_SUBFOLDER is None: # See if it's specified in the CONTROL namelist. control_dict = pwinputfile.namelists['CONTROL'] self._OUTPUT_SUBFOLDER = control_dict.get('outdir', None) if self._OUTPUT_SUBFOLDER is None: # See if the $ESPRESSO_TMPDIR is set. envar = open_transport.exec_command_wait( 'echo $ESPRESSO_TMPDIR')[1] if len(envar.strip()) > 0: self._OUTPUT_SUBFOLDER = envar.strip() else: # Use the default dir--the dir job was submitted in. self._OUTPUT_SUBFOLDER = self._get_remote_workdir() # Copy the pseudo files to the temp folder. for fnm in pwinputfile.atomic_species['pseudo_file_names']: remote_path = os.path.join(self._get_remote_workdir(), self._OUTPUT_SUBFOLDER, '{}.save/'.format(self._PREFIX), fnm) open_transport.get(remote_path, folder.abspath) # Make sure that ibrav = 0, since aiida doesn't support anything # else. if pwinputfile.namelists['SYSTEM']['ibrav'] != 0: raise FeatureNotAvailable( 'Found ibrav !=0 while parsing the input file. ' 'Currently, AiiDa only supports ibrav = 0.') # Create Dict node based on the namelist and link as input. # First, strip the namelist items that aiida doesn't allow or sets # later. # NOTE: ibrav = 0 is checked above. # NOTE: If any of the position or cell units are in alat or crystal # units, that will be taken care of by the input parsing tools, and # we are safe to fake that they were never there in the first place. parameters_dict = deepcopy(pwinputfile.namelists) for namelist, blocked_key in self._blocked_keywords: keys = list(parameters_dict[namelist].keys()) for this_key in parameters_dict[namelist].keys(): # take into account that celldm and celldm(*) must be blocked if re.sub('[(0-9)]', '', this_key) == blocked_key: parameters_dict[namelist].pop(this_key, None) parameters = Dict(dict=parameters_dict) self.use_parameters(parameters) # Initialize the dictionary for settings parameter data for possible # use later for gamma kpoint and fixed coordinates. settings_dict = {} # Create a KpointsData node based on the K_POINTS card block # and link as input. kpointsdata = pwinputfile.get_kpointsdata() self.use_kpoints(kpointsdata) # If only the gamma kpoint is used, add to the settings dictionary. if pwinputfile.k_points['type'] == 'gamma': settings_dict['gamma_only'] = True # Create a StructureData node based on the ATOMIC_POSITIONS, # CELL_PARAMETERS, and ATOMIC_SPECIES card blocks, and link as # input. structuredata = pwinputfile.get_structuredata() self.use_structure(structuredata) # Get or create a UpfData node for the pseudopotentials used for # the calculation. names = pwinputfile.atomic_species['names'] pseudo_file_names = pwinputfile.atomic_species['pseudo_file_names'] for name, fnm in zip(names, pseudo_file_names): local_path = os.path.join(folder.abspath, fnm) pseudo, created = UpfData.get_or_create(local_path) self.use_pseudo(pseudo, kind=name) # If there are any fixed coordinates (i.e. force modification # present in the input file, create a Dict node for these # special settings. fixed_coords = pwinputfile.atomic_positions['fixed_coords'] # NOTE: any() only works for 1-dimensional lists. if any((any(fc_xyz) for fc_xyz in fixed_coords)): settings_dict['FIXED_COORDS'] = fixed_coords # If the settings_dict has been filled in, create a Dict # node from it and link as input. if settings_dict: self.use_settings(Dict(dict=settings_dict)) self.set_attribute('input_nodes_created', True)
def _parse_file(self, inputs): from aiida.orm.nodes.data.dict import Dict result = {} for quantity in ExampleFileParser.PARSABLE_ITEMS: result[quantity] = Dict(dict={}) return result