示例#1
0
 def __init__(self, *args, **kwargs):
     super(Job, self).__init__(*args, **kwargs)
     self.input = kwargs.setdefault('inputfiles', {})
     self.output = kwargs.setdefault('outputfiles', {})
     self.filenames = joindicts(self.input, self.output)
     self.variables = kwargs.setdefault('variables', {})
     kwargs.setdefault('startdir', None)
     kwargs.setdefault('stagedir', None)
     self.queuingsystem = 'Local'
示例#2
0
 def __init__(self,*args,**kwargs):
     super(Job,self).__init__(*args,**kwargs)
     self.input = kwargs.setdefault('inputfiles',{})
     self.output = kwargs.setdefault('outputfiles',{})
     self.filenames = joindicts(self.input,self.output)
     self.variables = kwargs.setdefault('variables',{})
     kwargs.setdefault('startdir',None)
     kwargs.setdefault('stagedir',None)
     self.queuingsystem = 'Local'
示例#3
0
    def __init__(self, *args, **kwargs):
        """Set up SGE job.

        :Arguments:

           inputfiles
                            dict of input files (with relative path to startdir);
                            globs are not supported.
           outputfiles
                            dict of result files or glob patterns (relative to
                            stagedir == relative to startdir)
           variables
                            key/value pairs that can be used in the script as 
                            Job.variables[key]
           startdir
                            path to the directory where the input can be found
                            (must be nfs-mounted on node)
           stagedir
                            local scratch directory on node; all input files are copied
                            there. The default should be ok.

           JOB_NAME
                            unique identifier (only set this if this NOT submitted through
                            the Gridengine queuing system AND if the files should be copied
                            to a scratch disk (i.e. staging proceeds as it would for a
                            SGE-submitted job).)
           SGE_TASK_ID
                           fake a task id (use with JOB_NAME)

        """
        self.__MODE__ = "init"  # current state, for self.msg
        super(Job, self).__init__(*args, **kwargs)
        self.input = kwargs.setdefault('inputfiles', {})
        self.output = kwargs.setdefault('outputfiles', {})

        self.variables = kwargs.setdefault('variables', {})
        # where we find input files and copy back results
        self.startdir = self.startdir_name(kwargs.setdefault('startdir', None))
        # local directory on node
        self.stagedir = self.stagedir_name(kwargs.setdefault('stagedir', None))
        # normalized filenames (always under stagedir)
        self.filenames = {
            k: pathjoin(self.stagedir, path, refdir=self.startdir)
            for k, path in joindicts(self.input, self.output).items()
        }

        self.statusmessage()
示例#4
0
    def __init__(self,*args,**kwargs):
        """Set up SGE job.

        :Arguments:

           inputfiles
                            dict of input files (with relative path to startdir);
                            globs are not supported.
           outputfiles
                            dict of result files or glob patterns (relative to
                            stagedir == relative to startdir)
           variables
                            key/value pairs that can be used in the script as 
                            Job.variables[key]
           startdir
                            path to the directory where the input can be found
                            (must be nfs-mounted on node)
           stagedir
                            local scratch directory on node; all input files are copied
                            there. The default should be ok.

           JOB_NAME
                            unique identifier (only set this if this NOT submitted through
                            the Gridengine queuing system AND if the files should be copied
                            to a scratch disk (i.e. staging proceeds as it would for a
                            SGE-submitted job).)
           SGE_TASK_ID
                           fake a task id (use with JOB_NAME)

        """
        self.__MODE__ = "init"   # current state, for self.msg
        super(Job,self).__init__(*args,**kwargs)
        self.input = kwargs.setdefault('inputfiles',{})
        self.output = kwargs.setdefault('outputfiles',{})
        
        self.variables = kwargs.setdefault('variables',{})
        # where we find input files and copy back results
        self.startdir = self.startdir_name(kwargs.setdefault('startdir',None))
        # local directory on node
        self.stagedir = self.stagedir_name(kwargs.setdefault('stagedir',None))
        # normalized filenames (always under stagedir)
        self.filenames = dict([
            (k,pathjoin(self.stagedir,path,refdir=self.startdir))
             for k,path in joindicts(self.input,self.output).items()] )
        
        self.statusmessage()
示例#5
0
    def __init__(self, *args, **kwargs):
        """Set up the Job:

        job = Job(inputfiles=dict(...),outputfiles=dict(...),variables=dict(...),**kwargs)

        inputfiles and outputfiles are dictionaries with arbitrary
        keys; each item is a path to a file relative to the startdir
        (which by default is the directory from which the SGE job
        starts --- use the #$ -cwd flag!). If the files are not
        relative to the start dir then new directories are constructed
        under the stage dir; in this instance it uis important that
        the user script ONLY uses the filenames in self.filenames:
        These have the proper paths of the local (staged) files for
        the script to operate on.

        With

          job.stage()
        
        inputfiles are copied to the stagedir on the node's scratch
        dir and sub directories are created as necessary; directories
        mentioned as part of the outputfiles are created, too.

          job.unstage()

        copies back all files mentioned in output files (again, use
        directories as part of the path as necessary) and create the
        directories in the startdir if needed. For the outputfiles one
        can also use shell-style glob patterns, e.g. outfiles =
        {'all_dcd': '*.dcd', 'last_data':'*[5-9].dat'}

        Sensible defaults are automatically selected for startdir
        (cwd) and stagedir (/scratch/USER/JOB_NAME.JOB_ID).
        
        If the script is not run through SGE (i.e. the environment
        variable JOB_NAME is not set) then the script is run without
        staging; this is pretty much equivalent to using

          from staging.Local import Job

        :Input:

        inputfiles       dict of input files (with relative path to startdir);
                         globs are not supported.
        outputfiles      dict of result files or glob patterns (relative to
                         stagedir == relative to startdir)
        variables        key/value pairs that can be used in the script as 
                         Job.variables[key]
        startdir         path to the directory where the input can be found
                         (must be nfs-mounted on node)
        stagedir         local scratch directory on node; all input files are copied
                         there. The default should be ok.
                         
        JOB_NAME         unique identifier (only set this if this NOT submitted through
                         the Gridengine queuing system AND if the files should be copied
                         to a scratch disk (i.e. staging proceeds as it would for a
                         SGE-submitted job).)
        SGE_TASK_ID      fake a task id (use with JOB_NAME)

        :Attributes:
      
        input            inputfiles dict  (relative to startdir or absolute)
        output           outputfiles dict (relative to startdir or absolute, can contain globs)
        filenames        merged dict of input and output, pointing to *staged* files
        variables        variables dict

        :Methods:
        
        stage()          setup job on the nodes in stagedir
        unstage()        retrieve results to startdir
        cleanup()        remove all files on the node (rm -rf stagedir)
        """
        self.__MODE__ = "init"  # current state, for self.msg
        super(Job, self).__init__(*args, **kwargs)
        self.input = kwargs.setdefault('inputfiles', {})
        self.output = kwargs.setdefault('outputfiles', {})

        self.variables = kwargs.setdefault('variables', {})
        # where we find input files and copy back results
        self.startdir = self.startdir_name(kwargs.setdefault('startdir', None))
        # local directory on node
        self.stagedir = self.stagedir_name(kwargs.setdefault('stagedir', None))
        # normalized filenames (always under stagedir)
        self.filenames = dict([
            (k, pathjoin(self.stagedir, path, refdir=self.startdir))
            for k, path in joindicts(self.input, self.output).items()
        ])

        self.statusmessage()