# the walltime for the job. Can be set individually or for the multijob (which would have to be done later) job.setWalltimeInSeconds(60) # adding the job to the multijob multiPartJob.addJob(job) multiPartJob.setDefaultNoCpus(1); multiPartJob.setDefaultWalltimeInSeconds(60); # now we add an input file that is common to all jobs #multiPartJob.addInputFile("/home/markus/sample input files/JavaTestJob.jar"); # we don't want to submit to tpac because it doesn't work #multiPartJob.setSitesToExclude(["uq", "hpsc", "auckland", "canterbury"]); try: print "Creating jobs on the backend and staging files..." multiPartJob.prepareAndCreateJobs(True) except (JobsException), error: for job in error.getFailures().keySet(): print "Job: "+job.getJobname()+", Error: "+error.getFailures().get(job).getLocalizedMessage() sys.exit() print "Job distribution:" for subLoc in multiPartJob.getOptimizationResult().keySet(): print subLoc + " : " +multiPartJob.getOptimizationResult().get(subLoc) print "Submitting jobs..." multiPartJob.submit() restarted = False
batchJob.addJob(job) # this should be set because it's used for the matchmaking/metascheduling batchJob.setDefaultNoCpus(1); batchJob.setDefaultWalltimeInSeconds(60); # now we add an input file that is common to all jobs batchJob.addInputFile(inputFile1Url); batchJob.addInputFile(inputFile2Url); batchJob.addInputFile(inputFile3Url); # we don't want to submit to tpac because it doesn't work #multiPartJob.setSitesToExclude(["uq", "hpsc", "auckland", "canterbury"]); try: print "Creating jobs on the backend and staging files..." batchJob.prepareAndCreateJobs(True) except (JobsException), error: for job in error.getFailures().keySet(): print "Job: "+job.getJobname()+", Error: "+error.getFailures().get(job).getLocalizedMessage() sys.exit() # this is not really needed print "Job distribution:" for subLoc in batchJob.getOptimizationResult().keySet(): print subLoc + " : " +batchJob.getOptimizationResult().get(subLoc) print "Submitting jobs..." batchJob.submit(True)