def __call__(self, pnode): if pnode.type != "CMSSW": return for dataset in pnode._OutputDatasets: primary = DBSWriterObjects.createPrimaryDataset(dataset, self.apiRef) mergeAlgo = DBSWriterObjects.createMergeAlgorithm(dataset, self.apiRef) DBSWriterObjects.createProcessedDataset(primary, mergeAlgo, dataset, self.apiRef) inputDataset = dataset.get("ParentDataset", None) if inputDataset == None: continue processedDataset = dataset["ProcessedDataset"] self.apiRef.insertMergedDataset(inputDataset, processedDataset, mergeAlgo) # algorithm used when process jobs produce merged files directly # doesnt contain pset content - taken from processing (same hash) mergeDirectAlgo = DBSWriterObjects.createAlgorithm(dataset, None, self.apiRef) self.apiRef.insertAlgoInPD(makeDSName2(dataset), mergeDirectAlgo) logging.debug("ProcessedDataset: %s" % processedDataset) logging.debug("inputDataset: %s" % inputDataset) logging.debug("mergeAlgo: %s" % mergeAlgo) return
def __call__(self, pnode): if pnode.type != "CMSSW": return datasets = self.getOutputDatasetsWithPSet(pnode) cfgMeta = None try: cfgInt = pnode.cfgInterface cfgMeta = cfgInt.configMetadata cfgMeta['Type'] = self.workflow.parameters["RequestCategory"] except Exception as ex: msg = "Unable to Extract cfg data from workflow" msg += str(ex) logging.error(msg) return for dataset in datasets: primary = DBSWriterObjects.createPrimaryDataset( dataset, self.apiRef) algo = DBSWriterObjects.createAlgorithm(dataset, cfgMeta, self.apiRef) processed = DBSWriterObjects.createProcessedDataset( primary, algo, dataset, self.apiRef) return
def __call__(self, pnode): if pnode.type != "CMSSW": return for dataset in pnode._OutputDatasets: primary = DBSWriterObjects.createPrimaryDataset( dataset, self.apiRef) mergeAlgo = DBSWriterObjects.createMergeAlgorithm( dataset, self.apiRef) DBSWriterObjects.createProcessedDataset(primary, mergeAlgo, dataset, self.apiRef) inputDataset = dataset.get('ParentDataset', None) if inputDataset == None: continue processedDataset = dataset["ProcessedDataset"] self.apiRef.insertMergedDataset(inputDataset, processedDataset, mergeAlgo) # algorithm used when process jobs produce merged files directly # doesnt contain pset content - taken from processing (same hash) mergeDirectAlgo = DBSWriterObjects.createAlgorithm( dataset, None, self.apiRef) self.apiRef.insertAlgoInPD(makeDSName2(dataset), mergeDirectAlgo) logging.debug("ProcessedDataset: %s" % processedDataset) logging.debug("inputDataset: %s" % inputDataset) logging.debug("mergeAlgo: %s" % mergeAlgo) return
def __call__(self, pnode): if pnode.type != "CMSSW": return datasets = self.getOutputDatasetsWithPSet(pnode) cfgMeta = None try: cfgInt = pnode.cfgInterface cfgMeta = cfgInt.configMetadata cfgMeta['Type'] = self.workflow.parameters["RequestCategory"] except Exception as ex: msg = "Unable to Extract cfg data from workflow" msg += str(ex) logging.error(msg) return for dataset in datasets: primary = DBSWriterObjects.createPrimaryDataset( dataset, self.apiRef) algo = DBSWriterObjects.createAlgorithm( dataset, cfgMeta, self.apiRef) processed = DBSWriterObjects.createProcessedDataset( primary, algo, dataset, self.apiRef) return