def resolve_consistency(self, data, seed, numberProcesses): print "Number of splits - ", numberProcesses splits = split_df(data.data, houseidCol=self.hidName, workers=numberProcesses) args = [ (split, seed, self.activityAttribs, self.dailyStatusAttribs, self.dependencyAttribs) for split in splits ] resultList = [] resultList += resolve_by_multiprocessing(func=clean_schedules, args=args, workers=numberProcesses) return DataArray(resultList, self.colNames) """
def resolve_consistency(self, data, seed, numberProcesses): #TODO: add the number of workers/processes as an input in the config file splits = split_df(data.data, houseidCol=self.hidName, workers=numberProcesses) args = [(split,seed, self.activityAttribs) for split in splits] resultList = [] resultList += resolve_by_multiprocessing(func=reconcile_schedules, args=args, workers=numberProcesses) return DataArray(resultList, self.colNames)
def resolve_consistency(self, data, seed, numberProcesses): print "Number of splits - ", numberProcesses splits = split_df(data.data, houseidCol=self.hidName, workers=numberProcesses) args = [(split, seed, self.activityAttribs, self.dailyStatusAttribs, self.dependencyAttribs) for split in splits] resultList = [] resultList += resolve_by_multiprocessing( func=clean_aggregate_schedules, args=args, workers=numberProcesses) return DataArray(resultList, self.colNames) """