def resolve_consistency(self, data, seed, numberProcesses):
        print "Number of splits - ", numberProcesses
        splits = split_df(data.data, houseidCol=self.hidName, workers=numberProcesses)
        args = [
            (split, seed, self.activityAttribs, self.dailyStatusAttribs, self.dependencyAttribs) for split in splits
        ]

        resultList = []
        resultList += resolve_by_multiprocessing(func=clean_schedules, args=args, workers=numberProcesses)

        return DataArray(resultList, self.colNames)
        """
    def resolve_consistency(self, data, seed, numberProcesses):
        #TODO: add the number of workers/processes as an input in the config file

        splits = split_df(data.data, houseidCol=self.hidName, 
                                   workers=numberProcesses)
        args = [(split,seed,  self.activityAttribs) for split in splits]        
        
        resultList = []
        resultList += resolve_by_multiprocessing(func=reconcile_schedules, 
                                                                        args=args,  
                                                                        workers=numberProcesses)

        return DataArray(resultList,  self.colNames)
    def resolve_consistency(self, data, seed, numberProcesses):
        print "Number of splits - ", numberProcesses
        splits = split_df(data.data,
                          houseidCol=self.hidName,
                          workers=numberProcesses)
        args = [(split, seed, self.activityAttribs, self.dailyStatusAttribs,
                 self.dependencyAttribs) for split in splits]

        resultList = []
        resultList += resolve_by_multiprocessing(
            func=clean_aggregate_schedules, args=args, workers=numberProcesses)

        return DataArray(resultList, self.colNames)
        """