Ejemplo n.º 1
0
    def test_expect_false_for_first_criterion_fail_fail_fail(self):
        behaviorChange = Behaviour_check()
        executions_test_base = []
        executions_test_left = []
        executions_test_right = []
        executions_test_merge = []

        #test1 fails on base, left, and merge commits.
        #So no test conflict is expected to be reported by the first criterion
        #test1 was not executed on right commit, so no conflict is expected for the right commit.
        executions_test_base.append(
            JUnitResult(set(), set(), 0, set(), {"test1"},
                        {"RegressionText.test1()"}, 0, set(), set(), set(),
                        set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_left.append(
            JUnitResult(set(), set(), 0, set(), {"test1"},
                        {"RegressionText.test1()"}, 0, set(), set(), set(),
                        set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_right.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), {"test1"}, set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_merge.append(
            JUnitResult(set(), set(), 0, set(), {"test1"},
                        {"RegressionText.test1()"}, 0, set(), set(), set(),
                        set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))

        conflict_for_left_first_criterion = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        conflict_for_right_first_criterion = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_right[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        conflict_for_second_criteria = behaviorChange.check_conflict_occurrence_for_second_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_right[0], executions_test_merge[0], ["", ""],
            "aaa", "bbb", "ccc", "ddd", "AUX")

        self.assertFalse(conflict_for_left_first_criterion[0])
        self.assertFalse(conflict_for_right_first_criterion[0])
        self.assertFalse(conflict_for_second_criteria[0])
Ejemplo n.º 2
0
    def test_expect_true_second_criterion_pass_fail_fail_fail_error(self):
        behaviorChange = Behaviour_check()
        executions_test_base = []
        executions_test_left = []
        executions_test_right = []
        executions_test_merge = []

        #test1 passes on base, left and right commits but fails on merge commit.
        #So a test conflict is expected to be reported by the second criterion
        executions_test_base.append(
            JUnitResult({"test1"}, set(), 1, set(), set(), set(), 0, set(),
                        set(), set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_left.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, {"test1"},
                        {"RegressionText.test1()"}, set(), set(), set(), set(),
                        set(), 0, Coverage(set(), set(), 0, dict()), False))
        executions_test_right.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, {"test1"},
                        {"RegressionText.test1()"}, set(), set(), set(), set(),
                        set(), 0, Coverage(set(), set(), 0, dict()), False))
        executions_test_merge.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, {"test1"},
                        {"RegressionText.test1()"}, set(), set(), set(), set(),
                        set(), 0, Coverage(set(), set(), 0, dict()), False))

        leu = behaviorChange.check_different_test_results_for_commit_pair(
            executions_test_base[0], executions_test_left[0], ["", ""], "aaa",
            "bbb", "AUX")
        #parent_one, parent_two, parent_tree, path_suite, commitBase, commitParentTestSuite, commitMerge, tool
        aux = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        #parent_base, parent_left, parent_right, parent_merge, path_suite, commitBase, commitParentTestSuite, commitParentOther, commitMerge, tool
        aux2 = behaviorChange.check_conflict_occurrence_for_second_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_right[0], executions_test_merge[0], ["", ""],
            "aaa", "bbb", "ccc", "ddd", "AUX")

        self.assertFalse(aux[0])
        self.assertTrue(aux2[0])
Ejemplo n.º 3
0
    def test_expect_false_for_second_criterion_fail_fail_fail_fail(self):
        behaviorChange = Behaviour_check()
        executions_test_base = []
        executions_test_left = []
        executions_test_right = []
        executions_test_merge = []

        #test1 fails on all merge scenario commits.
        #So no test conflict is expected to be reported by the second criterion
        executions_test_base.append(
            JUnitResult(set(), set(), 0, set(), {"test1"},
                        {"RegressionText.test1()"}, 0, set(), set(), set(),
                        set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_left.append(
            JUnitResult(set(), set(), 0, set(), {"test1"},
                        {"RegressionText.test1()"}, 0, set(), set(), set(),
                        set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_right.append(
            JUnitResult(set(), set(), 0, set(), {"test1"},
                        {"RegressionText.test1()"}, 0, set(), set(), set(),
                        set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_merge.append(
            JUnitResult(set(), set(), 0, set(), {"test1"},
                        {"RegressionText.test1()"}, 0, set(), set(), set(),
                        set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))

        aux = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        aux2 = behaviorChange.check_conflict_occurrence_for_second_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_right[0], executions_test_merge[0], ["", ""],
            "aaa", "bbb", "ccc", "ddd", "AUX")

        self.assertFalse(aux[0])
        self.assertFalse(aux2[0])
Ejemplo n.º 4
0
    def test_expect_false_for_first_criterion_not_executed(self):
        behaviorChange = Behaviour_check()
        executions_test_base = []
        executions_test_left = []
        executions_test_right = []
        executions_test_merge = []

        #test1 passes on base and merge commits, but it was not executed on left and right commits.
        #So no test conflict is expected to be reported by the first criterion
        executions_test_base.append(
            JUnitResult({"test1"}, set(), 1, set(), set(), set(), 0, set(),
                        set(), set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_left.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), {"test1"}, set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_right.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), {"test1"}, set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_merge.append(
            JUnitResult({"test1"}, set(), 1, set(), set(), set(), 0, set(),
                        set(), set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))

        conflict_for_left_first_criterion = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        conflict_for_right_first_criterion = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_right[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        conflict_for_second_criteria = behaviorChange.check_conflict_occurrence_for_second_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_right[0], executions_test_merge[0], ["", ""],
            "aaa", "bbb", "ccc", "ddd", "AUX")

        self.assertFalse(conflict_for_left_first_criterion[0])
        self.assertFalse(conflict_for_right_first_criterion[0])
        self.assertFalse(conflict_for_second_criteria[0])
Ejemplo n.º 5
0
    def test_expect_false_second_criterion_not_executed(self):
        behaviorChange = Behaviour_check()
        executions_test_base = []
        executions_test_left = []
        executions_test_right = []
        executions_test_merge = []

        #test1 passes on base, left and right commits, but it was not executed on merge commit.
        #So a test conflict is expected to be reported by the second criterion
        executions_test_base.append(
            JUnitResult({"test1"}, set(), 1, set(), set(), set(), 0, set(),
                        set(), set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_left.append(
            JUnitResult({"test1"}, set(), 1, set(), set(), set(), 0, set(),
                        set(), set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_right.append(
            JUnitResult({"test1"}, set(), 1, set(), set(), set(), 0, set(),
                        set(), set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_merge.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), {"test1"}, set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))

        #parent_one, parent_two, parent_tree, path_suite, commitBase, commitParentTestSuite, commitMerge, tool
        aux = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        #parent_base, parent_left, parent_right, parent_merge, path_suite, commitBase, commitParentTestSuite, commitParentOther, commitMerge, tool
        aux2 = behaviorChange.check_conflict_occurrence_for_second_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_right[0], executions_test_merge[0], ["", ""],
            "aaa", "bbb", "ccc", "ddd", "AUX")

        self.assertFalse(aux[0])
        self.assertFalse(aux2[0])

        #parent_one, parent_two, path_suite, commitOneSHA, commitTwoSHA, tool
        behavior_change = behaviorChange.check_different_test_results_for_commit_pair(
            executions_test_base[0], executions_test_left[0], ["", ""], "aaa",
            "bbb", "Aux")
        behavior_change_merge = behaviorChange.check_different_test_results_for_commit_pair(
            executions_test_merge[0], executions_test_left[0], ["", ""], "aaa",
            "bbb", "Aux")

        self.assertFalse(behavior_change[0])
        self.assertFalse(behavior_change_merge[0])
Ejemplo n.º 6
0
    def test_expect_false_for_both_criteria(self):
        behaviorChange = Behaviour_check()
        executions_test_base = []
        executions_test_left = []
        executions_test_right = []
        executions_test_merge = []

        #empty sets for all merge scenario commits
        executions_test_base.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_left.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_right.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))
        executions_test_merge.append(
            JUnitResult(set(), set(), 0, set(), set(), set(), 0, set(), set(),
                        set(), set(), set(), set(), set(), 0,
                        Coverage(set(), set(), 0, dict()), False))

        aux = behaviorChange.check_conflict_occurrence_for_first_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_merge[0], ["", ""], "aaa", "bbb", "ccc", "AUX")
        aux2 = behaviorChange.check_conflict_occurrence_for_second_criterion(
            executions_test_base[0], executions_test_left[0],
            executions_test_right[0], executions_test_merge[0], ["", ""],
            "aaa", "bbb", "ccc", "ddd", "AUX")

        self.assertFalse(aux[0])
        self.assertFalse(aux2[0])

        behavior_change = behaviorChange.check_different_test_results_for_commit_pair(
            executions_test_base[0], executions_test_left[0], ["", ""], "aaa",
            "bbb", "Aux")
        behavior_change_merge = behaviorChange.check_different_test_results_for_commit_pair(
            executions_test_base[0], executions_test_left[0], ["", ""], "aaa",
            "bbb", "Aux")
        self.assertFalse(behavior_change[0])
        self.assertFalse(behavior_change_merge[0])
Ejemplo n.º 7
0
 def __init__(self):
     self.test_suite = None
     self.tool_parameters = None
     self.behaviour_change = Behaviour_check()
Ejemplo n.º 8
0
class Setup_tool(ABC):

    def __init__(self):
        self.test_suite = None
        self.tool_parameters = None
        self.behaviour_change = Behaviour_check()

    def setup_for_partial_merge_scenario(self, evo, scenario, jarBase, jarParent=None, jarMerge=None):
        try:
            evo.project_dep.baseDir = jarBase
            evo.project_dep.parentReg = jarParent
            evo.project_dep.mergeDir = jarMerge
            evo.project_dep.sut_classes = scenario.merge_scenario.get_sut_classes()
            evo.project_dep.sut_class = scenario.merge_scenario.get_sut_class()
            evo.project_dep.sut_method = scenario.merge_scenario.get_sut_method()
        except:
            print("Some project versions could not be evaluated")

    def setup_for_full_merge_scenario(self, evo, scenario, jarBase, jarParentForGeneration, jarOtherParent, jarMerge):
        try:
            self.setup_for_partial_merge_scenario(evo, scenario, jarBase, jarParentForGeneration, jarMerge)
            evo.project_dep.parentNotReg = jarOtherParent
        except:
            print("Some project versions could not be evaluated")

    def run_tool_for_semantic_conflict_detection(self, evo, scenario, jarBase, jarParentLeft, jarParentRight, jarMerge,
                                                 commitBaseSha, commitParentLeft, commitParentRight, commitMergeSha,
                                                 tool, projectName=None):
        conflict_info = []
        try:
            self.setup_for_full_merge_scenario(evo, scenario, jarBase, jarParentLeft, jarParentRight, jarMerge)
            test_results_left = self.generate_and_run_test_suites(evo, scenario, commitBaseSha, commitParentLeft,
                                    commitParentRight, commitMergeSha, conflict_info, tool)
            if (len(test_results_left[0]) > 0):
                self.check_semantic_conflict_occurrence(test_results_left[0], test_results_left[1], test_results_left[2],
                                    test_results_left[3], test_results_left[4], commitBaseSha, commitParentLeft,
                                    commitParentRight, commitMergeSha, tool, conflict_info)
                self.check_behavior_change_commit_pair(test_results_left[0], test_results_left[1], test_results_left[2],
                                    test_results_left[4], commitBaseSha, commitParentLeft, commitMergeSha, tool, conflict_info)

            self.setup_for_full_merge_scenario(evo, scenario, jarBase, jarParentRight, jarParentLeft, jarMerge)
            test_results_right = self.generate_and_run_test_suites(evo, scenario, commitBaseSha, commitParentRight,
                                    commitParentLeft, commitMergeSha, conflict_info, tool)
            if (len(test_results_right[0]) > 0):
                self.check_semantic_conflict_occurrence(test_results_right[0], test_results_right[1], test_results_right[2],
                                    test_results_right[3], test_results_right[4], commitBaseSha, commitParentRight,
                                    commitParentLeft, commitMergeSha, tool, conflict_info)
                self.check_behavior_change_commit_pair(test_results_right[0], test_results_right[1], test_results_right[2],
                                    test_results_right[4], commitBaseSha, commitParentRight, commitMergeSha, tool, conflict_info)

        except:
            print("Some project versions could not be evaluated")

        return conflict_info

    def run_tool_for_behaviour_change_detection(self, evo, scenario, jarCommitOne, jarCommitTwo, commitOne, commitTwo, tool):
        conflict_info = []
        try:
            self.setup_for_partial_merge_scenario(evo, scenario, jarCommitOne, jarCommitTwo)
            test_results_left = self.generate_and_run_test_suites_for_commit_pair(evo, scenario, commitOne, commitTwo,
                                conflict_info, tool)
            if (len(test_results_left[0]) > 0):
                conflict_info.append(self.behaviour_change.check_different_test_results_for_commit_pair(test_results_left[1],
                                test_results_left[2], test_results_left[0], commitOne, commitTwo, tool))
        except:
            print("Some project versions could not be evaluated")

        return conflict_info

    def generate_and_run_test_suites(self, evo, scenario, commitBaseSha, commitParentTestSuite, commitOtherParent,
                                     commitMergeSha, conflict_info, tool):
        path_suite = []
        test_result_base = set()
        test_result_parent_test_suite = set()
        test_result_other_parent = set()
        test_result_merge = set()
        try:
            path_suite = self.generate_test_suite(scenario, evo.project_dep)

            test_result_base = self.run_test_suite(evo.project_dep.parentReg, evo.project_dep.sut_class,
                                                   evo.project_dep.baseDir, evo.project_dep)
            test_result_parent_test_suite = self.run_test_suite(evo.project_dep.parentReg, evo.project_dep.sut_class,
                                                                evo.project_dep.parentReg, evo.project_dep)
            test_result_other_parent = self.run_test_suite(evo.project_dep.parentReg, evo.project_dep.sut_class,
                                                           evo.project_dep.parentNotReg, evo.project_dep)
            test_result_merge = self.run_test_suite(evo.project_dep.parentReg, evo.project_dep.sut_class,
                                                    evo.project_dep.mergeDir, evo.project_dep)
        except:
            print("Some project versions could not be evaluated")
            conflict_info.append(["NONE", set(), "NO-INFORMATION", commitBaseSha, commitParentTestSuite, commitMergeSha,
                                    tool])

        return path_suite, test_result_base, test_result_parent_test_suite, test_result_other_parent, test_result_merge

    def generate_and_run_test_suites_for_commit_pair(self, evo, scenario, commitOne, commitTestSuite, conflict_info, tool):
        path_suite = []
        test_result_base = set()
        test_result_parent_test_suite = set()

        try:
            path_suite = self.generate_test_suite(scenario, evo.project_dep)

            test_result_base = self.run_test_suite(evo.project_dep.parentReg, evo.project_dep.sut_class,
                                                   evo.project_dep.baseDir, evo.project_dep)
            test_result_parent_test_suite = self.run_test_suite(evo.project_dep.parentReg, evo.project_dep.sut_class,
                                                                evo.project_dep.parentReg, evo.project_dep)
        except:
            print("Some project versions could not be evaluated")
            conflict_info.append(["NONE", set(), "NO-INFORMATION", commitOne, commitTestSuite, "",
                                  tool])

        return path_suite, test_result_base, test_result_parent_test_suite

    def check_semantic_conflict_occurrence(self, path_suite, test_result_base, test_result_parent_test_suite,
                                    test_result_other_parent, test_result_merge, commitBaseSha, commitParentTestSuite,
                                    commitOtherParent, commitMergeSha, tool, conflict_info):

        conflict_info.append(self.behaviour_change.check_conflict_occurrence_for_first_criterion(test_result_base,
                                    test_result_parent_test_suite, test_result_merge, path_suite, commitBaseSha,
                                    commitParentTestSuite, commitMergeSha, tool))
        conflict_info.append(self.behaviour_change.check_conflict_occurrence_for_second_criterion(test_result_base,
                                    test_result_parent_test_suite, test_result_other_parent, test_result_merge, path_suite,
                                    commitBaseSha, commitParentTestSuite, commitOtherParent, commitMergeSha, tool))

    def check_behavior_change_commit_pair(self, path_suite, test_result_base, test_result_parent_test_suite, test_result_merge,
                                    commitBaseSha, commitParentTestSuite, commitMergeSHa, tool, conflict_info):

        conflict_info.append(self.behaviour_change.check_different_test_results_for_commit_pair(test_result_base,
                                    test_result_parent_test_suite, path_suite, commitBaseSha, commitParentTestSuite, tool))
        conflict_info.append(self.behaviour_change.check_different_test_results_for_commit_pair(test_result_merge,
                                    test_result_parent_test_suite, path_suite, commitMergeSHa, commitParentTestSuite, tool))

    def run_test_suite(self, classes_dir, sut_class, mutant_dir, project_dep):
        junit = JUnit(java=project_dep.java, classpath=classes_dir)
        res = junit.run_with_mutant(self.test_suite, sut_class, mutant_dir)
        return res

    @abstractmethod
    def generate_test_suite(self, scenario, project_dep):
        pass

    def generate_and_run_test_suites_for_commit(self, evo, scenario, commitOne, conflict_info, tool):
        path_suite = []
        test_result_base = set()

        try:
            path_suite = self.generate_test_suite(scenario, evo.project_dep)

            test_result_base = self.run_test_suite(evo.project_dep.parentReg, evo.project_dep.sut_class,
                                                   evo.project_dep.baseDir, evo.project_dep)
        except:
            print("Some project versions could not be evaluated")
            conflict_info.append(["NONE", set(), "NO-INFORMATION", commitOne, "",
                                  tool])

        return path_suite, test_result_base

    def run_tool_for_commit(self, evo, scenario, jarCommit, commitSha, tool, projectName=None):
        conflict_info = []
        try:
            self.setup_for_partial_merge_scenario(evo, scenario, jarCommit, jarCommit, jarCommit)
            test_results_left = self.generate_and_run_test_suites_for_commit(evo, scenario, commitSha, conflict_info, tool)
            if (len(test_results_left[0]) > 0):
                conflict_info = test_results_left[0][1]

        except:
            print("Some project versions could not be evaluated")

        return conflict_info