예제 #1
0
    def test_add_to_tree_when_adding_tree_root(self):
        dt = DecisionTree(tree={'a': {'b': {'c': dict()}}})
        dt.add_to_tree([('a', 'b'), ('c', 'd')], 'e', True)

        self.assertEqual(dt.get_tree(),
                         {'a': {
                             'b': {
                                 'c': {
                                     'd': {
                                         'e': dict()
                                     }
                                 }
                             }
                         }})
예제 #2
0
    def test_add_to_tree_when_path_is_empty(self):
        dt = DecisionTree()
        dt.add_to_tree([], 'a', True)

        self.assertEqual(dt.get_tree(), {'a': dict()})
예제 #3
0
파일: id3.py 프로젝트: careywadawi/ANN-AI
class ID3:
    def __init__(self, in_attr_list, out_attr):
        self.in_attr_list = in_attr_list
        self.out_attr = out_attr

        self.tree = DecisionTree()

    def generate_decision_tree(self, data):
        self.id3(self.in_attr_list, data)
        return self.tree.get_tree()

    def id3(self,
            in_attr_list: list,
            data: pandas.DataFrame,
            grouping_value="",
            ancestors=[]):
        if data.empty:
            self.tree.add_to_tree(ancestors, 'Failed')
            return

        if data[self.out_attr].unique().size == 1:
            node = data[self.out_attr].unique()[0]
            self.tree.add_to_tree(ancestors, node)
            return

        if not in_attr_list:
            node = data[self.out_attr].value_counts().idxmax()
            self.tree.add_to_tree(ancestors, node)
            return

        largest_ig_attr = self.__largest_ig_attr(in_attr_list, self.out_attr,
                                                 data)
        self.tree.add_to_tree(ancestors, largest_ig_attr, True)

        temp_in_attr_list = in_attr_list.copy()
        temp_in_attr_list.remove(largest_ig_attr)
        for grouping_value, group in data.groupby([largest_ig_attr]):
            self.id3(in_attr_list=temp_in_attr_list,
                     data=group.drop([largest_ig_attr], axis=1),
                     grouping_value=grouping_value,
                     ancestors=ancestors + [(largest_ig_attr, grouping_value)])

        return

    @classmethod
    def __largest_ig_attr(cls, in_attr_list, out_attr, data):
        highest_ig_attr = in_attr_list[0]
        highest_ig = cls.information_gain(highest_ig_attr, out_attr, data)

        for attr in in_attr_list[1:]:
            cur_ig = cls.information_gain(attr, out_attr, data)
            if highest_ig < cur_ig:
                highest_ig_attr = attr
                highest_ig = cur_ig

        return highest_ig_attr

    @classmethod
    def information_gain(cls, in_attr, out_attr, data):
        return cls.entropy(data[out_attr]) - cls.in_attr_entropy(
            in_attr, out_attr, data)

    @staticmethod
    def entropy(data):
        row_count = data.count()
        entropy = 0

        for value, value_count in data.value_counts().items():
            probability = value_count / row_count
            entropy -= probability * log2(probability)

        return entropy

    @staticmethod
    def in_attr_entropy(in_attr, out_attr, data):
        row_count = len(data.index)
        entropy = 0

        for group_name, group in data.groupby([in_attr]):
            group_size = len(group.index)
            entropy_acc = 0

            for subgroup_name, subgroup in group.groupby([out_attr]):
                probability = len(subgroup.index) / group_size
                entropy_acc -= probability * log2(probability)

            entropy_acc *= group_size / row_count
            entropy += entropy_acc

        return entropy

    def classify(self, data):
        if not self.tree.get_tree():
            raise Exception('Decision tree not generated.')

        result = []
        for index, row in data.iterrows():
            result.append(self.classify_row(row, self.tree.get_tree()))

        return result

    def classify_row(self, row, tree):
        attr = next(iter(tree))
        row_val = row[attr]
        decision = tree[attr][row_val]

        if isinstance(decision, dict):
            return self.classify_row(row, tree[attr][row_val])
        else:
            return decision
예제 #4
0
    def test_add_to_tree_when_adding_decision_node(self):
        dt = DecisionTree(tree={'a': {'b': {'c': dict()}}})
        dt.add_to_tree([('a', 'b'), ('c', 'd')], 'e')

        self.assertEqual(dt.get_tree(), {'a': {'b': {'c': {'d': 'e'}}}})
예제 #5
0
파일: id3.py 프로젝트: johnGachihi/ID3
class ID3:
    def __init__(self, in_attr_list, out_attr):
        self.in_attr_list = in_attr_list
        self.out_attr = out_attr

        self.tree = DecisionTree()

    def generate_decision_tree(self, data):
        self.id3(self.in_attr_list, data)
        return self.tree.get_tree()

    def id3(self,
            in_attr_list: list,
            data: pandas.DataFrame,
            grouping_value="",
            ancestors=[]):
        """
        :param in_attr_list: A list of input attributes
        :param data: A pandas DataFrame containing all the data
        :param grouping_value: The value used to group the data
        :type grouping_value: str
        :param ancestors: A list of tuples in the form (node, edge)
        tracing a path from the decision tree root node to the last
         inserted node and edge
        """

        if data.empty:
            self.tree.add_to_tree(ancestors, 'Failed')
            return

        if data[self.out_attr].unique().size == 1:
            node = data[self.out_attr].unique()[0]
            self.tree.add_to_tree(ancestors, node)
            return

        if not in_attr_list:
            node = data[self.out_attr].value_counts().idxmax()
            self.tree.add_to_tree(ancestors, node)
            return

        largest_ig_attr = self.__largest_ig_attr(in_attr_list, self.out_attr,
                                                 data)
        self.tree.add_to_tree(ancestors, largest_ig_attr, True)

        temp_in_attr_list = in_attr_list.copy()
        temp_in_attr_list.remove(largest_ig_attr)
        for grouping_value, group in data.groupby([largest_ig_attr]):
            self.id3(in_attr_list=temp_in_attr_list,
                     data=group.drop([largest_ig_attr], axis=1),
                     grouping_value=grouping_value,
                     ancestors=ancestors + [(largest_ig_attr, grouping_value)])

        return

    @classmethod
    def __largest_ig_attr(cls, in_attr_list, out_attr, data):
        """
        Gets the attribute, from among @in_attr_list, that has
        the greatest Information Gain.

        :param in_attr_list: List of input attributes
        :param out_attr: The output attribute
        :param data: A pandas DataFrame with columns for all input attributes
         attributes and the output attribute
        :return:
        """
        highest_ig_attr = in_attr_list[0]
        highest_ig = cls.information_gain(highest_ig_attr, out_attr, data)

        for attr in in_attr_list[1:]:
            cur_ig = cls.information_gain(attr, out_attr, data)
            if highest_ig < cur_ig:
                highest_ig_attr = attr
                highest_ig = cur_ig

        return highest_ig_attr

    @classmethod
    def information_gain(cls, in_attr, out_attr, data):
        return cls.entropy(data[out_attr]) - cls.in_attr_entropy(
            in_attr, out_attr, data)

    @staticmethod
    def entropy(data):
        """
        Calculate entropy. (Not relative entropy!). Takes in
        a Series object with a single column.

        :param data: Single row column data
        :type data: pandas.core.series.Series
        :return:
        """
        row_count = data.count()
        entropy = 0

        for value, value_count in data.value_counts().items():
            probability = value_count / row_count
            entropy -= probability * log2(probability)

        return entropy

    @staticmethod
    def in_attr_entropy(in_attr, out_attr, data):
        """
        Calculates entropy of @in_attr in relation to @out_attr

        :param in_attr: Name of input attribute
        :type in_attr: str
        :param out_attr: Name of output attribute
        :type out_attr: str
        :param data: DataFrame containing columns for the input and output attributes
        :type data: pandas.core.frame.DataFrame
        """

        row_count = len(data.index)
        entropy = 0

        for group_name, group in data.groupby([in_attr]):
            group_size = len(group.index)
            entropy_acc = 0

            for subgroup_name, subgroup in group.groupby([out_attr]):
                probability = len(subgroup.index) / group_size
                entropy_acc -= probability * log2(probability)

            entropy_acc *= group_size / row_count
            entropy += entropy_acc

        return entropy

    def classify(self, data):
        """
        Uses generated decision tree to classify data.
        This method should be called after the 'generate_decision_tree'
        method has been called. If called before an exception will be thrown.
        The reason is that no decision tree will have been generated

        :param data: A pandas DataFrame containing all the input attributes
         columns and the output attribute column that were used to create the
         ID3 instance
        :return: Returns a list of the categories of all rows in the data
         as classified by the decision tree and in the order of the rows.
        """
        if not self.tree.get_tree():
            raise Exception('Decision tree not generated.')

        result = []
        for index, row in data.iterrows():
            result.append(self.classify_row(row, self.tree.get_tree()))

        return result

    def classify_row(self, row, tree):
        attr = next(iter(tree))
        row_val = row[attr]
        decision = tree[attr][row_val]

        if isinstance(decision, dict):
            return self.classify_row(row, tree[attr][row_val])
        else:
            return decision