summaryrefslogtreecommitdiff
path: root/assignment1.py
blob: 64b865d8109195d384b3bb810414adde46afa9a5 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
import time
import numpy as np

credit_data = np.genfromtxt('./credit_score.txt',
                            delimiter=',',
                            skip_header=True)

# age,married,house,income,gender,class
# [(22, 0, 0, 28, 1, 0)
#  (46, 0, 1, 32, 0, 0)
#  (24, 1, 1, 24, 1, 0)
#  (25, 0, 0, 27, 1, 0)
#  (29, 1, 1, 32, 0, 0)
#  (45, 1, 1, 30, 0, 1)
#  (63, 1, 1, 58, 1, 1)
#  (36, 1, 0, 52, 1, 1)
#  (23, 0, 1, 40, 0, 1)
#  (50, 1, 1, 28, 0, 1)]


class Node():
    """
    @todo: docstring for Node
    """
    def __init__(self, value=None):
        """@todo: Docstring for init method.

        /value=None/ @todo

        """
        self.value = value

    def add_split(self, left, right):
        """
        @todo: Docstring for add_split
        """
        self.left = left
        self.right = right


class Leaf:
    def __init__(self, value: int):
        self.value = value


class Tree():
    """
    @todo: docstring for Tree
    """
    def __init__(self, root_node_obj):
        """@todo: Docstring for init method.

        /root_node_obj/ @todo

        """
        self.tree = root_node_obj

    # def __repr__(self):
    #     nodelist = [self.tree]
    #     tree_str = ''
    #     while nodelist:
    #         current_node = nodelist.pop()
    #         # print(current_node.value)
    #         try:
    #             childs = [current_node.right, current_node.left]
    #             nodelist += childs
    #         except AttributeError:
    #             pass
    #         col, c = current_node.value
    #         tree_str += f"{col=}, {c=}"
    #     return tree_str


def impurity(array) -> int:
    """
    Assumes the argument array is a one dimensional vector of zeroes and ones.
    Computes the gini index impurity based on the relative frequency of ones in
    the vector.

    Example:

    >>> array=np.array([1,0,1,1,1,0,0,1,1,0,1])
    >>> array
    array([1,0,1,1,1,0,0,1,1,0,1])

    >>> impurity(array)
    0.23140495867768596
    """
    # Total labels
    n_labels = len(array)
    if n_labels == 0:
        print(
            "division by zero will happen, child node is pure, doesnt contain anything"
        )
        n_labels = 1
    # Number of tuples labeled 1
    n_labels_1 = array.sum()
    # Calculate the relative frequency of ones with respect to the total labels
    rel_freq_1 = n_labels_1 / n_labels
    # Use the symmetry around the median property to also calculate the
    # relative frequency of zeroes
    rel_freq_0 = 1 - rel_freq_1
    # Multiply the frequencies to get the gini index
    gini_index = rel_freq_1 * rel_freq_0
    return gini_index


def bestsplit(x, y, slices) -> int:
    """
    x = vector of num values
    y = vector of class labels ... array([{x: x is 0 or 1}]) ??

    Consider splits of type "x <= c" where "c" is the average of two consecutive
    values of x in the sorted order.

    x and y must be of the same length

    y[i] must be the class label of the i-th observation, and x[i] is the
    correspnding value of attribute x

    Example (best split on income):

    >>> bestsplit(credit_data[:,3],credit_data[:,5])
     36
    """
    x_sorted = np.sort(np.unique(x))
    if len(x_sorted) <= 2:
        # Allows for normal cat classes slicing
        split_points = [0.5]
    else:
        split_points = (x_sorted[:len(x_sorted) - 1] + x_sorted[1:]) / 2

    best_dict = None
    for split in split_points:
        x_slices = {
            # "left": [row for row in range(len(x)) if x[row] > split],
            # "right": [row for row in range(len(x)) if x[row] <= split]
            "left": np.index_exp[x > split],
            "right": np.index_exp[x <= split]
        }

        # delta_i formule
        delta_i = impurity(y) - (len(y[x_slices["left"]]) * impurity(
            y[x_slices["left"]]) + len(y[x_slices["right"]]) *
                                 impurity(y[x_slices["right"]])) / len(y)

        # this part is pretty bad
        if isinstance(slices, dict):
            x_slices = {
                "left": slices["left"][x_slices["left"]],
                "right": slices["right"][x_slices["right"]]
                # "left": np.index_exp[x > split],
                # "right": np.index_exp[x <= split]
            }
        else:
            x_slices = {
                "left": slices[x_slices["left"]],
                "right": slices[x_slices["right"]]
                # "left": np.index_exp[x > split],
                # "right": np.index_exp[x <= split]
            }

        print(f"{split=}, {delta_i=}")
        # slices = bool_array_2_row_number(x_slices, slices)
        if best_dict is not None:
            if delta_i > best_dict["delta_i"]:
                best_dict = {
                    # Make slices work regardless of np array dimensions with this list comprehension
                    "slices": x_slices,
                    "split": split,
                    "delta_i": delta_i
                }
        else:
            best_dict = {
                "slices": x_slices,
                "split": split,
                "delta_i": delta_i
            }
    return best_dict


#
#
# Put all helper functions above this comment!


def tree_grow(x=None,
              y=None,
              n_min=None,
              min_leaf=None,
              n_feat=None,
              **defaults) -> Tree:
    """
    @todo: Docstring for tree_grow
    """
    # store slice as variable
    slices = np.array([row for row in range(len(y))])
    # Initiate the nodelist with tuples of slice and class labels
    nodelist = [Node(value=slices)]
    tree = Tree(nodelist[0])
    while nodelist:
        current_node = nodelist.pop()
        slices = current_node.value
        node_classes = y[slices]
        # print(node_classes)

        # f'Current node will be leaf node if (( (number of data "tuples" in child node) < {n_min=} )) \n'
        # put stopping rules here before making a split
        if len(node_classes) < n_min:
            current_node.value = Leaf(
                np.argmax(np.bincount(node_classes.astype(int))))
            print(f"leaf node has majority clas:\n{current_node.value.value=}")
            continue

        if impurity(node_classes) > 0:
            # print(
            #     f"Exhaustive split search says, new node will check these rows for potential spliterinos:\n{x[slices]}"
            # )

            # If we arrive here ever we are splitting
            # bestsplit(col, node_labels) ->
            # {"slices": list[int], "split": numpyfloat, "best_delta_i": numpyfloat}

            # slices (list) used for knowing which rows (int) to consider in a node
            # best_split saved in current_node.value
            # best_delta_i used to find best split among x_columns
            best_dict = None
            for i, x_col in enumerate(x[slices].transpose()):
                print(
                    "\nExhaustive split search says; \"Entering new column\":")
                col_split_dict = bestsplit(x_col, node_classes, slices)

                if best_dict is not None:
                    if col_split_dict["delta_i"] > best_dict["delta_i"]:
                        best_dict = col_split_dict
                        best_dict["col"] = i
                else:
                    best_dict = col_split_dict
                    best_dict["col"] = i
            print("\nThe best split for current node:", best_dict)

            # Here we store the splitted data into Node objects
            current_node.value = best_dict["split"]
            current_node.col = best_dict["col"]
            # Split will not happen if (( (number of data "tuples" potential split) < {min_leaf=} ))\n'
            if min([len(x) for x in best_dict["slices"].values()]) < min_leaf:
                continue
            else:
                # Invert left and right because we want left to pop() first
                children = [
                    Node(value=best_dict["slices"]["right"]),
                    Node(value=best_dict["slices"]["left"])
                ]
                current_node.add_split(children[1], children[0])
                nodelist += children
        else:
            current_node.value = Leaf(
                np.argmax(np.bincount(node_classes.astype(int))))
            print(f"\n\nLEAF NODE has majority clas:\n{current_node.value.value=}")
            continue
    return tree


def predict(x, nodes) -> list:
    """
    @todo: Docstring for predict
    """
    # which row to drop
    # print(x)
    drop = 0
    while not set(nodes).issubset({0,1}):
        print(nodes)
        # print(x[drop])
        if isinstance(nodes[drop].value, Leaf):
            nodes[drop] = nodes[drop].value.value
            drop += 1
            continue

        print(nodes[drop].value)
        print(nodes[drop].col)
        # print(nodes[drop].col)
        if x[drop, nodes[drop].col] > nodes[drop].value:
            nodes[drop] = nodes[drop].left
        else:
            nodes[drop] = nodes[drop].right
    return np.array(nodes)


def tree_pred(x=None, tr=None, **defaults) -> np.array:
    """
    @todo: Docstring for tree_pred
    """
    nodes = [tr.tree] * len(x)
    # y = np.linspace(0, len(x), 0)
    # y = np.array(ele)
    y = predict(x, nodes)
    print(f"\n\nPredicted classes for {x=}\n\n are: {y=}")
    return y


if __name__ == '__main__':
    #### IMPURITY TEST
    # array=np.array([1,0,1,1,1,0,0,1,1,0,1])
    # print(impurity(array))
    # Should give 0.23....

    #### BESTSPLIT TEST
    # print(bestsplit(credit_data[:, 3], credit_data[:, 5]))
    # Should give 36

    #### TREE_GROW TEST
    tree_grow_defaults = {
        'x': credit_data[:, :5],
        'y': credit_data[:, 5],
        'n_min': 2,
        'min_leaf': 1,
        'n_feat': 5
    }

    # Calling the tree grow, unpacking default as argument
    # tree_grow(**tree_grow_defaults)

    #### TREE_PRED TEST
    tree_pred_defaults = {
        'x': credit_data[:, :5],
        'tr': tree_grow(**tree_grow_defaults)
    }

    tree_pred(**tree_pred_defaults)

start_time = time.time()
print("--- %s seconds ---" % (time.time() - start_time))