def loading_dataSet():
    file = open("res/dataset.txt", "r")
    data = file.read()
    file.close()
    docs = data.split("\n")
    types = []
    train = []
    for d in docs:
        d = d.split()
        if len(d) != 0:
            types.append(d[0])
    print('dataset Count = ' + str(len(types)))
    normalized_corpus = Normalizer.normalize_corpus(docs)
    normalized_corpus.remove('')
    counter = 0
    for x in normalized_corpus:
        train.append((x, types[counter]))
        counter = counter + 1
    return train
def classify_btn_clicked():
    def setClassification(type):
        if type == '1':
            classi_out.setPlainText('culture')
        elif type == '2':
            classi_out.setPlainText('sport')
        elif type == '3':
            classi_out.setPlainText('economy')
        elif type == '4':
            classi_out.setPlainText('international')
        elif type == '5':
            classi_out.setPlainText('local')
        elif type == '6':
            classi_out.setPlainText('religion')

    tester_doc = file_.toPlainText().strip()
    normalized_tester_doc = Normalizer.normalize_corpus([tester_doc])
    featuresets_test = [features(words) for words in normalized_tester_doc]
    predicted_label = classifier.classify_many(featuresets_test)
    setClassification(predicted_label[0])