コード例 #1
0
ファイル: lgs.py プロジェクト: kellman/GSP_for_NLP
def essay_graph_gen(raw_text):
    print raw_text
    exclude_list = stopwords.words('english')
    test_dict = dict_gen(raw_text,0,1,exclude_list,[])
    print test_dict
    # print_table(test_dict[__sent_dict_key],__sent_dict_key)
    # print_table(test_dict[__word_dict_key],__word_dict_key)
    num_of_sentences = len(test_dict[__sent_dict_key])
    # print test_dict[__sent_dict_key]
    nodes = test_dict[__word_dict_key].keys() 
    verts = [[0 for i in nodes] for j in nodes]
    PARTS = [nodes,verts]
    for i in xrange(1,num_of_sentences+1):
        PARTS = (sent_graph_gen(i,test_dict[__word_dict_key],test_dict[__sent_dict_key],PARTS[0],PARTS[1]))
    return PARTS
コード例 #2
0
ファイル: lgs.py プロジェクト: kellman/GSP_for_NLP
def class_graph_gen(raw_text,global_nodes,global_dict):
    exclude_list = stopwords.words('english')
    test_dict = dict_gen(raw_text,0,1,exclude_list,[])
    if test_dict == None:
        return None
    # now we use global nodes instead of test_dict words_list keys
    num_of_sentences = len(test_dict[__sent_dict_key])
    # print test_dict[__sent_dict_key]
    verts = [[0 for i in global_nodes] for j in global_nodes]
    PARTS = [global_nodes,verts]
    for i in xrange(1,num_of_sentences+1):
        PARTS = sent_graph_gen(i,test_dict[__word_dict_key],test_dict[__sent_dict_key],PARTS[0],PARTS[1])    # G = parts[0][1]
    # for i in xrange(1,num_of_sentences):
    #     G = add_graphs(parts[i][0],G,parts[i][0],parts[i][1])
    WEIGHTS = word_count(test_dict,global_dict)
    print len(PARTS), "PARTS"
    print len(WEIGHTS), "WEIGHTS"
    return (PARTS,WEIGHTS)