def test_verbalize_header_random(test_data, capsys):
    """Test verbalization from NLTKInterface."""
    verbalization_log_file = os.path.join(LOG_FOLDER,
                                          "verbalization_random_1.logs")
    verbalization_log = FileWriter(verbalization_log_file)
    verbalization_log_time_file = os.path.join(
        LOG_FOLDER, "verbalization_random_time_1.logs")
    verbalization_log_time = FileWriter(verbalization_log_time_file)
    error_log_file = os.path.join(LOG_FOLDER,
                                  "verbalization_random_error.logs")
    error_log = FileWriter(error_log_file)

    start_time = time.time()
    empty_line = "\n"
    for index, header in enumerate(test_data):
        verbalization_log.write(repr(header))
        verbalization_log.write(empty_line)
        try:
            verbalized_header = nltk.verbalize_header_random(header)
            verbalization_log.write(repr(verbalized_header))
            verbalization_log.write(empty_line)
            verbalization_log.write(empty_line)
        except:
            error_log.write(repr(header))
            continue
    end_time = time.time()
    verbalization_log_time.write("%s" % (end_time - start_time))

    verbalization_log.close()
    verbalization_log_time.close()
def test_verbalize_header_palmetto(test_data, capsys):
    """Test verbalization from NLTKInterface."""
    verbalization_log_file = os.path.join(
        LOG_FOLDER,
        "verbalization_palmetto.logs"
    )
    verbalization_log = FileWriter(verbalization_log_file)
    verbalization_log_time_file = os.path.join(
        LOG_FOLDER,
        "verbalization_palmetto_time.logs"
    )
    verbalization_log_time = FileWriter(verbalization_log_time_file)

    start_time = time.time()
    empty_line = "\n"
    for header in test_data:
        verbalization_log.write(repr(header))
        verbalization_log.write(empty_line)
        verbalized_header = nltk.verbalize_header_palmetto(header)
        verbalization_log.write(repr(verbalized_header))
        verbalization_log.write(empty_line)
        verbalization_log.write(empty_line)
    end_time = time.time()
    verbalization_log_time.write("%s" % (end_time - start_time))

    verbalization_log.close()
    verbalization_log_time.close()
def test_get_header_synsets(header):
    """Test get_header_synsets method."""
    synset_packs = nltk.get_header_synsets(header)
    for index, synset_pack in enumerate(synset_packs):
        (label, synsets) = synset_pack
        assert header[index] == label
        assert isinstance(synsets, list)
def test_get_header_synsets(header):
    """Test get_header_synsets method."""
    synset_packs = nltk.get_header_synsets(header)
    for index, synset_pack in enumerate(synset_packs):
        (label, synsets) = synset_pack
        assert header[index] == label
        assert isinstance(synsets, list)
def test_is_graph_converge_pass(subgraph_converge):
    """
    Test is_graph_converge method.

    Subgraph should converge.
    """
    is_converge = nltk.is_graph_converge(subgraph_converge, 5)
    assert is_converge is True
def test_is_graph_converge_fail(subgraph_does_not_converge):
    """
    Test is_graph_converge method.

    Subgraph should not converge.
    """
    is_converge = nltk.is_graph_converge(subgraph_does_not_converge, 5)
    assert is_converge is False
def test_build_weighted_graph(synset_packs):
    """Test build_weigthed_graph method."""
    synset_graph = nltk.build_weighted_graph(synset_packs)
    number_of_nodes = len(synset_packs)
    number_of_edges = number_of_nodes * (number_of_nodes - 1) / 2
    assert len(synset_graph) == number_of_edges
    for edge in synset_graph:
        assert len(edge) != 0
def test_is_graph_converge_fail(subgraph_does_not_converge):
    """
    Test is_graph_converge method.

    Subgraph should not converge.
    """
    is_converge = nltk.is_graph_converge(subgraph_does_not_converge, 5)
    assert is_converge is False
def test_is_graph_converge_pass(subgraph_converge):
    """
    Test is_graph_converge method.

    Subgraph should converge.
    """
    is_converge = nltk.is_graph_converge(subgraph_converge, 5)
    assert is_converge is True
def test_build_weighted_graph(synset_packs):
    """Test build_weigthed_graph method."""
    synset_graph = nltk.build_weighted_graph(synset_packs)
    number_of_nodes = len(synset_packs)
    number_of_edges = number_of_nodes * (number_of_nodes - 1) / 2
    assert len(synset_graph) == number_of_edges
    for edge in synset_graph:
        assert len(edge) != 0
def test_verbalize_random_fast(test_data):
    """Test verbalization from NLTKInterface."""

    verbalization_log_file = os.path.join(LOG_FOLDER,
                                          "verbalization_fast.logs")
    verbalization_log = FileWriter(verbalization_log_file)

    start_time = time.time()
    empty_line = "\n"
    for header in test_data:
        verbalization_log.write(repr(header))
        verbalization_log.write(empty_line)
        verbalized_header = nltk.verbalize_header_random(header)
        verbalization_log.write(repr(verbalized_header))
        verbalization_log.write(empty_line)
        verbalization_log.write(empty_line)
    end_time = time.time()
def test_verbalize_random_fast(test_data):
    """Test verbalization from NLTKInterface."""

    verbalization_log_file = os.path.join(
        LOG_FOLDER,
        "verbalization_fast.logs"
    )
    verbalization_log = FileWriter(verbalization_log_file)

    start_time = time.time()
    empty_line = "\n"
    for header in test_data:
        verbalization_log.write(repr(header))
        verbalization_log.write(empty_line)
        verbalized_header = nltk.verbalize_header_random(header)
        verbalization_log.write(repr(verbalized_header))
        verbalization_log.write(empty_line)
        verbalization_log.write(empty_line)
    end_time = time.time()
def test_verbalize_header_palmetto(test_data, capsys):
    """Test verbalization from NLTKInterface."""
    verbalization_log_file = os.path.join(LOG_FOLDER,
                                          "verbalization_palmetto.logs")
    verbalization_log = FileWriter(verbalization_log_file)
    verbalization_log_time_file = os.path.join(
        LOG_FOLDER, "verbalization_palmetto_time.logs")
    verbalization_log_time = FileWriter(verbalization_log_time_file)

    start_time = time.time()
    empty_line = "\n"
    for header in test_data:
        verbalization_log.write(repr(header))
        verbalization_log.write(empty_line)
        verbalized_header = nltk.verbalize_header_palmetto(header)
        verbalization_log.write(repr(verbalized_header))
        verbalization_log.write(empty_line)
        verbalization_log.write(empty_line)
    end_time = time.time()
    verbalization_log_time.write("%s" % (end_time - start_time))

    verbalization_log.close()
    verbalization_log_time.close()
def test_verbalize_header_random(test_data, capsys):
    """Test verbalization from NLTKInterface."""
    verbalization_log_file = os.path.join(
        LOG_FOLDER,
        "verbalization_random_1.logs"
    )
    verbalization_log = FileWriter(verbalization_log_file)
    verbalization_log_time_file = os.path.join(
        LOG_FOLDER,
        "verbalization_random_time_1.logs"
    )
    verbalization_log_time = FileWriter(verbalization_log_time_file)
    error_log_file = os.path.join(
        LOG_FOLDER,
        "verbalization_random_error.logs"
    )
    error_log = FileWriter(error_log_file)

    start_time = time.time()
    empty_line = "\n"
    for index, header in enumerate(test_data):
        verbalization_log.write(repr(header))
        verbalization_log.write(empty_line)
        try:
            verbalized_header = nltk.verbalize_header_random(header)
            verbalization_log.write(repr(verbalized_header))
            verbalization_log.write(empty_line)
            verbalization_log.write(empty_line)
        except:
            error_log.write(repr(header))
            continue
    end_time = time.time()
    verbalization_log_time.write("%s" % (end_time - start_time))

    verbalization_log.close()
    verbalization_log_time.close()
def synset_graph():
    """Load synset graph fixture."""
    header = ['label', 'type', 'subject', 'homepage', 'foundation']
    synset_packs = nltk.get_header_synsets(header)
    return nltk.build_weighted_graph(synset_packs)
def test_data():
    """Load test data fixture."""
    return nltk.load_test_data()
Example #17
0
import opentablebench.NLTKInterface as nltk

header = ['label', 'type', 'subject']
_synsets = nltk.get_header_synsets(header)

import ipdb
ipdb.set_trace()
Example #18
0
import opentablebench.NLTKInterface as nltk

header = ['label', 'type', 'subject']
_synsets = nltk.get_header_synsets(header)

import ipdb; ipdb.set_trace()
def test_data():
    """Load test data fixture."""
    return nltk.load_test_data()
def synset_packs():
    """Load header fixture."""
    header = ['label', 'type', 'subject', 'homepage', 'foundation']
    return nltk.get_header_synsets(header)
def synset_graph():
    """Load synset graph fixture."""
    header = ['label', 'type', 'subject', 'homepage', 'foundation']
    synset_packs = nltk.get_header_synsets(header)
    return nltk.build_weighted_graph(synset_packs)
def synset_packs():
    """Load header fixture."""
    header = ['label', 'type', 'subject', 'homepage', 'foundation']
    return nltk.get_header_synsets(header)