diff --git a/xtas/tests/test_corenlp.py b/xtas/tests/test_corenlp.py index 1f69342..a1864a6 100644 --- a/xtas/tests/test_corenlp.py +++ b/xtas/tests/test_corenlp.py @@ -61,7 +61,7 @@ def test_lemmatize_unicode(): annotators=['tokenize', 'ssplit', 'pos', 'lemma']) saf = stanford_to_saf(lines) assert_equal({t['lemma'] for t in saf['tokens']}, - {'Cesar', 'hit', 'Hovik'}) + {'Cesar', 'hit', 'Hovik'}) # def test_ner(): @@ -92,7 +92,7 @@ def test_parse(): assert_equal(corefs, {tuple(sorted(['John', 'himself']))}) -def test_multiple_sentences(): +def test_multiple_sentences(): # _check_corenlp() p = parse("John lives in Amsterdam. He works in London") saf = stanford_to_saf(p) diff --git a/xtas/tests/test_es.py b/xtas/tests/test_es.py index 309b03e..2abbae0 100644 --- a/xtas/tests/test_es.py +++ b/xtas/tests/test_es.py @@ -86,7 +86,7 @@ def test_query_batch(): assert_equal(set(b), {"test", "test2"}) -def test_store_get_result(): +def test_store_get_result(): # "test whether results can be stored and retrieved" from xtas.tasks.es import ( store_single, diff --git a/xtas/tests/test_frog.py b/xtas/tests/test_frog.py index 7968ffc..7169236 100644 --- a/xtas/tests/test_frog.py +++ b/xtas/tests/test_frog.py @@ -94,4 +94,4 @@ def test_frog_task(): assert_equal(tokens[0]['lemma'], 'dit') saf = frog("dit is een test", output='saf') assert_equal(len(saf['tokens']), 4) - assert_equal(saf['header']['processed'][0]['module'], 'frog') + assert_equal(saf['header']['processed'][0]['module'], 'frog') # diff --git a/xtas/tests/test_pipeline.py b/xtas/tests/test_pipeline.py index e2dc6c6..7af2371 100644 --- a/xtas/tests/test_pipeline.py +++ b/xtas/tests/test_pipeline.py @@ -44,7 +44,7 @@ def test_pipeline(): s = "cats are furry" expected = [('cats', 'NNS'), ('are', 'VBP'), ('furry', 'JJ')] result = pos_tag(tokenize(s), 'nltk') - assert_equal(result, expected) + assert_equal(result, expected) # with eager_celery(): # do we get correct result from pipeline? r = pipeline(s, [{"module": tokenize},