mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-11-04 01:48:04 +03:00 
			
		
		
		
	Xfail beam stuff. 9 failures
This commit is contained in:
		
							parent
							
								
									880182afdb
								
							
						
					
					
						commit
						7309e49286
					
				| 
						 | 
				
			
			@ -118,6 +118,7 @@ def test_beam_advance_too_few_scores(beam, scores):
 | 
			
		|||
        beam.advance(scores[:-1])
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@pytest.mark.xfail(reason="no beam parser yet")
 | 
			
		||||
def test_beam_parse(examples, beam_width):
 | 
			
		||||
    nlp = Language()
 | 
			
		||||
    parser = nlp.add_pipe("beam_parser")
 | 
			
		||||
| 
						 | 
				
			
			@ -128,6 +129,7 @@ def test_beam_parse(examples, beam_width):
 | 
			
		|||
    parser(doc)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@pytest.mark.xfail(reason="no beam parser yet")
 | 
			
		||||
@hypothesis.given(hyp=hypothesis.strategies.data())
 | 
			
		||||
def test_beam_density(moves, examples, beam_width, hyp):
 | 
			
		||||
    beam_density = float(hyp.draw(hypothesis.strategies.floats(0.0, 1.0, width=32)))
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -55,6 +55,8 @@ PARTIAL_DATA = [
 | 
			
		|||
    ),
 | 
			
		||||
]
 | 
			
		||||
 | 
			
		||||
PARSERS = ["parser"] # TODO: Test beam_parser when ready
 | 
			
		||||
 | 
			
		||||
eps = 0.1
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -215,7 +217,7 @@ def test_parser_set_sent_starts(en_vocab):
 | 
			
		|||
            assert token.head in sent
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@pytest.mark.parametrize("pipe_name", ["parser", "beam_parser"])
 | 
			
		||||
@pytest.mark.parametrize("pipe_name", PARSERS)
 | 
			
		||||
def test_incomplete_data(pipe_name):
 | 
			
		||||
    # Test that the parser works with incomplete information
 | 
			
		||||
    nlp = English()
 | 
			
		||||
| 
						 | 
				
			
			@ -241,7 +243,7 @@ def test_incomplete_data(pipe_name):
 | 
			
		|||
    assert doc[2].head.i == 1
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@pytest.mark.parametrize("pipe_name", ["parser", "beam_parser"])
 | 
			
		||||
@pytest.mark.parametrize("pipe_name", PARSERS)
 | 
			
		||||
def test_overfitting_IO(pipe_name):
 | 
			
		||||
    # Simple test to try and quickly overfit the dependency parser (normal or beam)
 | 
			
		||||
    nlp = English()
 | 
			
		||||
| 
						 | 
				
			
			@ -292,6 +294,7 @@ def test_overfitting_IO(pipe_name):
 | 
			
		|||
    assert_equal(batch_deps_1, no_batch_deps)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@pytest.mark.xfail(reason="no beam parser yet")
 | 
			
		||||
def test_beam_parser_scores():
 | 
			
		||||
    # Test that we can get confidence values out of the beam_parser pipe
 | 
			
		||||
    beam_width = 16
 | 
			
		||||
| 
						 | 
				
			
			@ -330,6 +333,7 @@ def test_beam_parser_scores():
 | 
			
		|||
            assert 0 - eps <= head_score <= 1 + eps
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@pytest.mark.xfail(reason="no beam parser yet")
 | 
			
		||||
def test_beam_overfitting_IO():
 | 
			
		||||
    # Simple test to try and quickly overfit the Beam dependency parser
 | 
			
		||||
    nlp = English()
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -287,6 +287,7 @@ def test_multiple_predictions():
 | 
			
		|||
    dummy_pipe(doc)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@pytest.mark.xfail(reason="no beam parser yet")
 | 
			
		||||
def test_issue4313():
 | 
			
		||||
    """ This should not crash or exit with some strange error code """
 | 
			
		||||
    beam_width = 16
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in New Issue
	
	Block a user