import pytest AF_BASIC_TOKENIZATION_TESTS = [ ( "Elkeen het die reg tot lewe, vryheid en sekuriteit van persoon.", [ "Elkeen", "het", "die", "reg", "tot", "lewe", ",", "vryheid", "en", "sekuriteit", "van", "persoon", ".", ], ), ] @pytest.mark.parametrize("text,expected_tokens", AF_BASIC_TOKENIZATION_TESTS) def test_af_tokenizer_basic(af_tokenizer, text, expected_tokens): tokens = af_tokenizer(text) token_list = [token.text for token in tokens if not token.is_space] assert expected_tokens == token_list