tests/tests_nlp/tests_stanza/test_stanza_got.py
# ----------------------------------------------------------------------
# Wordless: Tests - NLP - Stanza - Gothic
# Copyright (C) 2018-2024 Ye Lei (εΆη£)
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
# ----------------------------------------------------------------------
from tests.tests_nlp.tests_stanza import test_stanza
def test_stanza_got():
test_stanza.wl_test_stanza(
lang = 'got',
results_sentence_tokenize = ['π²πΏππ°ππ°πΆπ³π°, π²πΏπππ°πΆπ³π° π°πΉπΈπΈπ°πΏ π²πΏππΉππΊπ° ππ°πΆπ³π° πΉππ π²π°ππ
πΏπ»ππ°π½π° π°πΏππππ°π²π°πΉππΌπ°π½πΉππΊπ° ππ°πΆπ³π° πππ³πΉπ³π° πππ°πΌ π²πΏππ°πΌ. ππΉ πΉππ π°πΉπ½π°π·π π°πΏππππ°π²π°πΉπΌπ°π½πΉππΊπ° ππ°πΆπ³π° πππ΄πΉ π·π°π±π°πΉπΈ π²π°πΌπ΄π»π΄πΉπ½πΉπ½π', '.'],
results_word_tokenize = ['π²πΏππ°ππ°πΆπ³π°,', 'π²πΏπππ°πΆπ³π°', 'π°πΉπΈπΈπ°πΏ', 'π²πΏππΉππΊπ°', 'ππ°πΆπ³π°', 'πΉππ', 'π²π°ππ
πΏπ»ππ°π½π°', 'π°πΏππππ°π²π°πΉππΌπ°π½πΉππΊπ°', 'ππ°πΆπ³π°', 'πππ³πΉπ³π°', 'πππ°πΌ', 'π²πΏππ°πΌ', '.'],
results_pos_tag = [('π²πΏππ°ππ°πΆπ³π°,', 'Nb'), ('π²πΏπππ°πΆπ³π°', 'Nb'), ('π°πΉπΈπΈπ°πΏ', 'Pd'), ('π²πΏππΉππΊπ°', 'Pd'), ('ππ°πΆπ³π°', 'Pd'), ('πΉππ', 'Pd'), ('π²π°ππ
πΏπ»ππ°π½π°', 'Nb'), ('π°πΏππππ°π²π°πΉππΌπ°π½πΉππΊπ°', 'Nb'), ('ππ°πΆπ³π°', 'Pd'), ('πππ³πΉπ³π°', 'Pd'), ('πππ°πΌ', 'Pd'), ('π²πΏππ°πΌ', 'Nb'), ('.', 'Df')],
results_pos_tag_universal = [('π²πΏππ°ππ°πΆπ³π°,', 'NOUN'), ('π²πΏπππ°πΆπ³π°', 'NOUN'), ('π°πΉπΈπΈπ°πΏ', 'DET'), ('π²πΏππΉππΊπ°', 'DET'), ('ππ°πΆπ³π°', 'DET'), ('πΉππ', 'DET'), ('π²π°ππ
πΏπ»ππ°π½π°', 'NOUN'), ('π°πΏππππ°π²π°πΉππΌπ°π½πΉππΊπ°', 'NOUN'), ('ππ°πΆπ³π°', 'DET'), ('πππ³πΉπ³π°', 'DET'), ('πππ°πΌ', 'DET'), ('π²πΏππ°πΌ', 'NOUN'), ('.', 'ADV')],
results_lemmatize = ['π²πΏππ°ππ°πΆπ³π°,', 'π²πΏπππ°πΆπ³π°', 'π°πΉπΈπΈπ°πΏ', 'π²πΏππΉππΊπ°', 'ππ°πΆπ³π°', 'πΉππ', 'π²π°ππ
πΏπ»ππ°π½π°', 'π°πΏππππ°π²π°πΉππΌπ°π½πΉππΊπ°', 'ππ°πΆπ³π°', 'πππ³πΉπ³π°', 'πππ°πΌ', 'π²πΏππ°πΌ', '-uh'],
results_dependency_parse = [('π²πΏππ°ππ°πΆπ³π°,', 'π²πΏπππ°πΆπ³π°', 'vocative', 1), ('π²πΏπππ°πΆπ³π°', 'π²πΏπππ°πΆπ³π°', 'root', 0), ('π°πΉπΈπΈπ°πΏ', 'π²πΏπππ°πΆπ³π°', 'det', -1), ('π²πΏππΉππΊπ°', 'π²π°ππ
πΏπ»ππ°π½π°', 'det', 3), ('ππ°πΆπ³π°', 'π²π°ππ
πΏπ»ππ°π½π°', 'det', 2), ('πΉππ', 'π²π°ππ
πΏπ»ππ°π½π°', 'det', 1), ('π²π°ππ
πΏπ»ππ°π½π°', 'π²πΏπππ°πΆπ³π°', 'appos', -5), ('π°πΏππππ°π²π°πΉππΌπ°π½πΉππΊπ°', 'π²π°ππ
πΏπ»ππ°π½π°', 'nmod', -1), ('ππ°πΆπ³π°', 'π²πΏππ°πΌ', 'det', 3), ('πππ³πΉπ³π°', 'π²πΏππ°πΌ', 'det', 2), ('πππ°πΌ', 'π²πΏππ°πΌ', 'det', 1), ('π²πΏππ°πΌ', 'π°πΏππππ°π²π°πΉππΌπ°π½πΉππΊπ°', 'appos', -4), ('.', '.', 'root', 0)]
)
if __name__ == '__main__':
test_stanza_got()