X-Git-Url: https://git.openstreetmap.org./nominatim.git/blobdiff_plain/5a9fb6eaf73ebfc5a82e553744328ade60b50f8a..fa1c402b47bdd1f76b8528885c0e3a5e90731cdc:/test/python/tokenizer/test_legacy.py diff --git a/test/python/tokenizer/test_legacy.py b/test/python/tokenizer/test_legacy.py index a16756df..bf208c92 100644 --- a/test/python/tokenizer/test_legacy.py +++ b/test/python/tokenizer/test_legacy.py @@ -1,3 +1,9 @@ +# SPDX-License-Identifier: GPL-3.0-or-later +# +# This file is part of Nominatim. (https://nominatim.org) +# +# Copyright (C) 2024 by the Nominatim developer community. +# For a full list of authors see the git log. """ Test for legacy tokenizer. """ @@ -6,10 +12,10 @@ import re import pytest -from nominatim.indexer.place_info import PlaceInfo -from nominatim.tokenizer import legacy_tokenizer -from nominatim.db import properties -from nominatim.errors import UsageError +from nominatim_db.data.place_info import PlaceInfo +from nominatim_db.tokenizer import legacy_tokenizer +from nominatim_db.db import properties +from nominatim_db.errors import UsageError from mock_legacy_word_table import MockLegacyWordTable @@ -23,7 +29,7 @@ def word_table(temp_db_conn): def test_config(project_env, tmp_path): module_dir = tmp_path / 'module_src' module_dir.mkdir() - (module_dir / 'nominatim.so').write_text('TEST nomiantim.so') + (module_dir / 'nominatim.so').write_text('TEST nominatim.so') project_env.lib_dir.module = module_dir @@ -115,7 +121,7 @@ def test_init_new(tokenizer_factory, test_config, monkeypatch, outfile = test_config.project_dir / 'module' / 'nominatim.so' assert outfile.exists() - assert outfile.read_text() == 'TEST nomiantim.so' + assert outfile.read_text() == 'TEST nominatim.so' assert outfile.stat().st_mode == 33261 @@ -200,7 +206,7 @@ def test_migrate_database(tokenizer_factory, test_config, temp_db_conn, monkeypa outfile = test_config.project_dir / 'module' / 'nominatim.so' assert outfile.exists() - assert outfile.read_text() == 'TEST nomiantim.so' + assert outfile.read_text() == 'TEST nominatim.so' assert outfile.stat().st_mode == 33261 @@ -232,25 +238,32 @@ def test_check_database_bad_setup(test_config, tokenizer_factory, monkeypatch, assert tok.check_database(False) is not None -def test_update_statistics_reverse_only(word_table, tokenizer_factory): +def test_update_statistics_reverse_only(word_table, tokenizer_factory, test_config): tok = tokenizer_factory() - tok.update_statistics() + tok.update_statistics(test_config) -def test_update_statistics(word_table, table_factory, temp_db_cursor, tokenizer_factory): +def test_update_statistics(word_table, table_factory, temp_db_cursor, tokenizer_factory, test_config): word_table.add_full_word(1000, 'hello') table_factory('search_name', 'place_id BIGINT, name_vector INT[]', [(12, [1000])]) tok = tokenizer_factory() - tok.update_statistics() + tok.update_statistics(test_config) assert temp_db_cursor.scalar("""SELECT count(*) FROM word WHERE word_token like ' %' and search_name_count > 0""") > 0 +def test_update_word_tokens(tokenizer_factory): + tok = tokenizer_factory() + + # This is a noop and should just pass. + tok.update_word_tokens() + + def test_normalize(analyzer): assert analyzer.normalize('TEsT') == 'test' @@ -536,7 +549,7 @@ class TestPlaceAddress: def test_process_place_street_empty(self): info = self.process_address(street='🜵') - assert 'street' not in info + assert info['street'] == '{}' def test_process_place_place(self):