X-Git-Url: https://git.openstreetmap.org./nominatim.git/blobdiff_plain/8c7140d92b7a679ae5ef5bb3655c69bd39b7dfeb..2bab0ca060ff224091b20a0ac808a3febaba04b1:/test/python/api/conftest.py diff --git a/test/python/api/conftest.py b/test/python/api/conftest.py index cb7f324a..08b7422c 100644 --- a/test/python/api/conftest.py +++ b/test/python/api/conftest.py @@ -1,8 +1,8 @@ -# SPDX-License-Identifier: GPL-2.0-only +# SPDX-License-Identifier: GPL-3.0-or-later # # This file is part of Nominatim. (https://nominatim.org) # -# Copyright (C) 2023 by the Nominatim developer community. +# Copyright (C) 2024 by the Nominatim developer community. # For a full list of authors see the git log. """ Helper fixtures for API call tests. @@ -14,10 +14,11 @@ import datetime as dt import sqlalchemy as sa -import nominatim.api as napi -from nominatim.db.sql_preprocessor import SQLPreprocessor -from nominatim.tools import convert_sqlite -import nominatim.api.logging as loglib +import nominatim_api as napi +from nominatim_core.db.sql_preprocessor import SQLPreprocessor +from nominatim_api.search.query_analyzer_factory import make_query_analyzer +from nominatim_db.tools import convert_sqlite +import nominatim_api.logging as loglib class APITester: @@ -160,6 +161,22 @@ class APITester: """))) + def add_word_table(self, content): + data = [dict(zip(['word_id', 'word_token', 'type', 'word', 'info'], c)) + for c in content] + + async def _do_sql(): + async with self.api._async_api.begin() as conn: + if 'word' not in conn.t.meta.tables: + await make_query_analyzer(conn) + word_table = conn.t.meta.tables['word'] + await conn.connection.run_sync(word_table.create) + if data: + await conn.execute(conn.t.meta.tables['word'].insert(), data) + + self.async_to_sync(_do_sql()) + + async def exec_async(self, sql, *args, **kwargs): async with self.api._async_api.begin() as conn: return await conn.execute(sql, *args, **kwargs) @@ -190,17 +207,40 @@ def apiobj(temp_db_with_extensions, temp_db_conn, monkeypatch): @pytest.fixture(params=['postgres_db', 'sqlite_db']) def frontend(request, event_loop, tmp_path): + testapis = [] if request.param == 'sqlite_db': db = str(tmp_path / 'test_nominatim_python_unittest.sqlite') def mkapi(apiobj, options={'reverse'}): + apiobj.add_data('properties', + [{'property': 'tokenizer', 'value': 'icu'}, + {'property': 'tokenizer_import_normalisation', 'value': ':: lower();'}, + {'property': 'tokenizer_import_transliteration', 'value': "'1' > '/1/'; 'ä' > 'ä '"}, + ]) + + async def _do_sql(): + async with apiobj.api._async_api.begin() as conn: + if 'word' in conn.t.meta.tables: + return + await make_query_analyzer(conn) + word_table = conn.t.meta.tables['word'] + await conn.connection.run_sync(word_table.create) + + apiobj.async_to_sync(_do_sql()) + event_loop.run_until_complete(convert_sqlite.convert(Path('/invalid'), db, options)) - return napi.NominatimAPI(Path('/invalid'), - {'NOMINATIM_DATABASE_DSN': f"sqlite:dbname={db}", - 'NOMINATIM_USE_US_TIGER_DATA': 'yes'}) + outapi = napi.NominatimAPI(Path('/invalid'), + {'NOMINATIM_DATABASE_DSN': f"sqlite:dbname={db}", + 'NOMINATIM_USE_US_TIGER_DATA': 'yes'}) + testapis.append(outapi) + + return outapi elif request.param == 'postgres_db': def mkapi(apiobj, options=None): return apiobj.api - return mkapi + yield mkapi + + for api in testapis: + api.close()