]> git.openstreetmap.org Git - nominatim.git/blobdiff - test/python/api/conftest.py
Merge pull request #3328 from lonvia/word-count-into-new-table
[nominatim.git] / test / python / api / conftest.py
index cb7f324a393fa24e2ddb097e710b92662fc96bd2..05eaddf5fc0f182cfc501504e48b8865fdb9af95 100644 (file)
@@ -16,6 +16,7 @@ import sqlalchemy as sa
 
 import nominatim.api as napi
 from nominatim.db.sql_preprocessor import SQLPreprocessor
+from nominatim.api.search.query_analyzer_factory import make_query_analyzer
 from nominatim.tools import convert_sqlite
 import nominatim.api.logging as loglib
 
@@ -160,6 +161,22 @@ class APITester:
                                      """)))
 
 
+    def add_word_table(self, content):
+        data = [dict(zip(['word_id', 'word_token', 'type', 'word', 'info'], c))
+                for c in content]
+
+        async def _do_sql():
+            async with self.api._async_api.begin() as conn:
+                if 'word' not in conn.t.meta.tables:
+                    await make_query_analyzer(conn)
+                    word_table = conn.t.meta.tables['word']
+                    await conn.connection.run_sync(word_table.create)
+                if data:
+                    await conn.execute(conn.t.meta.tables['word'].insert(), data)
+
+        self.async_to_sync(_do_sql())
+
+
     async def exec_async(self, sql, *args, **kwargs):
         async with self.api._async_api.begin() as conn:
             return await conn.execute(sql, *args, **kwargs)
@@ -190,17 +207,40 @@ def apiobj(temp_db_with_extensions, temp_db_conn, monkeypatch):
 
 @pytest.fixture(params=['postgres_db', 'sqlite_db'])
 def frontend(request, event_loop, tmp_path):
+    testapis = []
     if request.param == 'sqlite_db':
         db = str(tmp_path / 'test_nominatim_python_unittest.sqlite')
 
         def mkapi(apiobj, options={'reverse'}):
+            apiobj.add_data('properties',
+                        [{'property': 'tokenizer', 'value': 'icu'},
+                         {'property': 'tokenizer_import_normalisation', 'value': ':: lower();'},
+                         {'property': 'tokenizer_import_transliteration', 'value': "'1' > '/1/'; 'ä' > 'ä '"},
+                        ])
+
+            async def _do_sql():
+                async with apiobj.api._async_api.begin() as conn:
+                    if 'word' in conn.t.meta.tables:
+                        return
+                    await make_query_analyzer(conn)
+                    word_table = conn.t.meta.tables['word']
+                    await conn.connection.run_sync(word_table.create)
+
+            apiobj.async_to_sync(_do_sql())
+
             event_loop.run_until_complete(convert_sqlite.convert(Path('/invalid'),
                                                                  db, options))
-            return napi.NominatimAPI(Path('/invalid'),
-                                     {'NOMINATIM_DATABASE_DSN': f"sqlite:dbname={db}",
-                                      'NOMINATIM_USE_US_TIGER_DATA': 'yes'})
+            outapi = napi.NominatimAPI(Path('/invalid'),
+                                       {'NOMINATIM_DATABASE_DSN': f"sqlite:dbname={db}",
+                                        'NOMINATIM_USE_US_TIGER_DATA': 'yes'})
+            testapis.append(outapi)
+
+            return outapi
     elif request.param == 'postgres_db':
         def mkapi(apiobj, options=None):
             return apiobj.api
 
-    return mkapi
+    yield mkapi
+
+    for api in testapis:
+        api.close()