"""
Tests for functions to import a new database.
"""
+from pathlib import Path
+from contextlib import closing
+
import pytest
import psycopg2
-import sys
-from pathlib import Path
from nominatim.tools import database_import
from nominatim.errors import UsageError
-@pytest.fixture
-def nonexistant_db():
- dbname = 'test_nominatim_python_unittest'
-
- conn = psycopg2.connect(database='postgres')
-
- conn.set_isolation_level(0)
- with conn.cursor() as cur:
- cur.execute('DROP DATABASE IF EXISTS {}'.format(dbname))
-
- yield dbname
+class TestDatabaseSetup:
+ DBNAME = 'test_nominatim_python_unittest'
- with conn.cursor() as cur:
- cur.execute('DROP DATABASE IF EXISTS {}'.format(dbname))
+ @pytest.fixture(autouse=True)
+ def setup_nonexistant_db(self):
+ conn = psycopg2.connect(database='postgres')
-@pytest.mark.parametrize("no_partitions", (True, False))
-def test_setup_skeleton(src_dir, nonexistant_db, no_partitions):
- database_import.setup_database_skeleton('dbname=' + nonexistant_db,
- src_dir / 'data', no_partitions)
+ try:
+ conn.set_isolation_level(0)
+ with conn.cursor() as cur:
+ cur.execute(f'DROP DATABASE IF EXISTS {self.DBNAME}')
- conn = psycopg2.connect(database=nonexistant_db)
+ yield True
- try:
- with conn.cursor() as cur:
- cur.execute("SELECT distinct partition FROM country_name")
- partitions = set([r[0] for r in list(cur)])
- if no_partitions:
- assert partitions == set([0])
- else:
- assert len(partitions) > 10
- finally:
- conn.close()
+ with conn.cursor() as cur:
+ cur.execute(f'DROP DATABASE IF EXISTS {self.DBNAME}')
+ finally:
+ conn.close()
+ @pytest.fixture
+ def cursor(self):
+ conn = psycopg2.connect(database=self.DBNAME)
-def test_create_db_success(nonexistant_db):
- database_import.create_db('dbname=' + nonexistant_db, rouser='www-data')
+ try:
+ with conn.cursor() as cur:
+ yield cur
+ finally:
+ conn.close()
- conn = psycopg2.connect(database=nonexistant_db)
- conn.close()
+ def conn(self):
+ return closing(psycopg2.connect(database=self.DBNAME))
-def test_create_db_already_exists(temp_db):
- with pytest.raises(UsageError):
- database_import.create_db('dbname=' + temp_db)
-
-
-def test_create_db_unsupported_version(nonexistant_db, monkeypatch):
- monkeypatch.setattr(database_import, 'POSTGRESQL_REQUIRED_VERSION', (100, 4))
- with pytest.raises(UsageError, match='PostgreSQL server is too old.'):
- database_import.create_db('dbname=' + nonexistant_db)
+ def test_setup_skeleton(self):
+ database_import.setup_database_skeleton(f'dbname={self.DBNAME}')
+ # Check that all extensions are set up.
+ with self.conn() as conn:
+ with conn.cursor() as cur:
+ cur.execute('CREATE TABLE t (h HSTORE, geom GEOMETRY(Geometry, 4326))')
-def test_create_db_missing_ro_user(nonexistant_db):
- with pytest.raises(UsageError, match='Missing read-only user.'):
- database_import.create_db('dbname=' + nonexistant_db, rouser='sdfwkjkjgdugu2;jgsafkljas;')
+ def test_unsupported_pg_version(self, monkeypatch):
+ monkeypatch.setattr(database_import, 'POSTGRESQL_REQUIRED_VERSION', (100, 4))
-def test_setup_extensions(temp_db_conn, table_factory):
- database_import.setup_extensions(temp_db_conn)
+ with pytest.raises(UsageError, match='PostgreSQL server is too old.'):
+ database_import.setup_database_skeleton(f'dbname={self.DBNAME}')
- # Use table creation to check that hstore and geometry types are available.
- table_factory('t', 'h HSTORE, geom GEOMETRY(Geometry, 4326)')
+ def test_create_db_missing_ro_user(self):
+ with pytest.raises(UsageError, match='Missing read-only user.'):
+ database_import.setup_database_skeleton(f'dbname={self.DBNAME}',
+ rouser='sdfwkjkjgdugu2;jgsafkljas;')
-def test_setup_extensions_old_postgis(temp_db_conn, monkeypatch):
- monkeypatch.setattr(database_import, 'POSTGIS_REQUIRED_VERSION', (50, 50))
- with pytest.raises(UsageError, match='PostGIS version is too old.'):
- database_import.setup_extensions(temp_db_conn)
+ def test_setup_extensions_old_postgis(self, monkeypatch):
+ monkeypatch.setattr(database_import, 'POSTGIS_REQUIRED_VERSION', (50, 50))
+ with pytest.raises(UsageError, match='PostGIS is too old.'):
+ database_import.setup_database_skeleton(f'dbname={self.DBNAME}')
-def test_import_base_data(dsn, src_dir, temp_db_with_extensions, temp_db_cursor):
- database_import.import_base_data(dsn, src_dir / 'data')
- assert temp_db_cursor.table_rows('country_name') > 0
+def test_setup_skeleton_already_exists(temp_db):
+ with pytest.raises(UsageError):
+ database_import.setup_database_skeleton(f'dbname={temp_db}')
-def test_import_base_data_ignore_partitions(dsn, src_dir, temp_db_with_extensions,
- temp_db_cursor):
- database_import.import_base_data(dsn, src_dir / 'data', ignore_partitions=True)
+def test_import_osm_data_simple(table_factory, osm2pgsql_options):
+ table_factory('place', content=((1, ), ))
- assert temp_db_cursor.table_rows('country_name') > 0
- assert temp_db_cursor.table_rows('country_name', where='partition != 0') == 0
+ database_import.import_osm_data(Path('file.pbf'), osm2pgsql_options)
-def test_import_osm_data_simple(table_factory, osm2pgsql_options):
+def test_import_osm_data_multifile(table_factory, tmp_path, osm2pgsql_options):
table_factory('place', content=((1, ), ))
+ osm2pgsql_options['osm2pgsql_cache'] = 0
+
+ files = [tmp_path / 'file1.osm', tmp_path / 'file2.osm']
+ for f in files:
+ f.write_text('test')
- database_import.import_osm_data('file.pdf', osm2pgsql_options)
+ database_import.import_osm_data(files, osm2pgsql_options)
def test_import_osm_data_simple_no_data(table_factory, osm2pgsql_options):
table_factory('place')
with pytest.raises(UsageError, match='No data.*'):
- database_import.import_osm_data('file.pdf', osm2pgsql_options)
+ database_import.import_osm_data(Path('file.pbf'), osm2pgsql_options)
def test_import_osm_data_drop(table_factory, temp_db_conn, tmp_path, osm2pgsql_options):
osm2pgsql_options['flatnode_file'] = str(flatfile.resolve())
- database_import.import_osm_data('file.pdf', osm2pgsql_options, drop=True)
+ database_import.import_osm_data(Path('file.pbf'), osm2pgsql_options, drop=True)
assert not flatfile.exists()
assert not temp_db_conn.table_exists('planet_osm_nodes')
@pytest.mark.parametrize("threads", (1, 5))
-def test_load_data(dsn, src_dir, place_row, placex_table, osmline_table,
+def test_load_data(dsn, place_row, placex_table, osmline_table,
word_table, temp_db_cursor, threads):
for func in ('precompute_words', 'getorcreate_housenumber_id', 'make_standard_name'):
temp_db_cursor.execute("""CREATE FUNCTION {} (src TEXT)
assert temp_db_cursor.table_rows('placex') == 30
assert temp_db_cursor.table_rows('location_property_osmline') == 1
-
-
-@pytest.mark.parametrize("languages", (None, ' fr,en'))
-def test_create_country_names(temp_db_with_extensions, temp_db_conn, temp_db_cursor,
- table_factory, tokenizer_mock, languages):
-
- table_factory('country_name', 'country_code varchar(2), name hstore',
- content=(('us', '"name"=>"us1","name:af"=>"us2"'),
- ('fr', '"name"=>"Fra", "name:en"=>"Fren"')))
-
- assert temp_db_cursor.scalar("SELECT count(*) FROM country_name") == 2
-
- tokenizer = tokenizer_mock()
-
- database_import.create_country_names(temp_db_conn, tokenizer, languages)
-
- assert len(tokenizer.analyser_cache['countries']) == 2
-
- result_set = {k: set(v) for k, v in tokenizer.analyser_cache['countries']}
-
- if languages:
- assert result_set == {'us' : set(('us', 'us1', 'United States')),
- 'fr' : set(('fr', 'Fra', 'Fren'))}
- else:
- assert result_set == {'us' : set(('us', 'us1', 'us2', 'United States')),
- 'fr' : set(('fr', 'Fra', 'Fren'))}