X-Git-Url: https://git.openstreetmap.org./nominatim.git/blobdiff_plain/077a8c1f95d2c0a18722f705df4b98d83d6f9385..c4fd94bd1a778785e67e28d9e77044c6b09a07ff:/nominatim/tools/database_import.py diff --git a/nominatim/tools/database_import.py b/nominatim/tools/database_import.py index 4ba4493c..25efedb9 100644 --- a/nominatim/tools/database_import.py +++ b/nominatim/tools/database_import.py @@ -11,13 +11,13 @@ from pathlib import Path import psutil import psycopg2 -from ..db.connection import connect, get_pg_env -from ..db import utils as db_utils -from ..db.async_connection import DBConnection -from ..db.sql_preprocessor import SQLPreprocessor -from .exec_utils import run_osm2pgsql -from ..errors import UsageError -from ..version import POSTGRESQL_REQUIRED_VERSION, POSTGIS_REQUIRED_VERSION +from nominatim.db.connection import connect, get_pg_env +from nominatim.db import utils as db_utils +from nominatim.db.async_connection import DBConnection +from nominatim.db.sql_preprocessor import SQLPreprocessor +from nominatim.tools.exec_utils import run_osm2pgsql +from nominatim.errors import UsageError +from nominatim.version import POSTGRESQL_REQUIRED_VERSION, POSTGIS_REQUIRED_VERSION LOG = logging.getLogger() @@ -179,29 +179,29 @@ def import_osm_data(osm_file, options, drop=False, ignore_errors=False): Path(options['flatnode_file']).unlink() -def create_tables(conn, config, sqllib_dir, reverse_only=False): +def create_tables(conn, config, reverse_only=False): """ Create the set of basic tables. When `reverse_only` is True, then the main table for searching will be skipped and only reverse search is possible. """ - sql = SQLPreprocessor(conn, config, sqllib_dir) + sql = SQLPreprocessor(conn, config) sql.env.globals['db']['reverse_only'] = reverse_only sql.run_sql_file(conn, 'tables.sql') -def create_table_triggers(conn, config, sqllib_dir): +def create_table_triggers(conn, config): """ Create the triggers for the tables. The trigger functions must already have been imported with refresh.create_functions(). """ - sql = SQLPreprocessor(conn, config, sqllib_dir) + sql = SQLPreprocessor(conn, config) sql.run_sql_file(conn, 'table-triggers.sql') -def create_partition_tables(conn, config, sqllib_dir): +def create_partition_tables(conn, config): """ Create tables that have explicit partitioning. """ - sql = SQLPreprocessor(conn, config, sqllib_dir) + sql = SQLPreprocessor(conn, config) sql.run_sql_file(conn, 'partition-tables.src.sql') @@ -287,7 +287,7 @@ def load_data(dsn, data_dir, threads): cur.execute('ANALYSE') -def create_search_indices(conn, config, sqllib_dir, drop=False): +def create_search_indices(conn, config, drop=False): """ Create tables that have explicit partitioning. """ @@ -303,7 +303,7 @@ def create_search_indices(conn, config, sqllib_dir, drop=False): cur.execute('DROP INDEX "{}"'.format(idx)) conn.commit() - sql = SQLPreprocessor(conn, config, sqllib_dir) + sql = SQLPreprocessor(conn, config) sql.run_sql_file(conn, 'indices.sql', drop=drop) @@ -330,7 +330,7 @@ def create_country_names(conn, config): sql_statement = "{} IN (".format(sql_statement) delim = '' for language in languages.split(','): - sql_statement = "{}{}'name:{}'".format(sql_statement,delim,language) + sql_statement = "{}{}'name:{}'".format(sql_statement, delim, language) delim = ', ' sql_statement = '{})'.format(sql_statement) else: