X-Git-Url: https://git.openstreetmap.org./nominatim.git/blobdiff_plain/195f9f5ef3f79ea4c77ee1226f8aa7e7e8f4a73f..79d55357e89c6f98e3ec9240d40fcebe6cd31054:/nominatim/clicmd/replication.py diff --git a/nominatim/clicmd/replication.py b/nominatim/clicmd/replication.py index 554dbc4f..c75322d9 100644 --- a/nominatim/clicmd/replication.py +++ b/nominatim/clicmd/replication.py @@ -6,9 +6,9 @@ import logging import socket import time -from ..db import status -from ..db.connection import connect -from ..errors import UsageError +from nominatim.db import status +from nominatim.db.connection import connect +from nominatim.errors import UsageError LOG = logging.getLogger() @@ -17,17 +17,6 @@ LOG = logging.getLogger() # Using non-top-level imports to make pyosmium optional for replication only. # pylint: disable=E0012,C0415 -def _osm2pgsql_options_from_args(args, default_cache, default_threads): - """ Set up the standard osm2pgsql from the command line arguments. - """ - return dict(osm2pgsql=args.osm2pgsql_path, - osm2pgsql_cache=args.osm2pgsql_cache or default_cache, - osm2pgsql_style=args.config.get_import_style_file(), - threads=args.threads or default_threads, - dsn=args.config.get_libpq_dsn(), - flatnode_file=args.config.FLATNODE_FILE) - - class UpdateReplication: """\ Update the database using an online replication service. @@ -40,17 +29,17 @@ class UpdateReplication: help='Initialise the update process') group.add_argument('--no-update-functions', dest='update_functions', action='store_false', - help="""Do not update the trigger function to - support differential updates.""") + help=("Do not update the trigger function to " + "support differential updates.")) group = parser.add_argument_group('Arguments for updates') group.add_argument('--check-for-updates', action='store_true', help='Check if new updates are available and exit') group.add_argument('--once', action='store_true', - help="""Download and apply updates only once. When - not set, updates are continuously applied""") + help=("Download and apply updates only once. When " + "not set, updates are continuously applied")) group.add_argument('--no-index', action='store_false', dest='do_index', - help="""Do not index the new data. Only applicable - together with --once""") + help=("Do not index the new data. Only applicable " + "together with --once")) group.add_argument('--osm2pgsql-cache', metavar='SIZE', type=int, help='Size of cache to be used by osm2pgsql (in MB)') group = parser.add_argument_group('Download parameters') @@ -62,13 +51,11 @@ class UpdateReplication: from ..tools import replication, refresh LOG.warning("Initialising replication updates") - conn = connect(args.config.get_libpq_dsn()) - replication.init_replication(conn, base_url=args.config.REPLICATION_URL) - if args.update_functions: - LOG.warning("Create functions") - refresh.create_functions(conn, args.config, args.data_dir, - True, False) - conn.close() + with connect(args.config.get_libpq_dsn()) as conn: + replication.init_replication(conn, base_url=args.config.REPLICATION_URL) + if args.update_functions: + LOG.warning("Create functions") + refresh.create_functions(conn, args.config, True, False) return 0 @@ -76,10 +63,8 @@ class UpdateReplication: def _check_for_updates(args): from ..tools import replication - conn = connect(args.config.get_libpq_dsn()) - ret = replication.check_for_updates(conn, base_url=args.config.REPLICATION_URL) - conn.close() - return ret + with connect(args.config.get_libpq_dsn()) as conn: + return replication.check_for_updates(conn, base_url=args.config.REPLICATION_URL) @staticmethod def _report_update(batchdate, start_import, start_index): @@ -99,7 +84,7 @@ class UpdateReplication: from ..tools import replication from ..indexer.indexer import Indexer - params = _osm2pgsql_options_from_args(args, 2000, 1) + params = args.osm2pgsql_options(default_cache=2000, default_threads=1) params.update(base_url=args.config.REPLICATION_URL, update_interval=args.config.get_int('REPLICATION_UPDATE_INTERVAL'), import_file=args.project_dir / 'osmosischange.osc', @@ -122,13 +107,12 @@ class UpdateReplication: recheck_interval = args.config.get_int('REPLICATION_RECHECK_INTERVAL') while True: - conn = connect(args.config.get_libpq_dsn()) - start = dt.datetime.now(dt.timezone.utc) - state = replication.update(conn, params) - if state is not replication.UpdateState.NO_CHANGES: - status.log_status(conn, start, 'import') - batchdate, _, _ = status.get_status(conn) - conn.close() + with connect(args.config.get_libpq_dsn()) as conn: + start = dt.datetime.now(dt.timezone.utc) + state = replication.update(conn, params) + if state is not replication.UpdateState.NO_CHANGES: + status.log_status(conn, start, 'import') + batchdate, _, _ = status.get_status(conn) if state is not replication.UpdateState.NO_CHANGES and args.do_index: index_start = dt.datetime.now(dt.timezone.utc) @@ -137,10 +121,9 @@ class UpdateReplication: indexer.index_boundaries(0, 30) indexer.index_by_rank(0, 30) - conn = connect(args.config.get_libpq_dsn()) - status.set_indexed(conn, True) - status.log_status(conn, index_start, 'index') - conn.close() + with connect(args.config.get_libpq_dsn()) as conn: + status.set_indexed(conn, True) + status.log_status(conn, index_start, 'index') else: index_start = None @@ -154,8 +137,6 @@ class UpdateReplication: LOG.warning("No new changes. Sleeping for %d sec.", recheck_interval) time.sleep(recheck_interval) - return state.value - @staticmethod def run(args): @@ -167,4 +148,5 @@ class UpdateReplication: if args.check_for_updates: return UpdateReplication._check_for_updates(args) - return UpdateReplication._update(args) + UpdateReplication._update(args) + return 0