From: Sarah Hoffmann Date: Mon, 8 Feb 2021 16:23:05 +0000 (+0100) Subject: split cli.py by subcommands X-Git-Tag: v3.7.0~38^2~19 X-Git-Url: https://git.openstreetmap.org./nominatim.git/commitdiff_plain/195f9f5ef3f79ea4c77ee1226f8aa7e7e8f4a73f?ds=sidebyside split cli.py by subcommands Reduces file size below 1000 lines. --- diff --git a/nominatim/cli.py b/nominatim/cli.py index 37bcaffb..0bca03a3 100644 --- a/nominatim/cli.py +++ b/nominatim/cli.py @@ -2,31 +2,19 @@ Command-line interface to the Nominatim functions for import, update, database administration and querying. """ -import datetime as dt +import logging import os -import socket import sys -import time import argparse -import logging from pathlib import Path from .config import Configuration -from .tools.exec_utils import run_legacy_script, run_api_script, run_php_server -from .db.connection import connect -from .db import status +from .tools.exec_utils import run_legacy_script, run_php_server from .errors import UsageError +from . import clicmd LOG = logging.getLogger() -def _num_system_cpus(): - try: - cpus = len(os.sched_getaffinity(0)) - except NotImplementedError: - cpus = None - - return cpus or os.cpu_count() - class CommandlineParser: """ Wraps some of the common functions for parsing the command line @@ -105,16 +93,6 @@ class CommandlineParser: return 1 -def _osm2pgsql_options_from_args(args, default_cache, default_threads): - """ Set up the stanadrd osm2pgsql from the command line arguments. - """ - return dict(osm2pgsql=args.osm2pgsql_path, - osm2pgsql_cache=args.osm2pgsql_cache or default_cache, - osm2pgsql_style=args.config.get_import_style_file(), - threads=args.threads or default_threads, - dsn=args.config.get_libpq_dsn(), - flatnode_file=args.config.FLATNODE_FILE) - ##### Subcommand classes # # Each class needs to implement two functions: add_args() adds the CLI parameters @@ -237,153 +215,6 @@ class SetupSpecialPhrases: return run_legacy_script('specialphrases.php', '--wiki-import', nominatim_env=args) -class UpdateReplication: - """\ - Update the database using an online replication service. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('Arguments for initialisation') - group.add_argument('--init', action='store_true', - help='Initialise the update process') - group.add_argument('--no-update-functions', dest='update_functions', - action='store_false', - help="""Do not update the trigger function to - support differential updates.""") - group = parser.add_argument_group('Arguments for updates') - group.add_argument('--check-for-updates', action='store_true', - help='Check if new updates are available and exit') - group.add_argument('--once', action='store_true', - help="""Download and apply updates only once. When - not set, updates are continuously applied""") - group.add_argument('--no-index', action='store_false', dest='do_index', - help="""Do not index the new data. Only applicable - together with --once""") - group.add_argument('--osm2pgsql-cache', metavar='SIZE', type=int, - help='Size of cache to be used by osm2pgsql (in MB)') - group = parser.add_argument_group('Download parameters') - group.add_argument('--socket-timeout', dest='socket_timeout', type=int, default=60, - help='Set timeout for file downloads.') - - @staticmethod - def _init_replication(args): - from .tools import replication, refresh - - socket.setdefaulttimeout(args.socket_timeout) - - LOG.warning("Initialising replication updates") - conn = connect(args.config.get_libpq_dsn()) - replication.init_replication(conn, base_url=args.config.REPLICATION_URL) - if args.update_functions: - LOG.warning("Create functions") - refresh.create_functions(conn, args.config, args.data_dir, - True, False) - conn.close() - return 0 - - - @staticmethod - def _check_for_updates(args): - from .tools import replication - - conn = connect(args.config.get_libpq_dsn()) - ret = replication.check_for_updates(conn, base_url=args.config.REPLICATION_URL) - conn.close() - return ret - - @staticmethod - def _report_update(batchdate, start_import, start_index): - def round_time(delta): - return dt.timedelta(seconds=int(delta.total_seconds())) - - end = dt.datetime.now(dt.timezone.utc) - LOG.warning("Update completed. Import: %s. %sTotal: %s. Remaining backlog: %s.", - round_time((start_index or end) - start_import), - "Indexing: {} ".format(round_time(end - start_index)) - if start_index else '', - round_time(end - start_import), - round_time(end - batchdate)) - - @staticmethod - def _update(args): - from .tools import replication - from .indexer.indexer import Indexer - - params = _osm2pgsql_options_from_args(args, 2000, 1) - params.update(base_url=args.config.REPLICATION_URL, - update_interval=args.config.get_int('REPLICATION_UPDATE_INTERVAL'), - import_file=args.project_dir / 'osmosischange.osc', - max_diff_size=args.config.get_int('REPLICATION_MAX_DIFF'), - indexed_only=not args.once) - - # Sanity check to not overwhelm the Geofabrik servers. - if 'download.geofabrik.de'in params['base_url']\ - and params['update_interval'] < 86400: - LOG.fatal("Update interval too low for download.geofabrik.de.\n" - "Please check install documentation " - "(https://nominatim.org/release-docs/latest/admin/Import-and-Update#" - "setting-up-the-update-process).") - raise UsageError("Invalid replication update interval setting.") - - if not args.once: - if not args.do_index: - LOG.fatal("Indexing cannot be disabled when running updates continuously.") - raise UsageError("Bad argument '--no-index'.") - recheck_interval = args.config.get_int('REPLICATION_RECHECK_INTERVAL') - - while True: - conn = connect(args.config.get_libpq_dsn()) - start = dt.datetime.now(dt.timezone.utc) - state = replication.update(conn, params) - if state is not replication.UpdateState.NO_CHANGES: - status.log_status(conn, start, 'import') - batchdate, _, _ = status.get_status(conn) - conn.close() - - if state is not replication.UpdateState.NO_CHANGES and args.do_index: - index_start = dt.datetime.now(dt.timezone.utc) - indexer = Indexer(args.config.get_libpq_dsn(), - args.threads or 1) - indexer.index_boundaries(0, 30) - indexer.index_by_rank(0, 30) - - conn = connect(args.config.get_libpq_dsn()) - status.set_indexed(conn, True) - status.log_status(conn, index_start, 'index') - conn.close() - else: - index_start = None - - if LOG.isEnabledFor(logging.WARNING): - UpdateReplication._report_update(batchdate, start, index_start) - - if args.once: - break - - if state is replication.UpdateState.NO_CHANGES: - LOG.warning("No new changes. Sleeping for %d sec.", recheck_interval) - time.sleep(recheck_interval) - - return state.value - - @staticmethod - def run(args): - try: - import osmium # pylint: disable=W0611 - except ModuleNotFoundError: - LOG.fatal("pyosmium not installed. Replication functions not available.\n" - "To install pyosmium via pip: pip3 install osmium") - return 1 - - if args.init: - return UpdateReplication._init_replication(args) - - if args.check_for_updates: - return UpdateReplication._check_for_updates(args) - - return UpdateReplication._update(args) - class UpdateAddData: """\ Add additional data from a file or an online source. @@ -434,118 +265,6 @@ class UpdateAddData: return run_legacy_script(*params, nominatim_env=args) -class UpdateIndex: - """\ - Reindex all new and modified data. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('Filter arguments') - group.add_argument('--boundaries-only', action='store_true', - help="""Index only administrative boundaries.""") - group.add_argument('--no-boundaries', action='store_true', - help="""Index everything except administrative boundaries.""") - group.add_argument('--minrank', '-r', type=int, metavar='RANK', default=0, - help='Minimum/starting rank') - group.add_argument('--maxrank', '-R', type=int, metavar='RANK', default=30, - help='Maximum/finishing rank') - - @staticmethod - def run(args): - from .indexer.indexer import Indexer - - indexer = Indexer(args.config.get_libpq_dsn(), - args.threads or _num_system_cpus() or 1) - - if not args.no_boundaries: - indexer.index_boundaries(args.minrank, args.maxrank) - if not args.boundaries_only: - indexer.index_by_rank(args.minrank, args.maxrank) - - if not args.no_boundaries and not args.boundaries_only \ - and args.minrank == 0 and args.maxrank == 30: - conn = connect(args.config.get_libpq_dsn()) - status.set_indexed(conn, True) - conn.close() - - return 0 - - -class UpdateRefresh: - """\ - Recompute auxiliary data used by the indexing process. - - These functions must not be run in parallel with other update commands. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('Data arguments') - group.add_argument('--postcodes', action='store_true', - help='Update postcode centroid table') - group.add_argument('--word-counts', action='store_true', - help='Compute frequency of full-word search terms') - group.add_argument('--address-levels', action='store_true', - help='Reimport address level configuration') - group.add_argument('--functions', action='store_true', - help='Update the PL/pgSQL functions in the database') - group.add_argument('--wiki-data', action='store_true', - help='Update Wikipedia/data importance numbers.') - group.add_argument('--importance', action='store_true', - help='Recompute place importances (expensive!)') - group.add_argument('--website', action='store_true', - help='Refresh the directory that serves the scripts for the web API') - group = parser.add_argument_group('Arguments for function refresh') - group.add_argument('--no-diff-updates', action='store_false', dest='diffs', - help='Do not enable code for propagating updates') - group.add_argument('--enable-debug-statements', action='store_true', - help='Enable debug warning statements in functions') - - @staticmethod - def run(args): - from .tools import refresh - - if args.postcodes: - LOG.warning("Update postcodes centroid") - conn = connect(args.config.get_libpq_dsn()) - refresh.update_postcodes(conn, args.data_dir) - conn.close() - - if args.word_counts: - LOG.warning('Recompute frequency of full-word search terms') - conn = connect(args.config.get_libpq_dsn()) - refresh.recompute_word_counts(conn, args.data_dir) - conn.close() - - if args.address_levels: - cfg = Path(args.config.ADDRESS_LEVEL_CONFIG) - LOG.warning('Updating address levels from %s', cfg) - conn = connect(args.config.get_libpq_dsn()) - refresh.load_address_levels_from_file(conn, cfg) - conn.close() - - if args.functions: - LOG.warning('Create functions') - conn = connect(args.config.get_libpq_dsn()) - refresh.create_functions(conn, args.config, args.data_dir, - args.diffs, args.enable_debug_statements) - conn.close() - - if args.wiki_data: - run_legacy_script('setup.php', '--import-wikipedia-articles', - nominatim_env=args, throw_on_fail=True) - # Attention: importance MUST come after wiki data import. - if args.importance: - run_legacy_script('update.php', '--recompute-importance', - nominatim_env=args, throw_on_fail=True) - if args.website: - run_legacy_script('setup.php', '--setup-website', - nominatim_env=args, throw_on_fail=True) - - return 0 - - class AdminCheckDatabase: """\ Check that the database is complete and operational. @@ -662,246 +381,6 @@ class AdminServe: def run(args): run_php_server(args.server, args.project_dir / 'website') -STRUCTURED_QUERY = ( - ('street', 'housenumber and street'), - ('city', 'city, town or village'), - ('county', 'county'), - ('state', 'state'), - ('country', 'country'), - ('postalcode', 'postcode') -) - -EXTRADATA_PARAMS = ( - ('addressdetails', 'Include a breakdown of the address into elements.'), - ('extratags', """Include additional information if available - (e.g. wikipedia link, opening hours)."""), - ('namedetails', 'Include a list of alternative names.') -) - -DETAILS_SWITCHES = ( - ('addressdetails', 'Include a breakdown of the address into elements.'), - ('keywords', 'Include a list of name keywords and address keywords.'), - ('linkedplaces', 'Include a details of places that are linked with this one.'), - ('hierarchy', 'Include details of places lower in the address hierarchy.'), - ('group_hierarchy', 'Group the places by type.'), - ('polygon_geojson', 'Include geometry of result.') -) - -def _add_api_output_arguments(parser): - group = parser.add_argument_group('Output arguments') - group.add_argument('--format', default='jsonv2', - choices=['xml', 'json', 'jsonv2', 'geojson', 'geocodejson'], - help='Format of result') - for name, desc in EXTRADATA_PARAMS: - group.add_argument('--' + name, action='store_true', help=desc) - - group.add_argument('--lang', '--accept-language', metavar='LANGS', - help='Preferred language order for presenting search results') - group.add_argument('--polygon-output', - choices=['geojson', 'kml', 'svg', 'text'], - help='Output geometry of results as a GeoJSON, KML, SVG or WKT.') - group.add_argument('--polygon-threshold', type=float, metavar='TOLERANCE', - help="""Simplify output geometry. - Parameter is difference tolerance in degrees.""") - - -class APISearch: - """\ - Execute API search query. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('Query arguments') - group.add_argument('--query', - help='Free-form query string') - for name, desc in STRUCTURED_QUERY: - group.add_argument('--' + name, help='Structured query: ' + desc) - - _add_api_output_arguments(parser) - - group = parser.add_argument_group('Result limitation') - group.add_argument('--countrycodes', metavar='CC,..', - help='Limit search results to one or more countries.') - group.add_argument('--exclude_place_ids', metavar='ID,..', - help='List of search object to be excluded') - group.add_argument('--limit', type=int, - help='Limit the number of returned results') - group.add_argument('--viewbox', metavar='X1,Y1,X2,Y2', - help='Preferred area to find search results') - group.add_argument('--bounded', action='store_true', - help='Strictly restrict results to viewbox area') - - group = parser.add_argument_group('Other arguments') - group.add_argument('--no-dedupe', action='store_false', dest='dedupe', - help='Do not remove duplicates from the result list') - - - @staticmethod - def run(args): - if args.query: - params = dict(q=args.query) - else: - params = {k : getattr(args, k) for k, _ in STRUCTURED_QUERY if getattr(args, k)} - - for param, _ in EXTRADATA_PARAMS: - if getattr(args, param): - params[param] = '1' - for param in ('format', 'countrycodes', 'exclude_place_ids', 'limit', 'viewbox'): - if getattr(args, param): - params[param] = getattr(args, param) - if args.lang: - params['accept-language'] = args.lang - if args.polygon_output: - params['polygon_' + args.polygon_output] = '1' - if args.polygon_threshold: - params['polygon_threshold'] = args.polygon_threshold - if args.bounded: - params['bounded'] = '1' - if not args.dedupe: - params['dedupe'] = '0' - - return run_api_script('search', args.project_dir, - phpcgi_bin=args.phpcgi_path, params=params) - -class APIReverse: - """\ - Execute API reverse query. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('Query arguments') - group.add_argument('--lat', type=float, required=True, - help='Latitude of coordinate to look up (in WGS84)') - group.add_argument('--lon', type=float, required=True, - help='Longitude of coordinate to look up (in WGS84)') - group.add_argument('--zoom', type=int, - help='Level of detail required for the address') - - _add_api_output_arguments(parser) - - - @staticmethod - def run(args): - params = dict(lat=args.lat, lon=args.lon) - if args.zoom is not None: - params['zoom'] = args.zoom - - for param, _ in EXTRADATA_PARAMS: - if getattr(args, param): - params[param] = '1' - if args.format: - params['format'] = args.format - if args.lang: - params['accept-language'] = args.lang - if args.polygon_output: - params['polygon_' + args.polygon_output] = '1' - if args.polygon_threshold: - params['polygon_threshold'] = args.polygon_threshold - - return run_api_script('reverse', args.project_dir, - phpcgi_bin=args.phpcgi_path, params=params) - - -class APILookup: - """\ - Execute API reverse query. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('Query arguments') - group.add_argument('--id', metavar='OSMID', - action='append', required=True, dest='ids', - help='OSM id to lookup in format (may be repeated)') - - _add_api_output_arguments(parser) - - - @staticmethod - def run(args): - params = dict(osm_ids=','.join(args.ids)) - - for param, _ in EXTRADATA_PARAMS: - if getattr(args, param): - params[param] = '1' - if args.format: - params['format'] = args.format - if args.lang: - params['accept-language'] = args.lang - if args.polygon_output: - params['polygon_' + args.polygon_output] = '1' - if args.polygon_threshold: - params['polygon_threshold'] = args.polygon_threshold - - return run_api_script('lookup', args.project_dir, - phpcgi_bin=args.phpcgi_path, params=params) - - -class APIDetails: - """\ - Execute API lookup query. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('Query arguments') - objs = group.add_mutually_exclusive_group(required=True) - objs.add_argument('--node', '-n', type=int, - help="Look up the OSM node with the given ID.") - objs.add_argument('--way', '-w', type=int, - help="Look up the OSM way with the given ID.") - objs.add_argument('--relation', '-r', type=int, - help="Look up the OSM relation with the given ID.") - objs.add_argument('--place_id', '-p', type=int, - help='Database internal identifier of the OSM object to look up.') - group.add_argument('--class', dest='object_class', - help="""Class type to disambiguated multiple entries - of the same object.""") - - group = parser.add_argument_group('Output arguments') - for name, desc in DETAILS_SWITCHES: - group.add_argument('--' + name, action='store_true', help=desc) - group.add_argument('--lang', '--accept-language', metavar='LANGS', - help='Preferred language order for presenting search results') - - @staticmethod - def run(args): - if args.node: - params = dict(osmtype='N', osmid=args.node) - elif args.way: - params = dict(osmtype='W', osmid=args.node) - elif args.relation: - params = dict(osmtype='R', osmid=args.node) - else: - params = dict(place_id=args.place_id) - if args.object_class: - params['class'] = args.object_class - for name, _ in DETAILS_SWITCHES: - params[name] = '1' if getattr(args, name) else '0' - - return run_api_script('details', args.project_dir, - phpcgi_bin=args.phpcgi_path, params=params) - - -class APIStatus: - """\ - Execute API status query. - """ - - @staticmethod - def add_args(parser): - group = parser.add_argument_group('API parameters') - group.add_argument('--format', default='text', choices=['text', 'json'], - help='Format of result') - - @staticmethod - def run(args): - return run_api_script('status', args.project_dir, - phpcgi_bin=args.phpcgi_path, - params=dict(format=args.format)) - def nominatim(**kwargs): """\ @@ -912,7 +391,7 @@ def nominatim(**kwargs): parser.add_subcommand('import', SetupAll) parser.add_subcommand('freeze', SetupFreeze) - parser.add_subcommand('replication', UpdateReplication) + parser.add_subcommand('replication', clicmd.UpdateReplication) parser.add_subcommand('check-database', AdminCheckDatabase) parser.add_subcommand('warm', AdminWarm) @@ -920,18 +399,18 @@ def nominatim(**kwargs): parser.add_subcommand('special-phrases', SetupSpecialPhrases) parser.add_subcommand('add-data', UpdateAddData) - parser.add_subcommand('index', UpdateIndex) - parser.add_subcommand('refresh', UpdateRefresh) + parser.add_subcommand('index', clicmd.UpdateIndex) + parser.add_subcommand('refresh', clicmd.UpdateRefresh) parser.add_subcommand('export', QueryExport) parser.add_subcommand('serve', AdminServe) if kwargs.get('phpcgi_path'): - parser.add_subcommand('search', APISearch) - parser.add_subcommand('reverse', APIReverse) - parser.add_subcommand('lookup', APILookup) - parser.add_subcommand('details', APIDetails) - parser.add_subcommand('status', APIStatus) + parser.add_subcommand('search', clicmd.APISearch) + parser.add_subcommand('reverse', clicmd.APIReverse) + parser.add_subcommand('lookup', clicmd.APILookup) + parser.add_subcommand('details', clicmd.APIDetails) + parser.add_subcommand('status', clicmd.APIStatus) else: parser.parser.epilog = 'php-cgi not found. Query commands not available.' diff --git a/nominatim/clicmd/__init__.py b/nominatim/clicmd/__init__.py new file mode 100644 index 00000000..b7dfa47f --- /dev/null +++ b/nominatim/clicmd/__init__.py @@ -0,0 +1,8 @@ +""" +Subcommand definitions for the command-line tool. +""" + +from .replication import UpdateReplication +from .api import APISearch, APIReverse, APILookup, APIDetails, APIStatus +from .index import UpdateIndex +from .refresh import UpdateRefresh diff --git a/nominatim/clicmd/api.py b/nominatim/clicmd/api.py new file mode 100644 index 00000000..e50c00dc --- /dev/null +++ b/nominatim/clicmd/api.py @@ -0,0 +1,251 @@ +""" +Subcommand definitions for API calls from the command line. +""" +import logging + +from ..tools.exec_utils import run_api_script + +# Do not repeat documentation of subcommand classes. +# pylint: disable=C0111 + +LOG = logging.getLogger() + +STRUCTURED_QUERY = ( + ('street', 'housenumber and street'), + ('city', 'city, town or village'), + ('county', 'county'), + ('state', 'state'), + ('country', 'country'), + ('postalcode', 'postcode') +) + +EXTRADATA_PARAMS = ( + ('addressdetails', 'Include a breakdown of the address into elements.'), + ('extratags', """Include additional information if available + (e.g. wikipedia link, opening hours)."""), + ('namedetails', 'Include a list of alternative names.') +) + +DETAILS_SWITCHES = ( + ('addressdetails', 'Include a breakdown of the address into elements.'), + ('keywords', 'Include a list of name keywords and address keywords.'), + ('linkedplaces', 'Include a details of places that are linked with this one.'), + ('hierarchy', 'Include details of places lower in the address hierarchy.'), + ('group_hierarchy', 'Group the places by type.'), + ('polygon_geojson', 'Include geometry of result.') +) + +def _add_api_output_arguments(parser): + group = parser.add_argument_group('Output arguments') + group.add_argument('--format', default='jsonv2', + choices=['xml', 'json', 'jsonv2', 'geojson', 'geocodejson'], + help='Format of result') + for name, desc in EXTRADATA_PARAMS: + group.add_argument('--' + name, action='store_true', help=desc) + + group.add_argument('--lang', '--accept-language', metavar='LANGS', + help='Preferred language order for presenting search results') + group.add_argument('--polygon-output', + choices=['geojson', 'kml', 'svg', 'text'], + help='Output geometry of results as a GeoJSON, KML, SVG or WKT.') + group.add_argument('--polygon-threshold', type=float, metavar='TOLERANCE', + help="""Simplify output geometry. + Parameter is difference tolerance in degrees.""") + + +class APISearch: + """\ + Execute API search query. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('Query arguments') + group.add_argument('--query', + help='Free-form query string') + for name, desc in STRUCTURED_QUERY: + group.add_argument('--' + name, help='Structured query: ' + desc) + + _add_api_output_arguments(parser) + + group = parser.add_argument_group('Result limitation') + group.add_argument('--countrycodes', metavar='CC,..', + help='Limit search results to one or more countries.') + group.add_argument('--exclude_place_ids', metavar='ID,..', + help='List of search object to be excluded') + group.add_argument('--limit', type=int, + help='Limit the number of returned results') + group.add_argument('--viewbox', metavar='X1,Y1,X2,Y2', + help='Preferred area to find search results') + group.add_argument('--bounded', action='store_true', + help='Strictly restrict results to viewbox area') + + group = parser.add_argument_group('Other arguments') + group.add_argument('--no-dedupe', action='store_false', dest='dedupe', + help='Do not remove duplicates from the result list') + + + @staticmethod + def run(args): + if args.query: + params = dict(q=args.query) + else: + params = {k : getattr(args, k) for k, _ in STRUCTURED_QUERY if getattr(args, k)} + + for param, _ in EXTRADATA_PARAMS: + if getattr(args, param): + params[param] = '1' + for param in ('format', 'countrycodes', 'exclude_place_ids', 'limit', 'viewbox'): + if getattr(args, param): + params[param] = getattr(args, param) + if args.lang: + params['accept-language'] = args.lang + if args.polygon_output: + params['polygon_' + args.polygon_output] = '1' + if args.polygon_threshold: + params['polygon_threshold'] = args.polygon_threshold + if args.bounded: + params['bounded'] = '1' + if not args.dedupe: + params['dedupe'] = '0' + + return run_api_script('search', args.project_dir, + phpcgi_bin=args.phpcgi_path, params=params) + +class APIReverse: + """\ + Execute API reverse query. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('Query arguments') + group.add_argument('--lat', type=float, required=True, + help='Latitude of coordinate to look up (in WGS84)') + group.add_argument('--lon', type=float, required=True, + help='Longitude of coordinate to look up (in WGS84)') + group.add_argument('--zoom', type=int, + help='Level of detail required for the address') + + _add_api_output_arguments(parser) + + + @staticmethod + def run(args): + params = dict(lat=args.lat, lon=args.lon) + if args.zoom is not None: + params['zoom'] = args.zoom + + for param, _ in EXTRADATA_PARAMS: + if getattr(args, param): + params[param] = '1' + if args.format: + params['format'] = args.format + if args.lang: + params['accept-language'] = args.lang + if args.polygon_output: + params['polygon_' + args.polygon_output] = '1' + if args.polygon_threshold: + params['polygon_threshold'] = args.polygon_threshold + + return run_api_script('reverse', args.project_dir, + phpcgi_bin=args.phpcgi_path, params=params) + + +class APILookup: + """\ + Execute API reverse query. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('Query arguments') + group.add_argument('--id', metavar='OSMID', + action='append', required=True, dest='ids', + help='OSM id to lookup in format (may be repeated)') + + _add_api_output_arguments(parser) + + + @staticmethod + def run(args): + params = dict(osm_ids=','.join(args.ids)) + + for param, _ in EXTRADATA_PARAMS: + if getattr(args, param): + params[param] = '1' + if args.format: + params['format'] = args.format + if args.lang: + params['accept-language'] = args.lang + if args.polygon_output: + params['polygon_' + args.polygon_output] = '1' + if args.polygon_threshold: + params['polygon_threshold'] = args.polygon_threshold + + return run_api_script('lookup', args.project_dir, + phpcgi_bin=args.phpcgi_path, params=params) + + +class APIDetails: + """\ + Execute API lookup query. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('Query arguments') + objs = group.add_mutually_exclusive_group(required=True) + objs.add_argument('--node', '-n', type=int, + help="Look up the OSM node with the given ID.") + objs.add_argument('--way', '-w', type=int, + help="Look up the OSM way with the given ID.") + objs.add_argument('--relation', '-r', type=int, + help="Look up the OSM relation with the given ID.") + objs.add_argument('--place_id', '-p', type=int, + help='Database internal identifier of the OSM object to look up.') + group.add_argument('--class', dest='object_class', + help="""Class type to disambiguated multiple entries + of the same object.""") + + group = parser.add_argument_group('Output arguments') + for name, desc in DETAILS_SWITCHES: + group.add_argument('--' + name, action='store_true', help=desc) + group.add_argument('--lang', '--accept-language', metavar='LANGS', + help='Preferred language order for presenting search results') + + @staticmethod + def run(args): + if args.node: + params = dict(osmtype='N', osmid=args.node) + elif args.way: + params = dict(osmtype='W', osmid=args.node) + elif args.relation: + params = dict(osmtype='R', osmid=args.node) + else: + params = dict(place_id=args.place_id) + if args.object_class: + params['class'] = args.object_class + for name, _ in DETAILS_SWITCHES: + params[name] = '1' if getattr(args, name) else '0' + + return run_api_script('details', args.project_dir, + phpcgi_bin=args.phpcgi_path, params=params) + + +class APIStatus: + """\ + Execute API status query. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('API parameters') + group.add_argument('--format', default='text', choices=['text', 'json'], + help='Format of result') + + @staticmethod + def run(args): + return run_api_script('status', args.project_dir, + phpcgi_bin=args.phpcgi_path, + params=dict(format=args.format)) diff --git a/nominatim/clicmd/index.py b/nominatim/clicmd/index.py new file mode 100644 index 00000000..ca3f9dee --- /dev/null +++ b/nominatim/clicmd/index.py @@ -0,0 +1,58 @@ +""" +Implementation of the 'index' subcommand. +""" +import os + +from ..db import status +from ..db.connection import connect + +# Do not repeat documentation of subcommand classes. +# pylint: disable=C0111 +# Using non-top-level imports to avoid eventually unused imports. +# pylint: disable=E0012,C0415 + +def _num_system_cpus(): + try: + cpus = len(os.sched_getaffinity(0)) + except NotImplementedError: + cpus = None + + return cpus or os.cpu_count() + + +class UpdateIndex: + """\ + Reindex all new and modified data. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('Filter arguments') + group.add_argument('--boundaries-only', action='store_true', + help="""Index only administrative boundaries.""") + group.add_argument('--no-boundaries', action='store_true', + help="""Index everything except administrative boundaries.""") + group.add_argument('--minrank', '-r', type=int, metavar='RANK', default=0, + help='Minimum/starting rank') + group.add_argument('--maxrank', '-R', type=int, metavar='RANK', default=30, + help='Maximum/finishing rank') + + @staticmethod + def run(args): + from ..indexer.indexer import Indexer + + indexer = Indexer(args.config.get_libpq_dsn(), + args.threads or _num_system_cpus() or 1) + + if not args.no_boundaries: + indexer.index_boundaries(args.minrank, args.maxrank) + if not args.boundaries_only: + indexer.index_by_rank(args.minrank, args.maxrank) + + if not args.no_boundaries and not args.boundaries_only \ + and args.minrank == 0 and args.maxrank == 30: + conn = connect(args.config.get_libpq_dsn()) + status.set_indexed(conn, True) + conn.close() + + return 0 diff --git a/nominatim/clicmd/refresh.py b/nominatim/clicmd/refresh.py new file mode 100644 index 00000000..4e2ad1db --- /dev/null +++ b/nominatim/clicmd/refresh.py @@ -0,0 +1,88 @@ +""" +Implementation of 'refresh' subcommand. +""" +import logging +from pathlib import Path + +from ..db.connection import connect +from ..tools.exec_utils import run_legacy_script + +# Do not repeat documentation of subcommand classes. +# pylint: disable=C0111 +# Using non-top-level imports to avoid eventually unused imports. +# pylint: disable=E0012,C0415 + +LOG = logging.getLogger() + +class UpdateRefresh: + """\ + Recompute auxiliary data used by the indexing process. + + These functions must not be run in parallel with other update commands. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('Data arguments') + group.add_argument('--postcodes', action='store_true', + help='Update postcode centroid table') + group.add_argument('--word-counts', action='store_true', + help='Compute frequency of full-word search terms') + group.add_argument('--address-levels', action='store_true', + help='Reimport address level configuration') + group.add_argument('--functions', action='store_true', + help='Update the PL/pgSQL functions in the database') + group.add_argument('--wiki-data', action='store_true', + help='Update Wikipedia/data importance numbers.') + group.add_argument('--importance', action='store_true', + help='Recompute place importances (expensive!)') + group.add_argument('--website', action='store_true', + help='Refresh the directory that serves the scripts for the web API') + group = parser.add_argument_group('Arguments for function refresh') + group.add_argument('--no-diff-updates', action='store_false', dest='diffs', + help='Do not enable code for propagating updates') + group.add_argument('--enable-debug-statements', action='store_true', + help='Enable debug warning statements in functions') + + @staticmethod + def run(args): + from ..tools import refresh + + if args.postcodes: + LOG.warning("Update postcodes centroid") + conn = connect(args.config.get_libpq_dsn()) + refresh.update_postcodes(conn, args.data_dir) + conn.close() + + if args.word_counts: + LOG.warning('Recompute frequency of full-word search terms') + conn = connect(args.config.get_libpq_dsn()) + refresh.recompute_word_counts(conn, args.data_dir) + conn.close() + + if args.address_levels: + cfg = Path(args.config.ADDRESS_LEVEL_CONFIG) + LOG.warning('Updating address levels from %s', cfg) + conn = connect(args.config.get_libpq_dsn()) + refresh.load_address_levels_from_file(conn, cfg) + conn.close() + + if args.functions: + LOG.warning('Create functions') + conn = connect(args.config.get_libpq_dsn()) + refresh.create_functions(conn, args.config, args.data_dir, + args.diffs, args.enable_debug_statements) + conn.close() + + if args.wiki_data: + run_legacy_script('setup.php', '--import-wikipedia-articles', + nominatim_env=args, throw_on_fail=True) + # Attention: importance MUST come after wiki data import. + if args.importance: + run_legacy_script('update.php', '--recompute-importance', + nominatim_env=args, throw_on_fail=True) + if args.website: + run_legacy_script('setup.php', '--setup-website', + nominatim_env=args, throw_on_fail=True) + + return 0 diff --git a/nominatim/clicmd/replication.py b/nominatim/clicmd/replication.py new file mode 100644 index 00000000..554dbc4f --- /dev/null +++ b/nominatim/clicmd/replication.py @@ -0,0 +1,170 @@ +""" +Implementation of the 'replication' sub-command. +""" +import datetime as dt +import logging +import socket +import time + +from ..db import status +from ..db.connection import connect +from ..errors import UsageError + +LOG = logging.getLogger() + +# Do not repeat documentation of subcommand classes. +# pylint: disable=C0111 +# Using non-top-level imports to make pyosmium optional for replication only. +# pylint: disable=E0012,C0415 + +def _osm2pgsql_options_from_args(args, default_cache, default_threads): + """ Set up the standard osm2pgsql from the command line arguments. + """ + return dict(osm2pgsql=args.osm2pgsql_path, + osm2pgsql_cache=args.osm2pgsql_cache or default_cache, + osm2pgsql_style=args.config.get_import_style_file(), + threads=args.threads or default_threads, + dsn=args.config.get_libpq_dsn(), + flatnode_file=args.config.FLATNODE_FILE) + + +class UpdateReplication: + """\ + Update the database using an online replication service. + """ + + @staticmethod + def add_args(parser): + group = parser.add_argument_group('Arguments for initialisation') + group.add_argument('--init', action='store_true', + help='Initialise the update process') + group.add_argument('--no-update-functions', dest='update_functions', + action='store_false', + help="""Do not update the trigger function to + support differential updates.""") + group = parser.add_argument_group('Arguments for updates') + group.add_argument('--check-for-updates', action='store_true', + help='Check if new updates are available and exit') + group.add_argument('--once', action='store_true', + help="""Download and apply updates only once. When + not set, updates are continuously applied""") + group.add_argument('--no-index', action='store_false', dest='do_index', + help="""Do not index the new data. Only applicable + together with --once""") + group.add_argument('--osm2pgsql-cache', metavar='SIZE', type=int, + help='Size of cache to be used by osm2pgsql (in MB)') + group = parser.add_argument_group('Download parameters') + group.add_argument('--socket-timeout', dest='socket_timeout', type=int, default=60, + help='Set timeout for file downloads.') + + @staticmethod + def _init_replication(args): + from ..tools import replication, refresh + + LOG.warning("Initialising replication updates") + conn = connect(args.config.get_libpq_dsn()) + replication.init_replication(conn, base_url=args.config.REPLICATION_URL) + if args.update_functions: + LOG.warning("Create functions") + refresh.create_functions(conn, args.config, args.data_dir, + True, False) + conn.close() + return 0 + + + @staticmethod + def _check_for_updates(args): + from ..tools import replication + + conn = connect(args.config.get_libpq_dsn()) + ret = replication.check_for_updates(conn, base_url=args.config.REPLICATION_URL) + conn.close() + return ret + + @staticmethod + def _report_update(batchdate, start_import, start_index): + def round_time(delta): + return dt.timedelta(seconds=int(delta.total_seconds())) + + end = dt.datetime.now(dt.timezone.utc) + LOG.warning("Update completed. Import: %s. %sTotal: %s. Remaining backlog: %s.", + round_time((start_index or end) - start_import), + "Indexing: {} ".format(round_time(end - start_index)) + if start_index else '', + round_time(end - start_import), + round_time(end - batchdate)) + + @staticmethod + def _update(args): + from ..tools import replication + from ..indexer.indexer import Indexer + + params = _osm2pgsql_options_from_args(args, 2000, 1) + params.update(base_url=args.config.REPLICATION_URL, + update_interval=args.config.get_int('REPLICATION_UPDATE_INTERVAL'), + import_file=args.project_dir / 'osmosischange.osc', + max_diff_size=args.config.get_int('REPLICATION_MAX_DIFF'), + indexed_only=not args.once) + + # Sanity check to not overwhelm the Geofabrik servers. + if 'download.geofabrik.de'in params['base_url']\ + and params['update_interval'] < 86400: + LOG.fatal("Update interval too low for download.geofabrik.de.\n" + "Please check install documentation " + "(https://nominatim.org/release-docs/latest/admin/Import-and-Update#" + "setting-up-the-update-process).") + raise UsageError("Invalid replication update interval setting.") + + if not args.once: + if not args.do_index: + LOG.fatal("Indexing cannot be disabled when running updates continuously.") + raise UsageError("Bad argument '--no-index'.") + recheck_interval = args.config.get_int('REPLICATION_RECHECK_INTERVAL') + + while True: + conn = connect(args.config.get_libpq_dsn()) + start = dt.datetime.now(dt.timezone.utc) + state = replication.update(conn, params) + if state is not replication.UpdateState.NO_CHANGES: + status.log_status(conn, start, 'import') + batchdate, _, _ = status.get_status(conn) + conn.close() + + if state is not replication.UpdateState.NO_CHANGES and args.do_index: + index_start = dt.datetime.now(dt.timezone.utc) + indexer = Indexer(args.config.get_libpq_dsn(), + args.threads or 1) + indexer.index_boundaries(0, 30) + indexer.index_by_rank(0, 30) + + conn = connect(args.config.get_libpq_dsn()) + status.set_indexed(conn, True) + status.log_status(conn, index_start, 'index') + conn.close() + else: + index_start = None + + if LOG.isEnabledFor(logging.WARNING): + UpdateReplication._report_update(batchdate, start, index_start) + + if args.once: + break + + if state is replication.UpdateState.NO_CHANGES: + LOG.warning("No new changes. Sleeping for %d sec.", recheck_interval) + time.sleep(recheck_interval) + + return state.value + + + @staticmethod + def run(args): + socket.setdefaulttimeout(args.socket_timeout) + + if args.init: + return UpdateReplication._init_replication(args) + + if args.check_for_updates: + return UpdateReplication._check_for_updates(args) + + return UpdateReplication._update(args) diff --git a/nominatim/tools/replication.py b/nominatim/tools/replication.py index afc1af47..cb201b1e 100644 --- a/nominatim/tools/replication.py +++ b/nominatim/tools/replication.py @@ -6,13 +6,18 @@ from enum import Enum import logging import time -from osmium.replication.server import ReplicationServer -from osmium import WriteHandler - from ..db import status from .exec_utils import run_osm2pgsql from ..errors import UsageError +try: + from osmium.replication.server import ReplicationServer + from osmium import WriteHandler +except ModuleNotFoundError as exc: + logging.getLogger().fatal("pyosmium not installed. Replication functions not available.\n" + "To install pyosmium via pip: pip3 install osmium") + raise UsageError("replication tools not available") from exc + LOG = logging.getLogger() def init_replication(conn, base_url): diff --git a/test/python/test_cli.py b/test/python/test_cli.py index 702a4b74..983b792b 100644 --- a/test/python/test_cli.py +++ b/test/python/test_cli.py @@ -11,6 +11,8 @@ import pytest import time import nominatim.cli +import nominatim.clicmd.api +import nominatim.clicmd.refresh import nominatim.indexer.indexer import nominatim.tools.refresh import nominatim.tools.replication @@ -45,12 +47,6 @@ def mock_run_legacy(monkeypatch): monkeypatch.setattr(nominatim.cli, 'run_legacy_script', mock) return mock -@pytest.fixture -def mock_run_api(monkeypatch): - mock = MockParamCapture() - monkeypatch.setattr(nominatim.cli, 'run_api_script', mock) - return mock - def test_cli_help(capsys): """ Running nominatim tool without arguments prints help. @@ -110,7 +106,10 @@ def test_index_command(monkeypatch, temp_db_cursor, params, do_bnds, do_ranks): ('importance', ('update.php', '--recompute-importance')), ('website', ('setup.php', '--setup-website')), ]) -def test_refresh_legacy_command(mock_run_legacy, temp_db, command, params): +def test_refresh_legacy_command(monkeypatch, temp_db, command, params): + mock_run_legacy = MockParamCapture() + monkeypatch.setattr(nominatim.clicmd.refresh, 'run_legacy_script', mock_run_legacy) + assert 0 == call_nominatim('refresh', '--' + command) assert mock_run_legacy.called == 1 @@ -131,7 +130,10 @@ def test_refresh_command(monkeypatch, temp_db, command, func): assert func_mock.called == 1 -def test_refresh_importance_computed_after_wiki_import(mock_run_legacy, temp_db): +def test_refresh_importance_computed_after_wiki_import(monkeypatch, temp_db): + mock_run_legacy = MockParamCapture() + monkeypatch.setattr(nominatim.clicmd.refresh, 'run_legacy_script', mock_run_legacy) + assert 0 == call_nominatim('refresh', '--importance', '--wiki-data') assert mock_run_legacy.called == 2 @@ -233,7 +235,10 @@ def test_serve_command(monkeypatch): ('details', '--place_id', '10001'), ('status',) ]) -def test_api_commands_simple(mock_run_api, params): +def test_api_commands_simple(monkeypatch, params): + mock_run_api = MockParamCapture() + monkeypatch.setattr(nominatim.clicmd.api, 'run_api_script', mock_run_api) + assert 0 == call_nominatim(*params) assert mock_run_api.called == 1