X-Git-Url: https://git.openstreetmap.org./nominatim.git/blobdiff_plain/79da96b369aa86fdcec21c4d0eb8465b6ed54fc5..f3c557bf684a0079e4bc54b622cc5d766f3a6b56:/nominatim/clicmd/setup.py?ds=sidebyside diff --git a/nominatim/clicmd/setup.py b/nominatim/clicmd/setup.py index 014c9dc9..73095468 100644 --- a/nominatim/clicmd/setup.py +++ b/nominatim/clicmd/setup.py @@ -1,3 +1,9 @@ +# SPDX-License-Identifier: GPL-2.0-only +# +# This file is part of Nominatim. (https://nominatim.org) +# +# Copyright (C) 2022 by the Nominatim developer community. +# For a full list of authors see the git log. """ Implementation of the 'import' subcommand. """ @@ -8,18 +14,22 @@ import psutil from nominatim.db.connection import connect from nominatim.db import status, properties -from nominatim.version import NOMINATIM_VERSION +from nominatim.version import version_str # Do not repeat documentation of subcommand classes. # pylint: disable=C0111 # Using non-top-level imports to avoid eventually unused imports. -# pylint: disable=E0012,C0415 +# pylint: disable=C0415 LOG = logging.getLogger() class SetupAll: """\ Create a new Nominatim database from an OSM file. + + This sub-command sets up a new Nominatim database from scratch starting + with creating a new database in Postgresql. The user running this command + needs superuser rights on the database. """ @staticmethod @@ -28,7 +38,7 @@ class SetupAll: group = group_name.add_mutually_exclusive_group(required=True) group.add_argument('--osm-file', metavar='FILE', action='append', help='OSM file to be imported' - ' (repeat for importing multiple files.') + ' (repeat for importing multiple files)') group.add_argument('--continue', dest='continue_at', choices=['load-data', 'indexing', 'db-postprocess'], help='Continue an import that was interrupted') @@ -43,19 +53,22 @@ class SetupAll: group.add_argument('--no-updates', action='store_true', help="Do not keep tables that are only needed for " "updating the database later") + group.add_argument('--offline', action='store_true', + help="Do not attempt to load any additional data from the internet") group = parser.add_argument_group('Expert options') group.add_argument('--ignore-errors', action='store_true', help='Continue import even when errors in SQL are present') group.add_argument('--index-noanalyse', action='store_true', - help='Do not perform analyse operations during index') + help='Do not perform analyse operations during index (expert only)') @staticmethod - def run(args): - from ..tools import database_import, refresh, postcodes, freeze, country_info + def run(args): # pylint: disable=too-many-statements + from ..data import country_info + from ..tools import database_import, refresh, postcodes, freeze from ..indexer.indexer import Indexer - country_info.setup_country_config(args.config.config_dir / 'country_settings.yaml') + country_info.setup_country_config(args.config) if args.continue_at is None: files = args.get_osm_file_list() @@ -116,22 +129,20 @@ class SetupAll: drop=args.no_updates) LOG.warning('Create search index for default country names.') country_info.create_country_names(conn, tokenizer, - args.config.LANGUAGES) - conn.commit() + args.config.get_str_list('LANGUAGES')) if args.no_updates: freeze.drop_update_tables(conn) tokenizer.finalize_import(args.config) + LOG.warning('Recompute word counts') + tokenizer.update_statistics() webdir = args.project_dir / 'website' LOG.warning('Setup website at %s', webdir) with connect(args.config.get_libpq_dsn()) as conn: refresh.setup_website(webdir, args.config, conn) - with connect(args.config.get_libpq_dsn()) as conn: - SetupAll._set_database_date(conn) - properties.set_property(conn, 'database_version', - '{0[0]}.{0[1]}.{0[2]}-{0[3]}'.format(NOMINATIM_VERSION)) + SetupAll._finalize_database(args.config.get_libpq_dsn(), args.offline) return 0 @@ -147,7 +158,7 @@ class SetupAll: refresh.create_functions(conn, config, False, False) LOG.warning('Create tables') database_import.create_tables(conn, config, reverse_only=reverse_only) - refresh.load_address_levels_from_file(conn, Path(config.ADDRESS_LEVEL_CONFIG)) + refresh.load_address_levels_from_config(conn, config) LOG.warning('Create functions (2nd pass)') refresh.create_functions(conn, config, False, False) LOG.warning('Create table triggers') @@ -186,20 +197,24 @@ class SetupAll: LOG.warning('Creating support index') if tablespace: tablespace = 'TABLESPACE ' + tablespace - cur.execute("""CREATE INDEX idx_placex_pendingsector - ON placex USING BTREE (rank_address,geometry_sector) - {} WHERE indexed_status > 0 - """.format(tablespace)) + cur.execute(f"""CREATE INDEX idx_placex_pendingsector + ON placex USING BTREE (rank_address,geometry_sector) + {tablespace} WHERE indexed_status > 0 + """) conn.commit() @staticmethod - def _set_database_date(conn): + def _finalize_database(dsn, offline): """ Determine the database date and set the status accordingly. """ - try: - dbdate = status.compute_database_date(conn) - status.set_status(conn, dbdate) - LOG.info('Database is at %s.', dbdate) - except Exception as exc: # pylint: disable=broad-except - LOG.error('Cannot determine date of database: %s', exc) + with connect(dsn) as conn: + if not offline: + try: + dbdate = status.compute_database_date(conn) + status.set_status(conn, dbdate) + LOG.info('Database is at %s.', dbdate) + except Exception as exc: # pylint: disable=broad-except + LOG.error('Cannot determine date of database: %s', exc) + + properties.set_property(conn, 'database_version', version_str())