X-Git-Url: https://git.openstreetmap.org./nominatim.git/blobdiff_plain/f5e56d1ef1e2631c67cc95bd124403410248fbdd..452021ef0c9ac746949a2ef7fd12db4c2d5fee35:/nominatim/clicmd/refresh.py diff --git a/nominatim/clicmd/refresh.py b/nominatim/clicmd/refresh.py index e696e7b6..82a61f54 100644 --- a/nominatim/clicmd/refresh.py +++ b/nominatim/clicmd/refresh.py @@ -17,8 +17,16 @@ class UpdateRefresh: """\ Recompute auxiliary data used by the indexing process. - These functions must not be run in parallel with other update commands. + This sub-commands updates various static data and functions in the database. + It usually needs to be run after changing various aspects of the + configuration. The configuration documentation will mention the exact + command to use in such case. + + Warning: the 'update' command must not be run in parallel with other update + commands like 'replication' or 'add-data'. """ + def __init__(self): + self.tokenizer = None @staticmethod def add_args(parser): @@ -32,7 +40,7 @@ class UpdateRefresh: group.add_argument('--functions', action='store_true', help='Update the PL/pgSQL functions in the database') group.add_argument('--wiki-data', action='store_true', - help='Update Wikipedia/data importance numbers.') + help='Update Wikipedia/data importance numbers') group.add_argument('--importance', action='store_true', help='Recompute place importances (expensive!)') group.add_argument('--website', action='store_true', @@ -43,38 +51,40 @@ class UpdateRefresh: group.add_argument('--enable-debug-statements', action='store_true', help='Enable debug warning statements in functions') - @staticmethod - def run(args): + + def run(self, args): from ..tools import refresh, postcodes - from ..tokenizer import factory as tokenizer_factory from ..indexer.indexer import Indexer - tokenizer = tokenizer_factory.get_tokenizer_for_db(args.config) if args.postcodes: - LOG.warning("Update postcodes centroid") - postcodes.update_postcodes(args.config.get_libpq_dsn(), - args.project_dir, tokenizer) - indexer = Indexer(args.config.get_libpq_dsn(), tokenizer, - args.threads or 1) - indexer.index_postcodes() + if postcodes.can_compute(args.config.get_libpq_dsn()): + LOG.warning("Update postcodes centroid") + tokenizer = self._get_tokenizer(args.config) + postcodes.update_postcodes(args.config.get_libpq_dsn(), + args.project_dir, tokenizer) + indexer = Indexer(args.config.get_libpq_dsn(), tokenizer, + args.threads or 1) + indexer.index_postcodes() + else: + LOG.error("The place table doesn't exist. " + "Postcode updates on a frozen database is not possible.") if args.word_counts: - LOG.warning('Recompute frequency of full-word search terms') - refresh.recompute_word_counts(args.config.get_libpq_dsn(), args.sqllib_dir) + LOG.warning('Recompute word statistics') + self._get_tokenizer(args.config).update_statistics() if args.address_levels: - cfg = Path(args.config.ADDRESS_LEVEL_CONFIG) - LOG.warning('Updating address levels from %s', cfg) + LOG.warning('Updating address levels') with connect(args.config.get_libpq_dsn()) as conn: - refresh.load_address_levels_from_file(conn, cfg) + refresh.load_address_levels_from_config(conn, args.config) if args.functions: LOG.warning('Create functions') with connect(args.config.get_libpq_dsn()) as conn: refresh.create_functions(conn, args.config, args.diffs, args.enable_debug_statements) - tokenizer.update_sql_functions(args.config) + self._get_tokenizer(args.config).update_sql_functions(args.config) if args.wiki_data: data_path = Path(args.config.WIKIPEDIA_DATA_PATH @@ -96,4 +106,14 @@ class UpdateRefresh: LOG.warning('Setting up website directory at %s', webdir) with connect(args.config.get_libpq_dsn()) as conn: refresh.setup_website(webdir, args.config, conn) + return 0 + + + def _get_tokenizer(self, config): + if self.tokenizer is None: + from ..tokenizer import factory as tokenizer_factory + + self.tokenizer = tokenizer_factory.get_tokenizer_for_db(config) + + return self.tokenizer