+
+def load_address_levels_from_config(conn: Connection, config: Configuration) -> None:
+ """ Replace the `address_levels` table with the content as
+ defined in the given configuration. Uses the parameter
+ NOMINATIM_ADDRESS_LEVEL_CONFIG to determine the location of the
+ configuration file.
+ """
+ cfg = config.load_sub_configuration('', config='ADDRESS_LEVEL_CONFIG')
+ load_address_levels(conn, 'address_levels', cfg)
+
+
+def create_functions(conn: Connection, config: Configuration,
+ enable_diff_updates: bool = True,
+ enable_debug: bool = False) -> None:
+ """ (Re)create the PL/pgSQL functions.
+ """
+ sql = SQLPreprocessor(conn, config)
+
+ sql.run_sql_file(conn, 'functions.sql',
+ disable_diff_updates=not enable_diff_updates,
+ debug=enable_debug)
+
+
+
+WEBSITE_SCRIPTS = (
+ 'deletable.php',
+ 'details.php',
+ 'lookup.php',
+ 'polygons.php',
+ 'reverse.php',
+ 'search.php',
+ 'status.php'
+)
+
+# constants needed by PHP scripts: PHP name, config name, type
+PHP_CONST_DEFS = (
+ ('Database_DSN', 'DATABASE_DSN', str),
+ ('Default_Language', 'DEFAULT_LANGUAGE', str),
+ ('Log_DB', 'LOG_DB', bool),
+ ('Log_File', 'LOG_FILE', Path),
+ ('NoAccessControl', 'CORS_NOACCESSCONTROL', bool),
+ ('Places_Max_ID_count', 'LOOKUP_MAX_COUNT', int),
+ ('PolygonOutput_MaximumTypes', 'POLYGON_OUTPUT_MAX_TYPES', int),
+ ('Search_BatchMode', 'SEARCH_BATCH_MODE', bool),
+ ('Search_NameOnlySearchFrequencyThreshold', 'SEARCH_NAME_ONLY_THRESHOLD', str),
+ ('Use_US_Tiger_Data', 'USE_US_TIGER_DATA', bool),
+ ('MapIcon_URL', 'MAPICON_URL', str),
+)
+
+
+def import_wikipedia_articles(dsn: str, data_path: Path, ignore_errors: bool = False) -> int:
+ """ Replaces the wikipedia importance tables with new data.
+ The import is run in a single transaction so that the new data
+ is replace seamlessly.
+
+ Returns 0 if all was well and 1 if the importance file could not
+ be found. Throws an exception if there was an error reading the file.
+ """
+ datafile = data_path / 'wikimedia-importance.sql.gz'
+
+ if not datafile.exists():
+ return 1
+
+ pre_code = """BEGIN;
+ DROP TABLE IF EXISTS "wikipedia_article";
+ DROP TABLE IF EXISTS "wikipedia_redirect"
+ """
+ post_code = "COMMIT"
+ execute_file(dsn, datafile, ignore_errors=ignore_errors,
+ pre_code=pre_code, post_code=post_code)
+
+ return 0
+
+def import_secondary_importance(dsn: str, data_path: Path, ignore_errors: bool = False) -> int:
+ """ Replaces the secondary importance raster data table with new data.
+
+ Returns 0 if all was well and 1 if the raster SQL file could not
+ be found. Throws an exception if there was an error reading the file.
+ """
+ datafile = data_path / 'secondary_importance.sql.gz'
+ if not datafile.exists():
+ return 1
+
+ with connect(dsn) as conn:
+ postgis_version = conn.postgis_version_tuple()
+ if postgis_version[0] < 3:
+ LOG.error('PostGIS version is too old for using OSM raster data.')
+ return 2
+
+ execute_file(dsn, datafile, ignore_errors=ignore_errors)
+
+ return 0
+
+def recompute_importance(conn: Connection) -> None:
+ """ Recompute wikipedia links and importance for all entries in placex.
+ This is a long-running operations that must not be executed in
+ parallel with updates.