]> git.openstreetmap.org Git - nominatim.git/blobdiff - test/python/dummy_tokenizer.py
Merge pull request #2401 from lonvia/port-add-data-to-python
[nominatim.git] / test / python / dummy_tokenizer.py
index 013016c8d0ba7c2e84bddab61bfb3428944a205a..69202bc322ffd88e103f60f8ced809bfa8e82fd3 100644 (file)
@@ -13,20 +13,26 @@ class DummyTokenizer:
         self.dsn = dsn
         self.data_dir = data_dir
         self.init_state = None
+        self.analyser_cache = {}
 
 
-    def init_new_db(self, config):
-        assert self.init_state == None
+    def init_new_db(self, *args, **kwargs):
+        assert self.init_state is None
         self.init_state = "new"
 
 
     def init_from_project(self):
-        assert self.init_state == None
+        assert self.init_state is None
         self.init_state = "loaded"
 
 
+    @staticmethod
+    def finalize_import(_):
+        pass
+
+
     def name_analyzer(self):
-        return DummyNameAnalyzer()
+        return DummyNameAnalyzer(self.analyser_cache)
 
 
 class DummyNameAnalyzer:
@@ -38,15 +44,28 @@ class DummyNameAnalyzer:
         self.close()
 
 
+    def __init__(self, cache):
+        self.analyser_cache = cache
+        cache['countries'] = []
+
+
     def close(self):
-        """ Free all resources used by the analyzer.
-        """
         pass
 
-    def process_place(self, place):
-        """ Determine tokenizer information about the given place.
+    @staticmethod
+    def normalize_postcode(postcode):
+        return postcode
+
+    @staticmethod
+    def update_postcodes_from_db():
+        pass
+
+    def update_special_phrases(self, phrases, should_replace):
+        self.analyser_cache['special_phrases'] = phrases
+
+    def add_country_names(self, code, names):
+        self.analyser_cache['countries'].append((code, names))
 
-            Returns a JSON-serialisable structure that will be handed into
-            the database via the token_info field.
-        """
+    @staticmethod
+    def process_place(place):
         return {}