mirror of
				https://github.com/peeringdb/peeringdb.git
				synced 2024-05-11 05:55:09 +00:00 
			
		
		
		
	
		
			
				
	
	
		
			212 lines
		
	
	
		
			7.0 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			212 lines
		
	
	
		
			7.0 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| from django.db.models.signals import post_save, pre_delete
 | |
| from django.db.models import Q
 | |
| import peeringdb_server.rest
 | |
| from peeringdb_server.models import (UTC, InternetExchange, Network, Facility)
 | |
| import re
 | |
| import time
 | |
| import datetime
 | |
| import unidecode
 | |
| 
 | |
| def unaccent(v):
 | |
|     return unidecode.unidecode(v).lower()
 | |
| 
 | |
| # SEARCH INDEX BE STORED HERE
 | |
| 
 | |
| SEARCH_CACHE = {"search_index": {}, "time": 0}
 | |
| 
 | |
| # We want to hook searchable objects into save and delete signals
 | |
| # so we can update the search index as the data changes without having
 | |
| # to reload the entire thing all the time
 | |
| 
 | |
| 
 | |
| def hook_save(sender, **kwargs):
 | |
|     obj = kwargs.get("instance")
 | |
|     tag = obj._handleref.tag
 | |
|     idx = SEARCH_CACHE.get("search_index")
 | |
|     if obj.status == "ok":
 | |
|         if tag not in idx:
 | |
|             idx[tag] = {}
 | |
|         idx.get(tag)[obj.id] = obj
 | |
| #        print "%d %s refreshed in search index" % (obj.id, tag)
 | |
|     else:
 | |
|         try:
 | |
|             del idx[tag][obj.id]
 | |
|         except KeyError:
 | |
|             pass
 | |
| 
 | |
| 
 | |
| #        print "%d %s delete from search index" % (obj.id, tag)
 | |
| 
 | |
| 
 | |
| def hook_delete(sender, **kwargs):
 | |
|     obj = kwargs.get("instance")
 | |
|     tag = obj._handleref.tag
 | |
|     try:
 | |
|         del SEARCH_CACHE.get["search_index"][tag][obj.id]
 | |
|     except TypeError:
 | |
|         pass
 | |
|     except KeyError:
 | |
|         pass
 | |
| 
 | |
| 
 | |
| #    print "%d %s deleted from search index " % (obj.id, tag)
 | |
| 
 | |
| searchable_models = [InternetExchange, Network, Facility]
 | |
| 
 | |
| for model in searchable_models:
 | |
|     post_save.connect(hook_save, sender=model)
 | |
|     pre_delete.connect(hook_delete, sender=model)
 | |
| 
 | |
| 
 | |
| def search(term):
 | |
|     """
 | |
|     Search searchable objects (ixp, network, facility ...) by term
 | |
| 
 | |
|     Returns result dict
 | |
|     """
 | |
| 
 | |
|     search_tags = ('fac', 'ix', 'net')
 | |
|     ref_dict = peeringdb_server.rest.ref_dict()
 | |
|     t = time.time()
 | |
| 
 | |
|     if not SEARCH_CACHE.get("search_index"):
 | |
| 
 | |
|         # whole db takes 5ish seconds, too slow to cache inline here
 | |
|         search_index = {
 | |
|             tag:
 | |
|             {obj.id: obj
 | |
|              for obj in model.objects.filter(status__in=["ok"])}
 | |
|             for tag, model in ref_dict.items() if tag in search_tags
 | |
|         }
 | |
| 
 | |
|         for typ, stor in search_index.items():
 | |
|             print "CACHED: %d items in %s" % (len(stor), typ)
 | |
| 
 | |
|         tag_id_re = re.compile('(' + "|".join(search_tags) + '|asn|as)(\d+)')
 | |
| 
 | |
|         # FIXME: for now lets force a flush every 120 seconds, might want to look
 | |
|         # at an event based update solution instead
 | |
|         SEARCH_CACHE.update(search_index=search_index, time=t, update_t=t,
 | |
|                             tag_id_re=tag_id_re)
 | |
|     else:
 | |
|         search_index = SEARCH_CACHE.get('search_index')
 | |
|         tag_id_re = SEARCH_CACHE.get('tag_id_re')
 | |
| 
 | |
|     # while we are using signals to make sure that the search index gets updated whenever
 | |
|     # a model is saved, right now we still have updates from external sources
 | |
|     # to which those signals cannot be easily connected (importer, fac_merge command etc.)
 | |
|     #
 | |
|     # in order to reflect search index changes made by external sources
 | |
|     # we need to find new / updated object regularily and update the
 | |
|     # search index from that
 | |
|     #
 | |
|     # FIXME: this can be taken out when we turn the importer off - or just leave it
 | |
|     # in as a fail-safe as it is fairly unobtrusive
 | |
|     ut = SEARCH_CACHE.get("update_t", 0)
 | |
|     if t - ut > 600:
 | |
|         dut = datetime.datetime.fromtimestamp(ut).replace(tzinfo=UTC())
 | |
|         print "Updating search index with newly created/updates objects"
 | |
|         search_index_update = {
 | |
|             tag: {
 | |
|                 obj.id: obj
 | |
|                 for obj in model.objects.filter(
 | |
|                     Q(created__gte=dut)
 | |
|                     | Q(updated__gte=dut)).filter(status="ok")
 | |
|             }
 | |
|             for tag, model in ref_dict.items() if tag in search_tags
 | |
|         }
 | |
|         for tag, objects in search_index_update.items():
 | |
|             if tag not in SEARCH_CACHE["search_index"]:
 | |
|                 SEARCH_CACHE["search_index"][tag] = dict(
 | |
|                     [(obj.id, obj)
 | |
|                      for obj in ref_dict[tag].objects.filter(status="ok")])
 | |
|             SEARCH_CACHE["search_index"][tag].update(objects)
 | |
| 
 | |
|         SEARCH_CACHE["update_t"] = t
 | |
| 
 | |
|     # FIXME: for some reason this gets unset sometimes - need to figure out
 | |
|     # why - for now just recreate when its missing
 | |
|     if not tag_id_re:
 | |
|         tag_id_re = re.compile('(' + "|".join(search_tags) + '|asn|as)(\d+)')
 | |
|         SEARCH_CACHE['tag_id_re'] = tag_id_re
 | |
| 
 | |
|     print "Search index retrieval took %.5f seconds" % (time.time() - t)
 | |
| 
 | |
|     result = {tag: [] for tag, model in ref_dict.items()}
 | |
| 
 | |
|     term = unaccent(term)
 | |
| 
 | |
|     # try to convert to int for numeric search matching
 | |
|     typed_q = {}
 | |
|     try:
 | |
|         typed_q['int'] = int(term)
 | |
|     except ValueError:
 | |
|         pass
 | |
| 
 | |
|     # check for ref tags
 | |
|     try:
 | |
|         match = tag_id_re.match(term)
 | |
|         if match:
 | |
|             typed_q[match.group(1)] = match.group(2)
 | |
| 
 | |
|     except ValueError:
 | |
|         pass
 | |
| 
 | |
|     # FIXME  model should have a search_fields attr on it
 | |
|     # this whole thing should be replaced with something more modular to get
 | |
|     # rid of all the ifs
 | |
|     for tag, index in search_index.items():
 | |
|         for id, data in index.items():
 | |
|             if unaccent(data.name).find(term) > -1:
 | |
|                 result[tag].append({
 | |
|                     "id": id,
 | |
|                     "name": data.search_result_name,
 | |
|                     "org_id": data.org_id
 | |
|                 })
 | |
|                 continue
 | |
| 
 | |
|             if hasattr(data,
 | |
|                        'name_long') and unaccent(data.name_long).find(term) > -1:
 | |
|                 result[tag].append({
 | |
|                     "id": id,
 | |
|                     "name": data.search_result_name,
 | |
|                     "org_id" : data.org_id
 | |
|                 })
 | |
|                 continue
 | |
| 
 | |
|             if hasattr(data, 'aka') and unaccent(data.aka).find(term) > -1:
 | |
|                 result[tag].append({
 | |
|                     "id": id,
 | |
|                     "name": data.search_result_name,
 | |
|                     "org_id": data.org_id
 | |
|                 })
 | |
|                 continue
 | |
| 
 | |
|             if typed_q:
 | |
|                 if tag in typed_q:
 | |
|                     if str(data.id).startswith(typed_q[tag]):
 | |
|                         result[tag].append({
 | |
|                             "id": id,
 | |
|                             "name": data.search_result_name,
 | |
|                             "org_id": data.org_id
 | |
|                         })
 | |
|                         continue
 | |
| 
 | |
|                 # search asn on everyting? probably just if asn in search
 | |
|                 # fields
 | |
|                 if hasattr(data, 'asn'):
 | |
|                     asn = typed_q.get('as',
 | |
|                                       typed_q.get('asn',
 | |
|                                                   str(typed_q.get('int', ''))))
 | |
|                     if asn and str(data.asn).startswith(asn):
 | |
|                         result[tag].append({
 | |
|                             "id": id,
 | |
|                             "name": data.search_result_name,
 | |
|                             "org_id": data.org_id
 | |
|                         })
 | |
| 
 | |
|     for k, items in result.items():
 | |
|         result[k] = sorted(items, key=lambda row: row.get("name"))
 | |
| 
 | |
|     return result
 |