ArchiveBox/archivebox/search/__init__.py

109 lines
3.3 KiB
Python
Raw Normal View History

from typing import List, Union
2020-11-17 23:42:57 +00:00
from pathlib import Path
from importlib import import_module
2020-11-17 23:42:57 +00:00
from django.db.models import QuerySet
2020-11-17 23:42:57 +00:00
from archivebox.index.schema import Link
from archivebox.util import enforce_types
from archivebox.config import stderr, OUTPUT_DIR, USE_INDEXING_BACKEND, USE_SEARCHING_BACKEND, SEARCH_BACKEND_ENGINE
2020-11-17 23:42:57 +00:00
2020-11-23 22:23:26 +00:00
from .utils import get_indexable_content, log_index_started
def indexing_enabled():
2020-11-19 13:06:13 +00:00
return USE_INDEXING_BACKEND
2020-11-17 23:42:57 +00:00
def search_backend_enabled():
2020-11-19 13:06:13 +00:00
return USE_SEARCHING_BACKEND
2020-11-17 23:42:57 +00:00
def get_backend():
2020-11-19 13:06:13 +00:00
return f'search.backends.{SEARCH_BACKEND_ENGINE}'
2020-11-17 23:42:57 +00:00
def import_backend():
backend_string = get_backend()
try:
backend = import_module(backend_string)
except Exception as err:
raise Exception("Could not load '%s' as a backend: %s" % (backend_string, err))
return backend
2020-11-17 23:42:57 +00:00
@enforce_types
def write_search_index(link: Link, texts: Union[List[str], None]=None, out_dir: Path=OUTPUT_DIR, skip_text_index: bool=False) -> None:
if not indexing_enabled():
return
2020-11-17 23:42:57 +00:00
if not skip_text_index and texts:
from core.models import Snapshot
2020-11-17 23:42:57 +00:00
snap = Snapshot.objects.filter(url=link.url).first()
backend = import_backend()
2020-11-17 23:42:57 +00:00
if snap:
try:
backend.index(snapshot_id=str(snap.id), texts=texts)
except Exception as err:
stderr()
stderr(
f'[X] The search backend threw an exception={err}:',
color='red',
)
@enforce_types
def query_search_index(query: str, out_dir: Path=OUTPUT_DIR) -> QuerySet:
from core.models import Snapshot
if search_backend_enabled():
backend = import_backend()
try:
snapshot_ids = backend.search(query)
except Exception as err:
stderr()
stderr(
f'[X] The search backend threw an exception={err}:',
color='red',
)
raise
else:
# TODO preserve ordering from backend
qsearch = Snapshot.objects.filter(pk__in=snapshot_ids)
return qsearch
return Snapshot.objects.none()
@enforce_types
def flush_search_index(snapshots: QuerySet):
if not indexing_enabled() or not snapshots:
return
backend = import_backend()
snapshot_ids=(str(pk) for pk in snapshots.values_list('pk',flat=True))
try:
backend.flush(snapshot_ids)
except Exception as err:
stderr()
stderr(
f'[X] The search backend threw an exception={err}:',
color='red',
)
@enforce_types
def index_links(links: Union[List[Link],None], out_dir: Path=OUTPUT_DIR):
if not links:
return
from core.models import Snapshot, ArchiveResult
for link in links:
snap = Snapshot.objects.filter(url=link.url).first()
if snap:
results = ArchiveResult.objects.indexable().filter(snapshot=snap)
2020-11-23 22:23:26 +00:00
log_index_started(link.url)
try:
texts = get_indexable_content(results)
except Exception as err:
stderr()
stderr(
f'[X] An Exception ocurred reading the indexable content={err}:',
color='red',
)
else:
2020-12-11 14:51:11 +00:00
write_search_index(link, texts, out_dir=out_dir)