2019-04-27 21:26:24 +00:00
|
|
|
__package__ = 'archivebox.index'
|
2019-04-18 02:00:54 +00:00
|
|
|
|
2019-04-17 06:25:28 +00:00
|
|
|
import os
|
2019-04-19 01:09:54 +00:00
|
|
|
import sys
|
2019-05-01 03:13:04 +00:00
|
|
|
import json as pyjson
|
2020-07-22 19:22:00 +00:00
|
|
|
from pathlib import Path
|
2019-04-17 06:25:28 +00:00
|
|
|
|
|
|
|
from datetime import datetime
|
2020-09-03 22:26:49 +00:00
|
|
|
from typing import List, Optional, Iterator, Any, Union
|
2019-04-17 06:25:28 +00:00
|
|
|
|
2020-11-28 18:11:15 +00:00
|
|
|
from .schema import Link
|
2019-05-01 03:13:04 +00:00
|
|
|
from ..system import atomic_write
|
|
|
|
from ..util import enforce_types
|
2019-04-17 06:25:28 +00:00
|
|
|
from ..config import (
|
|
|
|
VERSION,
|
|
|
|
OUTPUT_DIR,
|
2019-04-19 01:09:54 +00:00
|
|
|
FOOTER_INFO,
|
|
|
|
DEPENDENCIES,
|
|
|
|
JSON_INDEX_FILENAME,
|
2019-04-24 08:07:46 +00:00
|
|
|
ARCHIVE_DIR_NAME,
|
2020-07-22 22:08:32 +00:00
|
|
|
ANSI
|
2019-04-17 06:25:28 +00:00
|
|
|
)
|
2019-04-27 21:26:24 +00:00
|
|
|
|
2019-04-17 06:25:28 +00:00
|
|
|
|
2019-04-19 01:09:54 +00:00
|
|
|
MAIN_INDEX_HEADER = {
|
|
|
|
'info': 'This is an index of site data archived by ArchiveBox: The self-hosted web archive.',
|
2019-04-27 21:26:24 +00:00
|
|
|
'schema': 'archivebox.index.json',
|
2019-04-19 01:09:54 +00:00
|
|
|
'copyright_info': FOOTER_INFO,
|
|
|
|
'meta': {
|
|
|
|
'project': 'ArchiveBox',
|
|
|
|
'version': VERSION,
|
2021-02-16 01:42:33 +00:00
|
|
|
'git_sha': VERSION, # not used anymore, but kept for backwards compatibility
|
2019-04-19 01:09:54 +00:00
|
|
|
'website': 'https://ArchiveBox.io',
|
2020-11-23 07:04:39 +00:00
|
|
|
'docs': 'https://github.com/ArchiveBox/ArchiveBox/wiki',
|
|
|
|
'source': 'https://github.com/ArchiveBox/ArchiveBox',
|
|
|
|
'issues': 'https://github.com/ArchiveBox/ArchiveBox/issues',
|
2019-04-19 01:09:54 +00:00
|
|
|
'dependencies': DEPENDENCIES,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2020-11-28 17:28:39 +00:00
|
|
|
@enforce_types
|
|
|
|
def generate_json_index_from_links(links: List[Link], with_headers: bool):
|
|
|
|
if with_headers:
|
|
|
|
output = {
|
|
|
|
**MAIN_INDEX_HEADER,
|
|
|
|
'num_links': len(links),
|
|
|
|
'updated': datetime.now(),
|
|
|
|
'last_run_cmd': sys.argv,
|
|
|
|
'links': links,
|
|
|
|
}
|
|
|
|
else:
|
|
|
|
output = links
|
|
|
|
return to_json(output, indent=4, sort_keys=True)
|
|
|
|
|
2019-04-17 06:25:28 +00:00
|
|
|
|
|
|
|
@enforce_types
|
2020-09-03 22:26:49 +00:00
|
|
|
def parse_json_main_index(out_dir: Path=OUTPUT_DIR) -> Iterator[Link]:
|
2019-04-27 21:26:24 +00:00
|
|
|
"""parse an archive index json file and return the list of links"""
|
2019-04-17 06:25:28 +00:00
|
|
|
|
2020-09-30 19:21:41 +00:00
|
|
|
index_path = Path(out_dir) / JSON_INDEX_FILENAME
|
|
|
|
if index_path.exists():
|
2019-04-17 06:25:28 +00:00
|
|
|
with open(index_path, 'r', encoding='utf-8') as f:
|
2019-05-01 03:13:04 +00:00
|
|
|
links = pyjson.load(f)['links']
|
2019-04-17 06:25:28 +00:00
|
|
|
for link_json in links:
|
2020-07-22 19:22:00 +00:00
|
|
|
try:
|
|
|
|
yield Link.from_json(link_json)
|
|
|
|
except KeyError:
|
2020-07-22 22:08:32 +00:00
|
|
|
try:
|
2020-07-23 15:22:36 +00:00
|
|
|
detail_index_path = Path(OUTPUT_DIR) / ARCHIVE_DIR_NAME / link_json['timestamp']
|
2020-07-22 22:08:32 +00:00
|
|
|
yield parse_json_link_details(str(detail_index_path))
|
|
|
|
except KeyError:
|
2020-07-24 14:24:52 +00:00
|
|
|
# as a last effort, try to guess the missing values out of existing ones
|
|
|
|
try:
|
|
|
|
yield Link.from_json(link_json, guess=True)
|
|
|
|
except KeyError:
|
|
|
|
print(" {lightyellow}! Failed to load the index.json from {}".format(detail_index_path, **ANSI))
|
|
|
|
continue
|
2019-04-17 06:25:28 +00:00
|
|
|
return ()
|
|
|
|
|
|
|
|
### Link Details Index
|
|
|
|
|
|
|
|
@enforce_types
|
|
|
|
def write_json_link_details(link: Link, out_dir: Optional[str]=None) -> None:
|
|
|
|
"""write a json file with some info about the link"""
|
|
|
|
|
|
|
|
out_dir = out_dir or link.link_dir
|
2020-09-30 19:21:41 +00:00
|
|
|
path = Path(out_dir) / JSON_INDEX_FILENAME
|
|
|
|
atomic_write(str(path), link._asdict(extended=True))
|
2019-04-17 06:25:28 +00:00
|
|
|
|
|
|
|
|
|
|
|
@enforce_types
|
2020-09-03 22:26:49 +00:00
|
|
|
def parse_json_link_details(out_dir: Union[Path, str], guess: Optional[bool]=False) -> Optional[Link]:
|
2019-04-17 06:25:28 +00:00
|
|
|
"""load the json link index from a given directory"""
|
2020-09-30 19:21:41 +00:00
|
|
|
existing_index = Path(out_dir) / JSON_INDEX_FILENAME
|
|
|
|
if existing_index.exists():
|
2019-04-17 06:25:28 +00:00
|
|
|
with open(existing_index, 'r', encoding='utf-8') as f:
|
2019-04-24 15:38:13 +00:00
|
|
|
try:
|
2019-05-01 03:13:04 +00:00
|
|
|
link_json = pyjson.load(f)
|
2020-07-24 14:24:52 +00:00
|
|
|
return Link.from_json(link_json, guess)
|
2019-05-01 03:13:04 +00:00
|
|
|
except pyjson.JSONDecodeError:
|
2019-04-24 15:38:13 +00:00
|
|
|
pass
|
2019-04-17 06:25:28 +00:00
|
|
|
return None
|
2019-04-24 08:07:46 +00:00
|
|
|
|
2019-05-01 03:13:04 +00:00
|
|
|
|
2019-04-24 08:07:46 +00:00
|
|
|
@enforce_types
|
2020-09-03 22:26:49 +00:00
|
|
|
def parse_json_links_details(out_dir: Union[Path, str]) -> Iterator[Link]:
|
2019-04-24 08:07:46 +00:00
|
|
|
"""read through all the archive data folders and return the parsed links"""
|
|
|
|
|
2020-09-30 19:21:41 +00:00
|
|
|
for entry in os.scandir(Path(out_dir) / ARCHIVE_DIR_NAME):
|
2019-04-24 08:07:46 +00:00
|
|
|
if entry.is_dir(follow_symlinks=True):
|
2020-09-30 19:21:41 +00:00
|
|
|
if (Path(entry.path) / 'index.json').exists():
|
2020-07-22 22:08:32 +00:00
|
|
|
try:
|
|
|
|
link = parse_json_link_details(entry.path)
|
|
|
|
except KeyError:
|
|
|
|
link = None
|
2019-04-27 21:26:24 +00:00
|
|
|
if link:
|
|
|
|
yield link
|
2019-05-01 03:13:04 +00:00
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
### Helpers
|
|
|
|
|
|
|
|
class ExtendedEncoder(pyjson.JSONEncoder):
|
|
|
|
"""
|
|
|
|
Extended json serializer that supports serializing several model
|
|
|
|
fields and objects
|
|
|
|
"""
|
|
|
|
|
|
|
|
def default(self, obj):
|
|
|
|
cls_name = obj.__class__.__name__
|
|
|
|
|
|
|
|
if hasattr(obj, '_asdict'):
|
|
|
|
return obj._asdict()
|
|
|
|
|
|
|
|
elif isinstance(obj, bytes):
|
|
|
|
return obj.decode()
|
|
|
|
|
|
|
|
elif isinstance(obj, datetime):
|
|
|
|
return obj.isoformat()
|
|
|
|
|
|
|
|
elif isinstance(obj, Exception):
|
|
|
|
return '{}: {}'.format(obj.__class__.__name__, obj)
|
|
|
|
|
|
|
|
elif cls_name in ('dict_items', 'dict_keys', 'dict_values'):
|
|
|
|
return tuple(obj)
|
|
|
|
|
|
|
|
return pyjson.JSONEncoder.default(self, obj)
|
|
|
|
|
|
|
|
|
|
|
|
@enforce_types
|
|
|
|
def to_json(obj: Any, indent: Optional[int]=4, sort_keys: bool=True, cls=ExtendedEncoder) -> str:
|
|
|
|
return pyjson.dumps(obj, indent=indent, sort_keys=sort_keys, cls=ExtendedEncoder)
|
|
|
|
|