mirror of
https://github.com/ArchiveBox/ArchiveBox
synced 2024-11-10 06:34:16 +00:00
commit
e517ba0d04
2 changed files with 46 additions and 21 deletions
|
@ -30,7 +30,7 @@ apt update; apt install google-chrome-beta python3 wget
|
||||||
1. Download your pocket export file `ril_export.html` from https://getpocket.com/export
|
1. Download your pocket export file `ril_export.html` from https://getpocket.com/export
|
||||||
2. Download this repo `git clone https://github.com/pirate/pocket-archive-stream`
|
2. Download this repo `git clone https://github.com/pirate/pocket-archive-stream`
|
||||||
3. `cd pocket-archive-stream/`
|
3. `cd pocket-archive-stream/`
|
||||||
4. `./archive.py ~/Downloads/ril_export.html`
|
4. `./archive.py ~/Downloads/ril_export.html [pinboard|pocket]`
|
||||||
|
|
||||||
It produces a folder `pocket/` containing an `index.html`, and archived copies of all the sites,
|
It produces a folder `pocket/` containing an `index.html`, and archived copies of all the sites,
|
||||||
organized by timestamp. For each sites it saves:
|
organized by timestamp. For each sites it saves:
|
||||||
|
|
65
archive.py
65
archive.py
|
@ -7,8 +7,11 @@
|
||||||
import re
|
import re
|
||||||
import os
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
import json
|
||||||
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
|
import time
|
||||||
|
|
||||||
from subprocess import run, DEVNULL
|
from subprocess import run, DEVNULL
|
||||||
|
|
||||||
INDEX_TEMPLATE = 'index_template.html'
|
INDEX_TEMPLATE = 'index_template.html'
|
||||||
|
@ -42,7 +45,22 @@ def parse_pocket_export(html):
|
||||||
'title': match.group(4).replace(' — Readability', '').replace('http://www.readability.com/read?url=', ''),
|
'title': match.group(4).replace(' — Readability', '').replace('http://www.readability.com/read?url=', ''),
|
||||||
}
|
}
|
||||||
|
|
||||||
def dump_index(links):
|
def parse_pinboard_export(html):
|
||||||
|
json_content = json.load(html)
|
||||||
|
for line in json_content:
|
||||||
|
if line:
|
||||||
|
erg = line
|
||||||
|
yield {
|
||||||
|
'url': erg['href'].replace('http://www.readability.com/read?url=', ''),
|
||||||
|
'domain': erg['href'].replace('http://', '').replace('https://', '').split('/')[0],
|
||||||
|
'base_url': erg['href'].replace('https://', '').replace('http://', '').split('?')[0],
|
||||||
|
'time': datetime.fromtimestamp(time.mktime(time.strptime(erg['time'].split(',')[0],'%Y-%m-%dT%H:%M:%SZ'))),
|
||||||
|
'timestamp': time.mktime(time.strptime(erg['time'].split(',')[0],'%Y-%m-%dT%H:%M:%SZ')),
|
||||||
|
'tags': erg['tags'],
|
||||||
|
'title': erg['description'].replace(' — Readability', '').replace('http://www.readability.com/read?url=', ''),
|
||||||
|
}
|
||||||
|
|
||||||
|
def dump_index(links, service):
|
||||||
with open(INDEX_TEMPLATE, 'r') as f:
|
with open(INDEX_TEMPLATE, 'r') as f:
|
||||||
index_html = f.read()
|
index_html = f.read()
|
||||||
|
|
||||||
|
@ -59,7 +77,7 @@ def dump_index(links):
|
||||||
<td>🔗 <img src="https://www.google.com/s2/favicons?domain={domain}" height="16px"> <a href="{url}">{url}</a></td>
|
<td>🔗 <img src="https://www.google.com/s2/favicons?domain={domain}" height="16px"> <a href="{url}">{url}</a></td>
|
||||||
</tr>"""
|
</tr>"""
|
||||||
|
|
||||||
with open('pocket/index.html', 'w') as f:
|
with open(''.join((service,'/index.html')), 'w') as f:
|
||||||
article_rows = '\n'.join(
|
article_rows = '\n'.join(
|
||||||
link_html.format(**link) for link in links
|
link_html.format(**link) for link in links
|
||||||
)
|
)
|
||||||
|
@ -119,12 +137,12 @@ def fetch_favicon(out_dir, link, overwrite=False):
|
||||||
print(' √ Skipping favicon')
|
print(' √ Skipping favicon')
|
||||||
|
|
||||||
|
|
||||||
def dump_website(link, overwrite=False):
|
def dump_website(link, service, overwrite=False):
|
||||||
"""download the DOM, PDF, and a screenshot into a folder named after the link's timestamp"""
|
"""download the DOM, PDF, and a screenshot into a folder named after the link's timestamp"""
|
||||||
|
|
||||||
print('[+] [{time}] Archiving "{title}": {url}'.format(**link))
|
print('[+] [{time}] Archiving "{title}": {url}'.format(**link))
|
||||||
|
|
||||||
out_dir = 'pocket/archive/{timestamp}'.format(**link)
|
out_dir = ''.join((service, '/archive/{timestamp}')).format(**link)
|
||||||
if not os.path.exists(out_dir):
|
if not os.path.exists(out_dir):
|
||||||
os.makedirs(out_dir)
|
os.makedirs(out_dir)
|
||||||
|
|
||||||
|
@ -149,46 +167,53 @@ def dump_website(link, overwrite=False):
|
||||||
|
|
||||||
run(['chmod', '-R', '755', out_dir], timeout=1)
|
run(['chmod', '-R', '755', out_dir], timeout=1)
|
||||||
|
|
||||||
def create_archive(pocket_file, resume=None):
|
def create_archive(service_file, service, resume=None):
|
||||||
print('[+] [{}] Starting pocket archive from {}'.format(datetime.now(), pocket_file))
|
print('[+] [{}] Starting {} archive from {}'.format(datetime.now(), service, service_file))
|
||||||
|
|
||||||
if not os.path.exists('pocket'):
|
if not os.path.exists(service):
|
||||||
os.makedirs('pocket')
|
os.makedirs(service)
|
||||||
|
|
||||||
if not os.path.exists('pocket/archive'):
|
if not os.path.exists(''.join((service,'/archive'))):
|
||||||
os.makedirs('pocket/archive')
|
os.makedirs(''.join((service,'/archive')))
|
||||||
|
|
||||||
with open(pocket_file, 'r', encoding='utf-8') as f:
|
with open(service_file, 'r', encoding='utf-8') as f:
|
||||||
links = parse_pocket_export(f)
|
if service == "pocket":
|
||||||
|
links = parse_pocket_export(f)
|
||||||
|
elif service == "pinboard":
|
||||||
|
links = parse_pinboard_export(f)
|
||||||
links = list(reversed(sorted(links, key=lambda l: l['timestamp']))) # most recent first
|
links = list(reversed(sorted(links, key=lambda l: l['timestamp']))) # most recent first
|
||||||
if resume:
|
if resume:
|
||||||
links = [link for link in links if link['timestamp'] >= resume]
|
links = [link for link in links if link['timestamp'] >= resume]
|
||||||
|
|
||||||
if not links:
|
if not links:
|
||||||
print('[X] No links found in {}, is it a getpocket.com/export export?'.format(pocket_file))
|
if service == "pocket":
|
||||||
|
print('[X] No links found in {}, is it a getpocket.com/export export?'.format(serivce_file))
|
||||||
|
elif service == "pinboard":
|
||||||
|
print ('[X] No links found in {}, is it a pinboard.in/export/format:json/ export?'.format(service_file))
|
||||||
raise SystemExit(1)
|
raise SystemExit(1)
|
||||||
|
|
||||||
dump_index(links)
|
dump_index(links, service)
|
||||||
|
|
||||||
run(['chmod', '-R', '755', 'pocket'], timeout=1)
|
run(['chmod', '-R', '755', service], timeout=1)
|
||||||
|
|
||||||
print('[*] [{}] Created archive index.'.format(datetime.now()))
|
print('[*] [{}] Created archive index.'.format(datetime.now()))
|
||||||
|
|
||||||
check_dependencies()
|
check_dependencies()
|
||||||
|
|
||||||
for link in links:
|
for link in links:
|
||||||
dump_website(link)
|
dump_website(link, service)
|
||||||
|
|
||||||
print('[√] [{}] Archive complete.'.format(datetime.now()))
|
print('[√] [{}] Archive complete.'.format(datetime.now()))
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
pocket_file = 'ril_export.html'
|
service_file = 'ril_export.html'
|
||||||
resume = None
|
resume = None
|
||||||
try:
|
try:
|
||||||
pocket_file = sys.argv[1] # path to pocket export file
|
service_file = sys.argv[1] # path to export file
|
||||||
resume = sys.argv[2] # timestamp to resume dowloading from
|
service = sys.argv[2] or "pocket" # select service for file format select
|
||||||
|
resume = sys.argv[3] # timestamp to resume dowloading from
|
||||||
except IndexError:
|
except IndexError:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
create_archive(pocket_file, resume=resume)
|
create_archive(service_file, service, resume=resume)
|
||||||
|
|
Loading…
Reference in a new issue