ArchiveBox/archivebox/parse.py

316 lines
12 KiB
Python
Raw Normal View History

2017-10-23 09:58:41 +00:00
"""
2019-03-21 05:28:12 +00:00
Everything related to parsing links from input sources.
2017-10-23 09:58:41 +00:00
For a list of supported services, see the README.md.
2019-03-21 05:28:12 +00:00
For examples of supported import formats see tests/.
2017-10-23 09:58:41 +00:00
2019-03-21 05:28:12 +00:00
Link: {
2017-10-23 09:58:41 +00:00
'url': 'https://example.com/example/?abc=123&xyc=345#lmnop',
2019-03-21 05:28:12 +00:00
'timestamp': '1544212312.4234',
2017-10-23 09:58:41 +00:00
'title': 'Example.com Page Title',
'tags': 'abc,def',
2019-03-21 05:28:12 +00:00
'sources': [
'output/sources/ril_export.html',
'output/sources/getpocket.com-1523422111.txt',
'output/sources/stdin-234234112312.txt'
]
2017-10-23 09:58:41 +00:00
}
"""
import re
import json
2017-10-23 09:58:41 +00:00
from datetime import datetime
2019-02-22 17:56:36 +00:00
import xml.etree.ElementTree as etree
2019-03-21 05:28:12 +00:00
from config import TIMEOUT
from util import (
2017-10-23 09:58:41 +00:00
str_between,
2019-01-11 09:09:39 +00:00
URL_REGEX,
2019-03-21 05:28:12 +00:00
check_url_parsing_invariants,
TimedProgress,
)
2019-03-21 05:28:12 +00:00
def parse_links(source_file):
"""parse a list of URLs with their metadata from an
RSS feed, bookmarks export, or text file
"""
2019-02-05 06:07:48 +00:00
2019-03-21 05:28:12 +00:00
check_url_parsing_invariants()
PARSERS = (
# Specialized parsers
('Pocket HTML', parse_pocket_html_export),
('Pinboard RSS', parse_pinboard_rss_export),
('Shaarli RSS', parse_shaarli_rss_export),
('Medium RSS', parse_medium_rss_export),
# General parsers
('Netscape HTML', parse_netscape_html_export),
('Generic RSS', parse_rss_export),
('Generic JSON', parse_json_export),
# Fallback parser
('Plain Text', parse_plain_text_export),
)
timer = TimedProgress(TIMEOUT * 4)
2019-03-21 05:28:12 +00:00
with open(source_file, 'r', encoding='utf-8') as file:
for parser_name, parser_func in PARSERS:
try:
2019-03-21 05:28:12 +00:00
links = list(parser_func(file))
if links:
timer.end()
2019-03-21 05:28:12 +00:00
return links, parser_name
except Exception as err:
2019-03-21 05:28:12 +00:00
# Parsers are tried one by one down the list, and the first one
# that succeeds is used. To see why a certain parser was not used
# due to error or format incompatibility, uncomment this line:
2019-02-06 08:23:04 +00:00
# print('[!] Parser {} failed: {} {}'.format(parser_name, err.__class__.__name__, err))
pass
timer.end()
2019-03-23 01:38:24 +00:00
return [], 'Failed to parse'
2019-03-21 05:28:12 +00:00
### Import Parser Functions
2019-02-05 06:07:48 +00:00
def parse_pocket_html_export(html_file):
"""Parse Pocket-format bookmarks export files (produced by getpocket.com/export/)"""
html_file.seek(0)
2017-10-23 09:58:41 +00:00
pattern = re.compile("^\\s*<li><a href=\"(.+)\" time_added=\"(\\d+)\" tags=\"(.*)\">(.+)</a></li>", re.UNICODE)
for line in html_file:
2017-10-23 09:58:41 +00:00
# example line
# <li><a href="http://example.com/ time_added="1478739709" tags="tag1,tag2">example title</a></li>
match = pattern.search(line)
if match:
url = match.group(1).replace('http://www.readability.com/read?url=', '') # remove old readability prefixes to get original url
time = datetime.fromtimestamp(float(match.group(2)))
tags = match.group(3)
title = match.group(4).replace(' — Readability', '').replace('http://www.readability.com/read?url=', '')
yield {
'url': url,
2019-02-05 02:45:25 +00:00
'timestamp': str(time.timestamp()),
'title': title or None,
'tags': tags or '',
'sources': [html_file.name],
}
2019-03-21 05:28:12 +00:00
def parse_json_export(json_file):
2018-10-08 16:45:51 +00:00
"""Parse JSON-format bookmarks export files (produced by pinboard.in/export/, or wallabag)"""
2019-03-21 05:28:12 +00:00
json_file.seek(0)
2019-03-21 05:28:12 +00:00
links = json.load(json_file)
json_date = lambda s: datetime.strptime(s, '%Y-%m-%dT%H:%M:%S%z')
for link in links:
2017-10-23 09:58:41 +00:00
# example line
# {"href":"http:\/\/www.reddit.com\/r\/example","description":"title here","extended":"","meta":"18a973f09c9cc0608c116967b64e0419","hash":"910293f019c2f4bb1a749fb937ba58e3","time":"2014-06-14T15:51:42Z","shared":"no","toread":"no","tags":"reddit android"}]
2019-03-21 05:28:12 +00:00
if link:
# Parse URL
url = link.get('href') or link.get('url') or link.get('URL')
if not url:
raise Exception('JSON must contain URL in each entry [{"url": "http://...", ...}, ...]')
# Parse the timestamp
ts_str = str(datetime.now().timestamp())
if link.get('timestamp'):
# chrome/ff histories use a very precise timestamp
ts_str = str(link['timestamp'] / 10000000)
elif link.get('time'):
ts_str = str(json_date(link['time'].split(',', 1)[0]).timestamp())
elif link.get('created_at'):
ts_str = str(json_date(link['created_at']).timestamp())
elif link.get('created'):
ts_str = str(json_date(link['created']).timestamp())
elif link.get('date'):
ts_str = str(json_date(link['date']).timestamp())
elif link.get('bookmarked'):
ts_str = str(json_date(link['bookmarked']).timestamp())
elif link.get('saved'):
ts_str = str(json_date(link['saved']).timestamp())
# Parse the title
title = None
if link.get('title'):
title = link['title'].strip() or None
elif link.get('description'):
title = link['description'].replace(' — Readability', '').strip() or None
elif link.get('name'):
title = link['name'].strip() or None
yield {
2018-10-08 16:45:51 +00:00
'url': url,
2019-03-21 05:28:12 +00:00
'timestamp': ts_str,
'title': title,
'tags': link.get('tags') or '',
'sources': [json_file.name],
}
2019-02-22 17:56:36 +00:00
def parse_rss_export(rss_file):
"""Parse RSS XML-format files into links"""
rss_file.seek(0)
items = rss_file.read().split('<item>')
items = items[1:] if items else []
for item in items:
# example item:
# <item>
# <title><![CDATA[How JavaScript works: inside the V8 engine]]></title>
# <category>Unread</category>
# <link>https://blog.sessionstack.com/how-javascript-works-inside</link>
# <guid>https://blog.sessionstack.com/how-javascript-works-inside</guid>
# <pubDate>Mon, 21 Aug 2017 14:21:58 -0500</pubDate>
# </item>
trailing_removed = item.split('</item>', 1)[0]
leading_removed = trailing_removed.split('<item>', 1)[-1].strip()
rows = leading_removed.split('\n')
2017-10-23 09:58:41 +00:00
def get_row(key):
return [r for r in rows if r.strip().startswith('<{}>'.format(key))][0]
2017-10-23 09:58:41 +00:00
url = str_between(get_row('link'), '<link>', '</link>')
ts_str = str_between(get_row('pubDate'), '<pubDate>', '</pubDate>')
time = datetime.strptime(ts_str, "%a, %d %b %Y %H:%M:%S %z")
2019-03-21 05:28:12 +00:00
title = str_between(get_row('title'), '<![CDATA[', ']]').strip() or None
yield {
'url': url,
'timestamp': str(time.timestamp()),
2019-03-21 05:28:12 +00:00
'title': title,
'tags': '',
2019-02-05 05:18:42 +00:00
'sources': [rss_file.name],
}
2019-02-22 17:56:36 +00:00
2019-02-05 05:18:42 +00:00
def parse_shaarli_rss_export(rss_file):
"""Parse Shaarli-specific RSS XML-format files into links"""
rss_file.seek(0)
entries = rss_file.read().split('<entry>')[1:]
for entry in entries:
# example entry:
# <entry>
# <title>Aktuelle Trojaner-Welle: Emotet lauert in gefälschten Rechnungsmails | heise online</title>
# <link href="https://www.heise.de/security/meldung/Aktuelle-Trojaner-Welle-Emotet-lauert-in-gefaelschten-Rechnungsmails-4291268.html" />
# <id>https://demo.shaarli.org/?cEV4vw</id>
# <published>2019-01-30T06:06:01+00:00</published>
# <updated>2019-01-30T06:06:01+00:00</updated>
# <content type="html" xml:lang="en"><![CDATA[<div class="markdown"><p>&#8212; <a href="https://demo.shaarli.org/?cEV4vw">Permalink</a></p></div>]]></content>
# </entry>
trailing_removed = entry.split('</entry>', 1)[0]
leading_removed = trailing_removed.strip()
rows = leading_removed.split('\n')
def get_row(key):
return [r.strip() for r in rows if r.strip().startswith('<{}'.format(key))][0]
title = str_between(get_row('title'), '<title>', '</title>').strip()
url = str_between(get_row('link'), '<link href="', '" />')
ts_str = str_between(get_row('published'), '<published>', '</published>')
time = datetime.strptime(ts_str, "%Y-%m-%dT%H:%M:%S%z")
yield {
2019-02-05 05:18:42 +00:00
'url': url,
'timestamp': str(time.timestamp()),
'title': title or None,
'tags': '',
'sources': [rss_file.name],
}
2017-10-23 09:58:41 +00:00
2019-02-05 06:07:48 +00:00
def parse_netscape_html_export(html_file):
"""Parse netscape-format bookmarks export files (produced by all browsers)"""
html_file.seek(0)
pattern = re.compile("<a href=\"(.+?)\" add_date=\"(\\d+)\"[^>]*>(.+)</a>", re.UNICODE | re.IGNORECASE)
for line in html_file:
2017-10-23 09:58:41 +00:00
# example line
# <DT><A HREF="https://example.com/?q=1+2" ADD_DATE="1497562974" LAST_MODIFIED="1497562974" ICON_URI="https://example.com/favicon.ico" ICON="data:image/png;base64,...">example bookmark title</A>
match = pattern.search(line)
if match:
url = match.group(1)
time = datetime.fromtimestamp(float(match.group(2)))
yield {
'url': url,
2019-02-05 02:53:54 +00:00
'timestamp': str(time.timestamp()),
'title': match.group(3).strip() or None,
'tags': '',
'sources': [html_file.name],
}
2017-10-23 09:58:41 +00:00
2018-03-14 01:11:11 +00:00
2019-02-05 06:07:48 +00:00
def parse_pinboard_rss_export(rss_file):
2018-03-14 01:11:11 +00:00
"""Parse Pinboard RSS feed files into links"""
rss_file.seek(0)
root = etree.parse(rss_file).getroot()
items = root.findall("{http://purl.org/rss/1.0/}item")
for item in items:
url = item.find("{http://purl.org/rss/1.0/}link").text
2019-03-19 22:44:22 +00:00
tags = item.find("{http://purl.org/dc/elements/1.1/}subject").text if item.find("{http://purl.org/dc/elements/1.1/}subject") else None
title = item.find("{http://purl.org/rss/1.0/}title").text.strip() if item.find("{http://purl.org/rss/1.0/}title").text.strip() else None
ts_str = item.find("{http://purl.org/dc/elements/1.1/}date").text if item.find("{http://purl.org/dc/elements/1.1/}date").text else None
2018-04-17 07:22:59 +00:00
2018-03-14 01:11:11 +00:00
# Pinboard includes a colon in its date stamp timezone offsets, which
# Python can't parse. Remove it:
if ts_str and ts_str[-3:-2] == ":":
2018-03-14 01:11:11 +00:00
ts_str = ts_str[:-3]+ts_str[-2:]
if ts_str:
time = datetime.strptime(ts_str, "%Y-%m-%dT%H:%M:%S%z")
else:
time = datetime.now()
yield {
2018-03-14 01:11:11 +00:00
'url': url,
2019-02-05 02:53:54 +00:00
'timestamp': str(time.timestamp()),
'title': title or None,
'tags': tags or '',
2018-03-14 01:11:11 +00:00
'sources': [rss_file.name],
}
2019-02-05 06:07:48 +00:00
def parse_medium_rss_export(rss_file):
"""Parse Medium RSS feed files into links"""
rss_file.seek(0)
root = etree.parse(rss_file).getroot()
items = root.find("channel").findall("item")
for item in items:
url = item.find("link").text
2019-01-11 09:09:39 +00:00
title = item.find("title").text.strip()
ts_str = item.find("pubDate").text
time = datetime.strptime(ts_str, "%a, %d %b %Y %H:%M:%S %Z")
yield {
'url': url,
2019-02-05 02:53:54 +00:00
'timestamp': str(time.timestamp()),
'title': title or None,
'tags': '',
'sources': [rss_file.name],
}
2019-01-11 09:09:39 +00:00
2019-02-05 06:07:48 +00:00
def parse_plain_text_export(text_file):
2019-01-11 09:09:39 +00:00
"""Parse raw links from each line in a text file"""
text_file.seek(0)
2019-03-21 05:28:12 +00:00
for line in text_file.readlines():
urls = re.findall(URL_REGEX, line) if line.strip() else ()
for url in urls:
yield {
'url': url,
'timestamp': str(datetime.now().timestamp()),
'title': None,
'tags': '',
'sources': [text_file.name],
}