2020-07-07 14:10:36 +00:00
|
|
|
import subprocess
|
2020-07-07 14:49:28 +00:00
|
|
|
import json
|
2020-10-08 16:02:26 +00:00
|
|
|
import sqlite3
|
2020-07-07 14:10:36 +00:00
|
|
|
|
|
|
|
from .fixtures import *
|
|
|
|
|
2020-08-04 13:42:30 +00:00
|
|
|
def test_depth_flag_is_accepted(process, disable_extractors_dict):
|
|
|
|
arg_process = subprocess.run(["archivebox", "add", "http://127.0.0.1:8080/static/example.com.html", "--depth=0"],
|
|
|
|
capture_output=True, env=disable_extractors_dict)
|
2020-07-07 15:25:02 +00:00
|
|
|
assert 'unrecognized arguments: --depth' not in arg_process.stderr.decode("utf-8")
|
|
|
|
|
2020-08-18 08:52:56 +00:00
|
|
|
|
2020-08-04 13:42:30 +00:00
|
|
|
def test_depth_flag_fails_if_it_is_not_0_or_1(process, disable_extractors_dict):
|
2020-08-18 08:52:56 +00:00
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--depth=5", "http://127.0.0.1:8080/static/example.com.html"],
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
2020-07-07 15:25:02 +00:00
|
|
|
assert 'invalid choice' in arg_process.stderr.decode("utf-8")
|
2020-08-18 08:52:56 +00:00
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--depth=-1", "http://127.0.0.1:8080/static/example.com.html"],
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
2020-07-07 15:25:02 +00:00
|
|
|
assert 'invalid choice' in arg_process.stderr.decode("utf-8")
|
2020-07-07 14:49:28 +00:00
|
|
|
|
2020-08-18 08:52:56 +00:00
|
|
|
|
2020-08-04 13:42:30 +00:00
|
|
|
def test_depth_flag_0_crawls_only_the_arg_page(tmp_path, process, disable_extractors_dict):
|
2020-08-18 08:52:56 +00:00
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--depth=0", "http://127.0.0.1:8080/static/example.com.html"],
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
2020-07-07 14:49:28 +00:00
|
|
|
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
2021-03-27 05:01:29 +00:00
|
|
|
with open(archived_item_path / "index.json", "r", encoding='utf-8') as f:
|
2020-07-07 14:49:28 +00:00
|
|
|
output_json = json.load(f)
|
2020-07-17 21:55:56 +00:00
|
|
|
assert output_json["base_url"] == "127.0.0.1:8080/static/example.com.html"
|
2020-07-07 15:07:44 +00:00
|
|
|
|
2020-08-18 08:52:56 +00:00
|
|
|
|
2020-08-04 13:42:30 +00:00
|
|
|
def test_depth_flag_1_crawls_the_page_AND_links(tmp_path, process, disable_extractors_dict):
|
2020-08-18 08:52:56 +00:00
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--depth=1", "http://127.0.0.1:8080/static/example.com.html"],
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
2020-10-08 16:02:26 +00:00
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
urls = c.execute("SELECT url from core_snapshot").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
urls = list(map(lambda x: x[0], urls))
|
|
|
|
assert "http://127.0.0.1:8080/static/example.com.html" in urls
|
|
|
|
assert "http://127.0.0.1:8080/static/iana.org.html" in urls
|
2020-08-18 08:52:56 +00:00
|
|
|
|
|
|
|
|
|
|
|
def test_overwrite_flag_is_accepted(process, disable_extractors_dict):
|
|
|
|
subprocess.run(
|
|
|
|
["archivebox", "add", "--depth=0", "http://127.0.0.1:8080/static/example.com.html"],
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--overwrite", "http://127.0.0.1:8080/static/example.com.html"],
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
assert 'unrecognized arguments: --overwrite' not in arg_process.stderr.decode("utf-8")
|
|
|
|
assert 'favicon' in arg_process.stdout.decode('utf-8'), 'archive methods probably didnt run, did overwrite work?'
|
2020-08-28 16:08:03 +00:00
|
|
|
|
|
|
|
def test_add_updates_history_json_index(tmp_path, process, disable_extractors_dict):
|
|
|
|
subprocess.run(
|
|
|
|
["archivebox", "add", "--depth=0", "http://127.0.0.1:8080/static/example.com.html"],
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
2020-10-08 16:02:26 +00:00
|
|
|
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
|
|
|
|
2021-03-27 05:01:29 +00:00
|
|
|
with open(archived_item_path / "index.json", "r", encoding="utf-8") as f:
|
2020-08-28 16:08:03 +00:00
|
|
|
output_json = json.load(f)
|
2020-11-13 14:24:34 +00:00
|
|
|
assert output_json["history"] != {}
|
|
|
|
|
|
|
|
def test_extract_input_uses_only_passed_extractors(tmp_path, process):
|
|
|
|
subprocess.run(["archivebox", "add", "http://127.0.0.1:8080/static/example.com.html", "--extract", "wget"],
|
|
|
|
capture_output=True)
|
|
|
|
|
|
|
|
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
|
|
|
|
|
|
|
assert (archived_item_path / "warc").exists()
|
2021-03-27 05:01:29 +00:00
|
|
|
assert not (archived_item_path / "singlefile.html").exists()
|
2024-02-27 22:48:19 +00:00
|
|
|
|
|
|
|
def test_json(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example.json', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=json"],
|
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
urls = c.execute("SELECT url from core_snapshot").fetchall()
|
|
|
|
tags = c.execute("SELECT name from core_tag").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
urls = list(map(lambda x: x[0], urls))
|
|
|
|
assert "http://127.0.0.1:8080/static/example.com.html" in urls
|
2024-02-29 01:38:49 +00:00
|
|
|
assert "http://127.0.0.1:8080/static/iana.org.html" in urls
|
|
|
|
assert "http://127.0.0.1:8080/static/shift_jis.html" in urls
|
|
|
|
assert "http://127.0.0.1:8080/static/title_og_with_html" in urls
|
2024-02-27 22:48:19 +00:00
|
|
|
# if the following URL appears, we must have fallen back to another parser
|
|
|
|
assert not "http://www.example.com/should-not-exist" in urls
|
|
|
|
|
|
|
|
tags = list(map(lambda x: x[0], tags))
|
|
|
|
assert "Tag1" in tags
|
|
|
|
assert "Tag2" in tags
|
2024-02-29 01:38:49 +00:00
|
|
|
assert "Tag3" in tags
|
|
|
|
assert "Tag4 with Space" in tags
|
|
|
|
assert "Tag5" in tags
|
|
|
|
assert "Tag6 with Space" in tags
|
2024-02-27 22:48:19 +00:00
|
|
|
|
|
|
|
def test_json_with_leading_garbage(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example.json.bad', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=json"],
|
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
urls = c.execute("SELECT url from core_snapshot").fetchall()
|
|
|
|
tags = c.execute("SELECT name from core_tag").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
urls = list(map(lambda x: x[0], urls))
|
|
|
|
assert "http://127.0.0.1:8080/static/example.com.html" in urls
|
|
|
|
# if the following URL appears, we must have fallen back to another parser
|
|
|
|
assert not "http://www.example.com/should-not-exist" in urls
|
|
|
|
|
|
|
|
tags = list(map(lambda x: x[0], tags))
|
|
|
|
assert "Tag1" in tags
|
|
|
|
assert "Tag2" in tags
|
2024-03-01 19:27:59 +00:00
|
|
|
|
|
|
|
def test_generic_rss(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example.rss', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=rss"],
|
2024-03-01 19:22:28 +00:00
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
urls = c.execute("SELECT url from core_snapshot").fetchall()
|
|
|
|
tags = c.execute("SELECT name from core_tag").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
urls = list(map(lambda x: x[0], urls))
|
|
|
|
assert "http://127.0.0.1:8080/static/example.com.html" in urls
|
|
|
|
# if the following URL appears, we must have fallen back to another parser
|
2024-03-01 19:27:59 +00:00
|
|
|
assert not "http://purl.org/dc/elements/1.1/" in urls
|
|
|
|
|
|
|
|
tags = list(map(lambda x: x[0], tags))
|
|
|
|
assert "Tag1 Tag2" in tags
|
|
|
|
|
|
|
|
def test_pinboard_rss(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example.rss', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=pinboard_rss"],
|
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
tags = c.execute("SELECT name from core_tag").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
tags = list(map(lambda x: x[0], tags))
|
|
|
|
assert "Tag1" in tags
|
|
|
|
assert "Tag2" in tags
|
|
|
|
|
|
|
|
def test_atom(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example.atom', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=rss"],
|
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
urls = c.execute("SELECT url from core_snapshot").fetchall()
|
|
|
|
tags = c.execute("SELECT name from core_tag").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
urls = list(map(lambda x: x[0], urls))
|
|
|
|
assert "http://127.0.0.1:8080/static/example.com.html" in urls
|
|
|
|
# if the following URL appears, we must have fallen back to another parser
|
|
|
|
assert not "http://www.w3.org/2005/Atom" in urls
|
2024-03-01 19:22:28 +00:00
|
|
|
|
|
|
|
tags = list(map(lambda x: x[0], tags))
|
|
|
|
assert "Tag1" in tags
|
|
|
|
assert "Tag2" in tags
|
2024-03-01 02:15:06 +00:00
|
|
|
|
|
|
|
def test_jsonl(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example.jsonl', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=jsonl"],
|
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
urls = c.execute("SELECT url from core_snapshot").fetchall()
|
|
|
|
tags = c.execute("SELECT name from core_tag").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
urls = list(map(lambda x: x[0], urls))
|
|
|
|
assert "http://127.0.0.1:8080/static/example.com.html" in urls
|
|
|
|
assert "http://127.0.0.1:8080/static/iana.org.html" in urls
|
|
|
|
assert "http://127.0.0.1:8080/static/shift_jis.html" in urls
|
|
|
|
assert "http://127.0.0.1:8080/static/title_og_with_html" in urls
|
|
|
|
# if the following URL appears, we must have fallen back to another parser
|
|
|
|
assert not "http://www.example.com/should-not-exist" in urls
|
|
|
|
|
|
|
|
tags = list(map(lambda x: x[0], tags))
|
|
|
|
assert "Tag1" in tags
|
|
|
|
assert "Tag2" in tags
|
|
|
|
assert "Tag3" in tags
|
|
|
|
assert "Tag4 with Space" in tags
|
|
|
|
assert "Tag5" in tags
|
|
|
|
assert "Tag6 with Space" in tags
|
|
|
|
|
|
|
|
def test_jsonl_single(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example-single.jsonl', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=jsonl"],
|
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
conn = sqlite3.connect("index.sqlite3")
|
|
|
|
c = conn.cursor()
|
|
|
|
urls = c.execute("SELECT url from core_snapshot").fetchall()
|
|
|
|
tags = c.execute("SELECT name from core_tag").fetchall()
|
|
|
|
conn.commit()
|
|
|
|
conn.close()
|
|
|
|
|
|
|
|
urls = list(map(lambda x: x[0], urls))
|
|
|
|
assert "http://127.0.0.1:8080/static/example.com.html" in urls
|
|
|
|
# if the following URL appears, we must have fallen back to another parser
|
|
|
|
assert not "http://www.example.com/should-not-exist" in urls
|
|
|
|
|
|
|
|
tags = list(map(lambda x: x[0], tags))
|
|
|
|
assert "Tag1" in tags
|
|
|
|
assert "Tag2" in tags
|
|
|
|
|
|
|
|
# make sure that JSON parser rejects a single line of JSONL which is valid
|
|
|
|
# JSON but not our expected format
|
|
|
|
def test_json_single(tmp_path, process, disable_extractors_dict):
|
|
|
|
with open('../../mock_server/templates/example-single.jsonl', 'r', encoding='utf-8') as f:
|
|
|
|
arg_process = subprocess.run(
|
|
|
|
["archivebox", "add", "--index-only", "--parser=json"],
|
|
|
|
stdin=f,
|
|
|
|
capture_output=True,
|
|
|
|
env=disable_extractors_dict,
|
|
|
|
)
|
|
|
|
|
|
|
|
assert 'expects list of objects' in arg_process.stderr.decode("utf-8")
|