fix: Add change to calculate wget folder when there is a port present
This commit is contained in:
parent
58ac44c867
commit
23e6803f02
6 changed files with 19 additions and 16 deletions
|
@ -136,7 +136,6 @@ def wget_output_path(link: Link) -> Optional[str]:
|
||||||
|
|
||||||
See docs on wget --adjust-extension (-E)
|
See docs on wget --adjust-extension (-E)
|
||||||
"""
|
"""
|
||||||
|
|
||||||
if is_static_file(link.url):
|
if is_static_file(link.url):
|
||||||
return without_scheme(without_fragment(link.url))
|
return without_scheme(without_fragment(link.url))
|
||||||
|
|
||||||
|
@ -174,10 +173,9 @@ def wget_output_path(link: Link) -> Optional[str]:
|
||||||
full_path = without_fragment(without_query(path(link.url))).strip('/')
|
full_path = without_fragment(without_query(path(link.url))).strip('/')
|
||||||
search_dir = os.path.join(
|
search_dir = os.path.join(
|
||||||
link.link_dir,
|
link.link_dir,
|
||||||
domain(link.url),
|
domain(link.url).replace(":", "+"),
|
||||||
urldecode(full_path),
|
urldecode(full_path),
|
||||||
)
|
)
|
||||||
|
|
||||||
for _ in range(4):
|
for _ in range(4):
|
||||||
if os.path.exists(search_dir):
|
if os.path.exists(search_dir):
|
||||||
if os.path.isdir(search_dir):
|
if os.path.isdir(search_dir):
|
||||||
|
|
|
@ -42,7 +42,7 @@
|
||||||
<p>This domain is for use in illustrative examples in documents. You may use this
|
<p>This domain is for use in illustrative examples in documents. You may use this
|
||||||
domain in literature without prior coordination or asking for permission.</p>
|
domain in literature without prior coordination or asking for permission.</p>
|
||||||
<p>
|
<p>
|
||||||
<a href="http://localhost:8080/static/iana.org.html">More information...</a>
|
<a href="http://127.0.0.1:8080/static/iana.org.html">More information...</a>
|
||||||
</p>
|
</p>
|
||||||
</div>
|
</div>
|
||||||
</body>
|
</body>
|
||||||
|
|
|
@ -4,25 +4,25 @@ import json
|
||||||
from .fixtures import *
|
from .fixtures import *
|
||||||
|
|
||||||
def test_depth_flag_is_accepted(process):
|
def test_depth_flag_is_accepted(process):
|
||||||
arg_process = subprocess.run(["archivebox", "add", "http://localhost:8080/static/example.com.html", "--depth=0"], capture_output=True)
|
arg_process = subprocess.run(["archivebox", "add", "http://127.0.0.1:8080/static/example.com.html", "--depth=0"], capture_output=True)
|
||||||
assert 'unrecognized arguments: --depth' not in arg_process.stderr.decode("utf-8")
|
assert 'unrecognized arguments: --depth' not in arg_process.stderr.decode("utf-8")
|
||||||
|
|
||||||
def test_depth_flag_fails_if_it_is_not_0_or_1(process):
|
def test_depth_flag_fails_if_it_is_not_0_or_1(process):
|
||||||
arg_process = subprocess.run(["archivebox", "add", "http://localhost:8080/static/example.com.html", "--depth=5"], capture_output=True)
|
arg_process = subprocess.run(["archivebox", "add", "http://127.0.0.1:8080/static/example.com.html", "--depth=5"], capture_output=True)
|
||||||
assert 'invalid choice' in arg_process.stderr.decode("utf-8")
|
assert 'invalid choice' in arg_process.stderr.decode("utf-8")
|
||||||
arg_process = subprocess.run(["archivebox", "add", "http://localhost:8080/static/example.com.html", "--depth=-1"], capture_output=True)
|
arg_process = subprocess.run(["archivebox", "add", "http://127.0.0.1:8080/static/example.com.html", "--depth=-1"], capture_output=True)
|
||||||
assert 'invalid choice' in arg_process.stderr.decode("utf-8")
|
assert 'invalid choice' in arg_process.stderr.decode("utf-8")
|
||||||
|
|
||||||
def test_depth_flag_0_crawls_only_the_arg_page(tmp_path, process):
|
def test_depth_flag_0_crawls_only_the_arg_page(tmp_path, process):
|
||||||
arg_process = subprocess.run(["archivebox", "add", "http://localhost:8080/static/example.com.html", "--depth=0"], capture_output=True)
|
arg_process = subprocess.run(["archivebox", "add", "http://127.0.0.1:8080/static/example.com.html", "--depth=0"], capture_output=True)
|
||||||
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
||||||
with open(archived_item_path / "index.json", "r") as f:
|
with open(archived_item_path / "index.json", "r") as f:
|
||||||
output_json = json.load(f)
|
output_json = json.load(f)
|
||||||
assert output_json["base_url"] == "localhost:8080/static/example.com.html"
|
assert output_json["base_url"] == "127.0.0.1:8080/static/example.com.html"
|
||||||
|
|
||||||
def test_depth_flag_1_crawls_the_page_AND_links(tmp_path, process):
|
def test_depth_flag_1_crawls_the_page_AND_links(tmp_path, process):
|
||||||
arg_process = subprocess.run(["archivebox", "add", "http://localhost:8080/static/example.com.html", "--depth=1"], capture_output=True)
|
arg_process = subprocess.run(["archivebox", "add", "http://127.0.0.1:8080/static/example.com.html", "--depth=1"], capture_output=True)
|
||||||
with open(tmp_path / "index.json", "r") as f:
|
with open(tmp_path / "index.json", "r") as f:
|
||||||
archive_file = f.read()
|
archive_file = f.read()
|
||||||
assert "http://localhost:8080/static/example.com.html" in archive_file
|
assert "http://127.0.0.1:8080/static/example.com.html" in archive_file
|
||||||
assert "http://localhost:8080/static/iana.org.html" in archive_file
|
assert "http://127.0.0.1:8080/static/iana.org.html" in archive_file
|
||||||
|
|
5
tests/test_extractors.py
Normal file
5
tests/test_extractors.py
Normal file
|
@ -0,0 +1,5 @@
|
||||||
|
from .fixtures import *
|
||||||
|
|
||||||
|
def test_wget_broken_pipe(tmp_path, process):
|
||||||
|
add_process = subprocess.run(['archivebox', 'add', 'http://127.0.0.1:8080/static/example.com.html'], capture_output=True)
|
||||||
|
assert "TypeError chmod_file(..., path: str) got unexpected NoneType argument path=None" not in add_process.stdout.decode("utf-8")
|
|
@ -18,7 +18,7 @@ def test_update(tmp_path, process):
|
||||||
|
|
||||||
def test_add_link(tmp_path, process):
|
def test_add_link(tmp_path, process):
|
||||||
os.chdir(tmp_path)
|
os.chdir(tmp_path)
|
||||||
add_process = subprocess.run(['archivebox', 'add', 'http://localhost:8080/static/example.com.html'], capture_output=True)
|
add_process = subprocess.run(['archivebox', 'add', 'http://127.0.0.1:8080/static/example.com.html'], capture_output=True)
|
||||||
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
||||||
|
|
||||||
assert "index.json" in [x.name for x in archived_item_path.iterdir()]
|
assert "index.json" in [x.name for x in archived_item_path.iterdir()]
|
||||||
|
@ -34,7 +34,7 @@ def test_add_link(tmp_path, process):
|
||||||
def test_add_link_support_stdin(tmp_path, process):
|
def test_add_link_support_stdin(tmp_path, process):
|
||||||
os.chdir(tmp_path)
|
os.chdir(tmp_path)
|
||||||
stdin_process = subprocess.Popen(["archivebox", "add"], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
stdin_process = subprocess.Popen(["archivebox", "add"], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
||||||
stdin_process.communicate(input="http://localhost:8080/static/example.com.html".encode())
|
stdin_process.communicate(input="http://127.0.0.1:8080/static/example.com.html".encode())
|
||||||
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
archived_item_path = list(tmp_path.glob('archive/**/*'))[0]
|
||||||
|
|
||||||
assert "index.json" in [x.name for x in archived_item_path.iterdir()]
|
assert "index.json" in [x.name for x in archived_item_path.iterdir()]
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
from archivebox import util
|
from archivebox import util
|
||||||
|
|
||||||
def test_download_url_downloads_content():
|
def test_download_url_downloads_content():
|
||||||
text = util.download_url("http://localhost:8080/static/example.com.html")
|
text = util.download_url("http://127.0.0.1:8080/static/example.com.html")
|
||||||
assert "Example Domain" in text
|
assert "Example Domain" in text
|
Loading…
Reference in a new issue