1
0
Fork 0

add --dump-dom docs

This commit is contained in:
Nick Sweeting 2018-06-10 18:27:26 -04:00 committed by GitHub
parent 9abe4badf9
commit b5cc6ca65f
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23

View file

@ -15,7 +15,7 @@ Save an archived copy of all websites you bookmark (the actual *content* of each
- Shaarli, Delicious, Instapaper, Reddit Saved Posts, Wallabag, Unmark.it, and more! - Shaarli, Delicious, Instapaper, Reddit Saved Posts, Wallabag, Unmark.it, and more!
- Contiuous archiving of browsing history (every site, not just bookmarked) (coming soon!...) - Contiuous archiving of browsing history (every site, not just bookmarked) (coming soon!...)
Outputs browsable static html archives of each site, a PDF, a screenshot, and a link to a copy on archive.org, all indexed with nice html & json files. Outputs browsable static html archives of each site made using wget, a PDF, a screenshot, a DOM dump from headless chrome, and a link to a copy on archive.org, all indexed with nice html & json files.
[DEMO: archive.sweeting.me](https://archive.sweeting.me) [DEMO: archive.sweeting.me](https://archive.sweeting.me)
@ -143,6 +143,7 @@ env CHROME_BINARY=google-chrome-stable RESOLUTION=1440,900 FETCH_PDF=False ./arc
- fetch images/css/js with wget: `FETCH_WGET_REQUISITES` (True is highly recommended) - fetch images/css/js with wget: `FETCH_WGET_REQUISITES` (True is highly recommended)
- print page as PDF: `FETCH_PDF` - print page as PDF: `FETCH_PDF`
- fetch a screenshot of the page: `FETCH_SCREENSHOT` - fetch a screenshot of the page: `FETCH_SCREENSHOT`
- fetch a DOM dump of the page: `FETCH_DOM`
- fetch a favicon for the page: `FETCH_FAVICON` - fetch a favicon for the page: `FETCH_FAVICON`
- submit the page to archive.org: `SUBMIT_ARCHIVE_DOT_ORG` - submit the page to archive.org: `SUBMIT_ARCHIVE_DOT_ORG`
- screenshot: `RESOLUTION` values: [`1440,900`]/`1024,768`/`...` - screenshot: `RESOLUTION` values: [`1440,900`]/`1024,768`/`...`
@ -421,6 +422,7 @@ will run fast subsequent times because it only downloads new links that haven't
## Changelog ## Changelog
- support for chrome `--dump-dom` to output full page HTML after JS executes
- v0.0.3 released - v0.0.3 released
- support for chrome `--user-data-dir` to archive sites that need logins - support for chrome `--user-data-dir` to archive sites that need logins
- fancy individual html & json indexes for each link - fancy individual html & json indexes for each link