![]() Related to https://github.com/webrecorder/browsertrix-crawler/issues/848 Several users have had issues with disk utilization checks, including the values reported by `df` inside the crawler container having unexpected results for mounted volumes. The commonly recommended solution to this is to use `docker system ps`, but that is of course not available within the Docker container itself. This PR changes disk utilization checks to be an opt-in feature by setting the default value to `0` (disabled). |
||
---|---|---|
.github/workflows | ||
.husky | ||
config/policies | ||
docs | ||
html | ||
src | ||
tests | ||
.dockerignore | ||
.eslintignore | ||
.eslintrc.cjs | ||
.gitignore | ||
.pre-commit-config.yaml | ||
.prettierignore | ||
.prettierrc | ||
CHANGES.md | ||
docker-compose.yml | ||
docker-entrypoint.sh | ||
Dockerfile | ||
LICENSE | ||
NOTICE | ||
package.json | ||
README.md | ||
requirements.txt | ||
test-setup.js | ||
tsconfig.eslint.json | ||
tsconfig.json | ||
yarn.lock |
Browsertrix Crawler 1.x
Browsertrix Crawler is a standalone browser-based high-fidelity crawling system, designed to run a complex, customizable browser-based crawl in a single Docker container. Browsertrix Crawler uses Puppeteer to control one or more Brave Browser browser windows in parallel. Data is captured through the Chrome Devtools Protocol (CDP) in the browser.
For information on how to use and develop Browsertrix Crawler, see the hosted Browsertrix Crawler documentation.
For information on how to build the docs locally, see the docs page.
Support
Initial support for 0.x version of Browsertrix Crawler, was provided by Kiwix. The initial functionality for Browsertrix Crawler was developed to support the zimit project in a collaboration between Webrecorder and Kiwix, and this project has been split off from Zimit into a core component of Webrecorder.
Additional support for Browsertrix Crawler, including for the development of the 0.4.x version has been provided by Portico.