Merge pull request #501 from webrecorder/docs-minor-fixes

Docs: Minor fixes to edit link & clarifications
This commit is contained in:
Henry Wilkinson 2024-03-20 13:04:12 -04:00 committed by GitHub
commit 79e39ae2f0
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
3 changed files with 3 additions and 4 deletions

View file

@ -10,15 +10,14 @@ Welcome to the Browsertrix Crawler official documentation.
Browsertrix Crawler is a simplified browser-based high-fidelity crawling system, designed to run a complex, customizable browser-based crawl in a single Docker container. Browsertrix Crawler uses [Puppeteer](https://github.com/puppeteer/puppeteer) to control one or more [Brave Browser](https://brave.com/) browser windows in parallel. Data is captured through the [Chrome Devtools Protocol (CDP)](https://chromedevtools.github.io/devtools-protocol/) in the browser.
Browsertrix Crawler is a command line application responsible for the core features of [Browsertrix](https://browsertrix.com), Webrecorder's cloud-based web archiving service. See the [Browsertrix documentation] for more information about Browsertrix, the cloud platform.
!!! note
This documentation applies to Browsertrix Crawler versions 1.0.0 and above. Documentation for earlier versions of the crawler is available in the [Browsertrix Crawler Github repository](https://github.com/webrecorder/browsertrix-crawler)'s README file in older commits.
## Features
- Single-container, browser based crawling with a headless/headful browser running pages in multiple windows.
- Support for custom browser behaviors, using [Browsertrix Behaviors](https://github.com/webrecorder/browsertrix-behaviors) including autoscroll, video autoplay, and site-specific behaviors.
- YAML-based configuration, passed via file or via stdin.

View file

@ -1,6 +1,6 @@
# Browsertrix Crawler User Guide
Welcome to the Browsertrix User Guide. This page covers the basics of using Browsertrix Crawler, Webrecorder's browser-based high-fidelity crawling system, designed to run a complex, customizable browser-based crawl in a single Docker container.
Welcome to the Browsertrix Crawler User Guide. This page covers the basics of using Browsertrix Crawler, Webrecorder's browser-based high-fidelity crawling system, designed to run a complex, customizable, browser-based crawl in a single Docker container.
## Getting Started

View file

@ -1,7 +1,7 @@
site_name: Browsertrix Crawler Docs
repo_url: https://github.com/webrecorder/browsertrix-crawler/
repo_name: Browsertrix Crawler
edit_uri: edit/main/docs/
edit_uri: edit/main/docs/docs/
extra_css:
- stylesheets/extra.css
theme: