mirror of
https://github.com/webrecorder/browsertrix-crawler.git
synced 2025-10-19 14:33:17 +00:00
39 lines
1.3 KiB
JavaScript
39 lines
1.3 KiB
JavaScript
![]() |
import util from "util";
|
||
|
import {exec as execCallback } from "child_process";
|
||
|
|
||
|
const exec = util.promisify(execCallback);
|
||
|
|
||
|
|
||
|
test("ensure one invalid seed doesn't end crawl if failOnFailedSeed is not set", async () => {
|
||
|
let passed = true;
|
||
|
try {
|
||
|
await exec("docker run -v $PWD/test-crawls:/crawls webrecorder/browsertrix-crawler crawl --url https://www.iana.org/ --url example.com/invalid-seed --generateWACZ --limit 1 --collection invalidseed");
|
||
|
} catch (error) {
|
||
|
console.log(error);
|
||
|
passed = false;
|
||
|
}
|
||
|
expect(passed).toBe(true);
|
||
|
});
|
||
|
|
||
|
test("ensure one invalid seed fails crawl if failOnFailedSeed is set", async () => {
|
||
|
let passed = true;
|
||
|
try {
|
||
|
await exec("docker run -v $PWD/test-crawls:/crawls webrecorder/browsertrix-crawler crawl --url https://www.iana.org/ --url example.com/invalid-seed --generateWACZ --limit 1 --failOnFailedSeed --collection failseed");
|
||
|
}
|
||
|
catch (error) {
|
||
|
passed = false;
|
||
|
}
|
||
|
expect(passed).toBe(false);
|
||
|
});
|
||
|
|
||
|
test("ensure crawl fails if no valid seeds are passed", async () => {
|
||
|
let passed = true;
|
||
|
try {
|
||
|
await exec("docker run -v $PWD/test-crawls:/crawls webrecorder/browsertrix-crawler crawl --url iana.org/ --url example.com/invalid-seed --generateWACZ --limit 1 --collection allinvalidseeds");
|
||
|
}
|
||
|
catch (error) {
|
||
|
passed = false;
|
||
|
}
|
||
|
expect(passed).toBe(false);
|
||
|
});
|