Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
C
crawl
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Model registry
Operate
Environments
Monitor
Incidents
Service Desk
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
ale
crawl
Commits
59f3725f
Commit
59f3725f
authored
6 years ago
by
ale
Browse files
Options
Downloads
Patches
Plain Diff
Explicitly mention the crawler limitations
parent
66ce654d
No related branches found
Branches containing commit
No related tags found
No related merge requests found
Changes
1
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
README.md
+28
-5
28 additions, 5 deletions
README.md
with
28 additions
and
5 deletions
README.md
+
28
−
5
View file @
59f3725f
...
...
@@ -29,8 +29,8 @@ as arguments on the command line:
$ crawl http://example.com/
By default, the tool will store the output WARC file and its own
database in the current directory. This can be
controlled with the
*--output*
and
*--state*
command-line options.
temporary crawl
database in the current directory. This can be
controlled with the
*--output*
and
*--state*
command-line options.
The crawling scope is controlled with a set of overlapping checks:
...
...
@@ -44,6 +44,29 @@ The crawling scope is controlled with a set of overlapping checks:
If the program is interrupted, running it again with the same command
line from the same directory will cause it to resume crawling from
where it stopped. At the end of a successful crawl, the database will
be removed (unless you specify the
*--keep*
option, for debugging
purposes).
where it stopped. At the end of a successful crawl, the temporary
crawl database will be removed (unless you specify the
*--keep*
option, for debugging purposes).
It is possible to tell the crawler to exclude URLs matching specific
regex patterns by using the
*--exclude*
or
*--exclude-from-file*
options. These option may be repeated multiple times. The crawler
comes with its own builtin set of URI regular expressions meant to
avoid calendars, admin panels of common CMS applications, and other
well-known pitfalls. This list is sourced from the
[
ArchiveBot
](
https://github.com/ArchiveTeam/ArchiveBot
)
project.
## Limitations
Like most crawlers, this one has a number of limitations:
*
it completely ignores
*robots.txt*
. You can make such policy
decisions yourself by turning the robots.txt into a list of patterns
to be used with
*--exclude-file*
.
*
it does not embed a Javascript engine, so Javascript-rendered
elements will not be detected.
*
CSS parsing is limited (uses regular expressions), so some
*url()*
resources might not be detected.
*
it expects reasonably well-formed HTML, so it may fail to extract
links from particularly broken pages.
*
support for
\<
object
\>
and
\<
video
\>
tags is limited.
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment