You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The crawler has a documented --allowHashUrls option, but it doesn’t appear to do anything. Searching the codebase, I can’t find any references to it except for the argument parser, so it doesn’t seem to actually be used.
I had expected this to allow a seed I’d listed with a hash URL to get captured. For example, using the following config:
The crawler has a documented
--allowHashUrls
option, but it doesn’t appear to do anything. Searching the codebase, I can’t find any references to it except for the argument parser, so it doesn’t seem to actually be used.I had expected this to allow a seed I’d listed with a hash URL to get captured. For example, using the following config:
Is this something that’s just not hooked up, or maybe a vestigial feature that was supposed to be removed?
The workarounds I’m currently trying are:
or:
or:
(Side note: I’d hoped I could use
allowHash
on the seed withscopeType: page
at the top level, but it looks like that scope type always preventsallowHash
from being configured, which seems less than ideal.)The text was updated successfully, but these errors were encountered: