-
Notifications
You must be signed in to change notification settings - Fork 128
Conversation
## Linkchecker | ||
The "linkchecker" verifies that no dead links are present on the website. | ||
`linkchecker` is a python app that can be acquired through pip by running | ||
`pip install LinkChecker`. To verify that there aren't any dead links present |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
oof. how complicated is it? could we have a go replacement?
also, ipfs is reliable enough to start "vendoring" deps to ipfs itself.
(having to manually run pip is frustrating and complicates usage and CI)
Hm, how bout a nodejs solution, as the website is already built by node tools? Two options from a quick google: |
@harlantwood those sounds good to me. thoughts @dylanPowers ? |
I checked out the node implementations first and for the most part they weren't very feature complete and buggy.
This python implementation is the most mature and feature complete at this point in time. It's also a non-trivial app https://github.com/wummel/linkchecker/tree/master/linkcheck Using go would still introduce another dependency so I'm not sure that's a great solution either, but there are some options there as well that could be looked at. |
python is installed by default on pretty much every linux machine. I dont really think having it as a dependency is all that bad |
FWIW this is the patch to get
|
But admittedly its not mature. I agree that python dependency is not so bad, tho it would be elegant to have a node solution. |
I don't want to depend on pip. it can be a nightmare. can we check the deps we need into ipfs and seed them? (assume python, just the modules) |
Update on the node options:
|
Maybe this will help? From the simplecrawler docs:
|
From my interpretation of the docs, filterByDomain means that the crawler will crawl the entire external website. Also mentioned in simplecrawler/simplecrawler#114. Update: |
I found a tool in python that checks for dead or invalid links. I went ahead and wrote up some directions on how to use it while also setting up a small configuration file.
This fixes #13