Another little PHP script, this time to find dead links. It takes a URL and starts crawling for links, testing if the places they link to are active are not. If the linked URL is on the same domain, it crawls into that page too, recursively (to a limit). Pages that have already been tested in the query are not tested a second time. Only one instance of this script can run at a time, so you may get a weird error message.
Syntax: "http://0.clrhome.tk/b/links/?url=<INITIAL>[&limit=<DEEP>][&set=<DOMAIN>]"
INITIAL: The page to start on (and host)
DEEP: Max number of recursions
DOMAIN: "both" (default), "domestic," or "international," to display all links, internal ones only, or external ones only
I made this after tifreak mentioned (on IRC) that he needed a crawler to find dead links, but I got it done too late. (I'm still going to use it for myself though.) It's a bit slow, but mainly that's because there's a half-second delay between requests.
Syntax: "http://0.clrhome.tk/b/links/?url=<INITIAL>[&limit=<DEEP>][&set=<DOMAIN>]"
INITIAL: The page to start on (and host)
DEEP: Max number of recursions
DOMAIN: "both" (default), "domestic," or "international," to display all links, internal ones only, or external ones only
I made this after tifreak mentioned (on IRC) that he needed a crawler to find dead links, but I got it done too late. (I'm still going to use it for myself though.) It's a bit slow, but mainly that's because there's a half-second delay between requests.