I know similar questions have been asked, but I'm not sure about the answers (and I can't easily test all of them), so before I go crazy continuing to search, I want to ask: Is there an easy way to crawl all the pages on a website and check them for broken and invalid links automatically? Prefereably I'd like a solution that does not require an install or compile as I'm severely limited. Thanks.
You could use Xenu's Link Sleuth. It does require an install, but is light-weight and no compile is needed.