Displaying 1 to 6 from 6 results

html-proofer - Test your rendered HTML files to make sure they're accurate.

  •    Ruby

If you generate HTML files, then this tool might be for you. HTMLProofer is a set of tests to validate your HTML output. These tests check if your image references are legitimate, if they have alt tags, if your internal links are working, and so on. It's intended to be an all-in-one checker for your output.

broken-link-checker - Find broken links, missing images, etc in your HTML.

  •    Javascript

Find broken links, missing images, etc in your HTML. Node.js >= 0.10 is required; < 4.0 will need Promise and Object.assign polyfills.

w3c-link-validator - Command line tool, identifying broken links, validate basic html standards and reporting

  •    Javascript

Very good command line tool for W3C validation and broken link detection of your local development or production URL. Validation errors,warnings or suggestions will be reported to terminal. Install with the terminal.

grunt-link-checker - Run node-simple-crawler to discover broken links on your website

  •    Javascript

Run node-simple-crawler to discover broken links on your website. grunt-link-checker will by default find any broken internal links on the given site and will also find broken fragment identifiers by using cheerio to ensure that an element exists with the given identifier. You can figure more options that are available via node-simplecrawler.




htmltest - :white_check_mark: Test generated HTML for problems

  •    Go

If you generate HTML files, html-proofer might be the tool for you. If you can't be bothered with a Ruby environment or fancy something a bit faster, htmltest may be a better option. 🔍 htmltest runs your HTML output through a series of checks to ensure all your links, images, scripts references work, your alt tags are filled in, et cetera.

krawler - A web crawling framework written in Kotlin

  •    Kotlin

Krawler is a web crawling framework written in Kotlin. It is heavily inspired by crawler4j by Yasser Ganjisaffar. The project is still very new, and those looking for a mature, well tested crawler framework should likely still use crawler4j. For those who can tolerate a bit of turbulence, Krawler should serve as a replacement for crawler4j with minimal modifications to existing applications. Using the Krawler framework is fairly simple. Minimally, there are two methods that must be overridden in order to use the framework. Overriding the shouldVisit method dictates what should be visited by the crawler, and the visit method dictates what happens once the page is visited. Overriding these two methods is sufficient for creating your own crawler, however there are additional methods that can be overridden to privde more robust behavior.