urlook

command module
v0.0.0-...-6188d4b Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Oct 27, 2019 License: MIT Imports: 8 Imported by: 0

README

URLook

Build Status MIT license Go Report Card

Find and check http(s) links in the text file.

Install

$ go get github.com/olshevskiy87/urlook

Usage

Usage: urlook [--fail-on-duplicate] [--timeout TIMEOUT] [--white WHITE] [FILENAMES [FILENAMES ...]]

Positional arguments:
  FILENAMES              filenames with links to check

Options:
  --fail-on-duplicate    fail if there is a duplicate url
  --timeout TIMEOUT, -t TIMEOUT
                         request timeout in seconds [default: 10]
  --white WHITE, -w WHITE
                         white list url (can be specified multiple times)
  --help, -h             display this help and exit
  --version              display version and exit

Examples

$ urlook test_links
URLs to check: 7
  1. http://cb.vu/unixtoolbox.xhtml
  2. https://eax.me
  3. http://eax.me
  4. https://google.com
  5. https://github.com
  6. https://github.com/non-existent-url-non-existent-url-non-existent-url
  7. https://www.reddit.com
✓→✓→✓x
  1. http://eax.me [301, Moved Permanently] -> https://eax.me/
  2. https://google.com [302, Found] -> https://www.google.ru/?gfe_rd=cr&dcr=0&ei=EFSZWqWQAcGDtAGArrLYBw
  3. https://github.com/non-existent-url-non-existent-url-non-existent-url [404, Not Found]
issues found: 3
$ grep github test_links | urlook --white non-existent-url
URLs to check: 1
  1. https://github.com
✓
White listed URLs (1):
 - https://github.com/non-existent-url-non-existent-url-non-existent-url

no issues found
$ echo 'check these links: https://ya.ru, https://www.reddit.com' | urlook
URLs to check: 2
  1. https://ya.ru
  2. https://www.reddit.com
✓✓
no issues found
$ echo 'https://ya.ru https://github.com https://ya.ru' | urlook --fail-on-duplicate
URLs to check: 2
  1. https://github.com
  2. https://ya.ru
✓✓
Duplicates:
 - https://ya.ru (2)
duplicates found: 1

Todo

  • read input from stdin
  • remove duplicate urls
  • add cli-option to fail on duplicate url
  • add cli-option timeout
  • add white list option
  • read more than one file
  • add tests
  • add option to allow redirects

Motivations

Inspired by ruby gem awesome_bot

License

MIT. See LICENSE for details.

Documentation

The Go Gopher

There is no documentation for this package.

Directories

Path Synopsis

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL