mirror of
https://github.com/0xAX/linux-insides
synced 2024-11-19 15:25:41 +00:00
add one script to check one link is live or not
This commit is contained in:
parent
a7a2f3e1fb
commit
5de39c8539
13
Scripts/README.md
Normal file
13
Scripts/README.md
Normal file
@ -0,0 +1,13 @@
|
||||
# Scripts
|
||||
|
||||
## Description
|
||||
|
||||
`get_all_links.py` : justify one link is live or dead with network connection
|
||||
|
||||
## Usage
|
||||
|
||||
`get_all_links.py` :
|
||||
|
||||
```
|
||||
./get_all_links.py ../
|
||||
```
|
77
Scripts/get_all_links.py
Executable file
77
Scripts/get_all_links.py
Executable file
@ -0,0 +1,77 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
from __future__ import print_function
|
||||
from socket import timeout
|
||||
|
||||
import os
|
||||
import sys
|
||||
import codecs
|
||||
import re
|
||||
|
||||
import markdown
|
||||
|
||||
try:
|
||||
# compatible for python2
|
||||
from urllib2 import urlopen
|
||||
from urllib2 import HTTPError
|
||||
from urllib2 import URLError
|
||||
except ImportError:
|
||||
# compatible for python3
|
||||
from urllib.request import urlopen
|
||||
from urllib.error import HTTPError
|
||||
from urllib.error import URLError
|
||||
|
||||
def check_live_url(url):
|
||||
|
||||
result = False
|
||||
try:
|
||||
ret = urlopen(url, timeout=2)
|
||||
result = (ret.code == 200)
|
||||
except HTTPError as e:
|
||||
print(e, file=sys.stderr)
|
||||
except URLError as e:
|
||||
print(e, file=sys.stderr)
|
||||
except timeout as e:
|
||||
print(e, file=sys.stderr)
|
||||
except Exception as e:
|
||||
print(e, file=sys.stderr)
|
||||
|
||||
return result
|
||||
|
||||
|
||||
def main(path):
|
||||
|
||||
filenames = []
|
||||
for (dirpath, dnames, fnames) in os.walk(path):
|
||||
for fname in fnames:
|
||||
if fname.endswith('.md'):
|
||||
filenames.append(os.sep.join([dirpath, fname]))
|
||||
|
||||
urls = []
|
||||
|
||||
for filename in filenames:
|
||||
fd = codecs.open(filename, mode="r", encoding="utf-8")
|
||||
for line in fd.readlines():
|
||||
refs = re.findall(r'(?<=<a href=")[^"]*', markdown.markdown(line))
|
||||
for ref in refs:
|
||||
if ref not in urls:
|
||||
urls.append(ref)
|
||||
|
||||
#print(len(urls))
|
||||
|
||||
for url in urls:
|
||||
if not url.startswith("http"):
|
||||
print("markdown file name: " + url)
|
||||
continue
|
||||
if check_live_url(url):
|
||||
print(url)
|
||||
else:
|
||||
print(url, file=sys.stderr)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
|
||||
if len(sys.argv) == 2:
|
||||
main(sys.argv[1])
|
||||
else:
|
||||
print("Choose one path as argument one")
|
Loading…
Reference in New Issue
Block a user