forked from arthurdejong/python-stdnum
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathgetisil.py
executable file
·71 lines (60 loc) · 2.48 KB
/
getisil.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
#!/usr/bin/env python
# getisil.py - script to donwload ISIL agencies
#
# Copyright (C) 2011 Arthur de Jong
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
# 02110-1301 USA
"""This script downloads a web page from the ISIL Registration Authority
and screen-scrapes the national and non-national ISIL agencies and
code prefixes."""
import urllib
import BeautifulSoup
import re
spaces_re = re.compile('\s+', re.UNICODE)
# the web page that holds information on the ISIL authorities
download_url = 'http://biblstandard.dk/isil/'
def clean(s):
"""Cleans up the string removing unneeded stuff from it."""
return spaces_re.sub(' ', s.replace(u'\u0096', '')).strip().encode('utf-8')
def parse(f):
"""Parse the specified file."""
print '# generated from ISIL Registration Authority, downloaded from'
print '# %s' % download_url
soup = BeautifulSoup.BeautifulSoup(f, convertEntities='html')
# find all table rows
for tr in soup.findAll('tr'):
# find the rows with four columns of text
tds = tr.findAll('td', attrs={'class': 'text'}, recursive=False)
if len(tds) == 4:
props = {}
cc = clean(tds[0].string)
if tds[1].string:
props['country'] = clean(tds[1].contents[0])
ra_a = tds[2].find('a')
if ra_a:
props['ra'] = clean(ra_a.string)
props['ra_url'] = clean(ra_a['href'])
elif tds[2].string:
props['ra'] = clean(tds[2].string)
# we could also get the search urls from tds[3].findAll('a')
print '%s$ %s' % (
cc,
' '.join(['%s="%s"' % (x, y)
for x, y in props.iteritems()]))
if __name__ == '__main__':
#f = open('isil.html', 'r')
f = urllib.urlopen(download_url)
parse(f)