|
@@ -64,8 +64,8 @@ class WestminsterParser: |
|
|
|
|
|
|
|
|
# Now get the search page |
|
|
# Now get the search page |
|
|
|
|
|
|
|
|
# sys.stderr.write("Fetching: %s\n" %self.base_url) |
|
|
|
|
|
# sys.stderr.write("post data: %s\n" %post_data) |
|
|
|
|
|
|
|
|
sys.stderr.write("Fetching: %s\n" %self.base_url) |
|
|
|
|
|
sys.stderr.write("post data: %s\n" %post_data) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# This gives us something to use as the callback |
|
|
# This gives us something to use as the callback |
|
@@ -81,13 +81,13 @@ class WestminsterParser: |
|
|
|
|
|
|
|
|
curlobj.perform() |
|
|
curlobj.perform() |
|
|
|
|
|
|
|
|
# sys.stderr.write("Got it\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Got it\n") |
|
|
soup = BeautifulSoup(fakefile.getvalue()) |
|
|
soup = BeautifulSoup(fakefile.getvalue()) |
|
|
|
|
|
|
|
|
# We may as well free up the memory used by fakefile |
|
|
# We may as well free up the memory used by fakefile |
|
|
fakefile.close() |
|
|
fakefile.close() |
|
|
|
|
|
|
|
|
# sys.stderr.write("Created soup\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Created soup\n") |
|
|
|
|
|
|
|
|
results_form = soup.find("form", {"name": "currentsearchresultsNext"}) |
|
|
results_form = soup.find("form", {"name": "currentsearchresultsNext"}) |
|
|
|
|
|
|
|
@@ -95,21 +95,21 @@ class WestminsterParser: |
|
|
# If there is no next page then there will be no inputs in the form. |
|
|
# If there is no next page then there will be no inputs in the form. |
|
|
# In this case, post_data will be '', which is false. |
|
|
# In this case, post_data will be '', which is false. |
|
|
|
|
|
|
|
|
# sys.stderr.write("Found form containing results\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Found form containing results\n") |
|
|
|
|
|
|
|
|
post_data = urllib.urlencode([(x['name'], x['value']) for x in results_form.findAll("input")]) |
|
|
post_data = urllib.urlencode([(x['name'], x['value']) for x in results_form.findAll("input")]) |
|
|
|
|
|
|
|
|
# sys.stderr.write("Got post data\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Got post data\n") |
|
|
|
|
|
|
|
|
# Each result has one link, and they are the only links in the form |
|
|
# Each result has one link, and they are the only links in the form |
|
|
|
|
|
|
|
|
links = results_form.findAll("a") |
|
|
links = results_form.findAll("a") |
|
|
|
|
|
|
|
|
# sys.stderr.write("Got list of links\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Got list of links\n") |
|
|
|
|
|
|
|
|
for link in links: |
|
|
for link in links: |
|
|
|
|
|
|
|
|
# sys.stderr.write("Working on link: %s\n" %link['href']) |
|
|
|
|
|
|
|
|
sys.stderr.write("Working on link: %s\n" %link['href']) |
|
|
|
|
|
|
|
|
application = PlanningApplication() |
|
|
application = PlanningApplication() |
|
|
|
|
|
|
|
@@ -124,7 +124,7 @@ class WestminsterParser: |
|
|
|
|
|
|
|
|
# To get the comment url, we're going to have to go to each info url :-( |
|
|
# To get the comment url, we're going to have to go to each info url :-( |
|
|
|
|
|
|
|
|
# sys.stderr.write("Fetching: %s\n" %application.info_url) |
|
|
|
|
|
|
|
|
sys.stderr.write("Fetching: %s\n" %application.info_url) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
fakefile = StringIO.StringIO() |
|
|
fakefile = StringIO.StringIO() |
|
@@ -138,7 +138,7 @@ class WestminsterParser: |
|
|
|
|
|
|
|
|
curlobj.perform() |
|
|
curlobj.perform() |
|
|
|
|
|
|
|
|
# sys.stderr.write("Got it\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Got it\n") |
|
|
|
|
|
|
|
|
info_soup = BeautifulSoup(fakefile.getvalue()) |
|
|
info_soup = BeautifulSoup(fakefile.getvalue()) |
|
|
|
|
|
|
|
@@ -154,10 +154,10 @@ class WestminsterParser: |
|
|
|
|
|
|
|
|
self._results.addApplication(application) |
|
|
self._results.addApplication(application) |
|
|
|
|
|
|
|
|
# sys.stderr.write("Finished that link\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Finished that link\n") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
# sys.stderr.write("Finished while loop, returning stuff.\n") |
|
|
|
|
|
|
|
|
sys.stderr.write("Finished while loop, returning stuff.\n") |
|
|
|
|
|
|
|
|
return self._results |
|
|
return self._results |
|
|
|
|
|
|
|
|