| @@ -0,0 +1,568 @@ | |||||
| import urllib2 | |||||
| import urllib | |||||
| import urlparse | |||||
| import cgi | |||||
| import re | |||||
| import datetime | |||||
| from BeautifulSoup import BeautifulSoup | |||||
| from PlanningUtils import PlanningApplication, \ | |||||
| PlanningAuthorityResults, \ | |||||
| getPostcodeFromText | |||||
| # Date format to enter into search boxes | |||||
| date_format = "%d/%m/%Y" | |||||
| # Regex for getting the application code | |||||
| # (needed for the comments url, when it exists) | |||||
| app_code_regex = re.compile("PARAM0=(\d*)") | |||||
| class PlanningExplorerParser: | |||||
| # If this authority doesn't have a comments page, | |||||
| # then set this email_address to an address for the | |||||
| # planning department, and it will be used in lieu of | |||||
| # a comments url. | |||||
| comments_email_address = None | |||||
| # These are the directories where the info urls, and search urls, | |||||
| # usually live underneath the base_url. | |||||
| # If these are different for a particular | |||||
| # authority, then they can be overridden in a subclass. | |||||
| info_url_path = "MVM/Online/Generic/" | |||||
| search_url_path = "MVM/Online/PL/GeneralSearch.aspx" | |||||
| # This is the most common place for comments urls to live | |||||
| # The %s will be filled in with an application code | |||||
| comments_path = "MVM/Online/PL/PLComments.aspx?pk=%s" | |||||
| # Most authorities don't need the referer header on the post | |||||
| # request. If one does, override this in the subclass | |||||
| use_referer = False | |||||
| # Some authorities won't give us anything back if we use the | |||||
| # python urllib2 useragent string. In that case, override this | |||||
| # in a subclass to pretend to be firefox. | |||||
| use_firefox_user_agent = False | |||||
| # This is the most common css class of the table containing the | |||||
| # the search results. If it is different for a particular authority | |||||
| # it can be overridden in a subclass | |||||
| results_table_attrs = {"class": "ResultsTable"} | |||||
| # These are the most common column positions for the | |||||
| # council reference, the address, and the description | |||||
| # in the results table. | |||||
| # They should be overridden in subclasses if they are different | |||||
| # for a particular authority. | |||||
| reference_td_no = 0 | |||||
| address_td_no = 1 | |||||
| description_td_no = 2 | |||||
| def _modify_response(self, response): | |||||
| """For most sites, we have managed to get all the apps on a | |||||
| single page by choosing the right parameters. | |||||
| If that hasn't been possible, override this method to get a | |||||
| new response object which has all the apps in one page. | |||||
| (See, for example, Hackney). | |||||
| """ | |||||
| return response | |||||
| def _find_trs(self, results_table): | |||||
| """Normally, we just want a list of all the trs except the first one | |||||
| (which is usually a header). | |||||
| If the authority requires a different list of trs, override this method. | |||||
| """ | |||||
| return results_table.findAll("tr")[1:] | |||||
| def _sanitisePostHtml(self, html): | |||||
| """This method can be overriden in subclasses if the | |||||
| html that comes back from the post request is bad, and | |||||
| needs tidying up before giving it to BeautifulSoup.""" | |||||
| return html | |||||
| def _sanitiseInfoUrl(self, url): | |||||
| """If an authority has info urls which are for some reason full | |||||
| of crap (like Broadland does), then this method should be overridden | |||||
| in order to tidy them up.""" | |||||
| return url | |||||
| def _getHeaders(self): | |||||
| """If the authority requires any headers for the post request, | |||||
| override this method returning a dictionary of header key to | |||||
| header value.""" | |||||
| headers = {} | |||||
| if self.use_firefox_user_agent: | |||||
| headers["User-Agent"] = "Mozilla/5.0 (X11; U; Linux i686; en-GB; rv:1.8.1.10) Gecko/20071126 Ubuntu/7.10 (gutsy) Firefox/2.0.0.10" | |||||
| if self.use_referer: | |||||
| headers["Referer"] = self.search_url | |||||
| return headers | |||||
| def _getPostData(self, asp_args, search_date): | |||||
| """Accepts asp_args (a tuple of key value pairs of the pesky ASP | |||||
| parameters, and search_date, a datetime.date object for the day | |||||
| we are searching for. | |||||
| This seems to be the most common set of post data which is needed | |||||
| for PlanningExplorer sites. It won't work for all of them, so | |||||
| will sometimes need to be overridden in a subclass. | |||||
| The parameter edrDateSelection is often not needed. | |||||
| It is needed by Charnwood though, so I've left it in | |||||
| to keep things simple. | |||||
| """ | |||||
| year_month_day = search_date.timetuple()[:3] | |||||
| post_data = urllib.urlencode(asp_args + ( | |||||
| ("_ctl0", "DATE_RECEIVED"), | |||||
| ("rbGroup", "_ctl5"), | |||||
| ("_ctl7_hidden", urllib.quote('<DateChooser Value="%d%%2C%d%%2C%d"><ExpandEffects></ExpandEffects></DateChooser>' %year_month_day)), | |||||
| ("_ctl8_hidden", urllib.quote('<DateChooser Value="%d%%2C%d%%2C%d"><ExpandEffects></ExpandEffects></DateChooser>' %year_month_day)), | |||||
| ("edrDateSelection", "1"), | |||||
| ("csbtnSearch", "Search"), | |||||
| ("cboNumRecs", "99999"), | |||||
| )) | |||||
| return post_data | |||||
| def _getPostCode(self): | |||||
| """In most cases, the postcode can be got from the address in | |||||
| the results table. Some councils put the address there without the | |||||
| postcode. In this case we will have to go to the info page to get | |||||
| the postcode. This should be done by overriding this method with | |||||
| one that parses the info page.""" | |||||
| return getPostcodeFromText(self._current_application.address) | |||||
| def __init__(self, | |||||
| authority_name, | |||||
| authority_short_name, | |||||
| base_url, | |||||
| debug=False): | |||||
| self.authority_name = authority_name | |||||
| self.authority_short_name = authority_short_name | |||||
| self.base_url = base_url | |||||
| self.search_url = urlparse.urljoin(base_url, self.search_url_path) | |||||
| self.info_url_base = urlparse.urljoin(self.base_url, self.info_url_path) | |||||
| self.debug = debug | |||||
| self._results = PlanningAuthorityResults(self.authority_name, self.authority_short_name) | |||||
| def getResultsByDayMonthYear(self, day, month, year): | |||||
| search_date = datetime.date(year, month, day) | |||||
| # First do a get, to get some state | |||||
| get_request = urllib2.Request(self.search_url) | |||||
| get_response = urllib2.urlopen(get_request) | |||||
| html = get_response.read() | |||||
| # We need to find those ASP parameters such as __VIEWSTATE | |||||
| # so we can use them in the next POST | |||||
| asp_args_regex = re.compile('<input[^>]*name=\"(__[A-Z]*)\"[^>]*value=\"([^\"]*)\"[^>]*>') | |||||
| # re.findall gets us a list of key value pairs. | |||||
| # We want to concatenate it with a tuple, so we must | |||||
| # make it a tuple | |||||
| asp_args = tuple(re.findall(asp_args_regex, html)) | |||||
| # The post data needs to be different for different councils | |||||
| # so we have a method on each council's scraper to make it. | |||||
| post_data = self._getPostData(asp_args, search_date) | |||||
| headers = self._getHeaders() | |||||
| request = urllib2.Request(self.search_url, post_data, headers) | |||||
| post_response = urllib2.urlopen(request) | |||||
| # We have actually been returned here by an http302 object | |||||
| # moved, and the response called post_response is really a get. | |||||
| # In some cases, we can't get the page size set high | |||||
| # until now. In that case, override _modify_response | |||||
| # so that we get back a response with all the apps on one page. | |||||
| # We pass in headers so that any | |||||
| post_response = self._modify_response(post_response) | |||||
| html = self._sanitisePostHtml(post_response.read()) | |||||
| soup = BeautifulSoup(html) | |||||
| results_table = soup.find("table", attrs=self.results_table_attrs) | |||||
| # If there is no results table, then there were no apps on that day. | |||||
| if results_table: | |||||
| trs = self._find_trs(results_table) | |||||
| self._current_application = None | |||||
| # The first tr is just titles, cycle through the trs after that | |||||
| for tr in trs: | |||||
| self._current_application = PlanningApplication() | |||||
| # There is no need to search for the date_received, it's what | |||||
| # we searched for | |||||
| self._current_application.date_received = search_date | |||||
| tds = tr.findAll("td") | |||||
| for td_no in range(len(tds)): | |||||
| if td_no == self.reference_td_no: | |||||
| # This td contains the reference number and a link to details | |||||
| self._current_application.council_reference = tds[td_no].a.string | |||||
| relative_info_url = self._sanitiseInfoUrl(tds[td_no].a['href']) | |||||
| self._current_application.info_url = urlparse.urljoin(self.info_url_base, relative_info_url) | |||||
| # What about a comment url? | |||||
| # There doesn't seem to be one, so we'll use the email address | |||||
| if self.comments_email_address is not None: | |||||
| # We're using the email address, as there doesn't seem | |||||
| # to be a web form for comments | |||||
| self._current_application.comment_url = self.comments_email_address | |||||
| else: | |||||
| # This link contains a code which we need for the comments url | |||||
| # (on those sites that use it) | |||||
| application_code = app_code_regex.search(relative_info_url).groups()[0] | |||||
| relative_comments_url = self.comments_path %(application_code) | |||||
| self._current_application.comment_url = urlparse.urljoin(self.base_url, relative_comments_url) | |||||
| elif td_no == self.address_td_no: | |||||
| # If this td contains a div, then the address is the | |||||
| # string in there - otherwise, use the string in the td. | |||||
| if tds[td_no].div is not None: | |||||
| address = tds[td_no].div.string | |||||
| else: | |||||
| address = tds[td_no].string | |||||
| self._current_application.address = address | |||||
| self._current_application.postcode = self._getPostCode() | |||||
| elif td_no == self.description_td_no: | |||||
| if tds[td_no].div is not None: | |||||
| # Mostly this is in a div | |||||
| # Use the empty string if the description is missing | |||||
| description = tds[td_no].div.string or "" | |||||
| else: | |||||
| # But sometimes (eg Crewe) it is directly in the td. | |||||
| # Use the empty string if the description is missing | |||||
| description = tds[td_no].string or "" | |||||
| self._current_application.description = description | |||||
| self._results.addApplication(self._current_application) | |||||
| return self._results | |||||
| def getResults(self, day, month, year): | |||||
| return self.getResultsByDayMonthYear(int(day), int(month), int(year)).displayXML() | |||||
| class BroadlandLike: | |||||
| # FIXME - BroadlandLike authorities don't have postcodes on their site, but | |||||
| # they do have grid references. We should use these. | |||||
| results_table_attrs = {"class": "display_table"} | |||||
| info_url_path = "Northgate/PlanningExplorer/Generic/" | |||||
| search_url_path = "Northgate/PlanningExplorer/GeneralSearch.aspx" | |||||
| use_firefox_user_agent = True | |||||
| use_referer = True | |||||
| def _getPostData(self, asp_args, search_date): | |||||
| post_data = urllib.urlencode(asp_args + ( | |||||
| ("cboSelectDateValue", "DATE_RECEIVED"), | |||||
| ("rbGroup", "rbRange"), | |||||
| ("dateStart", search_date.strftime(date_format)), | |||||
| ("dateEnd", search_date.strftime(date_format)), | |||||
| ("cboNumRecs", "99999"), | |||||
| ("csbtnSearch", "Search"), | |||||
| )) | |||||
| return post_data | |||||
| def _sanitiseInfoUrl(self, url): | |||||
| """The broadland info urls arrive full of rubbish. This method tidies | |||||
| them up.""" | |||||
| # We need to | |||||
| # 1) Remove whitespace | |||||
| # 2) Remove 
 and 
 | |||||
| ws_re = re.compile("(?:(?:\s)|(?:&#x\w;))*") | |||||
| return ''.join(ws_re.split(url)) | |||||
| class BlackburnParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| class BroadlandParser(BroadlandLike, PlanningExplorerParser): | |||||
| # FIXME - is http://secure.broadland.gov.uk/mvm/Online/PL/GeneralSearch.aspx | |||||
| # a better url for Broadland? | |||||
| def _sanitisePostHtml(self, html): | |||||
| """The page that comes back from the post for the broadland site | |||||
| has a broken doctype declaration. We need to tidy that up before | |||||
| giving it to BeautifulSoup.""" | |||||
| # This is what it looks like - note the missing close doublequote | |||||
| #<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd> | |||||
| # Split on the broken doctype and join with the doctype with | |||||
| # closing quote. | |||||
| html = '<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">'.join(html.split('<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd>')) | |||||
| return html | |||||
| class CamdenParser(BroadlandLike, PlanningExplorerParser): | |||||
| comments_path = "Northgate/PlanningExplorer/PLComments.aspx?pk=%s" | |||||
| class CharnwoodParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| class CreweParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| address_td_no = 4 | |||||
| def _getPostData(self, asp_args, search_date): | |||||
| year_month_day = search_date.timetuple()[:3] | |||||
| post_data = urllib.urlencode(asp_args + ( | |||||
| ("drDateReceived:_ctl0_hidden", urllib.quote('<DateChooser Value="%d%%2C%d%%2C%d"><ExpandEffects></ExpandEffects></DateChooser>' %year_month_day)), | |||||
| ("drDateReceivedxxctl0_input", search_date.strftime(date_format)), | |||||
| ("drDateReceived:_ctl1_hidden", urllib.quote('<DateChooser Value="%d%%2C%d%%2C%d"><ExpandEffects></ExpandEffects></DateChooser>' %year_month_day)), | |||||
| ("drDateReceivedxxctl1_input", search_date.strftime(date_format)), | |||||
| ("cboNumRecs", "99999"), | |||||
| ("csbtnSearch", "Search"), | |||||
| )) | |||||
| return post_data | |||||
| class EastStaffsParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| address_td_no = 4 | |||||
| description_td_no = 1 | |||||
| class EppingForestParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| address_td_no = 3 | |||||
| description_td_no = 1 | |||||
| class ForestHeathParser(BroadlandLike, PlanningExplorerParser): | |||||
| pass | |||||
| class HackneyParser(PlanningExplorerParser): | |||||
| # FIXME - This will only get the first ten records on this | |||||
| # day. Need to deal with paging. | |||||
| use_firefox_user_agent = True | |||||
| address_td_no = 6 | |||||
| description_td_no = 5 | |||||
| def _modify_response(self, response): | |||||
| # In order to make sure we don't have to worry about any paging, | |||||
| # We'll fetch this url again with PS=99999. | |||||
| real_url_tuple = urlparse.urlsplit(response.geturl()) | |||||
| query_string = real_url_tuple[3] | |||||
| # Get the query as a list of key, value pairs | |||||
| parsed_query_list = list(cgi.parse_qsl(query_string)) | |||||
| # Go through the query string replacing any PS parameters | |||||
| # with PS=99999 | |||||
| for i in range(len(parsed_query_list)): | |||||
| key, value = parsed_query_list[i] | |||||
| if key == "PS": | |||||
| value = "99999" | |||||
| parsed_query_list[i] = (key, value) | |||||
| new_query_string = urllib.urlencode(parsed_query_list) | |||||
| new_url_tuple = real_url_tuple[:3] + (new_query_string,) + real_url_tuple[4:] | |||||
| new_url = urlparse.urlunsplit(new_url_tuple) | |||||
| new_request = urllib2.Request(new_url, None, self._getHeaders()) | |||||
| new_response = urllib2.urlopen(new_request) | |||||
| return new_response | |||||
| def _getPostData(self, asp_args, search_date): | |||||
| post_data = urllib.urlencode(asp_args + ( | |||||
| ("ctl00", "DATE_RECEIVED"), | |||||
| ("rbGroup", "ctl05"), | |||||
| ("ctl07_input", search_date.strftime(date_format)), | |||||
| ("ctl08_input", search_date.strftime(date_format)), | |||||
| ("edrDateSelection", "1"), | |||||
| ("csbtnSearch", "Search"), | |||||
| )) | |||||
| return post_data | |||||
| class KennetParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| address_td_no = 3 | |||||
| class LincolnParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| class LiverpoolParser(PlanningExplorerParser): | |||||
| comments_email_address = "planningandbuildingcontrol@liverpool.gov.uk" | |||||
| use_firefox_user_agent = True | |||||
| use_referer = True | |||||
| results_table_attrs = {"xmlns:mvm":"http://www.mvm.co.uk"} | |||||
| info_url_path = "mvm/" | |||||
| search_url_path = "mvm/planningsearch.aspx" | |||||
| def _find_trs(self, results_table): | |||||
| """In this case we are after all trs except the first two which have a | |||||
| class attribute row0 or row1.""" | |||||
| return results_table.findAll("tr", {"class":["row0", "row1"]})[3:] | |||||
| def _getPostData(self, asp_args, search_date): | |||||
| post_data = urllib.urlencode(asp_args + ( | |||||
| ("dummy", "dummy field\tused for custom\tvalidator"), | |||||
| ("drReceived$txtStart", search_date.strftime(date_format)), | |||||
| ("drReceived$txtEnd", search_date.strftime(date_format)), | |||||
| ("cboNumRecs", "99999"), | |||||
| ("cmdSearch", "Search"), | |||||
| )) | |||||
| return post_data | |||||
| def _sanitiseInfoUrl(self, url): | |||||
| """The liverpool info urls arrive full of rubbish. This method tidies | |||||
| them up.""" | |||||
| # We need to | |||||
| # 1) Remove whitespace | |||||
| # 2) Remove 
 and 
 | |||||
| ws_re = re.compile("(?:(?:\s)|(?:&#x\w;))*") | |||||
| return ''.join(ws_re.split(url)) | |||||
| # FIXME - Merton, Shrewsbury, and South Norfolk need to be done here. | |||||
| # All are down today... | |||||
| class SouthNorfolkParser(PlanningExplorerParser): | |||||
| use_firefox_user_agent = True | |||||
| class SouthShropshireParser(PlanningExplorerParser): | |||||
| comments_email_address = "planning@southshropshire.gov.uk" | |||||
| use_firefox_user_agent = True | |||||
| info_url_path = "MVM/Online/PL/" | |||||
| def _getPostData(self, asp_args, search_date): | |||||
| local_date_format = "%d-%m-%Y" | |||||
| year, month, day = search_date.timetuple()[:3] | |||||
| post_data = urllib.urlencode(asp_args + ( | |||||
| ("edrDateSelection:htxtRange", "radRangeBetween"), | |||||
| ("cboDateList", "DATE_RECEIVED"), | |||||
| ("edrDateSelection:txtStart", search_date.strftime(local_date_format)), | |||||
| ("edrDateSelection:txtEnd", search_date.strftime(local_date_format)), | |||||
| ("edrDateSelection:txtDateReceived", "%(day)d-%(month)d-%(year)d~%(day)d-%(month)d-%(year)d" %({"day":day, "month":month, "year":year})), | |||||
| ("cboNumRecs", "99999"), | |||||
| ("csbtnSearch", "Search"), | |||||
| )) | |||||
| return post_data | |||||
| class SouthTynesideParser(BroadlandLike, PlanningExplorerParser): | |||||
| # Unlike the other BroadlandLike sites, there are postcodes :-) | |||||
| pass | |||||
| class StockportParser(PlanningExplorerParser): | |||||
| comments_email_address = "admin.dc@stockport.gov.uk" | |||||
| info_url_path = "MVM/Online/PL/" | |||||
| def _getPostData(self, asp_args, search_date): | |||||
| post_data = urllib.urlencode(asp_args + ( | |||||
| ("drDateReceived:txtStart", search_date.strftime(date_format)), | |||||
| ("drDateReceived:txtEnd", search_date.strftime(date_format)), | |||||
| ("cboNumRecs", "99999"), | |||||
| ("csbtnSearch", "Search"),), | |||||
| ) | |||||
| return post_data | |||||
| # FIXME - should add Swansea, but it is currently down | |||||
| class TamworthParser(PlanningExplorerParser): | |||||
| comments_email_address = "planningadmin@tamworth.gov.uk" | |||||
| use_firefox_user_agent = True | |||||
| info_url_path = "MVM/Online/PL/" | |||||
| class TraffordParser(PlanningExplorerParser): | |||||
| # There are no postcodes on the Trafford site. | |||||
| use_firefox_user_agent = True | |||||
| address_td_no = 3 | |||||
| class WestOxfordshireParser(PlanningExplorerParser): | |||||
| address_td_no = 3 | |||||
| description_td_no = 1 | |||||
| use_firefox_user_agent = True | |||||
| if __name__ == '__main__': | |||||
| # NOTE - 04/11/2007 is a sunday | |||||
| # I'm using it to test that the scrapers behave on days with no apps. | |||||
| #parser = BlackburnParser("Blackburn With Darwen Borough Council", "Blackburn", "http://195.8.175.6/") | |||||
| #parser = BroadlandParser("Broadland Council", "Broadland", "http://www.broadland.gov.uk/") | |||||
| #parser = CamdenParser("London Borough of Camden", "Camden", "http://planningrecords.camden.gov.uk/") | |||||
| #parser = CharnwoodParser("Charnwood Borough Council", "Charnwood", "http://portal.charnwoodbc.gov.uk/") | |||||
| #parser = CreweParser("Crewe and Nantwich Borough Council", "Crewe and Nantwich", "http://portal.crewe-nantwich.gov.uk/") | |||||
| #parser = EastStaffsParser("East Staffordshire Borough Council", "East Staffs", "http://www2.eaststaffsbc.gov.uk/") | |||||
| #parser = EppingForestParser("Epping Forest District Council", "Epping Forest", "http://plan1.eppingforestdc.gov.uk/") | |||||
| #parser = ForestHeathParser("Forest Heath District Council", "Forest Heath", "http://195.171.177.73/") | |||||
| parser = HackneyParser("London Borough of Hackney", "Hackney", "http://www.hackney.gov.uk/servapps/") | |||||
| #parser = KennetParser("Kennet District Council", "Kennet", "http://mvm-planning.kennet.gov.uk/") | |||||
| #parser = LincolnParser("Lincoln City Council", "Lincoln", "http://online.lincoln.gov.uk/") | |||||
| #parser = LiverpoolParser("Liverpool City Council", "Liverpool", "http://www.liverpool.gov.uk/") | |||||
| #parser = SouthNorfolkParser("South Norfolk Council", "South Norfolk", "http://planning.south-norfolk.gov.uk/") | |||||
| #parser = SouthShropshireParser("South Shropshire District Council", "South Shropshire", "http://194.201.44.102/") | |||||
| #parser = SouthTynesideParser("South Tyneside Council", "South Tyneside", "http://poppy.southtyneside.gov.uk/") | |||||
| #parser = StockportParser("Stockport Metropolitan District Council", "Stockport", "http://s1.stockport.gov.uk/council/eed/dc/planning/") | |||||
| #parser = TamworthParser("Tamworth Borough Council", "Tamworth", "http://80.1.64.77/") | |||||
| #parser = TraffordParser("Trafford Council", "Trafford", "http://planning.trafford.gov.uk/") | |||||
| #parser = WestOxfordshireParser("West Oxfordshire District Council", "West Oxfordshire", "http://planning.westoxon.gov.uk/") | |||||
| print parser.getResults(1, 11, 2007) | |||||
| # To Do | |||||
| # Sort out paging: | |||||
| # South Shropshire - pages on 6 | |||||
| # Investigate catching unavailable message: | |||||
| # Charnwood | |||||