Editing Parsing WhitePages Search Results HTML

From PeformIQ Upgrade
Jump to navigation Jump to search

Searching WhitePages

The following Python script uses HTTPLib to search against whitepages.com.au (via a proxy).

 #!/usr/bin/env python
 #
 #
 #-------------------------------------------------------------------------------
 
 import re
 import sys
 import base64
 import pprint
 import urllib
 import httplib
 
 from copy import copy
 
 #-------------------------------------------------------------------------------
 
 PROXY       = 'PROXY:8080'
 SITE        = 'www.whitepages.com.au'
 
 connection  = None
 
 #===== Headers =================================================================
 
 #    'Accept' : 'text/plain, text/html',
 
 get_headers = {
    'Accept-Encoding'    : 'gzip, deflate',
    'Accept'             : '*/*',
    'Accept-Language'    : 'en-au',
    'Host'               : SITE,
    'Connection'         : 'Keep-Alive',
    'User-Agent'         : 'Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1; .NET CLR 1.1.4322; .NET CLR 2.0.50727'
 }
 
 post_headers = {
    'Content-type'       : 'application/x-www-form-urlencoded',
    'Accept'             : 'text/plain'
 }
 
 idx      = 0
 
 #===== Logging =================================================================
 
 def log_req_header(idx, hdr):
    of = open('log/%04d.req' % idx, 'w')
    of.write("%s\n" % pprint.pformat(hdr))
    of.close()
 
 #-------------------------------------------------------------------------------
 
 def log_resp_header(idx, resp):
    of = open('log/%04d.hdr' % idx, 'w')
 
    of.write("resp.__dict__ ->\n%s\n\n" % pprint.pformat(resp.__dict__))
    of.write("Status %s  Reason [%s]\n" % (resp.status, resp.reason))
    of.write("Msg ->\n%s\n\n" % resp.msg)
    of.write("Msg.__dict__ ->\n%s\n\n" % pprint.pformat(resp.msg.__dict__))
 
    of.close()
 
 #-------------------------------------------------------------------------------
 
 def log_resp_body(idx, resp_body):
    of = open('log/%04d.bdy' % idx, 'w')
    of.write(resp_body);
    of.close()
 
 #===== Encapsulate the request code ============================================
 
 def request(method, url, params, headers):
    global idx
 
    print '>>>> %s %s <<<<' % (method, url)
 
    connection.request(method, url, params, headers)
 
    resp = connection.getresponse()
 
    log_req_header(idx, headers)
    log_resp_header(idx, resp)
 
    resp_body = resp.read()
 
    log_resp_body(idx, resp_body)
 
    print resp_body
 
    idx += 1
 
    return resp
 
 #===============================================================================
 
 def do():
    global connection
    connection  = httplib.HTTPConnection(PROXY)
 
    BASE_URL    = 'http://%s' % SITE
 
    #------------------------------------------------------------------------
 
    DO       = 'GET'
    URL      = BASE_URL + '/'
 
    headers = copy(get_headers)
 
    request(DO, URL, None, headers)
 
 
    #------------------------------------------------------------------------
 
    DO  = 'GET'
    URL = BASE_URL + '/wp/index.jsp'
 
    headers = copy(get_headers)
 
    resp = request(DO, URL, None, headers)
 
    m = re.search('JSESSIONID=(.*);', resp.msg.__dict__['dict']['set-cookie'])
 
    if m:
       print m.group(1)
       JSESSIONID = m.group(1)
 
    print JSESSIONID
 
    #---------------------------------------------------------------------
 
    DO  = 'POST'
    URL = BASE_URL + '/wp/busSearch.do;jsessionid=%s' % JSESSIONID
 
    headers = copy(post_headers)
 
    form_data = {
       'subscriberName' : 'Hard',
       'state'          : 'VIC',
       'suburb'         : '',
       'street'         : '',
       'Search'         : 'Search'
    }
 
    params = urllib.urlencode(form_data)
 
    headers['Content-Length'] = len(params)
    headers['Cookie']         = 'JSESSIONID=%s' % JSESSIONID
 
    request(DO, URL, params, headers)
 
    #---------------------------------------------------------------------
 
    URL = BASE_URL + '/wp/busSearch.do'
 
    form_data = {
       'subscriberName' : 'Hard',
       'state'          : 'VIC',
       'page'           : '2'
    }
 
    params = urllib.urlencode(form_data)
 
    headers['Content-Length'] = len(params)
    headers['Cookie']         = 'JSESSIONID=%s' % JSESSIONID
 
    request(DO, URL, params, headers)
  
 #===============================================================================
 
 do()
 
 #-------------------------------------------------------------------------------

This script writes the search results into files (page 1 => 'log/0002.bdy' and page 2 => 'log/0003.bdy'). Amend the above code handle more pages of search results being produced.

Parsing the Search Results

The search results HTML looks as follows:

<div class="encap_result" id="result-10"><ul><li id='res10-ln0'><h4><span class='blackboldcaps'>Hard ...</ul></li></ul><div class="clearMe"> </div></div><div class="encap_result" id="result-11"><ul><li>

Results of Search

Are wriiten to a file in a block of XML.

A more Complex Script

This script loops through all the available pages in the search results and parses out the search results using the BeautifulSoup HTTP parsing module.

wp_get.py

#!/usr/bin/env python
#
#
#-------------------------------------------------------------------------------

import re
import sys
import base64
import pprint
import urllib
import httplib

from copy import copy


#-------------------------------------------------------------------------------

from wp_parser import parse


#-------------------------------------------------------------------------------

PROXY          = 'PROXY:8080'
SITE           = 'www.whitepages.com.au'

connection     = None
next_page_pat  = None

results        = ''


#-------------------------------------------------------------------------------

#    'Accept' : 'text/plain, text/html',

get_headers = {
   'Accept-Encoding'    : 'gzip, deflate',
   'Accept'             : '*/*',
   'Accept-Language'    : 'en-au',
   'Host'               : SITE,
   'Connection'         : 'Keep-Alive',
   'User-Agent'         : 'Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1; .NET CLR 1.1.4322; .NET CLR 2.0.50727'
}

post_headers = {
   'Content-type'     : 'application/x-www-form-urlencoded',
   'Accept'           : 'text/plain'
}

idx      = 0

#-------------------------------------------------------------------------------

def log_req_header(idx, hdr):
   of = open('log/%04d.req' % idx, 'w')
   of.write("%s\n" % pprint.pformat(hdr))
   of.close()

#-------------------------------------------------------------------------------

def log_resp_header(idx, resp):
   of = open('log/%04d.hdr' % idx, 'w')

   of.write("resp.__dict__ ->\n%s\n\n" % pprint.pformat(resp.__dict__))
   of.write("Status %s  Reason [%s]\n" % (resp.status, resp.reason))
   of.write("Msg ->\n%s\n\n" % resp.msg)
   of.write("Msg.__dict__ ->\n%s\n\n" % pprint.pformat(resp.msg.__dict__))

   of.close()

#-------------------------------------------------------------------------------

def log_resp_body(idx, resp_body):
   of = open('log/%04d.bdy' % idx, 'w')
   of.write(resp_body);
   of.close()

#-------------------------------------------------------------------------------

def do(pattern, state):
   global connection
   global results

   print "Pattern ==> '%s'" % pattern

   connection  = httplib.HTTPConnection(PROXY)

   BASE_URL    = 'http://%s' % SITE

   #------------------------------------------------------------------------

   DO       = 'GET'
   URL      = BASE_URL + '/'

   headers = copy(get_headers)

   request(DO, URL, None, headers)


   #------------------------------------------------------------------------

   DO  = 'GET'
   URL = BASE_URL + '/wp/index.jsp'

   headers = copy(get_headers)

   r = request(DO, URL, None, headers)

   m = re.search('JSESSIONID=(.*);', r['response_header'].msg.__dict__['dict']['set-cookie'])

   if m:
      print m.group(1)
      JSESSIONID = m.group(1)

   print JSESSIONID

   #---------------------------------------------------------------------

   DO  = 'POST'
   URL = BASE_URL + '/wp/busSearch.do;jsessionid=%s' % JSESSIONID

   headers = copy(post_headers)

   form_data = {
      'subscriberName' : pattern,
      'state'          : state,
      'suburb'         : '',
      'street'         : '',
      'Search'         : 'Search'
   }

   params = urllib.urlencode(form_data)

   headers['Content-Length'] = len(params)
   headers['Cookie']         = 'JSESSIONID=%s' % JSESSIONID

   r = request(DO, URL, params, headers)

   xml = parse('log/%04d.bdy' % r['idx'])

   results += xml

   body = r['response_body']

   m = next_page_pat.findall(body)

   cnt = 0

   if m:
      no_pages = len(m)
      if no_pages > 0:
         for i in range(no_pages):
            print m[i]
            #print m[i].group(1)

         cnt = no_pages - 1

   #---------------------------------------------------------------------

   for i in range(cnt):

      URL = BASE_URL + '/wp/busSearch.do'

      form_data = {
         'subscriberName' : pattern,
         'state'          : state,
         'page'           : i + 1
      }

      params = urllib.urlencode(form_data)

      headers['Content-Length'] = len(params)
      headers['Cookie']         = 'JSESSIONID=%s' % JSESSIONID

      r = request(DO, URL, params, headers)

      xml = parse('log/%04d.bdy' % r['idx'])

      results += xml

   ofh = open('results.xml', 'a+')

   ofh.write(results)

   ofh.close()


#-------------------------------------------------------------------------------

def request(method, url, params, headers):
   global idx

   print '>>>> %s %s <<<<' % (method, url)

   connection.request(method, url, params, headers)

   resp = connection.getresponse()

   log_req_header(idx, headers)
   log_resp_header(idx, resp)

   resp_body = resp.read()

   log_resp_body(idx, resp_body)

   # print resp_body

   r = {'idx' : idx, 'request_header' : headers, 'response_header' : resp, 'response_body' : resp_body}

   idx += 1

   return r

#-------------------------------------------------------------------------------

def process(search_patterns):
   global next_page_pat


   next_page_pat = re.compile(r';(page=[0-9]*">[0-9]*<\/a>)')

   for search_pattern in search_patterns:
     do(search_pattern, 'VIC')

#-------------------------------------------------------------------------------

def searches():

   searches = []

   for i in range(4):
      ch = chr(ord('W') + i)
      for x in ('a', 'e', 'i', 'o', 'u'):
         pat  = ch + x
         searches.append(pat)

   return searches

#-------------------------------------------------------------------------------

def used():
   searches = []
   searches.append('Zeus')

#-------------------------------------------------------------------------------

def main():
   searches = []

   searches.append('Zany')
   searches.append('Zan')
   searches.append('Zen')
   searches.append('Zend')

   process(searches)

#-------------------------------------------------------------------------------

main()

#-------------------------------------------------------------------------------

wp_parse.py

#!/usr/bin/env python

import pprint

from BeautifulSoup import BeautifulSoup

#-------------------------------------------------------------------------------------

def parse(fname):
   doc = open(fname, 'r')

   soup = BeautifulSoup(doc)

   # print len(soup('table', { "class" : "table_style"}))

   # tables = soup.findAll('table', { "class" : "table_style"})
   objs = soup.findAll('div', { "class" : "encap_result"})

   pp = pprint.PrettyPrinter(3)

   xml = ''

   for obj in objs:
      t = obj.find(text=True)

      if t:
         xml +=  '<entry>\n'

         #print '[[%s]]\n\n' % obj.__dict__
         # print '[[%s]]\n\n' % obj


         f    = obj.findAll('span',  { 'class' : 'black'})

         for s in f:
            xml +=  '  <tag>%s</tag>\n' % s.find(text=True)


         f    = obj.findAll('input',  { "name" : 'placeName'})

         for s in f:
            # pp.pprint(s.__dict__)
            # print 'attrMap -> "%s"' % s.attrMap
            xml += '  <placeName>%s</placeName>\n' % s.attrMap['value']


         for s in obj.findAll('input',  { "name" : 'subscriberName'}):
            xml += '  <subscriberName>%s</subscriberName>\n' % s.attrMap['value']

         for s in obj.findAll('input',  { "name" : 'address'}):
            xml += '  <address>%s</address>\n' % s.attrMap['value']

         for s in obj.findAll('input',  { "name" : 'streetNumber'}):
            xml += '  <streetNumber>%s</streetNumber>\n' % s.attrMap['value']

         for s in obj.findAll('input',  { "name" : 'streetName'}):
            xml += '  <streetName>%s</streetName>\n' % s.attrMap['value']

         for s in obj.findAll('input',  { "name" : 'streetType'}):
            xml += '  <streetType>%s</streetType>\n' % s.attrMap['value']

         for s in obj.findAll('input',  { "name" : 'locality'}):
            xml += '  <locality>%s</locality>\n' % s.attrMap['value']

http://www.magusco.com/kb/index.php/Parsing_WhitePages_Search_Results_HTML
         lis       = obj.findAll('li',  { "class" : None})

         for li in lis:
            xml += '  <li>%s</li>\n' % li.find(text=True)


         for address in obj.findAll('li',  { "class" : "entryData address"}):
            xml += '  <addr>%s</addr>\n' % address.find(text=True)


         for phone in obj.findAll('li',  { "class" : "entryData phoneNumber"}):
	    xml += '  <phoneNumber>%s</phoneNumber>\n' % phone.find(text=True)

         xml += '</entry>\n\n'

   return xml

#-------------------------------------------------------------------------------------

def test():
   # xml = parse('html/0002.html')

   # print xml

#-------------------------------------------------------------------------------------

Results of Search

Are written to a file as XML.