Giter Site home page Giter Site logo

lexicon-crawling's Introduction

Web Crawling/Scraping (Python 3.5.1)

Here is an example of Python crawling. The task is to get all of the terms and their explanations from lexicon.ft.com. After taking a look at its website I found that the term pages are indexed alphabetically. Now let's get started.

##Initial plan: After checking the url of the term pages, the followings are found:

So the implementation is ####1. fetch each index page (A-Z) using requests, and find and store the urls for each term via bueatiful soup After investigating the source code of the index page, we can notice that all of the term urls are stored in <a href="Term?term=***"> , so just loop all hrefs and find the ones with "Term?term=":

    for link in soup.find_all('a'):
        termlink = link.get('href')
        # find term link from them
        if 'Term?term=' in termlink:
            if 'http://lexicon.ft.com/' not in termlink:
                    termlink = 'http://lexicon.ft.com/'+termlink
            termlinklist.append(termlink)    

####2. fetch each term page using request, and find and store the term and the explanation via bueatiful soup In its souce code, all terms are stored in <div class="moreItemsHeader"> and explanation is stored in <div class="definition">, so just get element.text via bueatiful soup (do not use .string).

    # open term link
    termHTML = requests.get(termlink).text
    termSoup = BeautifulSoup(termHTML, 'html.parser')
    # get term title div        
    title = termSoup.find_all('div',{'class':'moreItemsHeader'})
    # get title div string
    titleString = title[0].text
    # get title term, delete "Difinition of " totally 14 words
    term = titleString[14:]
    # get term explanation
    explanation = termSoup.find_all('div',{'class':'definition'})[0].text  
    termlist.append([term,explanation])

####3. The result list is saved to csv file.

# save to csv file
with open('output.csv', "w") as f:
    writer = csv.writer(f)
    writer.writerows(termlist)    
    print('Results are saved to file output.csv.');
    f.close()

Special characters like \xa0 \xa3 will cause problem in saving a text file. Thus this line is added to remove them:

explanation = re.sub('[^a-zA-Z0-9-_*.,;:+\/\&\@\"\'\(\)\$\%\[\]]', ' ', explanation)

This formed the initial version. However the biggest problem is that only 10 terms are shown on each index page. It turns out to be a http request. Seems like this is not as easy as I thought to be in the beginning.

##Upgraded plan: The solution is to mimick the action taken by clicking the "next" button. Open the developing tool in chrome and click any of the page link, we can see that in Network page, there is a new POST - "UpdateResultTerms". Click it, and in its header we can find the request url and the header data.

searchText:A
searchBy:Letter
ResetPaging:false
startRow:11
ODA_Parent:{"category":"Lexicon","name":"Search","label":"Search"}

Notice that there is a searhcText letter A, and startRow 11. Actually the startRow is the most important one: if it exceeds the max number of letter A it will automatically fetch the rows for letter B, and so on until letter Z. the task becomes to post the request to the url with different startRows until blank response is received.

First of all, we need to mimick the same data format to post request to the server to get the response.

  postdata={'searchTex':'A',
              'searchBy':'Letter',
              'ResetPaging':'false',
              'startRow':startRow,
              'ODA_Parent':'%7B%22category%22%3A%22Lexicon%22%2C%22name%22%3A%22Search%22%2C%22label%22%3A%22Search%22%7D'
            }
  response = requests.post(
        url='http://lexicon.ft.com/UpdateResultTerms',
        data=postdata,
        headers={ 'X-Requested-With': 'XMLHttpRequest'}
    )

The response is a json string:

{"html":"\u003ctbody data-ajax-content=\"1\"\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=absolute-return\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabsolute return\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eThe term absolute in a mathematical sense refers to the numerical value of a number without its sign. For example, the absolute value of -7...\u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=absolute-return-fund\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabsolute return fund\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eThis fund aims to deliver positive returns in all market conditions, with low volatility.Absolute return funds sprang up following the new...\u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=absolute-title\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabsolute title\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eA right of ownership of land that cannot be doubted and where there is no risk of anyone else claiming to be the owner.  \u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=absorb\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabsorb\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eIf a large organization absorbs a smaller one, it takes control of it and makes it part of the organization.If a business absorbs costs, it...\u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=absorption\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabsorption\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eWhen a larger organization takes control of a smaller one.  \u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=absorption-costing\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabsorption costing\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eA way of calculating the cost of a product, including the cost of producing it and also the general costs of running the business or...\u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=absorption-rate\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabsorption rate\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eThe speed at which new houses, offices etc are sold or rented.  \u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=abstract-of-title\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabstract of title\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eA legal document that proves someone\u0027s right to own a particular property, listing all the documents relating to the ownership of the...\u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=abundance-gaps\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabundance gaps\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eA term used in positive organisational scholarship.\r\nConsider a continuum that depicts a state of normal performance in the middle, a...\u003c/td\u003e\u003c/tr\u003e\u003ctr\u003e\u003cth\u003e\u003ca href=\"Term?term=abusive-tax-shelter\" data-oda=\"{&quot;category&quot;:&quot;Link click through&quot;,&quot;name&quot;:&quot;Lexicon Term&quot;,&quot;label&quot;:&quot;Page: Lexicon Search ; Module: Search Results Table&quot;}\"\u003eabusive tax shelter\u003c/a\u003e\u003c/th\u003e\u003ctd\u003eAn attempt to avoid tax by fraudulent means, for instance by manipulating the value of acquired assets.\u003c/td\u003e\u003c/tr\u003e\u003c/tbody\u003e","json":{"paging":{"rowsPerPage":10,"totalRows":699}}}

and it has two parameters:

  • html: html content to update the search result. This is what we wanted.
  • json: it provides the max number of rows per page and the max number of pages.

Actually the html parameter has provided us all info we need to obtain the term links, thus it is not necessary to request the index page anymore.

Also we wrap the codes into functions to make it more organized.

##Final plan: It seems the upgraded plan works well. When running through all terms, however, it is found that this version may take couple of hours. Is there any way to enhance the speed? The answer is to parallelly post the http requests via multi-threading. This may cause the order issue, but this is much easier to handle.

import threading
def parallelProcess(targetfun,source):
    threads = [threading.Thread(target=targetfun, args=(i,)) for i in source]
    for thread in threads:
        thread.start()
    for thread in threads:
        thread.join()

Then wrap the two functions and set the range for them:

# get term link list
parallelProcess(requestSearchResults,range(0,maxNumPage*termsPerPage)[0::termsPerPage])
    
# get terms and explanations
parallelProcess(parseTermPage,termlinklist)

Although my laptop only has 4 cores, the process has been speed up for much more than 4 times: the total fetching time has been lowered down to less than 15min. The reason is that there are always some links which is much slower to fetch. If there is only one thread, every other following fetching task will be blocked by those links; however with 4 threads, there will be fast lanes for those fast links to go through.

We can then sort the term list by the terms:

# sort
termlist = sorted(termlist,key=lambda x: x[0])

Then store the terms and explanations in csv file.

lexicon-crawling's People

Contributors

quartzing avatar

Watchers

James Cloos avatar Don Teo avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.