Get site ranking with python and beautifulsoup

To get information about site, we can use but sometimes we just need some pieces of information like ranking of site global or local country.

In this case i just need that information, when i looking on alexa i found so many information about the target site.

I have some knowledge about python and beautifulsoup library and i try to grep that information with them.

First time i analyzing structure of page of, find element of html tag which contains the data.

Based on screenshot, the data on the div tag with class data, first tag is global ranks, second is local ranks and then reputation and rating. So the script must be grep a div tag with specification then do a looping.

In the source i need somes python library like sys to get arguments, urllib2 to opening url, BeautifulSoup to parsing the html, re to remove empty character which produced by BeautifulSoup output.

Ok, this is the source to get the information.

    #!/usr/bin/env python  
    # -*- coding: utf8 -*-  
    # Ardi nooneDOTnu1ATgmailDOTcom  
    # GNU GPL  
    import sys  
    import urllib2  
    from bs4 import BeautifulSoup  
    from re import sub  
    ATTR = [  
        'Global ranks ',  
        'Local ranks  ',  
        'Reputation   ',  
        'Rating       ',  
    def openLink(url):  
        _url = ''+ url  
        opener = urllib2.build_opener()  
        opener.addheaders = [('User-agent', 'Mozilla/5.0')]  
    if __name__ == '__main__':  
            url = sys.argv[1]  
            print 'Error url'  
        html = openLink(url)  
        soup = BeautifulSoup(html)  
        tds = soup.find('tr', {'class':'data-row1'}).findAll('td')  
        for td in tds:  
            print ATTR[i] +' => '+ sub(r'\s+', ' ', td.find('div', {'class':'data'}).text.strip())  
            i = i + 1  

This script can be downloaded on

To execute script just type $ python url or copying into /usr/local/bin and adding execute permission to script (chmod +x).