Skip to content
Related Articles

Related Articles

Improve Article
Save Article
Like Article

Python Urllib Module

  • Difficulty Level : Medium
  • Last Updated : 13 Oct, 2021

Urllib package is the URL handling module for python. It is used to fetch URLs (Uniform Resource Locators). It uses the urlopen function and is able to fetch URLs using a variety of different protocols.

Urllib is a package that collects several modules for working with URLs, such as:

 Attention geek! Strengthen your foundations with the Python Programming Foundation Course and learn the basics.  

To begin with, your interview preparations Enhance your Data Structures concepts with the Python DS Course. And to begin with your Machine Learning Journey, join the Machine Learning - Basic Level Course

  • urllib.request for opening and reading.
  • urllib.parse for parsing URLs
  • urllib.error for the exceptions raised
  • urllib.robotparser for parsing robot.txt files

If urllib is not present in your environment, execute the below code to install it.

pip install urllib

Let’s see these in details.


This module helps to define functions and classes to open URLs (mostly HTTP). One of the most simple ways to open such URLs is :
We can see this in an example:

import urllib.request
request_url = urllib.request.urlopen('')
The source code of the URL i.e. Geeksforgeeks.


This module helps to define functions to manipulate URLs and their components parts, to build or break them. It usually focuses on splitting a URL into small components; or joining different URL components into URL strings.
We can see this from the below code:

from urllib.parse import * parse_url = urlparse(' / python-langtons-ant/')
unparse_url = urlunparse(parse_url)
ParseResult(scheme='https', netloc='', path='/python-langtons-ant/', params='', query='', fragment='')

Note:- The different components of a URL are separated and joined again. Try using some other URL for better understanding.

Different other functions of urllib.parse are :

urllib.parse.urlparseSeparates different components of URL
urllib.parse.urlunparseJoin different components of URL
urllib.parse.urlsplitIt is similar to urlparse() but doesn’t split the params
urllib.parse.urlunsplitCombines the tuple element returned by urlsplit() to form URL
urllib.parse.urldeflagIf URL contains fragment, then it returns a URL removing the fragment.

This module defines the classes for exception raised by urllib.request. Whenever there is an error in fetching a URL, this module helps in raising exceptions. The following are the exceptions raised :

  • URLError – It is raised for the errors in URLs, or errors while fetching the URL due to connectivity, and has a ‘reason’ property that tells a user the reason of error.
  • HTTPError – It is raised for the exotic HTTP errors, such as the authentication request errors. It is a subclass or URLError. Typical errors include ‘404’ (page not found), ‘403’ (request forbidden),
    and ‘401’ (authentication required).

We can see this in following examples :

# URL Error
import urllib.request
import urllib.parse
# trying to read the URL but with no internet connectivity
    x = urllib.request.urlopen('')
# Catching the exception generated     
except Exception as e :
URL Error: urlopen error [Errno 11001] getaddrinfo failed

# HTTP Error
import urllib.request
import urllib.parse
# trying to read the URL
    x = urllib.request.urlopen(' / search?q = test')
# Catching the exception generated    
except Exception as e :
HTTP Error 403: Forbidden

This module contains a single class, RobotFileParser. This class answers question about whether or not a particular user can fetch a URL that published robot.txt files. Robots.txt is a text file webmasters create to instruct web robots how to crawl pages on their website. The robot.txt file tells the web scraper about what parts of the server should not be accessed.
For example :

# importing robot parser class
import urllib.robotparser as rb
bot = rb.RobotFileParser()
# checks where the website's robot.txt file reside
x = bot.set_url(' / robot.txt')
# reads the files
y =
# we can crawl the main site
z = bot.can_fetch('*', '')
# but can not crawl the disallowed url
w = bot.can_fetch('*', ' / wp-admin/')

My Personal Notes arrow_drop_up
Recommended Articles
Page :

Start Your Coding Journey Now!