Open In App

Web Scraping – Amazon Customer Reviews

Last Updated : 08 Sep, 2021
Like Article

In this article, we are going to see how we can scrape the amazon customer review using Beautiful Soup in Python.

Module needed

  • bs4: Beautiful Soup(bs4) is a Python library for pulling data out of HTML and XML files. This module does not come built-in with Python. To install this type the below command in the terminal.

pip install bs4

  • requests: Request allows you to send HTTP/1.1 requests extremely easily. This module also does not come built-in with Python. To install this type the below command in the terminal.

pip install requests

To begin with web scraping, we first have to do some setup. Import all the required modules. Get the cookies data for making the request to amazon, without this you can not able to scrape. Create a header that contains your request cookies, without cookies you can not scrape amazon data it always shows some error. This website will provide you, specific user agent.

Pass the URL in the getdata() function(User Defined Function) to that will request to a URL, it returns a response. We are using get method to retrieve information from the given server using a given URL.


requests.get(url, args)

Convert that data into HTML code and then Parse the HTML content using bs4.

Syntax: soup = BeautifulSoup(r.content, ‘html5lib’)


  • r.content : It is the raw HTML content.
  • html.parser : Specifying the HTML parser we want to use.

Now filter the required data using soup.Find_all function.



# import module
import requests
from bs4 import BeautifulSoup
HEADERS = ({'User-Agent':
            'Mozilla/5.0 (Windows NT 10.0; Win64; x64) \
            AppleWebKit/537.36 (KHTML, like Gecko) \
            Chrome/90.0.4430.212 Safari/537.36',
            'Accept-Language': 'en-US, en;q=0.5'})
# user define function
# Scrape the data
def getdata(url):
    r = requests.get(url, headers=HEADERS)
    return r.text
def html_code(url):
    # pass the url
    # into getdata function
    htmldata = getdata(url)
    soup = BeautifulSoup(htmldata, 'html.parser')
    # display html code
    return (soup)
url = "\
soup = html_code(url)


Note: This is only HTML code or Raw data.

Now since the core setup is done let us see how scraping for a specific requirement can be done.

Scrape customer name

Now find the customer list with span tag where class_ = a-profile-name. You can open the webpage in the browser and inspect the relevant element by pressing right-click as shown in the figure.

You have to pass the tag name and attribute with its corresponding value to the find_all() function.



def cus_data(soup):
    # find the Html tag
    # with find()
    # and convert into string
    data_str = ""
    cus_list = []
    for item in soup.find_all("span", class_="a-profile-name"):
        data_str = data_str + item.get_text()
        data_str = ""
    return cus_list
cus_res = cus_data(soup)


[‘Amaze’, ‘Robert’, ‘D. Kong’, ‘Alexey’, ‘Charl’, ‘RBostillo’]

Scrape user review:

Now find the customer review as same above methods. Find the unique class name with a specific tag, here we use div tag.



def cus_rev(soup):
    # find the Html tag
    # with find()
    # and convert into string
    data_str = ""
    for item in soup.find_all("div", class_="a-expander-content \
    reviewText review-text-content a-expander-partial-collapse-content"):
        data_str = data_str + item.get_text()
    result = data_str.split("\n")
    return (result)
rev_data = cus_rev(soup)
rev_result = []
for i in rev_data:
    if i is "":


Scraping Production information

Here we will scrape product information like product name, ASIN number, Weight, dimension. By doing this we will use the span tag and with a specific unique class name.



def product_info(soup):
    # find the Html tag
    # with find()
    # and convert into string
    data_str = ""
    pro_info = []
    for item in soup.find_all("ul", class_="a-unordered-list a-nostyle\
    a-vertical a-spacing-none detail-bullet-list"):
        data_str = data_str + item.get_text()
        data_str = ""
    return pro_info
pro_result = product_info(soup)
# Filter the required data
for item in pro_result:
    for j in item:
        if j is "":


Scraping Review Image:

Here we will extract the image link from the review of the product using the same as the above methods. The tag name and attribute of the tag is passed to findAll() as above.



def rev_img(soup):
    # find the Html tag
    # with find()
    # and convert into string
    data_str = ""
    cus_list = []
    images = []
    for img in soup.findAll('img', class_="cr-lightbox-image-thumbnail"):
    return images
img_result = rev_img(soup)


Saving details into CSV file:

Here we will save the details into the CSV file, We will convert the data into dataframe and then export it into the CSV, Let us see how to export a Pandas DataFrame to a CSV file. We will be using the to_csv() function to save a DataFrame as a CSV file.

Syntax : to_csv(parameters)
Parameters :

  • path_or_buf : File path or object, if None is provided the result is returned as a string.



import pandas as pd
# initialise data of lists.
data = {'Name': cus_res,
        'review': rev_result}
# Create DataFrame
df = pd.DataFrame(data)
# Save the output.


Like Article
Suggest improvement
Share your thoughts in the comments

Similar Reads