What is a Webcrawler and where is it used?

Web Crawler is a bot that downloads the content from the internet and indexes it. The main purpose of this bot is to learn about the different web pages on the internet. This kind of bots is mostly operated by search engines. By applying the search algorithms to the data collected by the web crawlers, search engines can provide the relevant links as the response for the request requested by the user. In this article, let’s discuss how the web crawler is implemented.

Webcrawler is a very important application of the Breadth-First Search Algorithm. The idea is that the whole internet can be represented by a directed graph:

  • with vertices -> Domains/ URLs/ Websites.
  • edges -> Connections.


Approach: The idea behind the working of this algorithm is to parse the raw HTML of the website and look for other URL in the obtained data. If there is a URL, then add it to the queue and visit them in breadth-first search manner.

Note: This code will not work on an online IDE due to proxy issues. Try to run on your local computer.





// Java program to illustrate the WebCrawler
import java.io.BufferedReader;
import java.io.InputStreamReader;
import java.net.URL;
import java.util.HashSet;
import java.util.LinkedList;
import java.util.List;
import java.util.Queue;
import java.util.regex.Matcher;
import java.util.regex.Pattern;
// Class Contains the functions
// required for WebCrowler
class WebCrowler {
    // To store the URLs in the
    / /FIFO order required for BFS
    private Queue<String> queue;
    // To store visited URls
    private HashSet<String>
    // Constructor for initialzing the
    // required variables
    public WebCrowler()
            = new LinkedList<>();
            = new HashSet<>();
    // Function to start the BFS and
    // discover all URLs
    public void discover(String root)
        // Storing the root URL to
        // initiate BFS.
        // It will loop until queue is empty
        while (!queue.isEmpty()) {
            // To store the URL present in
            // the front of the queue
            String v = queue.remove();
            // To store the raw HTML of
            // the website
            String raw = readUrl(v);
            // Regular expression for a URL
            String regex
                = "https://(\\w+\\.)*(\\w+)";
            // To store the pattern of the
            // URL formed by regex
            Pattern pattern
                = Pattern.compile(regex);
            // To extract all the URL that
            // matches the pattern in raw
            Matcher matcher
                = pattern.matcher(raw);
            // It will loop until all the URLs
            // in the current website get stored
            // in the queue
            while (matcher.find()) {
                // To store the next URL in raw
                String actual = matcher.group();
                // It will check whether this URL is
                // visited or not
                if (!discovered_websites
                         .contains(actual)) {
                    // If not visited it will add
                    // this URL in queue, print it
                    // and mark it as visited
                        "Website found: "
                        + actual);
    // Function to return the raw HTML
    // of the current website
    public String readUrl(String v)
        // Initializing empty string
        String raw = "";
        // Use try-catch block to handle
        // any exceptions given by this code
        try {
            // Convert the string in URL
            URL url = new URL(v);
            // Read the HTML from website
            BufferedReader be
                = new BufferedReader(
                    new InputStreamReader(
            // To store the input
            // from the website
            String input = "";
            // Read the HTML line by line
            // and append it to raw
            while ((input
                    = br.readLine())
                   != null) {
                raw += input;
            // Close BufferedReader
        catch (Exception ex) {
        return raw;
// Driver code
public class Main {
    // Driver Code
    public static void main(String[] args)
        // Creating Object of WebCrawler
        WebCrowler web_crowler
            = new WebCrowler();
        // Given URL
        String root
            = "https:// www.google.com";
        // Method call



Website found: https://www.google.com
Website found: https://www.facebook.com
Website found: https://www.amazon.com
Website found: https://www.microsoft.com
Website found: https://www.apple.com

Applications: This kind of web crawler is used to acquire the important parameters of the web like:

  1. What are the frequently visited websites?
  2. What are the websites that are important in the network as a whole?
  3. Useful Information on social networks: Facebook, Twitter… etc.
  4. Who is the most popular person in a group of people?
  5. Who is the most important software engineer in a company?

Attention reader! Don’t stop learning now. Get hold of all the important Java and Collections concepts with the Fundamentals of Java and Java Collections Course at a student-friendly price and become industry ready.

My Personal Notes arrow_drop_up

Check out this Author's contributed articles.

If you like GeeksforGeeks and would like to contribute, you can also write an article using contribute.geeksforgeeks.org or mail your article to contribute@geeksforgeeks.org. See your article appearing on the GeeksforGeeks main page and help other Geeks.

Please Improve this article if you find anything incorrect by clicking on the "Improve Article" button below.