How to Scrape Data From Local HTML Files using Python?

The data of the local HTML file can be extracted using Beautiful Soup and Python file handling techniques. Beautiful Soup allows us to parse HTML documents and navigate their structure, while file handling enables us to fetch the HTML content from local files. By combining these tools, we can learn how to extract valuable data from HTML files stored on our computers. In this article, we will understand how we can scrape Data from Local HTML files using Python.


Before understanding how to scrape data from local HTML files, make sure you have Python installed on your machine. Additionally, it's recommended to have basic knowledge of Python programming and HTML structure.

Installing Python Libraries

To extract data from HTML files, we'll be using the following Python libraries:

  • Beautiful Soup − A powerful library for parsing HTML and XML files.

  • Requests − A library for making HTTP requests.

You can install these libraries using pip, the package installer for Python, by running the following commands in your terminal or command prompt:

pip install beautifulsoup4
pip install requests

Understanding HTML Structure

HTML files are structured using tags and attributes that define elements within the document. To scrape data effectively, we need to understand the structure and locate the relevant data within the HTML file. Familiarize yourself with HTML tags such as <div>, <p>, <table>, and attributes like class and id, as they will be crucial for extracting data.

Loading HTML Files in Python

Before extracting data, we need to load the HTML file into our Python script. The requests library allows us to fetch the HTML content from a local file. To do this, we use the get() method and pass the file path as the argument.

import requests

file_path = 'path/to/your/file.html'
response = requests.get('file://' + file_path)
html_content = response.text

Extracting Data from HTML Files

To extract data from HTML files, we'll be utilizing the Beautiful Soup library. Beautiful Soup provides an easy-to-use interface for parsing HTML and navigating through its elements. It allows us to search for specific tags, retrieve attributes, and extract textual data.

The first step is to create a Beautiful Soup object from the HTML content we fetched earlier. We do this by passing the HTML content and the parser library (usually 'html.parser') as arguments.

from bs4 import BeautifulSoup

soup = BeautifulSoup(html_content, 'html.parser')

Scraping Data from a Local HTML File

For example, Suppose we have an HTML file named demo.html' with the following structure.

    <div class="container">
      <h1>Scraping Example</h1>
        <li>Item 1</li>
        <li>Item 2</li>
        <li>Item 3</li>

We need to extract the text inside the <h1> tag and the list items (<li> tags) from this HTML file. Here's how we can achieve this:


In the below example, we first open the HTML file specified by the file_path variable and read its contents. The BeautifulSoup object is then created with the HTML content, allowing us to parse and navigate the HTML structure. The code extracts the text within the <h1> tag and prints it as the heading. It also finds all <li> tags, iterates over them, and prints the text of each list item. This process enables the extraction of specific data from the HTML file for further processing or analysis.

from bs4 import BeautifulSoup

file_path = 'demo.html'

# Open the HTML file and read its content
with open(file_path, 'r') as file:
    html_content =

# Create a Beautiful Soup object
soup = BeautifulSoup(html_content, 'html.parser')

# Extract the heading text
heading = soup.find('h1').text
print("Heading:", heading)

# Extract the list items
list_items = soup.find_all('li')
print("List Items:")
for item in list_items:


Heading: Scraping Example
List Items:
Item 1
Item 2
Item 3

Handling More Complex HTML Structures

The example above demonstrates basic HTML scraping. However, real-world HTML files can be more complex, with nested elements, attributes, and varying structures. To handle such scenarios, you may need to traverse the HTML structure using different methods provided by Beautiful Soup, such as find_all(), find_next(), and select(). Experiment with these methods and refer to the Beautiful Soup documentation for more advanced scraping techniques.


In this article, we discussed how we can scrape data from local HTML files using Beautiful Soup and the requests library of Python. By combining Python's file-handling capabilities and Beautiful Soup's HTML parsing functionalities, we can extract valuable information from HTML files stored on our local machines.

Updated on: 13-Oct-2023


Kickstart Your Career

Get certified by completing the course

Get Started