Start Your Project with Us

Whatever your project size is, we will handle it well with all the standards fulfilled! We are here to give 100% satisfaction.

  • Any feature, you ask, we develop
  • 24x7 support worldwide
  • Real-time performance dashboard
  • Complete transparency
  • Dedicated account manager
  • Customized solutions to fulfill data scraping goals

For job seekers, please visit our Career Page or send your resume to


Have you ever wished to know about discounted prices beforehand? This blog talks about creating a tool using web scraping techniques on a Raspberry Pi device to identify the best deals. You can easily make this device at home in just 10 minutes.


For the purpose of our use cases, a laptop or a Raspberry Pi can be used, but we will be using Raspberry Pi as a web scraping server that runs continuously. There are numerous Raspberry Pi projects available online, but most of them require some electrical engineering.

Python 3 is the language of choice for our application. It has a wide range of powerful libraries, and it is easy to get started and create a prototype. Since Python 2 will no longer be supported from January 1st, 2020, we will use Python 3.

Scrapy is among the finest open-source web extraction frameworks available in Python. It is a powerful and incredibly fast tool that is at the core of our set of tools. While new versions have been developed, the core components have remained largely unchanged. We will be using the latest version of Scrapy 2.0.1 on Python 3.6.10 in this article.

To inspect objects and extract HTML tags with ease, a modern browser with developer tools enabled is recommended.


To succeed in web scraping, it's important to choose a site with a high amount of traffic. Some websites that offer discounts and promo codes include SlickDeals, Dealnews, and DealMoon. For the purposes of this blog, we will be using SlickDeals as our chosen website to scrape data. While there will be different components on the HTML to extract, there are no restrictions on choosing a website that aligns with your interests.

Use a Developer Tool for Scraping Context

1. Go to SlickDeals website

2. To find the best bargains, check out the Frontpage Slickdeals section. Here, each item is accompanied by a product image, title, store/website, original price, current price, likes, and shipping details.


3. To extract data using Python's loop, start by opening the developer tool on the browser or inspecting an element on the website. Most developer tools will highlight your selection and focus on the HTML tag you choose. Look for a similar pattern to use in your loop. If you move to the next item, you may see the same tag again. For instance, a div tag with class "fpItem" is used for each item in this example - < div class="fpItem" >.


4. To retrieve additional data related to < div class="fpItem" >, we need to access its parent. You can obtain the names of all classes by following the same steps described earlier with the use of Developer Tools in your browser and extracting the necessary fields.


Scrapy Code

Once you have determined the appropriate class from which to extract data, you can create a Python Scrapy project and execute a test run. For additional information on Scrapy, please visit the following link.


The code shown is a file named located in Scrapy's Spider folder. To begin, we name the crawler "slickdeals." As previously mentioned, we use Selector to obtain a list item by calling it.


After obtaining the list, we can go through each item and gather the necessary information by utilizing XPath. We will verify if the class includes our desired keyword during this process.


After collecting the data, we save it in a CSV file for further analysis. If you prefer, you may also send an email with a specific keyword using Python's email module. Here's an example code without any content.


To test this program using a project root directory, just execute

scrapy crawl slickdeals

And the result will look something like this and you’d observe the fields which we have extracted.



To ensure our program runs continuously, it's best to use an energy-efficient Raspberry Pi. Once the code is confirmed to work, we can schedule the web crawler application to run automatically using Linux's crontab feature. To do this, open crontab with the command "crontab -e" and add the following command: "*/15 * * * *". This will execute the web crawler every 15 minutes.


Final thought

Great job! Your web scraping program is now up and running 24/7, just as you requested. Whether your aim is to find great deals, freebies, or coupons, our program is working tirelessly in the background to monitor and alert you of the best finds. We hope this blog has given you some insight into web scraping and the potential to build even more advanced programs on a small device like the Raspberry Pi.

For more details, contact Actowiz Solutions! You can also tell us your about your mobile app scraping or web scraping service requirements.

Recent Blog

View More

Restaurant Guru Data Scraping in Makati Philippines - A Comprehensive Guide

Explore comprehensive Restaurant Guru data scraping in Makati, Philippines, for valuable insights into the local restaurant scene and market trends..

How to Make Data Collection from Swiggy Instamart, Zepto, and Blinkit?

Learn how to collect data from Swiggy Instamart, Zepto, and Blinkit using web scraping techniques for market analysis and optimization.

Research And Report

View More

Actowiz Solutions Growth Report

Actowiz Solutions: Empowering Growth Through Innovative Solutions. Discover our latest achievements and milestones in our growth report.

Analysis of Trulia Housing Data

Comprehensive research report analyzing trends and insights from Trulia housing data for informed decision-making in real estate.

Case Studies

View More

Case Study - Empowering Price Integrity with Actowiz Solutions' MAP Monitoring Tools

This case study shows how Actowiz Solutions' tools facilitated proactive MAP violation prevention, safeguarding ABC Electronics' brand reputation and value.

Case Study - Revolutionizing Retail Competitiveness with Actowiz Solutions' Big Data Solutions

This case study exemplifies the power of leveraging advanced technology for strategic decision-making in the highly competitive retail sector.


View More

Maximize Growth with Price Sensitivity and Price Matching in 2024

Maximize growth in 2024 with insights on price sensitivity, price matching, price scraping, and effective pricing data collection techniques.

Unleash the power of e-commerce data scraping

Leverage the power of e-commerce data scraping to access valuable insights for informed decisions and strategic growth. Maximize your competitive advantage by unlocking crucial information and staying ahead in the dynamic world of online commerce.