Tips and Techniques


Interesting tips and articles about Web Scraping. How to successfully use automation to gather data from websites. Data extraction techniques and code are available in our tutorials

How To Optimize Playwright Web Scrapers Using Code Profiling Tools

How To Optimize Playwright Web Scrapers Using Code Profiling Tools

This article guides you to understand the method of code profiling. You also learn to optimize Playwright web scrapers using code profiling tools.

How Web Scraping With Excel Works: A Tutorial

How Web Scraping With Excel Works: A Tutorial

This tutorial deals with a detailed explanation of the Web Query feature in Excel and how web scraping with Excel is carried out using Web Query.

Web Scraping With Playwright in Python and JavaScript

Web Scraping With Playwright in Python and JavaScript

Learn about Playwright web scraping in Python and JavaScript by building and running web scrapers using a browser with Playwright.

Web Scraping using Urllib

Web Scraping using Urllib

Explore URL handling, parsing, quoting, and scraping with urllib. This step-by-step guide covers everything from basic URL operations to building a fully-fledged scraper using urllib.

What is Browser Fingerprinting? How to Bypass it?

What is Browser Fingerprinting? How to Bypass it?

Have you ever encountered the term “browser fingerprinting” while surfing the internet? With Browser fingerprinting, serves can uniquely identify clients and web scrapers, but there are ways to bypass this. Check out what browser fingerprinting entails and how you can avoid it.

Get Sales Leads From Google

Get Sales Leads From Google

In this tutorial we will show you how businesses can get sales leads from Google for free using Google Maps Crawler and Contact Detail Crawler available on ScrapeHero Cloud.

How To Make  Anonymous Requests using TorRequests and Python

How To Make Anonymous Requests using TorRequests and Python

Tor is quite useful when you have to use requests without revealing your IP address, especially when you are web scraping. This tutorial will use a wrapper in python that helps you with the same.

How To Install Python Packages for Web Scraping in Windows 10

How To Install Python Packages for Web Scraping in Windows 10

Web scraping using Python in Windows can be tough. In this tutorial follow the steps to setup python 3 and python packages on your Windows 10 computer for web scraping in Windows 10.

How to Solve Simple Captchas using Python Tesseract

How to Solve Simple Captchas using Python Tesseract

CAPTCHA stands for Completely Automated Public Turing test to tell Computers and Humans Apart. As the acronym suggests, it is a test used to determine whether the user is human or not. A typical captcha consists of a distorted test, which a computer program cannot interpret but a human can (hopefully) still read. This tutorial will […]

How to Parse Addresses using Python and Google GeoCoding API

How to Parse Addresses using Python and Google GeoCoding API

Web scraping can often lead to you having scraped address data which are unstructured. If you have come across a large number of freeform address as a single string, for example – “9 Downing St Westminster London SW1A, UK”,  you know how hard it would be to validate, compare and deduplicate these addresses. To start […]

The best data and file formats for scraped data

The best data and file formats for scraped data

The data we provide comes in various forms from the source and is largely text (barring rich media such as images and videos or proprietary file formats such as PDFs). Our customers need this data in various formats and the key to a successful and scalable solution that fits the best data formats for web […]

An API for every site using web scraping

An API for every site using web scraping

There is a lot of content available on the millions of websites on the Internet, and all of them involve some amount of programming to get them there, however, to get to all this content using a programmatic API isn’t really possible. If you need data scraped from a website in a specific format in […]

Turn the Internet into meaningful, structured and usable data