Screen scraping originally had nothing to do with websites.

The term comes from terminal emulation - the practice of reading character data from text-based display screens. In the 1970s and 1980s, business software ran on IBM mainframes and DEC minicomputers. The user interface was a green-screen terminal: rows and columns of characters, fixed positions, no graphics. No API. No file export. Just characters on a screen, and a user reading them.

If you needed to get data from one of these systems into another, your options were limited. The practical solution was software that read character positions directly from the terminal display and converted them into something a second application could use. Reading the screen. Scraping it, in the sense of collecting everything visible on its surface.

This was serious enterprise software. Companies built entire integration layers around it. The term appeared in IBM technical documentation throughout the 1980s and was well-established long before the web existed.

What happened when websites arrived

The web changed the surface being read while preserving the underlying problem. Instead of terminal screens - fixed character positions, no structure beyond row and column - there were web pages: HTML documents with nested tags, CSS classes, and data embedded in markup designed for browsers to display, not for software to extract.

Terminal screen scrapers and web scrapers share almost no technical DNA. One reads character grid positions from a terminal emulator. The other parses an HTML document tree. The mechanisms are different enough that the tools built for terminal extraction needed significant rethinking to handle web pages.

New tools emerged specifically for web data extraction. HTML parsers, pattern matchers, pagination handlers. These became known as web scrapers. The existing term "screen scraping" gradually extended to cover them - partly because the goal is similar (getting data off a visual display and into a usable format) and partly because terminology in software tends to lag the technology by a decade or more.

Today the term means different things depending on who's using it. Enterprise IT teams working with legacy mainframe systems use "screen scraping" to mean exactly what it meant in 1985: terminal emulation and character reading. No-code web scraping tools use it to mean extracting data from websites. The phrase sits across two entirely different technical domains, connected mainly by the word "screen."

How modern screen scraping tools work

The tools that come up in most current searches for "screen scraping" are browser-based extraction tools. Software that reads the structure of web pages and identifies data worth extracting - product names in heading tags, prices in elements with particular CSS classes, URLs in anchor tags within a specific container.

The underlying mechanism is HTML parsing, not screen reading. A data extraction tool working on a product listing page isn't looking at pixels. It's looking at the document structure: which elements repeat, what they contain, how they're related to each other. The "screen" metaphor survives because the output looks the same from the user's perspective - data that appeared visually is now in a spreadsheet - but the process underneath is document parsing, not image recognition.

No-code tools abstract this further. Instead of writing CSS selectors or XPath expressions, the user points at data on the page and the tool infers the pattern. Point at a product name. The tool identifies similar elements across the listing. Point at a price. The tool finds the same structure repeating through all the results. The complexity of HTML parsing disappears into the interface.

Where the original definition still applies

Some websites render their content entirely through JavaScript. The HTML that arrives from the server contains almost no data - the data appears on screen only after the JavaScript executes. Extracting from these pages requires something closer to the original screen scraping approach: run the JavaScript, wait for content to render visually, then read what's there.

This is a harder problem. Headless browsers - Puppeteer, Playwright - execute JavaScript like a real browser and then expose the resulting document for extraction. They're doing something technically closer to what terminal screen scrapers did: waiting for content to become visible, then collecting it from the rendered output.

For most everyday web data extraction - product listings, directories, search results pages - static HTML parsing is sufficient. The data is in the document before any JavaScript runs. Headless browsers add significant complexity and are only necessary when the target site renders content dynamically.

The three categories of tools

Screen scraping tools currently divide into three groups that rarely overlap.

Enterprise terminal emulators handle the original green-screen use case. Tools like Attachmate and Rocket Software are still actively maintained because industries built around legacy mainframe systems - banking, insurance, government - haven't migrated those systems, and the integration layer that connects mainframes to modern applications still runs on character-based screen reading.

Desktop and server-based web scrapers - Python libraries, Scrapy, Selenium, Playwright - handle complex web extraction at scale. These require technical setup and are typically used by developers building data pipelines or data teams running scheduled collection jobs. They're the right tool when volume, scheduling, or JavaScript rendering are non-negotiable requirements.

Browser extension tools handle the most common everyday extraction tasks: pulling structured data from a page you're already viewing, without setting up infrastructure or writing code. SiteScoop is in this category. The extraction logic runs in WebAssembly inside the browser tab - nothing leaves the machine - and handles the pattern detection and export that covers most no-code data collection use cases.

What the term tells you

A term that crosses several decades and two fundamentally different technical problems is usually a sign that the user need is consistent even when the technology isn't. The need that both terminal screen scraping and web data extraction address is the same: data is visible somewhere, getting it into a format you can work with is harder than it should be, and the manual alternative is copying it by hand.

The technology for doing this has changed completely. Green-screen character reading has almost nothing in common with WebAssembly-based HTML parsing. What hasn't changed is why people go looking for a screen scraping tool in the first place.