
The Superfund NPL Site Scraper automates the collection and standardization of data from U.S. EPA Superfund resources. Built in Python, the tool retrieves site-level information from EPA online tables, Microsoft Excel files, and individual site profile pages. It uses requests, BeautifulSoup, and pandas to parse structured and semi-structured content, extract cleanup milestones, and normalize outputs into consistent CSV schemas (e.g., site ID, site name, location, operational status, milestone history).The scraper is fully configurable, enabling users to add or modify target data fields without restructuring the codebase. Designed for repeated use, it supports research tracking, program reporting, and integration with Google Sheets and other database systems.
NPL (National Priorities List), Data parsing, Research data management, Web scraping
NPL (National Priorities List), Data parsing, Research data management, Web scraping
| selected citations These citations are derived from selected sources. This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | 0 | |
| popularity This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network. | Average | |
| influence This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically). | Average | |
| impulse This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network. | Average |
