No description
- Add encoding="utf-8" to all file writes - Catch requests.RequestException instead of bare except - Use raise_for_status() to also retry on HTTP errors - Use removeprefix/removesuffix instead of lstrip/rstrip - Use makedirs(exist_ok=True) - Remove unused common_suffix function and commonprefix import Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com> |
||
|---|---|---|
| .gitignore | ||
| crawler.py | ||
| flake.lock | ||
| flake.nix | ||
| pyproject.toml | ||
| Readme.md | ||
Bundescrawler
This repository contains the scraper (I just liked the name Bundescrawler), which collects the available information from the sites of the representatives of the German Parlament.
How to use
- Clone the repository
- Install the dependencies with
pip install . - Initialize a repository somewhere where you want to save the information.
- Run
python3 crawler.py -o <output directory>