Available at mefist.at
This is a personal side project. Issues and comments are welcome. I'm not looking for pull requests.
To remind myself how it works in future:
infodumpworkflow calls python moduleinfodump_tools.download, on a cron schedule- script checks "last updated" timestamp on Infodump homepage
- if a new Infodump is available, download files to
infodump/, calculate stats, and output tosrc/data/data.json - push json to repo
- a Personal Access Token must be stored in the
INFODUMP_ACCESS_TOKENrepo secret, so pushes trigger a deployment INFODUMP_USER_AGENTrepo secret should also be set
- pushes trigger the
deployworkflow topnpm buildthe site and upload to GitHub pages
-
create a venv, activate it, install requirements
python -m venv .env . .env/bin/activate pip install -r infodump_tools/requirements.txt -
optionally, set
INFODUMP_USER_AGENTenvironment variable to a string that identifies you (header is sent when downloading large Infodump files) -
run
python -m infodump_tools.download --dev infodump src/data/data.json- this downloads Infodump files to the
infodumpdirectory and outputs stats tosrc/data/data.json. with the-d|--devflag, we always regenerate the json, even if there is no new Infodump
- this downloads Infodump files to the
- Jupyter notebooks (in
notebooks/) are an easy way of developing and testing Polars expressions. They are not used to generate the live site. Install Jupyter kernel requirements fromnotebooks/requirements.txt. This includes packages needed by the Data Wrangler extension for VS Code.