Skip to content

Latest commit

 

History

History
37 lines (24 loc) · 1.33 KB

File metadata and controls

37 lines (24 loc) · 1.33 KB

What this program does and why

As already written in the description, this program downloads specified URLs, parses HTML and writes the parsed information to an atom-feed.

Thus, you do not need any account or login to be up to date what is happening on these sites. You just need a RSS-reader and an Internet-connection.

Dependencies

The only dependency you need to run this script is Python 3.x. If it is not installed already, you can simply do this for a Debian-like OS by running

sudo apt-get install python3

Installation

Simply download (or clone) this repository. You're done!

Usage

To run the program once, you can simply run

python Main.py

in a terminal.

For periodically usage – IMHO the main task – you can run it with the help of cron, runwhen, systemd or equivalent software. See their documentation on how to use it, please.

For further configuration – f.e. which sites should be parsed – see the file config/html2rss.cfg.default. This file contains comments that should document the file itself.

HTTP-Error-Handling

If the download of a page fails with f.e. a 503 “Service Unavailable”, an error message with the pageurl and status-code will be directly printed to stdout. Additionally, the page will be skipped in this run.

Run tests

To run the (basic) tests simply execute

python -m unittest