A simple Python scraper for s.is
First version, undocumented and very raw. Run the scraper:
$ ./scraper.py
and your current directory will fill up with route-XX.json files containing info for routes 1 - 44.
Depends on the Requests HTTP library.
When you have scraped all the routes you can manually inspect the data using srvr.py. It is a simple WSGI server that displays a timetable for the routes in the current directory. Run the server from the same directory as your json files are:
$ gunicorn srvr:app
Now you can browse the timetables on http://localhost:8000.
api.py contains a simple JSON API interface that uses the scraped data. Before running the API an SQLite DB needs to be created from the JSON files from the scraper. Create the DB:
$ ./create-db.py
and then run the API:
$ python api.py
It depends on Flask.