This section aims to briefly sketch out the process of installing the app for development.
Installation instructions with Vagrant¶
The suggested method includes Vagrant, a virtualization tool aimed at developers.
- Clone the github repository (duh)
- Navigate into the project dir cd OffenesParlament
- Setup and run the vagrant VM vagrant up. All requirements will be installed automatically inside the VM which may take a few minutes the first time.
- The script might ask you for your password as it will add offenesparlament.vm pointing to this VM to your hosts-file. It also automatically creates a django superuser admin with password admin.
- Log in to the running VM with vagrant ssh
- For the initial scraping instructions see below
- Run the server inside the VM (0.0.0.0 lets the server respond to requests from outside the VM - ie your physical machine where you probably run your browser)
cd offenesparlament python manage.py runserver 0.0.0.0:8000
- If you work on client files that have to be compiled (CSS, JS) you have to run grunt as well. ATM we have the tasks dev and reloading. dev watches and regenerates files when their sources change. (Remember that sources also change when you do a git pull and generated client files aren’t commited to git) And reloading does that and uses Browsersync to reload your browser when files change
cd /vagrant grunt dev
- To exit and shutdown the VM run
exit vagrant halt
Resetting the database¶
In case you need to reset the database (delete all migrations, flush the db content, recreate all objects etc.), run these commands in the django project folder ‘offenesparlament’:
Creating a Model-Diagram¶
It’s possible to view the current database model residing in the op_scraper app by calling:
A png-image will be generated as ignore/models.png.
There are currently four available scrapers, which should initially run in this order:
- llp (legislative periods)
- persons (for instance Rudolf Anschober <http://www.parlament.gv.at/WWER/PAD_00024/index.shtml>)
- administration (for instance, ‘Faymann II’ and all the Persons having a mandate for that administration)
- pre_laws (for instance Buchhaltungsagenturgesetz, Änderung (513/ME))
- laws_initiatives (for instance ÖBIB-Gesetz 2015 (458 d.B.))
To run a scraper, use the following command:
python manage.py scrape crawl <scraper_name>
python manage.py scrape crawl persons
The law_initiatives scraper also has an additional parameter to define, which legislative period to scan; per default, it scrapes the periods from XX to XV. This can be overriden this way:
python manage.py scrape crawl -a llp=21 laws_initiatives
to only scrape that period. Careful though: scraping of periods before the 20th legislative period is not possible as of yet (since there are no machine-readable documents available).
ElasticSearch and Re-Indexing¶
For now, reindexing (or updating the index, for that matter), is only done manually. To have all data indexed, just run:
python manage.py rebuild_index
for a full rebuild (wipes the indices first), or:
python manage.py update_index
to perform a simple update. For this to succeed, make sure ElasticSearch is up and running.