sudo ./import.sh
The target Elasticsearch-URL and index prefix can be configured at the top of the file.
By default the data is not persisted across container restarts. To enable this, you need to these actions, all from the base folder, not import-scripts:
-In the file ./docker-compose.yml
, uncomment this line: # - ./docker/elasticsearch/data:/usr/share/elasticsearch/data
.
-Create the folder ./docker/elasticsearch/data
, if it does not exist: From the base folder: sudo mkdir docker/elasticsearch/data
.
-Grant group write rights: sudo chmod g+w docker/elasticsearch/data
[old]
Before doing anything, you should do:
- fill the
config.json
file in their respective directories with lifeurl
: the URL to the WFS 1.1.0 aufgrabungen servereSurl
: the URL to the Elasticsearch server with index and typebaseUrl
: the URL stump to the WFS 2.0.0 for everything else
- run
yarn
in the respective directories before executing the scripts
Inserts the districts into an elasticsearch index indexed with geo_shape
Inserts the construction sites into the places index. Execute with node aufgrabungen-to-elasticsearch.js
Inserts the following feature classes into the places index.
- pool
- agencies
- bookbus
- library
- cemetery
- wc
- wifi
- school
- playground
- sport
- kindergarden
Execute with node index.js
Inserts the booking plans of sports facilities in Münster. You need to pass the environment parameter GOOGLE_API_KEY
(Google Maps API key) when running the import script import_hallenbelegung.py
To update the JSON file with new container data to the following:
- Go to https://overpass-turbo.eu
- Execute the script: [out:json][timeout:25]; // gather results ( // query part for: “"recycling:glass"=yes” node"recycling:glass"="yes"; way"recycling:glass"="yes"; relation"recycling:glass"="yes"; ); // print results out body;
; out skel qt;
- Find the JSON in the "data" tab