Adding export script stub, removing bulk data

This commit is contained in:
Will Bradley 2021-01-31 21:03:48 -08:00
parent aaa72ff924
commit bf1a80f177
4 changed files with 111 additions and 1 deletions

5
.gitignore vendored
View File

@ -7,7 +7,10 @@
*.xml *.xml
# backups # backups
original_data-backup.sql original_data*
# output
data/
# VIM stuff # VIM stuff
*.swp *.swp

107
export.sh Normal file
View File

@ -0,0 +1,107 @@
#!/bin/bash
#
# Based on https://github.com/codeforsanjose/OSM-SouthBay/blob/main/SJ_Buildings/start.sh
#
# Before running, download VTA TAZ data from Google Drive here:
# https://drive.google.com/file/d/0B098fXDVjQOhVHBFS0kwcDNGRlU/view
# and place into a folder named "data"
# (might need to rename VTATaz.dbf)
DBNAME=svosm
OGR2OSM=../../ogr2osm/ogr2osm.py
# DB setup
psql --echo-all --command="create extension if not exists hstore;" "${DBNAME}" postgres
psql --echo-all --command="create extension if not exists postgis;" "${DBNAME}" postgres
# Add ESRI:103240 to PostGIS
# from https://github.com/Esri/projection-engine-db-doc/
psql --echo-all --file="103240.sql" "${DBNAME}" postgres
echo "Importing TAZ"
shp2pgsql -d -D -s 103240 -I "data/VTATaz" | psql -d "${DBNAME}" >/dev/null
echo "Downloading Basemap"
curl "https://www.sanjoseca.gov/DocumentCenter/View/17141" --output "Basemap.zip"
unzip "Basemap.zip" "Parcel.*" -d "data"
echo "Importing Parcel"
shp2pgsql -d -D -s 103240 -t 2D -I "data/Parcel" | psql -d "${DBNAME}" >/dev/null
echo "Downloading Basemap_2"
curl "http://www.sanjoseca.gov/DocumentCenter/View/44895" --output "Basemap_2.zip"
unzip "Basemap_2.zip" "BuildingFootprint.*" "CondoParcel.*" "Site_Address_Points.*" \
-d "data"
echo "Importing BuildingFootprint"
shp2pgsql -d -D -s 103240 -I "data/BuildingFootprint" \
| psql -d "${DBNAME}" >/dev/null
echo "Importing CondoParcel"
shp2pgsql -d -D -s 103240 -I "data/CondoParcel" | psql -d "${DBNAME}" >/dev/null
echo "Importing Site_Address_Points"
shp2pgsql -d -D -s 2227 -k -I "data/Site_Address_Points" \
| psql -d "${DBNAME}" >/dev/null
# Download and import existing OSM data
echo "Downloading norcal-latest.osm.pbf"
curl "https://download.geofabrik.de/north-america/us/california/norcal-latest.osm.pbf" \
--output "data/norcal-latest.osm.pbf"
echo "Importing norcal-latest.osm.pbf"
osm2pgsql --database "${DBNAME}" --create \
--prefix osm \
--slim --hstore \
--latlong --multi-geometry \
--bbox "-122.038182903664,37.1409050504209,-121.593273327604,37.4640955052253" \
"data/norcal-latest.osm.pbf"
# Merge addresses to buildings
psql -v "ON_ERROR_STOP=true" --echo-queries --file="merge.sql" "${DBNAME}"
# Split into tasks
mkdir "out"
mkdir "out/intersecting"
mkdir "out/clean"
for intersects in false true; do
if ${intersects}; then
outdir="intersecting"
intersectsQuery="intersectsExisting"
else
outdir="clean"
intersectsQuery="not intersectsExisting"
fi
ogr2ogr -sql "select 'https://codeforsanjose.github.io/OSM-SouthBay/SJ_Buildings/out/${outdir}/buildings_' || key || '.osm' as import_url, ST_SimplifyPreserveTopology(geom, 4) from VTATaz" \
-t_srs EPSG:4326 \
"out/grouped_${outdir}_buildings_zones.geojson" \
"PG:dbname=${DBNAME} host=localhost"
sed -i 's/ //g' "out/grouped_${outdir}_buildings_zones.geojson"
for cid in {1153..2632}; do
# Skip empty TAZs
if [ $(psql --command="copy (select count(*) from VTATaz where key=${cid}) to stdout csv" ${DBNAME}) = 0 ]; then
continue
fi
output="out/${outdir}/buildings_${cid}.osm"
# Filter export data to each CID
for layer in "buildingfootprint" "Site_Address_Points" "mergedbuildings" "namedparcels"; do
psql -v "ON_ERROR_STOP=true" --echo-queries --command="create or replace view \"${layer}_filtered\" as select * from \"${layer}\" where cid=${cid} and ${intersectsQuery};" "${DBNAME}"
done
# Export to OSM
python "${OGR2OSM}" "PG:dbname=${DBNAME} host=localhost" -t "basemap.py" -f --no-memory-copy -o "${output}"
# Add sample region outline
#sed -i '3i<bounds minlat="37.2440898883458" minlon="-121.875007225253" maxlat="37.25775329679" maxlon="-121.855829662555" />' "${output}"
done
done

Binary file not shown.