govgis_nov2023 / to_mongo_4326reproj.py
joshuasundance's picture
Upload to_mongo_4326reproj.py
4f59df9
import asyncio
import json
import os
import random
import uuid
from glob import glob
from typing import AsyncGenerator, Optional
import aiofiles
import nest_asyncio
from motor.motor_asyncio import AsyncIOMotorClient
from motor.motor_tornado import MotorCollection
from pymongo.server_api import ServerApi
from pyproj import CRS, Proj, Transformer
from pyproj.exceptions import CRSError
from shapely.errors import GEOSException
from shapely.geometry import Polygon
nest_asyncio.apply()
async def safe_insert_many(collection: MotorCollection, documents: list[dict]) -> None:
if len(documents) > 0:
try:
# Attempt to insert the documents
await collection.insert_many(documents)
except OverflowError:
# If an OverflowError occurs, split the batch
if len(documents) > 1:
mid = len(documents) // 2
# Recursively attempt to insert each half
await safe_insert_many(collection, documents[:mid])
await safe_insert_many(collection, documents[mid:])
else:
# Handle the case where a single document is too large
raise ValueError(
"A document exceeds the maximum BSON size or dtype conflict.",
)
def create_uuid(input_str: str) -> str:
# Consistent random UUIDs based on input string
# https://nathanielknight.ca/articles/consistent_random_uuids_in_python.html
random.seed(input_str)
return str(
uuid.UUID(bytes=bytes(random.getrandbits(8) for _ in range(16)), version=4),
)
def validate_coordinate(lon: float, lat: float) -> tuple:
"""
Validate and adjust the longitude and latitude values to ensure they are within the valid range.
Parameters:
lon (float): Longitude value.
lat (float): Latitude value.
Returns:
tuple: Validated and potentially adjusted (longitude, latitude) pair.
"""
# Adjust longitude and latitude values to their valid range
if not -180 <= lon <= 180:
lon = min(max(lon, -180), 180)
if not -90 <= lat <= 90:
lat = min(max(lat, -90), 90)
return lon, lat
def reproject_to_4326_and_convert_to_geojson(
bbox: dict,
) -> dict:
"""
Reprojects a bounding box from EPSG:102100 (3857) to EPSG:4326 and converts it to GeoJSON.
Parameters:
bbox (dict): A dictionary containing the keys 'xmin', 'ymin', 'xmax', 'ymax', and 'spatialReference'.
Returns:
dict: A GeoJSON object representing the reprojected bounding box.
"""
# Define the source and destination coordinate systems
def get_src_proj() -> Proj:
sr = bbox["spatialReference"]
for wkid in ["latestWkid", "wkid"]:
if (sr_wkid := sr.get(wkid)) is not None:
for authority in ["EPSG", "ESRI"]:
try:
return Proj(f"{authority}:{sr_wkid}")
except CRSError:
pass
if (sr_wkt := sr.get("wkt")) is not None:
return Proj(CRS.from_wkt(sr_wkt))
raise ValueError("no spatialReference found")
src_proj = get_src_proj()
dst_proj = Proj("epsg:4326")
transformer = Transformer.from_proj(src_proj, dst_proj, always_xy=True)
# Extract coordinates
xmin, ymin, xmax, ymax = bbox["xmin"], bbox["ymin"], bbox["xmax"], bbox["ymax"]
# Transform the coordinates
xmin_trans, ymin_trans = validate_coordinate(*transformer.transform(xmin, ymin))
xmax_trans, ymax_trans = validate_coordinate(*transformer.transform(xmax, ymax))
# Create a polygon from the transformed coordinates
# Ensure that the polygon is closed by repeating the first point at the end
coords = [
(xmin_trans, ymin_trans),
(xmax_trans, ymin_trans),
(xmax_trans, ymax_trans),
(xmin_trans, ymax_trans),
(xmin_trans, ymin_trans),
]
if len(set(coords)) < 3:
raise ValueError("invalid extent")
polygon = Polygon(coords)
# # Convert the polygon to GeoJSON format
geojson = {
"type": "Polygon",
"coordinates": [list(polygon.exterior.coords)],
}
return geojson
keepkeys = {
"id",
"associatedlayers",
"domains",
"copyrighttext",
"description",
"documentinfo",
"fields",
"mapname",
"name",
"parentlayer",
"servicedescription",
"subLayers",
"tables",
"version",
"currentversion",
"geometrytype",
"extent",
"type",
"url",
"server",
"layers",
"service",
}
async def process_metadata(
metadata: dict,
additional_fields: Optional[dict] = None,
) -> dict:
additional_fields = additional_fields or {}
# Process metadata and add any additional fields
processed_md = {
k: v for k, v in metadata.items() if k not in ["folders", "services", "layers"]
}
processed_md.update(additional_fields)
processed_md["original_id"] = processed_md.get("id", None)
processed_md["id"] = processed_md["hash"]
del processed_md["hash"]
processed_md = {k: v for k, v in processed_md.items() if k in keepkeys}
if "extent" in processed_md:
if (
(extent := processed_md.get("extent")) is not None
and extent.get("spatialReference") is not None
and not any(
str(v).lower() in {"nan", "none", "null"} for v in extent.values()
)
):
try:
processed_md["extent"] = reproject_to_4326_and_convert_to_geojson(
extent,
)
except (ValueError, GEOSException, CRSError):
del processed_md["extent"]
# print(extent, e)
else:
del processed_md["extent"]
return processed_md
def get_type(layer: dict) -> str:
return layer.get("type", "unknown").lower().replace(" ", "_").strip()
async def read_data(jsonfiles: list[str]) -> AsyncGenerator[dict, None]:
# Async generator to yield file content one by one
for f in jsonfiles:
async with aiofiles.open(f, "r") as infile:
content = await infile.read()
yield json.loads(content)
# Modified process_server function
async def process_server(server: dict, layers_collection: MotorCollection):
server_services = server.pop("services")
server_md = await process_metadata(
server["metadata"],
{
"url": server["metadata"]["url"],
"hash": create_uuid(server["metadata"]["url"]),
},
)
layer_tasks = []
for service in server_services:
service_md = await process_metadata(
service["metadata"],
{
"url": service["url"],
"hash": create_uuid(service["url"]),
"server": server_md,
},
)
for layer in service["metadata"].pop("layers"):
task = asyncio.create_task(process_layer(layer, service_md))
layer_tasks.append(task)
layers_md = await asyncio.gather(*layer_tasks)
await safe_insert_many(layers_collection, layers_md)
# Modified process_layer function
async def process_layer(layer: dict, service: dict) -> dict:
# Embed service metadata into layer
layer_md = await process_metadata(
layer,
{
"url": layer["url"],
"hash": create_uuid(layer["url"]),
"service": {**service},
},
)
return layer_md
async def main() -> None:
output_dir = os.path.abspath("/home/appuser/restgdf_api/lab/output_tryagain")
jsonfiles = glob(os.path.join(output_dir, "*.json"))
client = AsyncIOMotorClient(
r"mongodb://root:example@mongo:27017/",
server_api=ServerApi("1"),
)
db = client["govgis-nov2023-slim-spatial"]
layers = db.layers
# Process each server concurrently
async for server in read_data(jsonfiles):
await process_server(server, layers)
if __name__ == "__main__":
asyncio.run(main())