1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
|
#----------------------------------------------------------------------
# Backend utilities for the Klimatanalys Norr project (common module)
# Copyright © 2024 Guilhem Moulin <info@guilhem.se>
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program. If not, see <https://www.gnu.org/licenses/>.
#----------------------------------------------------------------------
# pylint: disable=missing-module-docstring
import os
from os import path as os_path, curdir as os_curdir, pardir as os_pardir, sep as os_sep
import sys
from fnmatch import fnmatchcase
from pathlib import Path, PosixPath
from urllib.parse import urlparse
from stat import S_ISDIR
import math
import logging
from typing import Any, Optional, Never
from xdg.BaseDirectory import xdg_config_home
import yaml
def init_logger(app : str =__file__, level : int = logging.WARNING) -> logging.Logger:
"""Initialize the logger"""
log_fmt = logging.Formatter('%(levelname)s: %(message)s')
log = logging.getLogger()
log.setLevel(level)
pid = os.getenv('SYSTEMD_EXEC_PID', None)
if (pid is None or int(pid) != os.getpid()
or os.getenv('JOURNAL_STREAM', None) is None):
ch = logging.StreamHandler()
else:
# started in systemd, use journald for filtering incl. coloring
from systemd.journal import JournalHandler # pylint: disable=import-outside-toplevel
ch = JournalHandler(SYSLOG_IDENTIFIER=app)
ch.setFormatter(log_fmt)
log.addHandler(ch)
return log
class MissingConfiguration(Exception):
"""Exception raised when no configuration file could be found"""
def __init__(self, name : str) -> Never:
super().__init__(f'Could not find configuration file {name}')
def find_config(filename : str = 'config.yml', appname : str = 'webmap') -> Path:
"""Return the configuration file path"""
dirs = [
Path(),
Path(xdg_config_home).joinpath(appname),
PosixPath('/etc').joinpath(appname)
]
for d in dirs:
p = d.joinpath(filename)
if p.exists():
return p
raise MissingConfiguration(filename)
class BadConfiguration(Exception):
"""Exception raised when there is a bad configuration"""
def __init__(self, config_path : Path, message : str) -> Never:
super().__init__(str(config_path) + ': ' + message)
def parse_config(path : Optional[Path] = None,
groupnames : Optional[list[str]] = None) -> dict[str, Any]:
"""Parse configuration file"""
config_path = find_config() if path is None else path
with config_path.open(mode='r', encoding='utf-8') as fp:
config = yaml.safe_load(fp)
layers = config.get('layers', {})
# validate sources
destinations = {}
for name, layerdefs in layers.items():
if isinstance(layerdefs, dict) and 'sources' not in layerdefs:
layers[name] = { 'sources': [layerdefs] }
for k in ['description', 'create', 'publish']:
if k in layerdefs:
layers[name][k] = layerdefs.pop(k)
layerdefs = layers[name]
if 'sources' not in layerdefs:
# pylint: disable-next=broad-exception-raised
raise Exception(f'Layer "{name}" does not have any source recipe')
for sourcedef in layerdefs.get('sources', []):
source = sourcedef.get('source', None)
if source is None:
continue
download = source.get('download', None)
if download is None:
url = None
dl_module = None
elif isinstance(download, str):
url = download
dl_module = None
source['download'] = download = { 'url': url }
else:
url = download.get('url', None)
dl_module = download.get('module', None)
if url is None:
urlp = None
else:
urlp = urlparse(url)
if urlp is None:
# pylint: disable-next=broad-exception-raised
raise Exception(f'urlparse({url}) failed')
cache = source.get('cache', None)
if cache is None or isinstance(cache, str):
source['cache'] = { 'path': cache }
else:
cache = cache.get('path', None)
if cache is None or cache in ['', os_curdir, os_pardir] or cache.endswith(os_sep):
# infer filename from the source URL
if urlp is None or urlp.path is None or urlp.path == '' or urlp.path.endswith('/'):
# pylint: disable-next=broad-exception-raised
raise Exception(f'Layer "{name}": Could not infer filename from URL {url}')
p = PosixPath(urlp.path)
if p is None or p.name is None or p.name == '':
# pylint: disable-next=broad-exception-raised
raise Exception(f'Invalid PosixPath({urlp.path})')
if cache is None or cache == '':
cache = Path()
else:
cache = Path(cache)
cache = cache.joinpath(p.name)
else:
cache = Path(cache)
source['cache']['path'] = cache
v = { 'url': urlp, 'module': dl_module }
if cache in destinations and destinations[cache] != v:
# allow destination conflicts, but only when the source URL and module match
# pylint: disable-next=broad-exception-raised
raise Exception(f'Destination conflict for layer "{name}"')
destinations[cache] = v
# filter layers that are not of interest
if groupnames is not None:
layernames = []
layer_groups = config.get('layer-groups', {})
for groupname in groupnames:
if groupname not in layer_groups:
if groupname in layers:
# fallback to layer names
layernames.append(groupname)
else:
logging.error('Unknown group/layer name "%s"', groupname)
sys.exit(1)
else:
patterns = layer_groups[groupname]
if isinstance(patterns, str):
patterns = [patterns]
for pat in patterns:
has_match = False
for layername in layers:
if fnmatchcase(layername, pat):
if layername in layernames:
logging.debug('Layer "%s" was already added, skipping', layername)
else:
layernames.append(layername)
has_match = True
if has_match:
pass
elif pat in layers:
# fallback to exact match
if pat in layernames:
logging.debug('Layer "%s" was already added, skipping', pat)
else:
layernames.append(pat)
else:
logging.warning('Group name "%s" does not match anything', groupname)
layers = { name: layers[name] for name in layernames }
config['layers'] = layers
extent = config.get('extent', None)
if extent is not None:
if isinstance(extent, list):
config['extent'] = tuple(extent)
if config.get('SRS', None) is None:
# pylint: disable-next=broad-exception-raised
raise Exception('Configured extent without SRS')
return config
def format_bytes(n : int, threshold : int = 768, precision : int = 2) -> str:
"""Format a number of bytes to a SI unit"""
if n < threshold:
return f'{n}\u202FB'
if n < threshold * 1024:
return f'{n/1024:.{precision}f}\u202FkiB'
if n < threshold * 1048576:
return f'{n/1048576:.{precision}f}\u202FMiB'
return f'{n/1073741824:.{precision}f}\u202FGiB'
def format_time(ts : float, precision : int = 3) -> str:
"""Format a timestamp to HH:MM:SS.fff"""
w = 2 if precision == 0 else precision + 3
ts = round(ts, precision)
m = math.floor(ts/60.)
s = ts - 60. * m
h, m = divmod(m, 60)
return f'{h:02d}:{m:02d}:{s:0{w}.{precision}f}'
# pylint: disable-next=invalid-name, redefined-builtin
def gdalVersionMin(gdal, maj : int = 0, min : int = 0, rev : int = 0) -> bool:
"""Return a boolean indicating whether the installer GDAL version is
greater than or equal to the provider (maj, min, rev) triplet."""
if maj < 1 or (maj == 1 and min < 10):
# GDAL_VERSION_NUM() macro was changed in 1.10. That version
# was released in 2013 so we blindly assume the installer
# version is more recent
return True
version_cur = int(gdal.VersionInfo())
# cf. GDAL_COMPUTE_VERSION(maj,min,rev) in gcore/gdal_version.h.in
version_min = maj*1000000 + min*10000 + rev*100
return version_min <= version_cur
# pylint: disable-next=invalid-name
def gdalGetMetadataItem(obj, k : str) -> bool:
"""Wrapper around gdal.MajorObject.GetMetadataItem(name)."""
v = obj.GetMetadataItem(k)
if v is not None and isinstance(v, str):
return v.upper() == 'YES'
return False
def escape_identifier(identifier : str) -> str:
"""Escape the given identifier, cf.
swig/python/gdal-utils/osgeo_utils/samples/validate_gpkg.py:_esc_id()."""
if '\x00' in identifier:
# pylint: disable-next=broad-exception-raised
raise Exception(f'Invalid identifier "{identifier}"')
# SQL:1999 delimited identifier
return '"' + identifier.replace('"', '""') + '"'
# pylint: disable-next=invalid-name,dangerous-default-value
def gdalSetOpenExArgs(gdal, option_dict : Optional[dict[str, Any]] = {}, flags : int = 0):
"""Return a pair kwargs and driver to use with gdal.OpenEx()."""
kwargs = { 'nOpenFlags': gdal.OF_VECTOR | flags }
fmt = option_dict.get('format', None)
if fmt is None:
drv = None
else:
drv = gdal.GetDriverByName(fmt)
if drv is None:
# pylint: disable-next=broad-exception-raised
raise Exception(f'Unknown driver name "{fmt}"')
if not gdalGetMetadataItem(drv, gdal.DCAP_VECTOR):
# pylint: disable-next=broad-exception-raised
raise Exception(f'Driver "{drv.ShortName}" has no vector capabilities')
kwargs['allowed_drivers'] = [ drv.ShortName ]
oo = option_dict.get('open-options', None)
if oo is not None:
kwargs['open_options'] = [ k + '=' + str(v) for k, v in oo.items() ]
return kwargs, drv
# pylint: disable-next=invalid-name
def getSRS(osr, srs_str : Optional[str]):
"""Return the decoded Spatial Reference System."""
if srs_str is None:
return None
srs = osr.SpatialReference()
if srs_str.startswith('EPSG:'):
code = int(srs_str.removeprefix('EPSG:'))
srs.ImportFromEPSG(code)
else:
# pylint: disable-next=broad-exception-raised
raise Exception(f'Unknown SRS {srs_str}')
logging.debug('Default SRS: "%s" (%s)', srs.ExportToProj4(), srs.GetName())
return srs
# pylint: disable-next=invalid-name
def getExtent(extent : tuple[float, float, float, float] | list[float], srs = None):
"""Convert extent [minX, minY, maxX, maxY] into a polygon and assign the
given SRS. Return a pair with the densified and non-densified extent.
Like apps/ogr2ogr_lib.cpp, the former is obtained by segmentizing the
polygon to make sure it is sufficiently densified when transforming to
source layer SRS for spatial filtering."""
if extent is None:
return None, None
if not isinstance(extent, (list, tuple)) or len(extent) != 4:
# pylint: disable-next=broad-exception-raised
raise Exception(f'Invalid extent {extent}')
if srs is None:
# pylint: disable-next=broad-exception-raised
raise Exception('Configured extent but no SRS')
logging.debug('Configured extent in %s: %s',
srs.GetName(), ', '.join(map(str, extent)))
from osgeo import ogr, osr # pylint: disable=import-outside-toplevel
ogr.UseExceptions()
ring = ogr.Geometry(ogr.wkbLinearRing)
ring.AddPoint_2D(extent[0], extent[1])
ring.AddPoint_2D(extent[2], extent[1])
ring.AddPoint_2D(extent[2], extent[3])
ring.AddPoint_2D(extent[0], extent[3])
ring.AddPoint_2D(extent[0], extent[1])
polygon = ogr.Geometry(ogr.wkbPolygon)
polygon.AddGeometry(ring)
# we expressed extent as minX, minY, maxX, maxY (easting/northing
# ordered, i.e., in traditional GIS order)
srs2 = srs.Clone()
srs2.SetAxisMappingStrategy(osr.OAMS_TRADITIONAL_GIS_ORDER)
polygon.AssignSpatialReference(srs2)
if not srs2.IsSame(srs):
polygon.TransformTo(srs)
# densify the rectangle to avoid issues when reprojecting to the
# source layer SRS, cf. apps/ogr2ogr_lib.cpp:ApplySpatialFilter()
polygon_dense = polygon.Clone()
segment_distance_metre = 10 * 1000
if srs.IsGeographic():
# pylint: disable-next=invalid-name
dfMaxLength = segment_distance_metre / math.radians(srs.GetSemiMajor())
polygon_dense.Segmentize(dfMaxLength)
elif srs.IsProjected():
# pylint: disable-next=invalid-name
dfMaxLength = segment_distance_metre / srs.GetLinearUnits()
polygon_dense.Segmentize(dfMaxLength)
return polygon_dense, polygon
######
# The function definitions below are taken from cpython's source code
# and augmented with dir_fd.
# An alternative would be to use str(Path(f'/proc/self/fd/{dir_fd}').joinpath(path)).
def isdir(path : str, dir_fd : Optional[int] = None, follow_symlinks : bool = True) -> bool:
"""Is a path a directory? (From genericpath.py.)"""
try:
st = os.stat(path, dir_fd=dir_fd, follow_symlinks=follow_symlinks)
except (OSError, ValueError):
return False
return S_ISDIR(st.st_mode)
def exists(path : str, dir_fd : Optional[int] = None, follow_symlinks : bool = True) -> bool:
"""Does a path exist? (From genericpath.py.)"""
try:
os.stat(path, dir_fd=dir_fd, follow_symlinks=follow_symlinks)
except (OSError, ValueError):
return False
return True
def makedirs(name : str, mode : int = 0o777,
exist_ok : bool = False,
dir_fd : Optional[int] = None) -> None:
"""Create a leaf directory and all intermediate ones. (From os.py.)"""
head, tail = os_path.split(name)
if not tail:
head, tail = os_path.split(head)
if head and tail and not exists(head, dir_fd=dir_fd):
try:
makedirs(head, exist_ok=exist_ok, dir_fd=dir_fd)
except FileExistsError:
# Defeats race condition when another thread created the path
pass
cdir = os_curdir
if isinstance(tail, bytes):
cdir = bytes(os_curdir, 'ASCII')
if tail == cdir: # xxx/newdir/. exists if xxx/newdir exists
return
logging.debug('mkdir("%s", 0%o)', name, mode)
try:
os.mkdir(name, mode, dir_fd=dir_fd)
except OSError:
# Cannot rely on checking for EEXIST, since the operating system
# could give priority to other errors like EACCES or EROFS
if not exist_ok or not isdir(name, dir_fd=dir_fd):
raise
|