[postgis-users] Need better strategy for a query
Stephen Woodbridge
stephenwoodbridge37 at gmail.com
Tue Nov 19 13:33:28 PST 2019
Hi,
I have a global dataset in SRS EPSG:4326 that are ocean depth contours.
The problem is that these tend to be long and have huge bbox so spatial
index does not help when trying to render them in mapserver.
I plan was to chop these into shorter segments with a command like:
bathymetry=# explain select depth,
feet,
st_linesubstring(geom, 0.5*n/length,
case when 0.5*(n+1)<length then 0.5*(n+1)/length else 1 end) as
geom
from (
select a.depth,
-round(a.depth/0.3048) as feet,
st_linemerge(a.geom) as geom,
st_length(a.geom) as length
from c100 a
where st_length(a.geom)>0
) as t
cross join generate_series(0,10000) as n
where n*0.5/length < 1;
QUERY PLAN
-----------------------------------------------------------------------------------------------------------
Nested Loop (cost=0.00..549466002.55 rows=1345242667 width=1723)
Join Filter: (((((n.n)::numeric * 0.5))::double precision /
st_length(a.geom)) < '1'::double precision)
-> Seq Scan on c100 a (cost=0.00..1279615.77 rows=4035728 width=1719)
Filter: (st_length(geom) > '0'::double precision)
-> Function Scan on generate_series n (cost=0.00..10.00 rows=1000
width=4)
(5 rows)
but running a restricted test query on 10,000 lines took 1.5 hours and
with about 12M lines in the source table I estimate about 75 days to
compute :(
The 0.5 in the query is for 1/2 degree (~35 mile) max length. I can
adjust that to say 1 deg but I suspect that will only nominally impact
the run time (running a test to see).
So is there a faster way to do this?
Would smoothing the lines first help or would the cost of smooth and
then dividing be about the same?
-Steve W
More information about the postgis-users
mailing list