[GRASSLIST:9360] Re: [GRASS5] Re: Re: v.in.ascii problems

Roger Bivand Roger.Bivand at nhh.no
Wed Dec 7 10:09:03 EST 2005


On Wed, 7 Dec 2005, Roger Bivand wrote:

> On Thu, 8 Dec 2005, Hamish wrote:
> 
> > > There is an on-going discussion about this on the GRASS development
> > > list. >>From a simple test I ran last night, v.in.ascii -b (the -b
> > > flag is new in GRASS 6.1 CVS) does not build topology, and this
> > > removes one of the two humps in memory consumption. The other hump (>
> > > 200MB for a 1M point file with a single attribute) was associated with
> > > writing the dbf file (the file is 60MB), and is where things stick
> > > now. In addition, the -b flag leaves the vector data set at level 1
> > > topology (absent), and almost all vector commands need level 2. 
> > > 
> > > I do now know whether the use of a different database driver than the
> > > default would help. The dbf writing stage preceeds the topology
> > > building, so the two memory-intensive humps are separate, with
> > > topology being a little larger. Reading 1M points on a 1.5GHz P4 with
> > > topology took about 7 minutes, without about half that time.
> > 
> > 
> > Use the -z and -t flags to avoid making the table. (and the z= option)
> > If the input is just x,y,z data there is no need for a table.
> 
> For me in an x-y location my data with 1M points and -zbt now read in just
> under a minute; lidaratm2.txt in effectively the same time (64 rather than
> 57 seconds, z here is double not int) and v.in.ascii stays at a
> respectable 3.3MB size. d.vect works, but as you say prints a warning.

Emboldened by the 1M points case, I've tried 13.6M points, with:

GRASS > date; v.in.ascii -zbt input=gptsL.txt output=gptsLz x=1 y=2 z=3 ; date
Wed Dec  7 15:44:23 CET 2005
Maximum input row length: 37
Maximum number of columns: 3
Minimum number of columns: 3
Wed Dec  7 15:57:20 CET 2005

so scaling quite decently at about 1M points a minute. Once again the 
v.in.ascii process stayed at just over 3MB the whole time. So putting the 
single attribute of interest in the z column, not doing topology and not 
creating a database table seems to work - the coords file is now about 
480MB, so on 32-bit machines, its size may be a further limiting factor.

> 
> > 
> > At minimum, we need v.info, v.surf.rst, v.univar, v.out.ascii (points) 
> > and some sort of subsampling module (ie s.cellstats port) working with 
> > this data. d.vect works already (with a warning). maybe v.surf.idw too.
> > 
> > Probably not many more modules though? -- I think if Radim doesn't want 
> > this to be common-place use of the vector model then it probably shouldn't 
> > be. He knows it better than anyone.. So for now massive point datasets 
> > need to be treated as a special case to the vector model & only a 
> > work-around solution.
> > 
> > 
> > e.g. with the sample LIDAR data (GRASS downloads page)
> > 
> > G61> v.in.ascii -zbt in=lidaratm2.txt out=lidaratm2 x=1 y=2 z=3 fs=,
> > 
> > The first 250k points take about 20 seconds to load.
> > 
> > 
> > If I use the full million it gets stuck on the scanning step:
> > 
> > D3/3: row 374430 : 28 chars
> > 
> > Interesting, that line is the second value with elevation > 100.
> > 
> > changing the first z value to 500.054 it segfaults pretty quick:
> > 
> > D5/5: Vect_hist_write()
> > D4/5: G_getl2: ->-75.622346,35.949693,500.054<-
> > D3/5: row 1 : 28 chars
> > D4/5: token: -75.622346
> > D4/5: is_latlong north: -75.622346
> > D4/5: row 1 col 0: '-75.622346' is_int = 0 is_double = 1
> > D4/5: token: 35.949693
> > D4/5: is_latlong north: 35.949693
> > D4/5: row 1 col 1: '35.949693' is_int = 0 is_double = 1
> > D4/5: row 1 col 2: '500.054' is_int = 0 is_double = 1
> > D4/5: G_getl2: ->-75.629469,35.949693,11.962<-
> > D3/5: row 2 : 27 chars
> > D4/5: token: -75.629469
> > D4/5: is_latlong north: -75.629469
> > D4/5: row 2 col 0: 'H629469' is_int = 0 is_double = 0
> > Segmentation fault
> > 
> > where is 'H629469' coming from?
> 
> I was also seeing seg-faults with my data in a long-lat location, so 
> switched to x-y (current CVS 6.1).
> 
> Roger
> 
> > 
> > v.in.ascii/points.c
> >  tmp_token is getting corrupted, cascades from there
> > 
> > int points_analyse (){
> > ...
> >     char **tokens;
> > ...
> >     tmp_token=(char *) G_malloc(256);
> > ...
> >     while (1) {
> > ...
> >         tokens = G_tokenize (buf, fs);
> > ...
> >         for ( i = 0; i < ntokens; i++ ) {
> > ...
> > [*]                 sprintf(tmp_token, "%f", northing);
> > ...
> > 		    /* replace current DMS token by decimal degree */
> >                     tokens[i]=tmp_token;
> > 
> > BOOM. pointer abuse. (bug is new lat/lon scanning code, only in 6.1CVS)
> > 
> > [*] and if northing column is longer than 256 without hitting the fs, 
> >    buffer overflow??  add ,int maxlength, parameter to G_tokenize()?
> >    or can %f never be more than 256 bytes long?
> >     same %f effectively cutting down precision of lat/lon coords to 6 
> >    spots after the decimal place? (be that pretty small on the ground)
> > 
> > 
> > improvements come one bug at a time...
> > 
> > Hamish
> > 
> 
> 

-- 
Roger Bivand
Economic Geography Section, Department of Economics, Norwegian School of
Economics and Business Administration, Helleveien 30, N-5045 Bergen,
Norway. voice: +47 55 95 93 55; fax +47 55 95 95 43
e-mail: Roger.Bivand at nhh.no




More information about the grass-user mailing list