Presentation on theme: "Tony Rees Divisional Data Centre CSIRO Marine Research, Australia Application of c-squares spatial indexing to an archive of remotely."— Presentation transcript:
Tony Rees Divisional Data Centre CSIRO Marine Research, Australia (Tony.Rees@csiro.au) Application of c-squares spatial indexing to an archive of remotely sensed data
The starting point … Archive of c.60,000 satellite scenes, 1986- current, accumulating at another 15+ per day Region covered varies for each scene Often is requirement to retrieve scenes including a specific region or place Low-res GIF images (like examples shown) are available for last 24 months only Polygons can be calculated for each scene, but currently are not stored in any spatially searchable system.
Options for spatial indexing - 1 Bounding rectangles Mostly simple to generate/ store/ query Frequently a poor fit to the actual dataset extent Potentially become massively over-representative as data approaches the pole Do not cope well with oblique or curved boundaries to the dataset footprint (common in satellite images) Poor discrimination for fine- scale queries (many false positives returned) Can be problems with footprints which cross the date line, or include a pole Scene footprint bounding rectangle (follows lat, long grid lines)
Options for spatial indexing - 2 Bounding polygons OK to generate/ store Best fit to the actual dataset extent Need additional points as data approaches the pole (regions of high curvature) Computationally expensive to query (often use bounding rectangles as well, as pre-filter) Tend to require a dedicated GIS system to handle the spatial queries. bounding polygon inflection points
Options for spatial indexing - 3 Raster (tiled) approach Simple to generate/ store (as list of tiles intersected) Reasonable fit to the actual dataset extent; improves with decreasing unit square size Automatically increases resolution as data approaches the pole (regions of high curvature) Simple/ rapid to query, no computation or dedicated GIS required No ambiguity with footprints which extend across date line, or include a pole. * * * * * * * * * * * * * * * * * * * * * * * * * * * boundary of all tiles for this scene
C-squares basics Based on a tiled (gridded) representation of the earths surface, at choice of 10 x 10, 5 x 5, 1 x 1, 0.5 x 0.5 degree tiles, etc. (0.5 x 0.5 degree tiles are used in this example) Every tile has a unique, hierarchical alphanumeric ID (incorporates the ID of all parent, grandparent, etc. in every child tile ID) Dataset (=scene) extents are represented by a list of all the tiles included in, or intersected by the footprint Spatial search comprises looking for one or more tile IDs in the set associated with any dataset (= simple text search). (more details – see www.marine.csiro.au/csquares/ )www.marine.csiro.au/csquares
Spatial queries supported Retrieve all scenes which include all or part of a given tile (at 10 x 10, 5 x 5, 1 x 1, or 0.5 x 0.5 degrees), or set of tiles –optionally, filter also by other criteria e.g. date, satellite, etc. Retrieve all tiles associated with a given scene, with option to: –draw representation on a map (or range of maps) –export the list to a file or another database –compare the list with other lists (~ = polygon overlays) NB, choice of initial tile size is important: - too few tiles, only coarse spatial queries supported; - too many tiles, indexes get very large (and queries potentially slow) However, compression of blocks of contiguous tiles can be quite effective (1 code can replace 4, 25, 100, or 400 codes in certain cases).
A real-world example NOAA-12.. 21 Jun 2003 06:23 10 x 10 degree squares (28) (base level of hierarchy, cannot compress) 5 x 5 degree squares (99) = 36 after comp. 1 x 1 degree squares (1,982) = 515 after comp. 0.5 x 0.5 degree squares (7,691) = 704 after comp. 0.5 x 0.5 degree squares - detail
Practicalities for satellite data A single scene may measure (say) 40 x 50 degrees approx., = 2000 1 x 1 degree squares, or 8000 0.5 x 0.5 degree squares Quadtree-like compression reduces this to (e.g.) 500 codes at 1 x 1 degree resolution, 700 codes at 0.5 x 0.5 degree resolution Still require quite a lot of codes (e.g. 42 million) to represent a collection of 60,000 scenes Each code is 10 characters long, scene IDs are (say) 6 characters long, thus c.670 million bytes required for raw index data (before compiling secondary indexes, etc.) With secondary indexing, probably need around 2 Gb (+) to hold the spatial index.
Designing the index Option 1: a table of scenes, with string of codes comprising each scene footprint, 1 row per scene –comment: fast to retrieve all codes for a scene, slow for a spatial query. List of codes may present storage problems (e.g. in database field/s) owing to length. Option 2: a table of c-square codes, with list of scenes containing any code, 1 row per code (= an inverted index) –comment: fast for a spatial search, slow to retrieve all codes for a scene. List of scenes may present storage problems (e.g. in database field/s) owing to length. Harder to manage with respect to scene addition/deletion etc. (continues…)
Designing the index (contd) Option 3: a table of scene/c-square pairs, 1 row per pair, indexed on both columns –comment: OK to retrieve all codes for a scene, also for a spatial query. Row length is always constant, new rows simply created as required. However, table potentially becomes very long (e.g. 40 million rows), may become slow to query … Option 4: as option (3), however single table replaced by multiple smaller tables – e.g. split by year, blocks of squares, or satellite –comment: will tend to be tuned to favour queries which can be completed by accessing smallest number of tables. If queries are assumed to be by year (for example) in first instance, then split by year is reasonable (replace 1 table with 17 in this instance).
Present philosophy … Currently going with Option 3: a table of scenes/c- square pairs, 1 row per pair, indexed on both columns. However, will probably split this large table into a number of smaller ones (= Option 4) when loading is complete, to minimise time required for index rebuilds after new data are added (small indexes rebuild faster than larger ones).
Present table structure – 2 search tables only Scene details (1 row per scene) Scene/c-square pairs (600-700 rows per scene)
SQL for spatial search (example for 0.5 degree search square) select distinct A.scene_id, B.satellite, B.scene_date_time, B.image_location from satdata.satdata_csq_all A, satdata.scene_info B where ( (sqrsize = 0.5 and (A.csquare = search_csq -- e.g. 3414:100:1 (0.5 degree square) or A.csquare = substr(search_csq,1,8)|| ': * ' -- 1 level of compression or A.csquare = substr(search_csq,1,6)|| ' ** : * ' -- 2 levels of compression or A.csquare = substr(search_csq,1,4)||': *** : * ') -- 3 levels of compression ) -- (plus other supported search square size options go here) ) and (startdate is null or B.scene_date_time >= startdate) and (enddate is null or B.scene_date_time <= enddate) and (sat = 'any' or B.satellite = sat) and A.scene_id = B.scene_id order by B.scene_date_time, B.satellite;
C-square encoding algorithm First generation conversion code constructed January- May 2003, at CMR Hobart Version 2 constructed August 2003 - runs approx. 7 times faster, however minor compromises on ultimate accuracy (a few fine slivers may not be coded) Currently runs in Oracle PL/SQL, including compression of the c-square string from each scene Includes treatment for scenes crossing the date line as standard (scenes crossing the pole require a different algorithm)
Next steps … Index performance (i.e., retrieval times) seems fine as database size increases, however time to rebuild indexes slows down (required whenever more scenes are added). Probably would be best to split into a set of smaller tables, as previously discussed, to minimise time spent on index rebuilds. Need to assess how useful this approach is, cf. systems already in use elsewhere – and whether of interest to others?? Could investigate improving the encoding procedure speed further (currently, processing a large archive still requires several weeks of solid computation time).
Automated scene processing and upload 1.An automated routine prepares 2 sets of files: scene information, and scene polygons, in batches of 100-300 scenes 2.For the first batch, a script initiates file upload to the database, then file conversion (c-square generation) for each scene in the batch (initially processed 15-20 scenes/hr, revised code processes c.100 scenes/hr) 3.When all scenes have been converted, the c-squares are added to the master searchable table, and the scenes are flagged as upload completed 4.Close to the estimated completion time, a robot wakes up and checks at 10 minute intervals to detect when the job is complete (via a http call to the same procedure which creates the web scene count on the opening screen). 5.When the scene count returns 0 scenes currently being uploaded, the robot calls the script to start again from stage (2)-(5), with the next batch of scenes.
Your consent to our cookies if you continue to use this website.