Finding geodata can be frustrating. There is no Google for publicly available geodata, and even within our own lab data exists almost exclusively on individual workstations, fallow and unsearchable, like oil tragically buried under a wildlife refuge. However, the GIIF has the funding, resources, and collective brainpower to alleviate this problem (uh, with the geodata, not the oil). Behold, my vision of the future: a centralized geodata provision service (or G-P-S, not to be confused with Gap, Inc. stock). I see this as a centralized repository for data relevant to us, stored in a standardized, possibly version-controlled manner that enforces metadata creation, and searchable through a usable web interface that allows queries on metadata and spatial queries (i.e. show me all the transportaiton data in Alameda County created after 2003). Here are my thoughts. Does anyone else think this is a good idea or should I seek therapy? What other features would you like to see? Does this software already exist? Please comment!
Geodata Server
Who it would serve
- GIIF users
- CNR
- UC Berkeley
- general public
- desktop users
- GIS application developers (including webGIS projects)
Note this all implies some kind of access control
What it would serve
- our own data (VTM, SOD models, fire models, wetland maps, etc.)
- external data (TIGER, FRAP, CaSIL, US Census, anything that allows redistribution
- spatially indexed links to external data where redistribution is prohibited
- symbologies (recommended symbologies and/or visualizations for complex data, i.e. Mapserver layer definitions and ESRI .lyr files)
Geodata Portal Application
How it would serve data
How it would store data
- extents stored in a geodatabase for retrieval via spatial query (box selection, search by distance, etc)
- metadata stored in a database for retrieval by metadata query (by date, by source, by type, etc)
- actual data maintained in original format (versioning? how do we deal with data in external geodatabases?)
How users would interact with the data
- search for data via spatial and non-spatial queries (see above)
- automatically generated previews (see what you're getting)
- AJAX development techniques used whenever appropriate
- submission: users should be able to submit new data, or modified versions of existing data
- enforced metadata: all data subject must have certain kinds of metadata before being accepted (date created, source, accuracy. We can derive a lot automatically, like extent, file type, projection, file size, etc)
Other ideas
Possible models/inspirations
Possible software tools/components
Note extreme open-source bias... Definitely open to other suggestions.