Friday, August 12, 2016
Generally state agencies, some regional agencies, and for a fee, a plethora of data vendors.
The first step is aggregation, which is the process of taking many locally developed parcel data sets and standardizing each to a common format and in some rare cases providing quality control and/or spatially reconciling at the boundaries. A national standard for parcel attributes exists, (http://nationalcad.org/CadStandards/CadStand.html) but as with locally produced data, each state has its own standards to meet its state business needs. A recent review of state parcel standards found over 20 states had developed state parcel publication standards with many common attributes but no commonality of field names, types, or lengths. In all states reviewed the state aggregated data did have data definitions and was easier to understand and interpret than individual producer data sets.
The tools used to build aggregated data sets range from brute force to Safe Software’s FME, Esri’s Community Parcel tools, and customized state specific tools. Most states are either using or moving to web based processing for aggregation.
Update frequency is typically annually, some are twice a year, and a few are daily or continuous updates.
Some of the nuances and challenges for data aggregation are described in this article (http://www.esri.com/esri-news/arcnews/winter16articles/making-local-parcel-data-open-at-state-national-levels).
Distribution has typically been zip file download and a web based viewer. Files may be a single statewide file but often is individual files for each data provider, such as each town or each county. A noticeable trend for aggregated and distributed parcel data is the use of feature services. Many national parcel data in federal agencies require a data download to incorporate information into agency systems, but even federal agency applications are increasingly using feature services.
Paul Ramsey presented an intriguing twist to data aggregation and distribution in 2015. (http://s3.cleverelephant.ca.s3.amazonaws.com/2015-ccog.pdf). Mr. Ramsey discusses relevancy in terms of frequency of use. If data are not used it is less relevant than data that are used. Let’s go with the parcel data are important and has many uses, and the parcel data must be relevant. It must be available to be used and recognized as a useable source to be relevant.
The Moment of Opportunity, as described by Mr. Ramsey, is that small window when data (parcel data) can be provided in a way that developers can easily harvest and embed it in applications that can be seen and used by many on mobile devices. An interesting implication is that data needs to be distributed in a way that the data can be used and accessed by developers, rather than focusing on end user consumption.
This is an interesting perspective and important to consider. As Mr. Ramsey states “it just means that governments need to accept the way that the technology ecosystem is going to want to consume their data, and change their behavior to fit. The first step is to recommit to the idea of data as a public good. If this data (parcel data) is critical infrastructure, as we believe it to be, making it available to all members of civil society, without restriction, is a basic requirement. … Commit to simplicity in distribution. Follow the lead of NASA and publish raw data, with computer readable manifests, with stable URLs, close to the point of consumption on public cloud infrastructure”
Who distributes parcel data? Generally data aggregators, but we should all keep an eye on distributing our data in ways that will keep it relevant.