Parcel Data Onboarding and FAQ

Regrid Parcel Data Onboarding Documentation and Frequently Asked Questions

Where does your county data come from?

We source our data directly from each county or whom they designate as the official source for their parcel data.

Some counties are missing from your data set but are labeled as “Available”, what does that mean?

Available county data means that we have confirmed the data is available from the county, but for some reason we did not obtain the data. The usual reason is cost from the county, some counties price their parcel data very high. If you are interested in any missing county, please let us know and we will update our research to confirm costs or other changes and work something out with you to obtain and standardize that data.

How do you standardize county data generally?

The main way we make county data much easier to work with is by standardizing the column names of the raw data provided by each county. We do not standardize the values in most columns, most we keep those exactly as provided by the county, but we do make sure that every county in our system is converted to a standard table schema, with consistent column names across the nationwide dataset. Please see “What is the Regrid Parcel Schema?” question below.

In addition, we further standardize the parcel address fields using the US Postal Service database of addresses. For details on address specific standardization please see the “How do you standardize and normalize addresses?” question below.

How do you deliver bulk data?

All bulk data is provided via SFTP as zip files of each county in the format of your choice (GeoJSON, NDGeoJSON, SQL, Shapefile, FileGDB, GeoPackage, KML, CSV), using a pull model. We organize things on a county by county basis using the county’s FIPS code (geoid in our Regrid Parcel Schema column).

What is the Regrid Parcel Schema?

We standardize column names for easy access across counties in our nationwide dataset in to our Regrid Parcel Schema for tables.

We are currently at version 6 of our Regrid Parcel Schema and it standardizes the column names for approximately 80 county provided data columns, and 25 Regrid provided data columns. This schema is applied to 100% of our dataset. A data dictionary is available:

Why do your parcel value fields (‘parval’) not look the way I expect?

Our parcel value related fields are all directly from the county assessor’s data. We populate as directly from the assessor attributes as possible our ‘improvements value’ (improvval), ‘land value’ (landval), ‘parcel value’ (parval), ‘ag value’ (agval) and ‘parcel valuation method’ (parvaltype) attributes. However, while those are the most common value related attributes, every county has their own definition for those attributes and their own methods for how they calculate, record and display amounts for tax purposes. We can not answer questions about why the county records the values in those attributes. We suggest visting the county’s website or calling the assessor’s office directly to better understand those values. If, after contacting the county it appears we have an error in what values we have in those attribute fields, please send an email to and we will review our data asap.

Why do your parcel numbers (apn, pin, etc) not look the way I expect or are null?

County assessor parcel identification numbers (‘parcelnumb’) are well known for being complicated and often have variable punctuation or zeros (0) that can affect searching or matching by parcel id number. Counties do occasionally change their method for generating or assigning parcel numbers and that can lead to “new” and “old” parcel number situations. We always retain any identification number attributes as ‘County Custom Columns’ so it should be possible to match up our data with county data directly, even if our parcel number field is not the only identification number used by the county.

Also, sometimes a State GIS source will add their own unique id to a local parcel id number.

We suggest visiting the county’s website or calling the assessor’s office directly to better understand their parcel numbering system. If, after contacting the county it appears we have an error in what values we have in our parcel id attribute field, please send an email to and we will review our data asap.

Null and duplicated parcel numbers are common across the US. Counties do it for a variety of reasons, but the most common reason is they do not bother with parcel numbers for non taxable parcels like city parks, rights of way, etc. However, condos, timeshares, mineral or air rights, unassigned rights etc, all might use duplicated or null parcel numbers and the county uses some other attribute, combination of attributes or system for tracking them.

Do you have a specific attribute for a specific county?

A current, detailed list of every county in our data set and what data fields we have for each is always available at the following URL. This spreadsheet can be downloaded as a CSV for closer analysis:

Why do my files have columns not in the Regrid Parcel Schema?

Every county has our schema columns, so any code or process can rely on those columns (also called attributes in GIS software) being present. However, most counties also provide attributes that do not map in to one of our Regrid Parcel Schema columns, so we keep those attributes and include them on the end of our schema columns using whatever name is provided by the county. In many tools you can control what attributes are retained during import or merging, so we suggest folks working across multiple counties just keep the schema columns or a sub set of them and leave what we call “custom columns” off entirely to get a uniform set of columns. Going back and reviewing those custom columns can reveal interesting information provided by the counties, but not provided very often in other counties.

How can I explore the custom columns for each county?

We work with all of our counties in a PostgreSQL database, each county in its own table. That makes managing the custom columns from each county much easier. Most database servers provide a way to search the column names of the tables in a database. For example, in Postgres you would do it this way:

SELECT table_name, column_name FROM information_schema.columns 
  WHERE table_schema = public and 
        column_name ~ juri
order BY table_name, column_name;

Also, directly browsing the data on a place-by-place process of areas or regions you are interested in can be very useful. DBeaver is a cross platform, multi vendor database client that can render geographic data.

Why do Shapefile attribute names not match the Regrid Parcel Schema column names?

Some of our Regrid Parcel Schema column names are longer than the Esri Shapefile format allows and the column names in your attribute table will be truncated to the first 10 characters of the Regrid Parcel Schema column names.

When was your data last updated?

On average 94% of our parcels have been refreshed in the last 12 months, with most of those in the last 6 months. We work on a rolling update schedule, refreshing county data directly for 100 - 300 counties per month, usually grouped by state.

Monthly we share, both in machine readable format and via a monthly update email, what counties have been updated, and what states are in the pipeline for the upcoming quarter. An example of our Monthly Data Update Email is available.

Our USPS related attributes are updated monthly for our entire data set.

All data is tracked with the date of “last refresh” from the county.

A detailed listing of every county in our data set and the date we last refreshed directly from the county is always available. This spreadsheet can be downloaded as a CSV for closer analysis:

How do you provide data updates?

Monthly, we update every data tier with the refreshed counties for the month, usually 100 - 300 counties per month.

Monthly, premium attribute data tiers are updated for every county in the data set, because every county’s USPS related attributes have been updated for the month.

Quarterly, at the start of the calendar quarter, every data tier will have every county exported. We make improvements to the nationwide data set on a regular basis, often based on feedback from clients. Quarterly, full data set exports ensure every data tier has all of the improvements across the data set, even if the data was not refreshed directly from the county that quarter. We are working to provide an easily sharable changelog for data set wide fixes, but currently give a notice to significant updates in our monthly data update emails.

Each month we replace the existing county zip file in a client’s download directory with the refreshed county files. We do not provide ‘diff’ files or deltas for the data at this time.

All bulk data is provided via SFTP as zip files of each county in the format of your choice (GeoJSON, NDGeoJSON, SQL, Shapefile, FileGDB, GeoPackage, KML, CSV), using a pull model. We organize things on a county by county basis using the county’s FIPS code (geoid in our Regrid Parcel Schema column). We add or refresh 100 - 300 counties every month with data from the county source.

We send out an email each month listing the counties that were refreshed, as well as an updated listing of all the counties and their last_refresh date, which is the last date we updated the data directly from the source.

We also provide a CSV file of our VERSE table that lists the last_refresh date for each county.

On each parcel we provide a ll_uuid number that permanently and uniquely identifies each parcel across data refreshes and updates. The ll_uuid can be used to match any locally stored parcels with updated parcels in the county file.

ll_uuid is tied to the county provided parcel numbers. Parcel revisions like splits and other modifications that result in new parcel numbers, will get new ll_uuid’s. When the county retires a parcel number, we also retire our ll_uuid for that parcel.

We also make improvements to the data that does not come directly from the county, like standardizing addresses, cleaning out non-standard characters, adding additional data attributes, etc. We generally re-export our full dataset quarterly to reflect those changes. Improvements we make to data does NOT affect the last_refresh of a county, that is always the date we last pulled data directly from the county source. Notices of full dataset exports are also included in the monthly email update sent to all clients.

How do I keep my data up-to-date?

We provide our verse tracking table as a csv updated monthly. It has every county in the US’s state, county name, state+county fips (we call geoid), date we last pulled directly from the source (last_refresh column), and a filename_stem column that indicates the file name with no format .extension or .zip, just the basename of the file. Note: null in our verse table_name column means the county is not in our dataset.

Everyone’s software stack and internal systems’ environment are different, but generally we think the outline of steps is as follows:

  1. Pull a copy of our verse table from our SFTP server
  2. Use the last_refresh date in our verse table to determine which county or counties you need to update
  3. Use the filename_stem field in our verse table to pull the needed county or counties via SFTP from our server
  4. Once you have the file(s) locally, import it/them into a database (we use PostgreSQL/PostGIS, GDAL and ogr2ogr for this last step internally)

We do not track changes at any level below the county level, but with our ll_uuid it should be possible to determine if you need to update a row in your database if you do not want to replace the whole county.

That same ll_uuid should be used if you need a permanent, nationwide unique id for a parcel, and to match up with data you might attach to our parcel data in your usage.

How do you standardize and normalize addresses?

The county-provided parcel address data is transformed to populate the typical address fields in use, address line 1, address line 2, city, state, zip. Basic text cleaning is done at that same time to remove special or non-printing characters.

Where counties have not provided full addresses for parcels, we use US Census data to fill in missing zip codes and cities.

Parcel addresses are then passed through a USPS address verification system which further normalizes and standardizes the address to USPS guidelines where a match was made against the USPS database of addresses. Non matching address are left as they were sent by the county.

The original county-provided raw address data is always retained unaltered and included on the parcel record.

What software can I use to work with your data?

Editing or working with most of our data requires software for working with geographic and geospatial data The OSGeo project provides free and open source desktop software to work this kind of data called QGIS.

There are other free and paid software options for working with geospatial data, but all of it has a learning curve. We are not able to provide support for those 3rd party softare applications. For QGIS at least, the best options for learning are via the many community developed tutorial videos and texts.

We suggest starting with the GeoPackage (geoPKG, .gpkg) formatted files to use in QGIS or any geospatial software.

What about Google Earth?

We provide KML/KMZ options for Google Earth and Google Earth Pro, but neither of those applications support editing our data, only viewing the data. If you need to make changes to the data you get from us, you will need a desktop application like QGIS discussed above.

How large is the nationwide dataset?

The nationwide dataset is approximately 400-800 GB uncompressed, varying by file format, storage method, attribute tier, and other factors.

How do you load all of these files into a database?

We generally work in a GNU/Linux environment using the command line. Our internal workflow makes use of the OSGeo Foundation libraries and tools including GDAL/OGR and PostGIS for PostgreSQL. The OSGeo project also provides an MS Windows 10 installer for using the tools on a Windows machine named OSGeo4Win.

The ogr2ogr command line tool is the best way to import data into a PostgreSQL/PostGIS or MS SQL Server database.

Below is a typical command line to cycle through a set of GeoDB files and append them to a table in PostgreSQL:

for gdb in /path_to_files/*.gdb ; 
do echo "$gdb" ; 
  ogr2ogr -f "PostgreSQL" PG:"dbname=dbname" $gdb -progress --config PG_USE_COPY YES -nln public_schema.target_table_name -append ; 

Dealing with custom columns at scale is best accomplished by ignoring them during import. One approach is to use the -sql option with ogr2ogr to restrict what columns you load to all of, or a sub-set of, the columns in our Regrid Parcel Schema. We use this same method for exporting and it causes one thing to watch out for: The layer name to use in the sql, where a table name would normally go, is literally the exact string sql_statement

ogr2ogr -f 'PostgreSQL' PG:'dbname=dbname' st_county.gpkg -sql 'select wkb_geometry,geoid,etc from sql_statement' -nln st_county_imported

Using ogr2ogr to load parcel data into a MS SQL Server works the same way. Parcel data should use the geometry data type in MS SQL Server. A good example of how to do that is this blog post by Alastair Aitchison. They also cover installing the OSGeo4Win environment.

An example osgeo4w shell command to load a folder full of geopackages into MS SQL Server looks like this:

for /R %G in (*.gpkg) do ogr2ogr -progress -f "MSSQLSpatial" "MSSQL:server=localhost;database=alabama;trusted_connection=yes" %G -nln "%~nG"

The main item of the command line options are the database connection options. You will have to make sure the user name and password are available and that the client can actually connect to the database and has all the needed permissions. For PostgreSQL on GNU/Linux, there are standard PG_* environmental variables and the .pgpass file for storing credentials that will work with the ogr2ogr commands so they do not have to be included in the command line.

Why do some Shapefiles say ‘2GB_WARN’ in the file name?

The shapefile format itself has a ‘soft limit’ of 2 gigabytes (GB) of data. ‘Soft limit’ means it is just a rule of the format: “no data larger than 2GB”. There is no technical limit preventing more data being encoded as a shapefile. When we export large counties, our tools inform us the resulting shapefile is over that soft limit.

We can confirm that some software handles 2GB and larger shapefiles just fine (OSGeo tools like QGIS), but some software will just silently ignore attribute data above the 2GB limit (ArcGIS). A sincere thank you to the Regrid client who did a lot of in depth research and testing on this and shared their results with all of us.

Starting in July 2020, we made the following changes to help flag the counties where the data exceeds the 2GB soft limit. Please double check how you are handling these files.

  1. The filenames themselves will indicate the county generated the 2GB warning on export. A _2GB_WARN suffix is added to the file names so you can know just by checking the name.
  2. We added a column to our verse table, named shapefile_size_flag so you can check if a place needs a different format than Shapefile, or to generate a list of places you need to pull the alternate format for.
  3. Shapefiles that are larger than 2GB are not available through our data store.

We recommend the GeoPackage (GeoPKG) exports for situations where you would normally use a Shapefile. These can be opened by both Open Source and ArcGIS tools. ArcGIS uses ESRI’s “Data Interoperability” extension to work with GeoPackages.

What about multi unit parcels or parcels with secondary addresses?

Our data packages do not contain the list of every secondary unit on a parcel. If you need the full secondary address list per parcel, please contact us at

Our data attribute packages have the main, primary physical address for a parcel and the mailing address of record for the parcels. These do vary somewhat based on what the county provides, so please review our “Attribute Completeness Report”:

Regrid Attribute Completeness Report

What about condominium or cooperative buildings or parcels?

Why are some parcels duplicated or stacked parcels?

Our data is directly from county GIS and county assessor’s offices around the US. They are primarily focused on collecting taxes so recording ownership and mailing addresses is their main goal. They often use their software in creative ways to get things recorded so they can track the taxes. For parcels with multiple, individual owners the counties often ‘stack’ parcels so they can enter different owner and contact information for each owner. The 3 most common ways are:

  1. Identical polygons stacked, exactly the same size parcel, just 4 or 10 or 100 stacked, all exactly on top of each other, with different attributes for the different owners. By far this is the most common way and is widespread around the US.

  2. Puzzle pieces, ground parcels with exact cutouts for the footprint of the building. This is common for downtown buildings. They have no intentional overlap.

  3. Laying condo parcels on top of a ground parcel. This is a polygon the exact size of the building, but instead of ‘cut out’ of the ground parcels, it just lays on top of the ground parcel, like a stacked solution, but only 2 layers: one big ground parcel, with smaller parcels stacked on the ground parcel, but spread out.

In these cases, our dataset usually does contain all of the addresses associated with the parcel, as each owner’s ‘parcel’ usually has what is considered a primary address.

The vast majority of the counties create unique parcel numbers for each stacked parcel. The unique parcel number is a benefit of the stacked parcel approcach. However, some counties will duplicate the parcel number and use a secondary id field for the sub-parcels. We think this is much more rare, and we would retain any secondary parcel numbers as a custom column in our data.

Building Data

How do you determine the Regrid Building Count value (ll_bldg_count)?

We work with a 3rd pary data firm that collects and curates building footprint data sets directly from counties in the US. They provide one of the most comprehensive, authoritative building footprint nationwide layers available.

We then process that footprint data set with our parcel shapes data set to determine how many buildings are on the parcel.

Combining two large data sets, each of which is a combination of a few thousand smaller data sets, to derive a value is always a challenge and we do our best to be as accurate as possible, but some things to keep in mind when working with the building count data:

  1. Each county determines what buildings to track in their GIS system. These are considered by the county to be the significant structures on the parcel. Some counties will include every structure regardless of size, some will only track houses and garages for example.
  2. Much like county parcel data, counties refresh footprint data on their own individual schedules, and less frequently than parcel data in our experience. We do not yet have a way to track the ‘last_refresh’ date of the building footprint data by county.

Please contact us at if you find issues, or have any questions, related to buildings.

How do you calculate the parcel acres, parcel square feet, and building square feet (ll_gisacre, ll_gissqft, ll_bldg_footprint_sqft)?

We projected each parcel and building footprint into its UTM Zone SRS, calculated the area in meters and converted that to acres and sqft. This should provide a relatively uniform and consistent value across the US.

API Access

What is an API and why might I use one?

API stands for Application Programming Interface, and is for use by software developers to interact directly with our national dataset. They are usually used when someone wants to automate some process that interacts with our national parcel dataset. All APIs require programming work to make them useful for any specific application.

What APIs does Regrid make available?

We provide two different APIs for working with our data. Please remember, all APIs are intended for use by software developers and are not for end users’ direct use.

Tile Map Service (TMS) Layer - This is an interactive, vector and raster rendering of our dataset, for use in web or desktop applications where using the parcel shapes overlaid on other layers is useful. Regrid’s TMS layer is in Mapbox Vector Tile (.mvt) format.

RESTful Parcel API - This is a typical stateless, client/server based API that supports retrieving parcel shape and meta data using lat / lon coordinates, among other options. Please see for a complete reference to the API.


How do I download Regrid Parcel data?

We use the “Secure File Transfer Protocol”, also called SFTP. This is supported in most traditional FTP clients and SSH client software.

Both of the clients listed below are multi protocol and also support connecting to services like S3.

Windows users can use WinSCP, which also supports scripting and .NET/PowerShell integration.

MacOS users can use CyberDuck

Regrid Standardized Land Use Codes - Classification

Please visit our Regrid Standardized Land Use Codes specific documentation.