Introduction

Two important considerations in real estate assessment are the accuracy and equity of property assessments. These considerations can be measured in terms of the Coefficient of Dispersion (COD) and the Price-Related Differential (PRD), as described by the International Association of Assessing Officers (IAAO). “Uniformity” is measured with the COD (IAAO, 2013), while “vertical equity” is measured with the PRD (IAAO, 2003). From a statistical perspective, the percent absolute prediction error (PAPE) is a common measure of predictive accuracy.

There have been many studies in recent decades that consider spatial aspects to more uniformly and equitably estimate assessed values. Among others, these include the nonparametric and/or semi-parametric approaches of Geographically Weighted Regressions (GWR), also commonly referred to as Locally Weighted Regressions (LWR). The LWR methods were developed in Cleveland and Devlin (1988) and later used in urban economics by Meese and Wallace (1991) and McMillen (1996). Among a class of Automated Valuation Methods (AVM), including least squares regressions and fixed effects regression estimations, GWR has been shown to be a favorable alternative in general by Bidanset and Lombard (2014a, b), Borst and McCluskey (2008) and McCluskey et al. (2013), among others. Another set of semi-parametric approaches, known as local polynomial regressions as in Cohen et al. (2017), is intended to address the separate valuation of land and improvements; however, its computational intensity has deterred widespread use in practice.

But little attention has focused on incorporating the time dimension into GWR or LWR models in the context of AVM, such that when assessing properties, the importance of comparables will depend on exactly how long ago they sold. Bringing this time element into the GWR model for the purposes of estimating assessed values and examining the effects on uniformity and vertical equity is a critical contribution of our research. The goal of our paper is to provide the foundations of Time GWR as a method and then demonstrate its potential value for assessors.

Another advantage of our study is that we apply this new methodology to two extensive data sets. We have sales data for all arms-length transactions of residential properties in a sample of 50 municipalities in Connecticut for 1994–2013. We also have transactions of single-family homes for 145 towns in the Greater Boston Area (GBA) for 1987–2012.

An outline of the remainder of the paper is as follows. First, we provide a brief literature review of GWR and then present additional background on the GWR methodology. Then, we explain how we propose to modify the GWR framework to incorporate time variation, which we call Time-Geographically Weighted Regressions (TGWR). We also describe the approach to forecast the “estimated sales price” that is to be used in computing the COD, PRD and PAPE. Next, we describe the data we use for property sales in municipalities in Connecticut (CT) and Massachusetts (MA), followed by the analysis of these data using our TGWR approach. As a part of this analysis, we compare the estimates of COD, PRD and PAPE from using OLS/FE, GWR and TGWR to demonstrate which approach is more appropriate for use in property assessment. Finally, we conclude with a summary of the main findings and some suggestions for future research.

Methodology of GWR and TGWR in AVM

GWR’s popularization in urban economics can be traced back to Meese and Wallace (1991), and software has been developed to estimate GWR models in Stata (see https://www.staff.ncl.ac.uk/m.s.pearce/stbgwr.htm) and more recently there is software for LWR in R (see https://sites.google.com/site/danielpmcmillen/r-packages). The basic idea behind GWR (and LWR) is that it allows for a general functional form between the dependent variable in a regression and the explanatory variables. Such a nonparametric approach has been explored in other contexts, commonly known as kernel regression, but the case of GWR utilizes a specific form of the kernel weights based on geographic distances between a given target observation and the other observations in the sample. While McMillen and Redfearn (2010) note that the choice of the functional form of the kernel tends to have little impact on the GWR estimates, a more complicated consideration is the bandwidth parameter. There are various approaches to bandwidth selection, including the Silverman (1986) “Rule of Thumb” bandwidth, and other approaches such as cross-validation. McMillen and Redfearn (2010) and Cohen et al. (2017) provide GWR estimation results for two separate bandwidth choices. We use one bandwidth for the main analysis and then provide some robustness checks using other bandwidths.

First, consider a linear relationship between X (a matrix of observations on property characteristics and neighborhood amenities) and Y (a vector of observations on property sales price), so that:

$$ {\displaystyle \begin{array}{l}\mathrm{Y}=\mathrm{X}\upbeta +\mathrm{u},\\ {}\mathrm{u}=\mathrm{f}+\upvarepsilon, \\ {}\upvarepsilon \sim \mathrm{N}\left(0,\Omega \right).\end{array}} $$
(1)

Typically, Y is specified as the log of sales price. It is straightforward to estimate the vector of parameters β by OLS (where f denotes fixed effects), invoking the classical assumptions on Ω.

But if one believes that a linear relationship between property values and characteristics may be too restrictive, one could specify the following general (nonparametric) relationship:

$$ \mathrm{Y}=\mathrm{h}\left(\mathrm{X}\right)+\upvarepsilon, $$
(2)

where h is some unknown function. We can estimate the marginal effects of X on Y at any point in the sales dataset, by using GWR.

More specifically, GWR can be thought of as a form of weighted least squares (McMillen and Redfearn 2010). To estimate Eq. (2) using GWR we run a weighted least squares regression of wijYi against wijXi, where Xi = [x1i, x2i, …, xmi], m is the number of explanatory variables in the estimation, wij = [kij(•)]1/2, and kij(•) is a set of kernel weights. A common choice of kernel weights is the Gaussian form, where kij = (exp(−(dij/b)2) and dij is the geographical distance between observations i and j, and b is the bandwidth parameter.

The GWR approach can generate separate estimates of βi for each target point, i. GWR allows for properties that are further away from a subject property to be given less weight in the estimation, opposed to an arithmetic mean as in the OLS/FE case.

One critical assumption of GWR, however, is that all properties are sold at the same time, or, at the very least, that the time of sale is not a crucial factor to consider in the kernel weights. One might address this by including time fixed effects in Eq. (1). However, as an alternative approach, we propose a methodology similar to that suggested by Cohen et al. (2014), where the kernel weights function incorporates the date of sale of nearby properties. In such a case, the kernel weights might take the following form:

$$ {\mathrm{k}}_{\mathrm{ij}\mathrm{t}}=\exp \left(-{\left(\frac{{\mathrm{d}}_{\mathrm{ij}}}{\mathrm{b}}\right)}^2\right)\cdot \exp \left(-{\left(\frac{\tau_{\mathrm{ij}}}{\mathrm{b}}\right)}^2\right) $$
(3)

In this version of the kernel weights, τij represents the difference in the time between the sale of property i and property j (perhaps in terms of the number of days). More formally, if we order the dates of sale on a line so that the earliest property in our sample had a sale date, t1 = 1, and suppose the second property in our sample had a sales date 5 days later so that t2 = 6, etc., then we define:

$$ {\displaystyle \begin{array}{l}{\tau}_{ij}={t}_i-{t}_j, if\ {t}_i\ge {t}_j,\\ {}\kern0.95em =0\ \mathrm{otherwise}.\end{array}} $$
(4)

In other words, when determining the kernel for each target point, we only consider sales that occurred prior to the sale date of the target point, i. We define the TGWR estimator as a generalization of GWR where the kernel weights are given as kijt in Eq. (3).

We think the TGWR is an important improvement over GWR since this increases the number of comparable sales that can be used to generate assessed values. Furthermore, it is likely that units that sold at an earlier time but are spatially near the designated unit are likely to be closer in value than units that are sold at a similar time but farther away geographically. This is because location is such an important determinant of house value.

For the purposes of AVM, there are a number of recent studies that have demonstrated the superiority of GWR over OLS, including Lockwood and Rossini (2011) and Borst and McCluskey (2008). The only known studies that use GWR to consider equity and uniformity of assessments are Bidanset and Lombard (2014a, b). However, neither of these two studies incorporate time differentials of sales into the GWR framework. Given the importance in considering recent sales in the appraisal process, we propose the following approach to AVM in a space-time context.

But before introducing our approach, we first discuss some common assessment practices in Connecticut and Massachusetts. In performing revaluations, our informal conversations with several assessors indicate they often estimate revaluations for property i by comparing the ratio of sales to previously assessed values of other nearby properties, j, multiplied by the most recent assessed value of property i. One would expect the importance of nearby properties (j) to diminish with distance from property i, although it is not evident how assessors choose the relevant distance. The date of sale of nearby properties should be an important factor as well, however this might be determined somewhat arbitrarily.

In addition to a property’s characteristics, there may be other factors that are important determinants of a property’s value, such as school quality, crime, the presence of jobs, and the population density of the town, even for nearby properties that straddle boundaries of different districts. In a regression context, these factors can be denoted as X, where X is a matrix consisting of m vectors of variables, where m is the total number of characteristics and other factors.

One potential set of approaches to estimate these effects that addresses these issues is as follows. First, for all properties, i, that recently sold (with sales price, Yi), run a fixed effects regression as in (1) and estimate β. The fixed effects proxy for neighborhood characteristics.

Second, consider the nonparametric GWR and TGWR models. We can estimate the marginal effects of X on Y at any point in the sales dataset, by using GWR and TGWR. This will give us separate estimates of βi for each i. Then, for these same properties that sold recently, use the estimates of βi to predict the values of the same properties using eq. (2), separately for GWR and TGWR.

Note that when applying fixed effects, GWR, and TGWR, we exclude future sales from the sample since this is what realistically would be carried out for assessment purposes. When using fixed effects for units that sold in year t, we include sales in year t and t-1. When using GWR, we include sales in the previous two years (based on the month of sale). When using TGWR, we only use previous sales that are weighted by the kernel (eq. 3) and limited by the bandwidth.

We follow the approach of Bidanset and Lombard (2014a, b) and calculate the COD and PRD using OLS/FE, GWR, and TGWR. The COD and PRD are measurement standards from the International Association of Assessing Officers ( 2003). The COD measures, on average, how far each property’s ratio of estimated to actual sales price differs from the mean ratio of estimated to actual sales price. It is expressed as a percentage of the median. Thus, a smaller COD indicates a more uniform assessment. Values less than 15% (but not less than 5%) are viewed as acceptable.

Meanwhile, the PRD is a measure of whether high-value properties are assessed at the same ratio to market value as low-value properties. The numerator of the PRD is the ratio of the simple mean of the estimated to actual sales price, while the denominator is the weighted mean of the sum of the estimated prices relative to the sum of the actual sales prices. Values less the 0.98 indicate that high-value properties tend to be “overappraised” relative to low-value properties. Values exceeding1.03 indicate that high-value properties tend to “underappraised” relative to low-value ones. The goal for PRD is to be in the “neutral” range, i.e., close to 1.0.

The formulas for the COD and PRD are as follows:

$$ \mathrm{COD}=\frac{100}{\mathrm{N}}\ \frac{\sum_{\mathrm{i}=1}^{\mathrm{N}}\mid \frac{{\mathrm{Estimated}\ \mathrm{Price}}_{\mathrm{i}}}{{\mathrm{Sales}\ \mathrm{price}}_{\mathrm{i}}}-\mathrm{Median}\left(\frac{{\mathrm{Estimated}\ \mathrm{Price}}_{\mathrm{i}}}{{\mathrm{Sales}\ \mathrm{price}}_{\mathrm{i}}}\right)\mid }{\mathrm{Median}\left(\frac{{\mathrm{Estimated}\ \mathrm{Price}}_{\mathrm{i}}}{{\mathrm{Sales}\ \mathrm{price}}_{\mathrm{i}}}\right)} $$
(5)
$$ \mathrm{PRD}=\frac{\mathrm{Mean}\left(\frac{{\mathrm{Estimated}\ \mathrm{Price}}_{\mathrm{i}}}{{\mathrm{Sales}\ \mathrm{price}}_{\mathrm{i}}}\right)}{\frac{\sum {\mathrm{Estimated}\ \mathrm{Price}}_{\mathrm{i}}}{\sum {\mathrm{Sales}\ \mathrm{price}}_{\mathrm{i}}}} $$
(6)

We calculate the COD and PRD separately for each of the towns in Connecticut and Massachusetts that are in our dataset. Finally, we also calculate the PAPE as a measure of prediction accuracy.Footnote 1

Note that we run the regressions (Eq. 1) using the log of sales price as the dependent variable. To convert the predicted values back to levels we use the transformation.

$$ \hat{\mathrm{Y}}={\hat{\upalpha}}_0\exp \left(\mathrm{l}\hat{\mathrm{n}}\mathrm{Y}\right) $$
(7)

where

$$ {\hat{\upalpha}}_0=\overline{\exp \left({\hat{\mathrm{u}}}_{\mathrm{i}}\right)} $$
(8)

and ^i is the residual from the corresponding regression (Wooldridge 2016).

Data

We use sales data for all arms-length transactions of residential properties in a sample of 50 municipalities in Connecticut for 1994–2013. These data come from the Warren Group. An attractive feature of these data is that they may enable us to determine how the OLS/FE and TGWR approaches perform differently for properties in different towns within the state.

To filter out non-arms-length sales, we omit properties with sales price under $50,000. We eliminate outliers on the high end by dropping a small number of observations with sales price over $7,000,000. In our sample, as shown in Table 1, the average sales price was approximately $387,500, with an average lot size of 0.68 acres, living area of 1880 square feet, approximately 7 rooms, 3.19 bathrooms, and an average age at sale date of 47.62 years. The oldest house was 118 years, and there were some houses that were 0 years old (i.e., new construction). Unfortunately we do not have information broken down about the number of half baths; instead, the total baths variable includes the sum of full plus half baths, so that a property with a value of 4 total baths may have 4 full baths or 3 full plus 2 half baths, for example. We have the sale date and latitude/longitude for each property. There are 258,473 properties in this CT dataset.

Table 1 House-level summary statistics

The 50 municipalities included in the Connecticut sample and the 145 towns included in the Massachusetts sample can be seen in Fig. 1. For the Connecticut sample, there are several urban cities (e.g., Hartford and New Haven), and many suburban locations (such as West Hartford, Greenwich, etc.). The towns in the Massachusetts sample are all located in the Greater Boston Area. We also use transactions of single-family homes in the Greater Boston Area (GBA) for 1987–2012. The data are from the Warren Group for 1987–1994 and CoreLogic for 1995–2012 and cover towns in Bristol, Essex, Middlesex, Norfolk, Plymouth, and Suffolk Counties. These data include the month and year of sale and the exact location (latitude and longitude). Observations with missing sales date or location information are dropped.

Fig. 1
figure 1

COD values – OLS

Sales that were not standard market transactions such as foreclosures, bankruptcies, land court sales, and intra-family sales are excluded. Further, for each year, observations with the bottom and top 1% sales prices are excluded to further guard against non-arms-length sales and transcription errors. The data include typical house characteristics: age, living space, lot size, the number of bathrooms, bedrooms, and total rooms. The sample is limited to units with at least one bedroom and bathroom, 3 total rooms and 500 square feet of living space and no more than 10 bedrooms and 10 bathrooms, 25 total rooms, and 8000 square feet of living space, or 10 acres.

The second transaction is excluded for properties that sold twice within 6 months (similar to Case/Shiller) and for properties with two sales in the same calendar year with the same transaction price (likely duplicate records). Properties for which consecutive transactions occurred in the same year or in consecutive years and where the transaction price changed (in absolute value) by more than 100% are excluded. Similarly, properties where consecutive transactions were in year t and t + j and where the transaction price changed (in absolute value) by more than j00% were excluded for j = 2,…,12.

Thirty-two towns in Massachusetts with less than 100 total observations are dropped and 36 census tracts with less than 10 observations are excluded leaving a total of 145 towns and 833 census tracts for a total of 609,123 observations. Descriptive statistics for the data are provided in Table 1.

Results

First, we present OLS results for the entire samples from Connecticut and Massachusetts, including year and town fixed effects in Table 2 to see how the results compare across the two states.Footnote 2,Footnote 3 The dependent variable is the log of sales price. We include independent variables for log of living area square feet, log of lot size, and categorical variables for the number of total rooms number of bedrooms, number of bathrooms and house age.Footnote 4 In our sample of Connecticut towns, there are 258,473 observations between 1994 and 2013 and the R2 is 0.844. In our sample of MA towns, there are 609,123 observations between 1987 and 2012 and the R2 is 0.700.

Table 2 Town fixed effects results-dependent variable: logprice

Generally, the results for Connecticut and Massachusetts are similar although there are some interesting differences that we highlight below. In both regressions, the living square feet, lotsize, and bathroom dummies coefficients are all positive and highly statistically significant and the number of bedrooms has little impact on price. But in Massachusetts, the age of the house is negatively related to price whereas houses age 10 to 30 are worth the most in Connecticut. Finally, house prices increase with the number of rooms in Massachusetts but only matter (positively) for the very largest houses in Connecticut.

Given that our goal is to compare different methods for property value assessment, we estimate hedonic regressions using fixed effects, GWR, and TGWR separately for each town in Connecticut and Massachusetts. Because assessors canl only use current and past sales to make their assessments, we limit all samples to sales in the current and prior years of sale. For fixed effects, we estimate a separate regression for each year using the current and previous year’s sales. For GWR, we estimate a regression for each sale using sales in the same town in current and previous year. For TGWR, the time dimension of the kernel is limited to sales in the current and previous years.

We use a bandwidth of 200 when applying GWR and TGWR. We offer robustness checks using different bandwidths below. We provide summary statistics for the results using these three estimators in Tables 3 and 4. The means of the point estimates for the different explanatory variables are similar. Not surprisingly, the standard deviations for the GWR and TGWR estimates are larger than for the fixed effects estimates since GWR and TGWR provide different point estimates for each sale.

Table 3 Summary statistics for coefficient estimates for town-Level OLS, GWR, and TGWR: CT data, b = 200
Table 4 Summary statistics for coefficient estimates for town-level OLS, GWR, and TGWR: MA data, b = 200

We provide three statistics for evaluating the effectiveness of the three approaches to property value assessment: PAPE, COD, and PRD.

Table 5 provides the summary statistics for the COD and PRD statistics for OLS, GWR, and TGWR First, we discuss the COD, which is a measure of accuracy. With the OLS estimates, all 50 towns in Connecticut have COD below the suggested upper bound of 15.0, while all but 24 in Massachusetts are in the acceptable range. This can be seen in Fig. 1. The situation improves in Massachusetts for GWR and TGWR, shown in Figs. 2 and 3, as all towns are below the cutoff for GWR and all towns but one are below the cutoff for TGWR. Further, in both states many towns that were close to the cutoff with OLS are now further below the acceptable cutoff.

Table 5 Summary statistics for COD and PRD b = 200
Fig. 2
figure 2

COD values – GWR (200)

Fig. 3
figure 3

COD values – TGWR (200)

Next, we discuss the PRD estimates, which is a measure of equity or progressivity/regressivity. As seen in Fig. 4, the PRD for OLS in Fig. 4 is above the cutoff of 1.03 for all 145 municipalities in MA, implying regressivity throughout the state. In Connecticut, there are several towns with PRD above the cutoff for OLS (see Fig. 4). While in Fig. 5, we see that the situation is mixed in Connecticut when re-estimating with GWR; in Massachusetts PRD improves somewhat with GWR compared with OLS. There is a dramatic improvement with TGWR, as nearly all towns in both states below the PRD cutoff of 1.03, implying neutrality (see Fig. 6).

Fig. 4
figure 4

PRD values – OLS

Fig. 5
figure 5

PRD values – GWR (200)

Fig. 6
figure 6

PRD values – TGWR (200)

Finally, we consider the PAPE results; summary statistics for the full set of estimates are given in Table 6. We see that OLS preforms significantly worse than GWR and TGWR, whereas TGWR has a lower mean and standard deviation than GWR. The real difference is in the upper tail of the distribution where TGWR outperforms GWR. For example, at the town-level in MA, the 95th percentile is 50% PAPE for TGWR, 55% PAPE for GWR, and 88% PAPE for OLS. For CT, the corresponding percentages in the 95th percentile for TGWR, GWR, and OLS are 52%, 57%, and 140%.

Table 6 Summary statistics for percent absolute prediction error: b = 200

Next, we consider the town-level statistics for PAPE since assessors will only be carrying out assessments within their own town. Tables 7 and 8 provide the summary statistics for the town-level distributions of the PAPE. We find similar results as in Table 6; the TGWR outperforms both OLS and GWR. The advantage of TGWR over GWR is most evident at the extremes of the distributions. For an assessor, this could prove quite valuable since avoiding large errors is important.

Table 7 Summary statistics for town-level distribution of percent absolute prediction error: CT Data, b = 200
Table 8 Summary statistics for town-Level distribution of percent absolute prediction error: MA Data, b = 200

The PAPE metric provides a clear signal that TGWR is a preferred approach for estimating sales prices. This is underscored by the fact that the extreme values are the properties that in general have fewer sales comparisons for assessors to use in their valuations, so having a methodology such as TGWR that is demonstrated to perform well can be particularly valuable in estimating the values of these high-end properties.

Robustness Checks

As robustness checks, we consider several separate bandwidths.Footnote 5 For MA, we estimate TGWR using bandwidths of 200 and 500 (we do not consider other comparisons given the huge size of this dataset). The comparison of the PAPE calculated using the TGWR results for the 200 and 500 bandwidths are given in Tables 9 and 10. We can see that TGWR performs better at the lower bandwidth. The results are similar for CT, in the sense that the tails continue to perform better with TGWR.

Table 9 summary statistics for percent absolute prediction error: CT Data, TGWR
Table 10 Summary statistics for percent absolute prediction error: MA Data, TGWR

For CT we use bandwidths of 200, 500, and 1000. We examine the PAPE and find that the GWR density is virtually unchanged but the TGWR density is slightly worse at the lower percent errors but slightly improved for the larger percent errors. These differences are accentuated slightly further when increasing the bandwidth out to 1000. In other words, across these bandwidths, TGWR does a good job in mitigating large forecast errors that are present with GWR and OLS. As is the case for MA, the TGWR estimator performs better at the lower bandwidths (i.e., the PAPE is lower for b = 200 at the extreme values when compared with the extreme values of b = 500 and b = 1000).

For the b = 500 bandwidth, the mean of all of the coefficient estimates across the more than 250,000 weighted least squares regressions in CT and 600,000 observations in MA are similar to the magnitudes of the OLS estimates. But there is substantial variation in these estimates, which is apparent from viewing the standard deviations of the coefficients and their minimum and maximum values. For instance, all regressors have some observations for which the coefficients are negative and some for which the coefficients are positive. In many cases, the standard deviations are large relative to the mean of the coefficients.

One might hypothesize that including earlier observations with equal weight (i.e., OLS) will bias the estimates, especially in a sample like ours that spans 1994–2013 for CT and 1987–2012 for MA. Given these OLS, TGWR and GWR estimates, our objective is to use these to calculate predicted values of the sales prices that can be used to calculate the PAPE, COD and PRD, at the town level, separately for the OLS, TGWR and GWR models. We also calculate several separate sets of PAPE, COD and PRD for the GWR estimates, for several different bandwidth estimates. The GWR estimates are very insensitive to a broad range of bandwidths (including one as low as 0.5 and as high as 1000, with several in between). For TGWR, there is slightly more variation in COD, PRD, and PAPE when changing the bandwidth, but not much. Given this stability, along with the large sample sizes for the CT and MA data (250,000+ observations across 50 towns, and 600,000+ observations over 145 towns), and given that we find little variation in the estimates for GWR and TGWR across bandwidths, we present the results for GWR and TGWR for a bandwidth of 200 in the results tables in this paper.

To compare estimates across the two states, we present the COD and PRD at the town-level in Figs. 1, 2, 3, 4, 5, and 6. The PAPE estimates for b = 200 are presented in Tables 6, 7, and 8. All calculations are done at the town level. Regressions include the same variables as those using the full sample.

Conclusion

We develop the TGWR estimator for hedonic pricing models as a new procedure for calculating assessed values. We compare this new methodology to two previously used approaches; GWR and OLS. As a means of comparison, we use the Coefficient of Dispersion (COD) and the Price-Related Differential (PRD); statistics that are commonly employed by the IAAO to examine accuracy and regressivity/progressivity. As a measure of predictive accuracy, we also compare these three methodologies using the percent absolute prediction error (PAPE).

Based on the PAPE metric, TGWR performs better than GWR and OLS. This discrepancy in performance is particularly strong at the extreme percentiles. This is an especially important finding for assessors, because the extremepercentiles (e.g. 95th) tend to have the highest value properties that are often the most difficult to value because of the lack of comparable sales. Also, with the PRD and COD measures, there is variation across towns in terms of the progressivity/regressivity/neutrality. In many locations we see PRD estimates that indicate more municipalities are in the neutral range (i.e., less than 1.03) with TGWR than with GWR or OLS.

An important issue that is worthy of further attention is the bandwidth selection, since the COD and PRD results in some of the individual towns are sensitive to bandwidth selected. Ideally, an assessor would use cross-validation (CV) separately on every sale in their municipality to determine a separate optimal bandwidth for the local regression for each property. But, given approximately 900,000 observations in our study, this would be extremely time consuming to implement; as it is, for each bandwidth it takes more than 10 days to solve GWR (and separately, TGWR) for the 900,000+ observations. We recommend that individual assessors use CV on their own municipalities which would be much more feasible to implement in a TGWR or GWR setting. Also, while one possibility to reduce computation time might be to use a sub-sample of the sales in a municipality and interpolate sales prices for other properties, this results in the loss of some valuable information for assessors. Therefore, it is important for individual assessors to use all available data on sales when performing this type of analysis.

Also, TGWR is a very time-intensive set of computations, which is magnified when considering many different bandwidth options. Therefore, assessors should plan a lot of time to develop Time-GWR estimates since it can take many hours to run for larger cities for each bandwidth. If one finds, for instance, that the smaller bandwidths perform better, this implies that when the tails of the distribution shrink and the distribution of the kernel weights become more focused around the target point, one obtains more accurate assessments. But the tradeoff is that we need a sufficient number of observations to make the local regressions work, so an assessor would not want to have too few observations that are given positive weight. Therefore, a detailed bandwidth consideration, such as cross-validation, may be important for assessors in refining this approach. However, this adds additional layers of computational difficulty that need to be considered in future work.