Estimation of spatial econometric linear models with large datasets: How big can spatial Big Data be?

Publication date: Available online 14 February 2019Source: Regional Science and Urban EconomicsAuthor(s): G. Arbia, C. Ghiringhelli, A. MiraAbstractSpatial econometrics is currently experiencing the Big Data revolution both in terms of the volume of data and the velocity with which they are accumulated. Regional data, employed traditionally in spatial econometric modeling, can be very large, with information that are increasingly available at a very fine resolution level such as census tracts, local markets, town blocks, regular grids or other small partitions of the territory. When dealing with spatial microeconometric models referred to the granular observations of the single economic agent, the number of observations available can be a lot higher. This paper reports the results of a systematic simulation study on the limits of the current methodologies when estimating spatial models with large datasets. In our study we simulate a Spatial Lag Model (SLM), we estimate it using Maximum Likelihood (ML), Two Stages Least Squares (2SLS) and Bayesian estimator (B), and we test their performances for different sample sizes and different levels of sparsity of the weight matrices. We considered three performance indicators, namely: computing time, storage required and accuracy of the estimators. The results show that using standard computer capabilities the analysis becomes prohibitive and unreliable when the sample size is greater than 70,000 even for low levels of sparsity. This r...
Source: Regional Science and Urban Economics - Category: Science Source Type: research
More News: Computers | Economics | Science | Study