Keywords

1 Introduction

Affordable housing is emerging as a major requirement due to the growth in the need for creating equality in the living standards of people in our society [1]. Government is extending the collaboration towards public–private partnership, especially realtors to construct houses and apartments for people who are below the median line of income. The locations for construction of such houses are dependent on numerous real estate attributes, which include social, cultural, economic, physical, and governmental [2]. Some of these attributes include public transportation facilities, availability of public schools and colleges, availability of water and supportive weather conditions, availability of hotels and restaurants, and so on [3]. As the attribute number list grows, so does the complexity of decision-making in location choice. Identification of best locations is an important requirement from the perspective of not only house construction but also purchase of the existing house and renting.

Hence, to understand the trends and solve the above mentioned problems in real estate, many methods and tools are used which are derived from various fields like data science, network science, statistics, probability theory, estimation theory, and so on. Today, the availability of huge volumes of data has paved the path for researchers to use concepts and tools of data analytics to discover goal-oriented knowledge from the existing data. One such important application is hedonic modeling, where the dependency of the various attributes on the real estate price is computed using basic regression methods and machine learning techniques. Use of linear and logistic regression, clustering techniques, support vector machines, and artificial neural networks [4,5,6,7,8,9,10,11,12,13,14] on the real estate multiple listing service data is often encountered in the literature. Additionally, predicting the house price using data analytic tools is also a well-studied research area. Regression techniques, neural networks, and gradient boosting [15,16,17] are few tools that are often used. The increasing availability of voluminous data has attracted network science practitioners to infer on the relational status and mutual dependencies among the various data ingredients. Use of social media like Facebook, Twitter, LinkedIn, and various search engines is generating digital footsteps which are used to develop knowledge graphs. These graphs (networks) make us understand the behavior of entities in a network [18,19,20].

To the best of the authors’ knowledge, there has been no focused effort to use network science for the analysis of real estate networks and in specific for location identification. Majority of the software [21, 22] ask the user to enter an exact location, and based on the database query search it will identify suitable condominiums (apartment complexes) for investment. In addition, the existing literature on data analytics in real estate investment is based on the assumptions that a user already knows a location. There are many reasons why an investor may not know the specific location for investment. A simple reason may be that an investor is new to the city. A more involved reason is that even though an investor is native to the city, it is logically impossible to narrow down to a very specific location; at best a small geographical area can be identified. However, in big cities even a small area can easily compromise thousands of dwellings and commercial property; further, even the small area is often highly heterogeneous (in terms of people, establishments, facilities, etc.). Focusing only on price trends does not address the multiple concerns of an investor [15].

Choosing a good location for investment is very crucial since it is dependent on a large number of user’s requirements. It may be based on job availability, economic status of people, availability of restaurants, low criminal activities and safety, public transportation facility, availability of schools and shopping malls, and many more. These multiple attributes make a user’s decision to select a location more complex and difficult. Under the influence of this huge number of attributes, the location selection may tend towards suboptimal decisions. Hence, an intelligent way of choosing the locations is of greater need in real estate investment that also focuses on the selection of best attributes among that huge number of attributes. Moreover, the mutual influence of the attributes is not considered in the existing literature and attributes are assumed to be independent which is actually not true in general.

In this chapter, a novel algorithm has been presented that selects best attributes for a user and based on user’s choice the algorithm identifies the best locations (throughout this chapter, authors refer to condominium complexes as locations) for investment in real estate. In the proposed work, nearly 200 real estate attributes were considered and the best attributes were selected based on the metric called χ which is computed mainly based on the Pearson correlation coefficient [23]. The obtained attributes form a source of choice for a user and based on his/her selections layers of machine learning techniques are activated. In the first layer roads and streets are identified and in layer 2 condominiums (locations) in that street is spotted. For this purpose, statistical modeling with machine learning techniques is used which are taken from data science as analytic tools. However, applications of data analytics succeed well in identification and classification problems; a clearer inference on the relational status of the attributes cannot be obtained, especially when the entities are in huge number. Hence, a bipartite network is added as an extension to the machine learning layers that provides a relational status of the attributes and their influence on the various streets and roads (collectively we call them as landmarks hereafter), while selecting best location among the shortlisted locations by the machine learning layers, for investment. The network uses eigen centrality on a bipartite network of attributes and condominiums for selecting the best condominium for investment. The advantage of using the network layer is the selection of condominium based on the relationship status of all the other attributes with the condominiums in the network. An example simulation on the analysis of network dynamics is provided to leverage the advantages of network science by creating perturbations in the link weights of the network, to find the influential and stable attribute in the designed real estate bipartite network.

The location identification algorithm is tested on the data obtained from the official database called TerraFly [24] which is created and maintained by Florida International University (FIU). We have restricted this work to nine landmarks. Rest of this chapter is organized as follows: Sect. 2 deals with the applications of data and network science for smart governance; Sect. 3 discusses about the related works and the state-of-the-art comparison with the current work in the chapter; Sect. 4 deals with the data set used in this work and the assumptions on the work; Sect. 5 discusses the statistical modeling used to identify the best attributes, and discusses the use of decision trees and PCA and K-means clustering for location identification; Sect. 6 deals with the network science for location identification problem; Sect. 7 discusses the obtained results and discussions; Sect. 8 discusses the implications of the obtained result on the smart governance; and finally Sect. 9 discusses the conclusions of the work.

1.1 Scopes of This Work

  • Application of statistical modeling to obtain the best attributes from nearly 200 real estate attributes based on the metric χ which is computed using Pearson’s correlation coefficient.

  • Use stacks of machine learning algorithms to identify locations (condominiums) for investment.

  • Application of network science to obtain the best condominium based on centrality measures by constructing a real estate network.

  • Simulation study on the most consistent and influential attribute in the presence of link weight perturbations in the designed real estate network.

2 Data and Network Science Applications in Smart Governance

Technology has already made its way into our daily activities. Shopping, transportation, communication, education, health, and so many other things rely on smart devices that are driven by advanced computing and techniques. Usage of such gadgets has paved the way to accumulate a large amount of user data which enabled the era of big data analytics [25]. Analysis of and drawing useful inferences based on various applications can help in the betterment of society. However, operating on such voluminous and versatile data has to be carried out using sophisticated methods. Data science is one such solution enabler [26] which is an interdisciplinary area comprising the tools for statistics, mathematics, probability theory, artificial intelligence, and machine learning to a larger extent and provides us better understandability of the data for various applications and one such application is smart governance. Applications in smart governance include consumer behavioral analytics, natural calamity predictions, crime predictions, social service-related analytics, healthcare analytics, data security and privacy, finance, and bank analytics.

Network science is another interdisciplinary research area that draws the theories and methods from graph theory, mathematics, statistics, physics, sociology, and data mining. The current-day computer networks, Internet, and various communication networks are analyzed using network analysis. Similarly, there are other few interesting applications of the network science (complex network analysis) that find their profound applications in smart governance including biological networks [27], transportation networks, epidemic networks, social networks, financial networks, and so on. Readers who are further interested to know about complex networks are directed to read [28]. Even though this chapter highlights the use of data and network science to solve the real estate location identification for housing application, there are still numerous other tasks in smart cities and governance that can be addressed using the able usage of data and network sciences separately or by the combination of both. Two specific examples of smart governance applications are discussed in detail in Sect. 8.

3 Related Works and State-of-the-Art Comparison

The application of data analytics and network science to solve the real estate location identification problem is novel. The existing works in data analytics focus much on the prediction and modeling of the real estate price. Authors in [29] propose a method to predict the house price index (HSI) using data analytic techniques like clustering and principal component analysis for Kookmin bank data. Work in [4] discusses the use of linear regression to find the relation between the real estate price and the attributes; for this purpose authors use real estate data of Harbin city. In [30], authors propose a linear regression hedonic model to find the spatial dependency of the real estate price; for this purpose they have considered the real estate data of eight countries from the multiple listing service database.

In [31], authors discuss a framework of using fuzzy logic to find the selling price of a real estate property in the presence of incomplete information. They claim that the developed method helps in reducing the risk that arises from the uncertainties in the input. In [32], authors use fuzzy logic systems for hedonic house price modeling. In [33], authors use network science to determine the financial activities among different entities. This analysis would help in revealing financial crimes like terrorism, narcotic laundering, and so on.

Authors in [34] discuss the usage of network science to model the supply chain in the form of a network. Authors claim that such kind of structural studies helps in revealing the robustness of supply chains and reveal the topological behavior of the supply chains while overcoming the limitations of the existing supply chain networks. In [35], authors use complex network analysis in considering the micro- and macro-level attributes of revenue of UK stock market and forecast the stock value predictions using complex networks.

There have been limited applications of data science towards real estate location identification problem. The trend majorly includes hedonic price modeling that relates the real estate price with its attributes and prediction of real estate price using the attributes. There are no specific trends in network science applications for real estate investment-related problems.

By carefully traversing through the existing works, as quoted in Sect. 1 and in the current section, it is clear that all the existing works in data analytics focus more on the predictions of the price and hedonic modeling; the location identification for investment has not been explored much. In addition, there are no specific works of network science usage that focus on the issues in the real estate investment. There are few applications in the area of finance but there are no confined works that focus on the real estate. The existing software and companies of real estate investment [36,37,38] either focus on the construction of building, maintaining, and looking after the documentation works or suggest the investment locations based on some database queries. All these methods in the existing works do not apply any machine learning techniques that are based on the query approach. Hence, considering the real estate investment location identification problem the work proposed in this chapter is a novel attempt that combines data and network science under a single roof to find a robust solution for the given problem of location identification for affordable housing in real estate investment.

4 Data Set

The data is obtained from TerraFly database [24] managed and maintained by Florida International University (FIU) in collaboration with the US Government. The database is a big data platform and a query-based system with complete information regarding economic, social, physical, and governmental factors of selected countries.

The landmarks (we call streets, roads, and so on as landmarks in this work) of Miami Beach are divided into clusters. Preference is given to nearby landmarks while clustering, however, can also be random. For our initial work, single cluster with nine landmarks and their associated real estate data (available as multiple listing service (MLS) data) is considered. They are condominium (also called as condos) data of Alton Rd, Bay Rd, Collins Ave, Dade Blvd, James Ave, Lincoln Rd, Lincoln CT, Washington Ave, and West Ave. These landmarks belong to Miami Beach City of Miami Dade County, FL, USA. The approximate count of condominiums was obtained from the official database of Miami Beach, i.e., for Alton Rd-7000 condominiums, Bay Rd-7000, Collins Ave-9000, Dade Blvd-1500, James Ave-2000, Lincoln Rd-2000, Lincoln CT-2000, Washington Ave-4000, and West Ave-2000, respectively. The hierarchical view of the data is shown in Fig. 8.1. For our analysis from every landmark, 500 condominium data were randomly picked for training and 500 for validation. The processes of training and validation were repeated in five sets and the average validation accuracy is quoted, which will be discussed in detail in the results section.

Fig. 8.1
figure 1

Hierarchical view of the available data (clustered)

4.1 Assumptions in This Work

It is assumed that a user is not fully aware of the city location details. He/she has a very little idea about the locations, but do not know whether it is best or not for investment. In addition, a user has assumed a set of attributes; however, they need not be optimal. However, a user should at least know which cluster of landmarks should be opted for his/her investment.

5 Identification of Location Using Data Analytics

In this section, a detailed analogy of the attribute selection, stacked layers of machine learning for location identification, and real estate network with its related analytics is discussed.

5.1 Attribute Selection

Real estate comprises a large list of attributes that can be broadly classified into economic, social, physical, and governmental [24]. The current framework is bound to the real estate attributes; however, the same method can be extended for other factors as well.

Out of a large number of attributes, the best set of attributes are selected based on the value χ, which is a representative of the strength of an attribute in a landmark in real estate investment. The calculation of χ is as follows:

$$ \chi ={w}_1C+{w}_2N $$
(8.1)

where w1and w2 are constants and called the weights, C is the Pearson coefficient of an attribute with the real estate price, and N is the number of data sample points available in an attribute after cleansing. Here, χ is an identity number assigned to every attribute in a landmark, based on which a top attribute is selected. For this calculation, the Pearson coefficient is used as an initial choice. However, there are other correlation metrics such as Spearman and Kendall coefficients [39] that can also be used instead. The selection algorithm is as follows:

Algorithm 1

  1. 1.

    Start

  2. 2.

    Collect the condominium data of all landmarks in a cluster. Initialize w1 and w2.

  3. 3.

    For first landmark, find the parameter χ which determines the relation between the first attribute of a condominium and real estate price.

  4. 4.

    Repeat the experiment for all the attributes. Select the top k number attributes from every condominium. Select the top u number of attributes based on the number of occurrences in a landmark.

  5. 5.

    Repeat steps 2 and 3 for all landmarks.

  6. 6.

    Combine all u and select top y attributes based on number of occurrences. This set is the optimal attribute set for that cluster of landmarks.

  7. 7.

    Repeat this process for all clusters of landmarks.

  8. 8.

    End.

For simulations, the cluster of nine landmarks was chosen and the parameters were set to k = 10 and u = 10, i.e., choosing top ten attributes every time. However, y = 9 is the count of the number of attributes for the entire cluster. The simulation is repeated for five iterations with each time 500 condominiums selected in random from every landmark (training data set). The top nine attributes obtained are as follows:

  • Number of beds: Number of bedrooms available in the unit of a condominium building.

  • Number of full baths: Number of full bathrooms (tub, shower, sink, and toilet) available in the unit.

  • Living area in sq. ft.: The space of the property where people are living.

  • Number of garage spaces: Number of spaces available for parking vehicles.

  • List price: Selling price of the property (land + assets) to the public.

  • Application fee: Fee paid for owners’ associations.

  • Year built: Year in which the condominium/apartment complex is built.

  • Family limited property total value 1: The property value accounted for taxation after all exemptions. This is for the district that does not contain schools and other facilities.

  • Tax amount: The amount paid as tax for the property every year.

  • When a user chooses this cluster of nine landmarks, the above attributes will be given to him/her as choices. A user can select and set the magnitudes to the attributes according to his/her wish. It is not mandatory that all attributes need to be filled. These attributes are passed onto two layers of machine learning: in the first layer decision trees and in the second layer principal component analysis with K-means clustering. In addition, the rationale behind the choice of a multilayer classification model is provided in [40]; interested readers are suggested to refer to that article.

5.2 Decision Trees in Layer 1

In this section, the use of decision trees [41, 42] to select the best landmark for investment is dealt with detail. The working of tree follows the naive ID3 algorithm [43].

The attribute set used in a tree may change depending upon the landmarks in that cluster selected. The top nine attributes of any cluster have χ values in every landmark obtained by averaging the χ values of all the condominiums in that landmark. These χ values are compared with the other landmarks and the highest χ value is retained with its respective landmark. The result in Table 8.1 is the output of Algorithm 1 on the training set (500 condominiums selected randomly from every landmark). The process was repeated for five iterations and in all iterations the landmarks and the attribute pair remained same likewise shown in Table 8.1. The χ values shown are the average of the five iterations. Same results were obtained for the validation set as well. This table serves as a backbone for the decision tree operation.

Table 8.1 χ values of attributes input to decision tree

A user selects a particular cluster followed by choosing attributes of his/her interest and setting suitable magnitudes for them. For a decision tree, only a user interest vector is used. For example, suppose a user is interested in the number of beds and number of full baths in the same order as in Sect. 5.1; then the user vector is 110000000. This is fed into the decision tree.

An example decision tree with three attributes which is symmetric and binary is as shown in Fig. 8.2. In the figure, every time when the tree traversal hops from one node to another, it considers the χ value (if it is “YES” case) of the current attribute and compares with the χ value of the previous and selects the landmark with the highest χ value. The process continues until one landmark is retained at the end. This technique is called the highest magnitude win approach. The obtained truth table is shown in Table 8.2.

Table 8.2 Truth table for decision tree operation

Notice the last column in Table 8.2. The rows are the different cases (shown are just four cases) in which a user can enter attributes. The generated user vector is the row. For the case “011,” which is the third row in Table 8.2, it means a user is interested in number of full baths and list price. The magnitude of χ for these attributes is 1.380 and 1.894, respectively. According to the highest magnitude win strategy, James Ave wins in this case, which is placed in the output landmark column in the corresponding row. This is the leaf of that branch in the tree. The tree and its traversal are shown in Fig. 8.2.

Fig. 8.2
figure 2

Decision tree and its traversal

Even though the tree output does not change, the position of the tree node (order of the columns in Table 8.2) matters for the fact that the time taken for the tree to converge to a landmark depends on the node information richness. To identify the positions of the nodes, the ID3 algorithm is adapted and in turn helps us to choose an optimal root attribute. For a truth table that is binary in nature and having all possibilities in it, all attributes have the same information gain according to ID3 and any attribute can be a root. However, if the truth table has chosen possibilities of truth and false, ID3 will help to choose the best root attribute. A detailed procedure on the usage of ID3 for root node selection is provided in APPENDIX-B of [40]. In addition, a decision tree need not be always as shown in Fig. 8.2, where the children nodes are identical, but they can differ as well, based on the landmarks and the cluster considered.

5.3 Principal Component Analysis (PCA) and K-Means Clustering in Layer 2

The previous section discussed the use of decision trees to identify the best landmark among the chosen cluster of landmarks. In this section, layer 2 will be analyzed in detail, which helps to find locations in the landmark, output by layer 1. In the previous section, only the user’s interest vector was considered, but in this layer the entered magnitudes are considered for location identification. The steps adapted in the layer 2 are discussed further in detail.

5.3.1 Finding Principal Components

Every landmark has condominiums which have the attributes mentioned as per Sect. 5 (in our study there are nine top attributes). Let us consider the algorithm below:

Algorithm 2

  1. 1.

    Start.

  2. 2.

    Consider first landmark from a cluster.

  3. 3.

    For every condominium in that landmark, find the principal components for the top attributes.

  4. 4.

    Consider the first principal component among the available components and average all the first components over a landmark.

  5. 5.

    Repeat steps 2 and 3 for all landmarks.

  6. 6.

    End.

Let the principal components [44] of a landmark be PC. Using PC of a particular landmark, principal scores of units in that condominium are calculated using (8.2). Every top attribute in that condominium is multiplied by PC:

$$ \mathrm{PC}\mathrm{s}=\sum \limits_{i=1}^y\mathrm{attribute}\kern0.28em (i)\ast \mathrm{PC}\kern0.28em (i). $$
(8.2)

where y is the count of top attributes of a cluster of landmarks. On averaging PCS of all the condominium units, a PC score for a condominium is obtained. Repeat the process for all the landmarks and their associated condominiums. Averaging PC scores of units in a condominium will result in PC score for a condominium. Once the PC scores are available, K-means clustering [45] is applied to the scores of the condominiums of a landmark dividing the condominiums into K groups each having its own centroid.

Layer 2 mainly operates on the magnitude entered by a user which was not considered in layer 1. Since the landmark is already chosen by layer 1, the average PC of that landmark is known. Hence, the user-entered magnitude and the average PC are multiplied to get a PC score using (8.2). This score is compared with the centroids of the groups created by K-means clustering and the group with shortest Euclidean distance is selected as the best condominium group and its ingredient condominiums are rated as the best for a user according to his/her entered choices.

6 Network Science Approach for Location Identification

Data analytics uses systematic ways of modeling and learning the data, which is the digital trace in the ongoing world. However, to understand the network prospects of the data with their mutual influences, combining data with a network is very essential [46]. This interdisciplinary setup makes complex systems like real estate investment more understandable.

In this context, a network structure is rigged for the condominiums obtained from the layer 2 machine learning technique. This network view helps a user to decide the best condominium in the presence of mutual relationships among other condominiums with their attributes. The system model is shown in Fig. 8.3.

Fig. 8.3
figure 3

Extended network layer

In Fig. 8.3, a network of condominiums is constructed. These condominiums were the list of condominiums output from layer 2 that comprised of PCA and K-means clustering. The obtained graph is a bipartite network which comprises two parties, viz. attributes and condominiums. It is to be observed that there is no link between the same party members and the link always flows from one party to the other. The attributes are shown in a square shape and the condominiums are shown in a triangular shape. On applying various centrality measures like eigen centrality [47], alpha centrality, closeness, and so on [48], it is able to draw various conclusions on the important condominiums in the network. In this work, eigen centrality is used to obtain the most influential condominium. In terms of real estate, eigenvalue depicts the amplification factor for the influence and the eigenvector infers the direction of the influence in the network.

Addition of network at the end of the machine learning layers is just a demonstration; however, layer 1 or layer 2 can also be visualized as a network, where suitable centrality measures will infer for landmark and location selection in that particular layer. As an example, in layer 1 eigen centrality can be used to select a landmark and in layer 2 alpha centrality to select the set of condominiums.

7 Results and Discussions

This section briefs the obtained results of the proposed methodology and its related discussions in detail.

The attribute selection (Algorithm 1) discussed in Sect. 5.1 was executed on five training and validation data sets. In each case, for the entire nine landmarks, unique attributes were listed. The attribute set was compared between the training and validation set and the number of mismatches was accounted. The obtained results for five datasets selected in random (i.e., five iterations) are available in Table. 8.3. It was observed that the system remains consistent with the training and validation result matching with an average of 96.86% accuracy.

Table 8.3 Accuracy of best attributes selected

The same process is repeated for the decision trees of layer 1. In the case of the tree, as it is explained already in Sect. 5.2, the tree traverses from one node to another considering the value of χ of every attribute. Hence, the attribute and landmark corresponding to the highest value of χ play a major role in decision tree outputting a landmark. The top attributes as per Sect. 5.1 are listed with its χ value averaged over a landmark in Table 8.4 for five iterations (both training and validation). It is observed that the winning landmarks with highest χ for an attribute remain the same in training and validation, which is the accuracy of 100% and in turn defines the accuracy of the decision tree.

Table 8.4 Attributes and their average χ values in every landmark

From Table 8.4, the attribute with the landmark having highest χ value remains consistent throughout five iterations (as per both training and validation data sets), with 100% validation accuracy. The highest χ value is highlighted in bold in the table.

In layer 2, principal component analysis and K-means clustering were used to find the best condominiums. These techniques were applied to the training and the validation data set. The centroids of the groups of condominiums obtained after clustering were compared in both training and validation over five iteration data sets. The deviation error (mean absolute error) was noted in each case and the obtained results are available in Table 8.5.

Table 8.5 Deviation error of centroids

The average validation accuracy of layer 2 from Table 8.5 is 90.25%.

The obtained condominiums from layer 2 were used to construct a bipartite network with attributes and the condominiums as the two parties and χ values linking them. From Table 8.4, it is evident that every attribute has χ value linking landmarks. Hence, a complex network of the attributes and condominiums is constructed and eigen centrality is applied to check the best condominium.Let us consider following values for the attributes: number of garage spaces = 3, application fee = 400, number of full bathrooms = 3, number of bedrooms = 2, built year = 1986, taxable property value = 1,942,446, living area = 1007 Sq. ft, tax amount = 8633, and list price = 2,000,000. It was observed that the landmark selected by layer 1 was James Ave, and 401 condominiums were selected by layer 2 in that landmark. After applying eigen centrality, out of 401 condominiums condominium-1701 was selected as the most central condominium. The obtained complex network of real estate scenario including attributes and landmarks is shown in Fig. 8.4, in which the blue-colored circles are the attributes and the red-colored triangles are the condominiums. The links are having varying colors based on their weights (χ magnitude). If the link weight is more than 1 then the color is green; else it is blue.

Fig. 8.4
figure 4

Complex real estate network of condominiums of James Ave

One of the important purposes of using network science is to study the relationship between the condominiums and the attributes, and their influences on each other. Consider another example, where the attributes are set for simulations like the following: number of beds = 2; number of garage spaces = 2; number of full bathrooms = 2; and application fee = 126. According to the proposed algorithm, Alton Rd was the result of the decision tree and 169 condominiums were selected by layer 2. According to eigen centrality condominium-6487 was selected as the best condominium. The obtained network is as shown in Fig. 8.5.

Fig. 8.5
figure 5

Complex real estate network of condominiums of Alton Rd

To know which attribute remains consistent under the link weight variations, the link weights of the network in the Fig. 8.5 are varied (such that percent of the link weight was added to the weight as random noise) and each time the eigen centrality values were noted. At 0% variation (or wihtout variations) of link weights, the obtained centrality values are as shown in Table 8.6. The link weights were added in steps of 10% of their existing weights and the changes in the centrality measures of the network nodes were observed.

Table 8.6 Eigen centrality values at 0% change in the link weight

It was observed that as the link weight increases, F's centrality value also increases and at a point of 30% increase L loses its central position and F becomes the more central attribute (Fig. 8.6). Hence, it was concluded that adjusting the weight of the links controls the centrality of the nodes in the network. This also implies that the more the correlation of an attribute with the real estate price of a condominium in a landmark, the more that attribute will become central in the network. The same explanation holds true for the case when the link weights are decreased. In another experiment, the weights associated with all attributes were increased to a maximum 10% of their values’ existing weights randomly, to check the most stable attribute. This helps us to understand the most consistent attribute due to sudden uncertain inflations. This simulation indicated that during sudden changes in the correlation between an attribute and real estate price, i.e., χ value, which may be due to natural calamities, inflation, and so on, list price attribute remains stable attribute by being most influential in the real estate investment. The same analogy can be drawn on condominiums as well and the most consistent condominium can be found.

Fig. 8.6
figure 6

Effect of the link weight change on the centrality values of the top two attributes

8 Implications of Results on Smart Governance

In this section, two major applications of the results obtained due to this research study are highlighted. Firstly, the impact of the various factors on major public utility centers in a city (Fig. 8.7): The public utilities may be bus stations, airports, train stations, shopping malls, highways, and harbors and the factors include weather, ongoing protest/strikes, and road traffic. It is clear that in this network, there are two parties: the public utilities and the factors influencing them. Hence, a bipartite network will give a relationship between these parties. The thickness of the edges indicates the intensity of impact on the entity by that respective factor. From the sample network it is clear that the shopping malls in the city are highly impacted by all these factors and the harbor is least affected. A detailed analysis of how the various shopping malls are affected by the factors will give a detailed analysis of the relationship. These kind of inferences can be drawn in the same lines as that of the real estate network analysis. All factors influencing the entities may be considered or just top factors/attributes depending on the requirements of the study. If there are many attributes influencing the entities, the best attributes can be selected using the technique mentioned in the above sections. 

Fig. 8.7
figure 7

Impact network (W weather, A airport, T train station, B bus station, P ongoing protest, H harbor, S shopping mall, Y highway, R road traffic)

Another interesting application is the crime studies. In Fig. 8.8, there are five criminals namely C1–C5 and the criminal activities in which they were indulged include murder, robbery, rape, and molestation. The network studies on the graph inference that murder is the most central activity performed by these criminals in that city and among all C5 turns out to be central. Knowing the inference we can go a step deeper and analyze the murder activity in the city in more detail. These are two among numerous scenarios that could be studied using a combination of data and network science.

Fig. 8.8
figure 8

Crime network (N molestation, M murder, R rape, P robbery)

9 Conclusion

Identification of location has always been a complex task for a user in real estate investment for affordable housing. In this context, the work proposed in this chapter uses the concepts of data science like statistical modeling which uses the Pearson correlation as the means to identify best attributes, and stacked machine learning techniques like decision trees, PCA, and K-means clustering for identification of location. Use of these machine learning algorithms is just a demonstration use case; however other techniques like artificial neural networks, deep learning networks, support vector machines, and so on can also be used. For the locations obtained from the machine learning layers, a bipartite network is constructed and the best location is selected in the presence of the influence of other attributes using eigen centrality. Combining of data and network analytics to obtain more insight into the location identification problem has not been explored much in the existing literature. Hence, this combination provides a more comprehensive approach to affordable housing in real estate investment. In addition, the methodology and the results obtained can be adapted for solving other issues in smart governance.