Search Page

Showing 1 - 10 of 10
DM-85 - Point, Line, and Area Generalization

Generalization is an important and unavoidable part of making maps because geographic features cannot be represented on a map without undergoing transformation. Maps abstract and portray features using vector (i.e. points, lines and polygons) and raster (i.e pixels) spatial primitives which are usually labeled. These spatial primitives are subjected to further generalization when map scale is changed. Generalization is a contradictory process. On one hand, it alters the look and feel of a map to improve overall user experience especially regarding map reading and interpretive analysis. On the other hand, generalization has documented quality implications and can sacrifice feature detail, dimensions, positions or topological relationships. A variety of techniques are used in generalization and these include selection, simplification, displacement, exaggeration and classification. The techniques are automated through computer algorithms such as Douglas-Peucker and Visvalingam-Whyatt in order to enhance their operational efficiency and create consistent generalization results. As maps are now created easily and quickly, and used widely by both experts and non-experts owing to major advances in IT, it is increasingly important for virtually everyone to appreciate the circumstances, techniques and outcomes of generalizing maps. This is critical to promoting better map design and production as well as socially appropriate uses.

AM-94 - Machine Learning Approaches

Machine learning approaches are increasingly used across numerous applications in order to learn from data and generate new knowledge discoveries, advance scientific studies and support automated decision making. In this knowledge entry, the fundamentals of Machine Learning (ML) are introduced, focusing on how feature spaces, models and algorithms are being developed and applied in geospatial studies. An example of a ML workflow for supervised/unsupervised learning is also introduced. The main challenges in ML approaches and our vision for future work are discussed at the end.

AM-21 - The Evolution of Geospatial Reasoning, Analytics, and Modeling

The field of geospatial analytics and modeling has a long history coinciding with the physical and cultural evolution of humans. This history is analyzed relative to the four scientific paradigms: (1) empirical analysis through description, (2) theoretical explorations using models and generalizations, (3) simulating complex phenomena and (4) data exploration. Correlations among developments in general science and those of the geospatial sciences are explored. Trends identify areas ripe for growth and improvement in the fourth and current paradigm that has been spawned by the big data explosion, such as exposing the ‘black box’ of GeoAI training and generating big geospatial training datasets. Future research should focus on integrating both theory- and data-driven knowledge discovery.

AM-97 - An Introduction to Spatial Data Mining

The goal of spatial data mining is to discover potentially useful, interesting, and non-trivial patterns from spatial data-sets (e.g., GPS trajectory of smartphones). Spatial data mining is societally important having applications in public health, public safety, climate science, etc. For example, in epidemiology, spatial data mining helps to nd areas with a high concentration of disease incidents to manage disease outbreaks. Computational methods are needed to discover spatial patterns since the volume and velocity of spatial data exceed the ability of human experts to analyze it. Spatial data has unique characteristics like spatial autocorrelation and spatial heterogeneity which violate the i.i.d (Independent and Identically Distributed) assumption of traditional statistic and data mining methods. Therefore, using traditional methods may miss patterns or may yield spurious patterns, which are costly in societal applications. Further, there are additional challenges such as MAUP (Modiable Areal Unit Problem) as illustrated by a recent court case debating gerrymandering in elections. In this article, we discuss tools and computational methods of spatial data mining, focusing on the primary spatial pattern families: hotspot detection, collocation detection, spatial prediction, and spatial outlier detection. Hotspot detection methods use domain information to accurately model more active and high-density areas. Collocation detection methods find objects whose instances are in proximity to each other in a location. Spatial prediction approaches explicitly model the neighborhood relationship of locations to predict target variables from input features. Finally, spatial outlier detection methods find data that differ from their neighbors. Lastly, we describe future research and trends in spatial data mining.

AM-107 - Spatial Data Uncertainty

Although spatial data users may not be aware of the inherent uncertainty in all the datasets they use, it is critical to evaluate data quality in order to understand the validity and limitations of any conclusions based on spatial data. Spatial data uncertainty is inevitable as all representations of the real world are imperfect. This topic presents the importance of understanding spatial data uncertainty and discusses major methods and models to communicate, represent, and quantify positional and attribute uncertainty in spatial data, including both analytical and simulation approaches. Geo-semantic uncertainty that involves vague geographic concepts and classes is also addressed from the perspectives of fuzzy-set approaches and cognitive experiments. Potential methods that can be implemented to assess the quality of large volumes of crowd-sourced geographic data are also discussed. Finally, this topic ends with future directions to further research on spatial data quality and uncertainty.

DM-90 - Hydrographic Geospatial Data Standards

Coastal nations, through their dedicated Hydrographic Offices (HOs), have the obligation to provide nautical charts for the waters of national jurisdiction in support of safe maritime navigation. Accurate and reliable charts are essential to seafarers whether for commerce, defense, fishing, or recreation. Since navigation can be an international activity, mariners often use charts published from different national HOs. Standardization of data collection and processing, chart feature generalization methods, text, symbology, and output validation becomes essential in providing mariners with consistent and uniform products regardless of the region or the producing nation. Besides navigation, nautical charts contain information about the seabed and the coastal environment useful in other domains such as dredging, oceanography, geology, coastal modelling, defense, and coastal zone management. The standardization of hydrographic and nautical charting activities is achieved through various publications issued by the International Hydrographic Organization (IHO). This chapter discusses the purpose and importance of nautical charts, the establishment and role of the IHO in coordinating HOs globally, the existing hydrographic geospatial data standards, as well as those under development based on the new S-100 Universal Hydrographic Data Model.

DM-20 - Entity-based Models

As we translate real world phenomena into data structures that we can store in a computer, we must determine the most appropriate spatial representation and how it relates to the characteristics of such a phenomenon. All spatial representations are derivatives of graph theory and should therefore be described in such terms. This then helps to understand the principles of low-level GIS operations. A constraint-driven approach allows the reader to evaluate implementations of the geo-relational principle in terms of the hierarchical level of mathematical space adopted.

AM-10 - Spatial Interaction

Spatial interaction (SI) is a fundamental concept in the GIScience literature, and may be defined in numerous ways. SI often describes the "flow" of individuals, commodities, capital, and information over (geographic) space resulting from a decision process. Alternatively, SI is sometimes used to refer to the influence of spatial proximity of places on the intensity of relations between those places. SI modeling as a separate research endeavor developed out of a need to mathematically model and understand the underlying determinants of these flows/influences. Proponents of SI modeling include economic geographers, regional scientists, and regional planners, as well as climate scientists, physicists, animal ecologists, and even some biophysical/environmental researchers. Originally developed from theories of interacting particles and gravitational forces in physics, SI modeling has developed through a series of refinements in terms of functional form, conceptual representations of distances, as well as a range of analytically rigorous technical improvements.
 

AM-106 - Error-based Uncertainty

The largest contributing factor to spatial data uncertainty is error. Error is defined as the departure of a measure from its true value. Uncertainty results from: (1) a lack of knowledge of the extent and of the expression of errors and  (2) their propagation through analyses. Understanding error and its sources is key to addressing error-based uncertainty in geospatial practice. This entry presents a sample of issues related to error and error based uncertainty in spatial data. These consist of (1) types of error in spatial data, (2) the special case of scale and its relationship to error and (3) approaches to quantifying error in spatial data.

DM-70 - Problems of Large Spatial Databases

Large spatial databases often labeled as geospatial big data exceed the capacity of commonly used computing systems as a result of data volume, variety, velocity, and veracity. Additional problems also labeled with V’s are cited, but the four primary ones are the most problematic and focus of this chapter (Li et al., 2016, Panimalar et al., 2017).  Sources include satellites, aircraft and drone platforms, vehicles, geosocial networking services, mobile devices, and cameras. The problems in processing these data to extract useful information include query, analysis, and visualization. Data mining techniques and machine learning algorithms, such as deep convolutional neural networks, often are used with geospatial big data. The obvious problem is handling the large data volumes, particularly for input and output operations, requiring parallel read and write of the data, as well as high speed computers, disk services, and network transfer speeds. Additional problems of large spatial databases include the variety and heterogeneity of data requiring advanced algorithms to handle different data types and characteristics, and integration with other data. The velocity at which the data are acquired is a challenge, especially using today’s advanced sensors and the Internet of Things that includes millions of devices creating data on short temporal scales of micro seconds to minutes. Finally, the veracity, or truthfulness of large spatial databases is difficult to establish and validate, particularly for all data elements in the database.