# Inverse distance weighting

Inverse Distance Weighting (IDW) is a type of deterministic method for multivariate interpolation with a known scattered set of points. The assigned values to unknown points are calculated with a weighted average of the values available at the known points.

The name given to this type of methods was motivated by the weighted average applied, since it resorts to the inverse of the distance to each known point ("amount of proximity") when assigning weights.

## Definition of the Problem

The expected result is a discrete assignment of the unknown function $u$ in a study region:

The set of $N$ known data points can be described as a list of tuples:

The function is to be "smooth" (continuous and once differentiable), to be exact ($u(x_{i})=u_{i}$ ) and to meet the user's intuitive expectations about the phenomenon under investigation. Furthermore, the function should be suitable for a computer application at a reasonable cost (nowadays, a basic implementation will probably make use of parallel resources ).

## Shepard's method

### Historical Reference

At the Harvard Laboratory for Computer Graphics and Spatial Analysis, beginning in 1965, a varied collection of scientists converged to rethink, among other things, what we now call geographic information systems.

The motive force behind the Laboratory, Howard Fisher, conceived an improved computer mapping program that he called SYMAP, which, from the start, Fisher wanted to improve on the interpolation. He showed Harvard College freshmen his work on SYMAP, and many of them participated in Laboratory events. One freshman, Donald Shepard, decided to overhaul the interpolation in SYMAP, resulting in his famous article from 1968.

Shepard’s algorithm was also influenced by the theoretical approach of William Warntz and others at the Lab who worked with spatial analysis. He conducted a number of experiments with the exponent of distance, deciding on something closer to the gravity model (exponent of -2). Shepard implemented not just basic inverse distance weighting, but also he allowed barriers (permeable and absolute) to interpolation.

Other research centers were working on interpolation at this time, particularly University of Kansas and their SURFACE II program. Still, the features of SYMAP were state-of-the-art, even though programmed by an undergraduate.

### Basic Form

$u(\mathbf {x} )={\begin{cases}{\frac \sum _{i=1}^{N}{w_{i}(\mathbf {x} )u_{i}}}\sum _{i=1}^{N}{w_{i}(\mathbf {x} )}}},&{\text{if }}d(\mathbf {x} ,\mathbf {x} _{i})\neq 0{\text{ for all }}i\\u_{i},&{\text{if }}d(\mathbf {x} ,\mathbf {x} _{i})=0{\text{ for some }}i\end{cases}}$ where

$w_{i}({\mathbf {x} })={\frac {1}{d({\mathbf {x} },{\mathbf {x} }_{i})^{p}}}$ is a simple IDW weighting function, as defined by Shepard, x denotes an interpolated (arbitrary) point, xi is an interpolating (known) point, $d$ is a given distance (metric operator) from the known point xi to the unknown point x, N is the total number of known points used in interpolation and $p$ is a positive real number, called the power parameter.

Here weight decreases as distance increases from the interpolated points. Greater values of $p$ assign greater influence to values closest to the interpolated point, with the result turning into a mosaic of tiles (a Voronoi diagram) with nearly constant interpolated value for large values of p. For two dimensions, power parameters $p\leq 2$ cause the interpolated values to be dominated by points far away, since with a density $\rho$ of data points and neighboring points between distances $r_{0}$ to $R$ , the summed weight is approximately

$\sum _{j}w_{j}\approx \int _{r_{0}}^{R}{\frac {2\pi r\rho dr}{r^{p}}}=2\pi \rho \int _{r_{0}}^{R}r^{1-p}dr,$ which diverges for $R\rightarrow \infty$ and $p\leq 2$ . For N dimensions, the same argument holds for $p\leq N$ . For the choice of value for p, one can consider the degree of smoothing desired in the interpolation, the density and distribution of samples being interpolated, and the maximum distance over which an individual sample is allowed to influence the surrounding ones.

Shepard's method is a consequence of minimization of a functional related to a measure of deviations between tuples of interpolating points {x, u} and i tuples of interpolated points {xi, ui}, defined as:

$\phi ({\mathbf {x} },u)=\left(\sum _{i=0}^{N}{\frac {(u-u_{i})^{2}}{d({\mathbf {x} },{\mathbf {x} }_{i})^{p}}}\right)^{\frac {1}{p}},$ derived from the minimizing condition:

${\frac {\partial \phi ({\mathbf {x} },u)}{\partial u}}=0.$ The method can easily be extended to other dimensional spaces and it is in fact a generalization of Lagrange approximation into a multidimensional spaces. A modified version of the algorithm designed for trivariate interpolation was developed by Robert J. Renka and is available in Netlib as algorithm 661 in the toms library.

### Łukaszyk-Karmowski metric

Yet another modification of the Shepard's method was proposed by Łukaszyk also in applications to experimental mechanics. The proposed weighting function had the form:

$w_{k}({\mathbf {x} })={\frac {1}{(D_{**}({\mathbf {x} },{\mathbf {x} }_{k}))^{\frac {1}{2}}}},$ where $D_{**}({\mathbf {x} },{\mathbf {x} }_{k})$ is the Łukaszyk–Karmowski metric chosen also with regard to the statistical error probability distributions of measurement of the interpolated points.

### Modified Shepard's Method

Another modification of Shepard's method calculates interpolated value using only nearest neighbors within R-sphere (instead of full sample). Weights are slightly modified in this case:

$w_{k}({\mathbf {x} })=\left({\frac {\max(0,R-d({\mathbf {x} },{\mathbf {x} }_{k}))}{Rd({\mathbf {x} },{\mathbf {x} }_{k})}}\right)^{2}$ When combined with fast spatial search structure (like kd-tree) it becomes efficient N*logN interpolation method suitable for large-scale problems.