Scale-space theory is a framework for multi-scale signal representation developed by the computer vision, image processing and signal processing communities with complementary motivations from physics and biological vision. It is a formal theory for handling image structures at different scales, by representing an image as a one-parameter family of smoothed images, the scale-space representation, parametrized by the size of the smoothing kernel used for suppressing fine-scale structures.[1][2][3][4][5][6][7][8] The parameter in this family is referred to as the scale parameter, with the interpretation that image structures of spatial size smaller than about have largely been smoothed away in the scale-space level at scale .
The main type of scale space is the linear (Gaussian) scale space, which has wide applicability as well as the attractive property of being possible to derive from a small set of scale-space axioms. The corresponding scale-space framework encompasses a theory for Gaussian derivative operators, which can be used as a basis for expressing a large class of visual operations for computerized systems that process visual information. This framework also allows visual operations to be made scale invariant, which is necessary for dealing with the size variations that may occur in image data, because real-world objects may be of different sizes and in addition the distance between the object and the camera may be unknown and may vary depending on the circumstances.[9][10]
wit83
was invoked but never defined (see the help page).koe84
was invoked but never defined (see the help page).lin94
was invoked but never defined (see the help page).lin94review
was invoked but never defined (see the help page).flo97
was invoked but never defined (see the help page).sporring-etal-1997
was invoked but never defined (see the help page).rom03
was invoked but never defined (see the help page).lin08
was invoked but never defined (see the help page).lin14compvisrefguide
was invoked but never defined (see the help page).