The Gabor transform, named after Dennis Gabor, is a special case of the short-time Fourier transform. It is used to determine the sinusoidal frequency and phase content of local sections of a signal as it changes over time. The function to be transformed is first multiplied by a Gaussian function, which can be regarded as a window function, and the resulting function is then transformed with a Fourier transform to derive the time-frequency analysis.[1] The window function means that the signal near the time being analyzed will have higher weight. The Gabor transform of a signal x(t) is defined by this formula:
The Gaussian function has infinite range and it is impractical for implementation. However, a level of significance can be chosen (for instance 0.00001) for the distribution of the Gaussian function.
Outside these limits of integration () the Gaussian function is small enough to be ignored. Thus the Gabor transform can be satisfactorily approximated as
This simplification makes the Gabor transform practical and realizable.
The window function width can also be varied to optimize the time-frequency resolution tradeoff for a particular application by replacing the with for some chosen .