Paradoxically, signal detection is concerned with fitting models to supposedly random series similarly to mathematical proofs by reductio ad absurdum of the antithesis. That is, the hypothesis (antithesis) is made, that the observed series has properties of a pure noise series, . Then, a model is fitted and series and are obtained. If the quality of the model fits to the observations does not significantly differ from the quality of a fit to pure noise , then is true and we say that contains no signal but noise. In the opposite case of model fitting significantly better than , we reject and say that the model signal was detected in . The difference is significant (at some level) if it is not likely (at this level) to occur between two different realizations of the noise .
The quality of the fit is evaluated using a function S of the series , , and . A function of random variables, such as , is a random variable itself and is called a statistic. A random variable S is characterized by its probability distribution function. Following we use the distribution of S for pure noise signal , and , to be denoted or simply . Precisely, we shall use the cumulative probability distribution function which for a given critical value of the statistic supplies the probability for the observed S to fall on one side of the .
The observed value of the statistic and its probability distribution, and respectively, are used to obtain the probability of being true. That is, if p turns out small, , is improbable and has no properties of . Then we say that the model signal has been detected at the confidence level . The smaller is, the more convincing (significant) is the detection. The special realization of a random series which consists of independent variables of common (gaussian) distribution is called (gaussian) white noise. We assume here that the noise is white noise. Note that in the signal detection process, frequency and lag l are considered independent variables and do not count as parameters.
Summarizing, the basis for the determination of the properties of a an observed time series is a test statistic, S, with known probability distribution for (white) noise, .
Let consist of random variables and let a given model have parameters. Then the modeled series corresponds to a combination of random variables and the residual series corresponds to a combination of random variables. The proof rests on the observation that orthogonal transformations convert vectors of independent variables into vectors of independent variables. Let us consider an approximately linear model with matrix so that , where P is a vector of parameters. Then spans a vector space with no more than orthogonal vectors (dimensions). The numbers , and are called the numbers of degrees of freedom of the observations, the model fit, and the residuals, respectively.