Mathematical Motivation Complex Valued Cnn Lecun
Introduction
Convolutional networks (convnets) have become increasingly important to artificial intelligence in recent years, as reviewed by LeCun et al. (2015). The present paper presents a theoretical argument for complex-valued convnets and their remarkable performance; complex-valued convnets turn out to calculate 'data-driven multiscale windowed spectra' characterizing certain stochastic processes common in the modeling of time series (such as audio) and natural images (including patterns and textures). We motivate the construction of such multiscale spectra via 'local averages of multiwavelet absolute values' or, more generally, 'nonlinear multiwavelet packets.'
A textbook treatment of all concepts and terms used above and below is given by Mallat (2008). Further information is available in the original work of Daubechies (1992), Meyer (1993), Coifman et al. (1994), Coifman and Donoho (1995), Simoncelli and Freeman (1995), Meyer and Coifman (1997), LeCun et al. (1998), Donoho et al. (2003), Srivastava et al. (2003), Rabiner and Schafer (2007), and Mallat (2008), for example. The work of Haensch and Hellwich (2010), Mallat (2010), Poggio et al. (2012), Bruna and Mallat (2013), Bruna et al. (2015), and Chintala et al. (2015) also develops complex-valued convnets, providing copious applications and numerical experiments. A related, more sophisticated connection (to renormalization group theory) is given by Mehta and Schwab (2014). Our exposition relies on nothing but the basic signal processing treated by Mallat (2008). Via the connections discussed below, the rich, rigorous mathematical analysis surveyed by Daubechies (1992), Meyer (1993), Mallat (2008), and others applies directly to complex-valued convnets.
Citing such connections, the present paper's anonymous reviews suggested viewing complexvalued convnets as a kind of baseline architecture for much of the deep learning reviewed by LeCun et al. (2015). Section 6 presents numerical analyses corroborating this viewpoint. Having such a theoretical basis for deep learning could help in paring down the combinatorial explosion of possibilities for future developments, while probably also illuminating further possibilities.
The present paper proceeds as follows: Section 2 reviews stationary stochastic processes and their spectra. Section 3 reviews locally stationary stochastic processes and the connection of their spectra to stages in a complex-valued convnet. Section 4 introduces multiscale (multiple stages in a convnet). Section 5 describes the fitting/learning/training that the connection to convnets facilitates. Section 6 briefly compares on a common benchmark the accuracies for the complexvalued convnets of Chintala et al. (2015) to those for the scattering transforms of Mallat (2010) and for the standard real-valued convnets of Krizhevsky et al. (2012). Section 7 generalizes and summarizes the aforementioned sections.
Stationary stochastic processes
For simplicity, we first limit consideration to the special case of a doubly infinite sequence of nonnegative random variables X k , where k ranges over the integers. This input data will be the result of convolving an unmeasured independent and identically distributed (i.i.d.) sequence Z k , where k ranges over the integers, with an unknown sequence of real numbers f k , where k ranges over the integers (this latter sequence is known as a 'filter,' whereas the i.i.d. sequence is known as 'white noise'):
$$
$$
for any integer j . Such a sequence X k , with k ranging over the integers, is a (strictly) 'stationary stochastic process.' The terminology 'strictly stationary' refers to the fact that lagging or shifting the process preserves the probability distribution of the process: indeed, for any integer l , the shift Y k = X k -l , where k ranges over the integers, satisfies
$$
$$
The associated 'absolute spectrum' is
$$
$$
for any real number ω (usually we consider not just any, but instead restrict consideration to a sequence running from 0 to about 2 π ). Please note that lagging or shifting the process changes neither the probability distribution of the process (since the process is stationary) nor the absolute spectrum: for any integer l , the shift Y k = X k -l yields ˜ Y ( ω ) = ˜ X ( ω ) for any real number ω , due to the absolute value in equation 3.
$$
$$
for any real number ω ; there is an extra squaring under the expectation in equation 4 compared to equation 3. Again, lagging or shifting the process changes neither the probability distribution of the process nor the power spectrum: for any integer l , the shift Y k = X k -l yields ˜ ˜ Y ( ω ) = ˜ ˜ X ( ω ) for any real number ω , due to the absolute value in equation 4. The remainder of the present paper focuses on the absolute spectrum; most of the discussion applies to the power spectrum, too.
Remark 1. The absolute spectrum can be more robust than the power spectrum, in the same sense that the mean absolute deviation can be more robust than the variance or standard deviation. The power spectrum is more fundamental in a certain sense, yet the absolute spectrum may be preferable for applications to machine learning. We conjecture that both can work about the same. We focus on the absolute spectrum to simplify the exposition.
.
Locally stationary stochastic processes
In practice, the input data is seldom strictly stationary, but usually only locally stationary, that is, equation 1 becomes
$$
$$
for any integer j , where f ( j ) k changes much more slowly when changing j than when changing k . To accommodate such data, we introduce windowed spectra; for any even nonnegative-valued sequence g k , with k ranging through the integers - this sequence could be samples of a Gaussian or any other window suitable for Gabor analysis (the data itself will determine g during training) - we consider for any integer l , with some positive integer n . The extra summation in equation 6 averages away noise and is a kind of approximation to the expected value in equation 3. Usually g k is fairly close to 1 for k = -n , -n + 1, . . . , n -1, n , and g k is fairly close to 0 for | k | > n , making a reasonably smooth transition between 0 and 1. The most important difference between equation 3 and equation 6 is the absence of a limit in the latter (hence the terminology, 'local' spectrum).
$$
$$
Due to the absolute value, equation 6 is equivalent to
$$
$$
for any even nonnegative-valued sequence g k , with k ranging through the integers, where
$$
$$
for any integer k ('even' means that g -k = g k for every integer k ). Please note that the right-hand side of equation 7 is just a convolution followed by the absolute value followed by local averaging; this will facilitate fitting/learning/training using data - enabling a 'data-driven' approach - in Section 5.
Multiscale
In most cases, the ideal choices of n and width of the window in equation 7, that is, the ideal number of indices for which g k is substantially nonzero, are far from obvious. Often, in fact, multiple widths are relevant (say, wider for lower-frequency variations than for higher frequency). Not knowing the ideal a priori, we use multiple windows on multiple scales. An especially efficient multiscale implementation processes the results of the lowest-frequency channels recursively. For the lowest frequency, ω = 0, and when X k is nonnegative for every integer k (for example, the input X k could be the ˜ X k arising from previous processing), equation 7 simplifies to
$$
$$
$$
$$
for any integer l , and again g j , with j ranging through the integers, is an even sequence of nonnegative real numbers ('even' means that g -j = g j for every integer j ). The result of equation 9 is simply a convolution with the input sequence, and further convolutions - say via recursive processing of the form in equation 7 - can undo this convolution and set the effective window however desired in later stages. The deconvolution and subsequent convolution with the windowed exponential of a later stage is numerically stable if the later window is wider than the preceding. In particular, recursively processing the zero-frequency channels in this way can implement a 'wavelet transform' (if each recursive stage considers only two values for ω , one zero and one nonzero - see Figure 1) or a 'multiwavelet transform' (if each recursive stage considers multiple values for ω , with one of the values being zero - see Figure 2). For multidimensional signals, multiwavelets detect local directionality beyond what wavelets provide. If we recursively process the higher-frequency channels, too, then we obtain a 'nonlinear wavelet packet transform' or a 'nonlinear multiwavelet packet transform' - a kind of nonlinear iterated filter bank - see Figure 3. Linearly recombining the different frequency channels may help realize local rotation-invariance and other potentially desirable properties (indeed, Mallat (2010) did this for rotations and other transformations) including generating harmonics when processing audio signals. The transforms just discussed are undecimated, but interleaving appropriate decimation or subsampling applied to the sequences yields the usual decimated transforms.
Remark 2. In practice, decimation or subsampling is important to avoid overfitting in the datadriven approach discussed below, by limiting the number of degrees of freedom appropriately. Even when the signal is not a strictly stationary stochastic process, the averaging in equation 7 - the leftmost summation - performs the 'cycle spinning' of Coifman and Donoho (1995) to avoid artifacts that would otherwise arise due to windows' partitioning after subsampling. The averaging reduces the variance; wider averaging would further reduce the variance.
Remark 3. Sequences that are finite rather than doubly infinite provide only enough information for estimating a smoothed version of the spectrum. Alternatively, a finite amount of data provides information for estimating multiscale windowed spectra yielding time-frequency (or space-Fourier) resolution similar to the multiresolution analysis of wavelets.
Remark 4. SIFT, HOG, SURF, etc. of Lowe (1999), Lowe (2004), Dalal and Triggs (2005), Bay et al. (2008), and others are more analogous to the multiwavelet architecture of Figure 2 than to the more general wavelet-packet architecture of Figure 3.
for any integer l , where
.
.
.
Fitting/learning/training
The 'multiwavelet transform' constitutes a desirable baseline model. We can easily adapt to the data the choices of windows and indeed the whole recursive structure of the processing (whether restricting the recursion to the zero-frequency channels, or also allowing the recursive processing of higher-frequency channels). Viewing the convolutional filters in equation 7 that serve as windowed exponentials as parameters, the desirable baseline is just one member of a parametric family of models. This parametric family is known as a 'complex-valued convolutional network.' We can fit (that is, learn or train) the parameters to the data via optimization procedures such as stochastic gradient descent in conjunction with 'backpropagation' (backpropagation is the chain rule of Calculus applied to calculate gradients of our recursively composed operations). For 'supervised learning,' we optimize according to a specified objective, usually using the multiscale spectra as inputs to a scheme for classification or regression, as detailed by LeCun et al. (1998), for example.
Remark 5. In consonance with the 'best-basis' approach of Coifman et al. (1994) and Saito and Coifman (1995), a potentially more efficient possibility is to restrict the convolutional filters in equation 7 to be windowed exponentials that are designed completely a priori, aside from one overall scaling factor per filter, fitting/learning/training only the scaling factors. How best to effect this approach is an open question.
.
Numerical experiments
The following reports the classification accuracies for the complex-valued convnets of Chintala et al. (2015), the standard real-valued convnets of Krizhevsky et al. (2012), and the scattering transforms of Oyallon and Mallat (2015), on a benchmark data set, 'CIFAR-10,' from Krizhevsky (2009) (CIFAR-10 contains 50,000 images in its training set and 10,000 images in its testing set; each image falls into one of ten classes, is full-color, and consists of a 32 × 32 grid of pixels): According to Table 4 of Oyallon and Mallat (2015), the scattering transforms attain an error rate of 18% on the test set, after training their classifiers on the training set. According to Section 3.3 of Krizhevsky et al. (2012), a standard real-valued convnet attains an error rate of 13% on the test set without the 'local response normalization' of that Section 3.3, and attains 11% with the local response normalization. The complex-valued convnets detailed in Chintala et al. (2015) attain an error rate of 12% on the test set, at least when using a larger net and training with enough iterations for the test error to settle down and converge (for complex-valued convnets, accuracy seems to improve as the net becomes larger - for the error rate of 12%, a net eight times the size of that reported in Table 1 of Chintala et al. (2015) was sufficient, using the same kernel sizes and other parameter settings as for Table 1). Augmenting the training images with their mirror images improved convergence to the reported accuracies. All in all, the extensively trained real- and complex-valued convnets yielded similar error rates, which are about a third less than those which scattering transforms attained. Of course, the fitting/learning/training involved for classification with the scattering transforms is much less extensive.
Conclusion
While the above concerns X k , where k ranges over the integers, extension to analyzing X j,k , where j and k range over the integers, is straightforward - the latter could be a 'locally homogeneous random field.' Also, the infinite range of the integers is far from essential; implementations on computers obviously use only finite sequences. Moreover, the above construction is appropriate
for processing any locally stationary stochastic process, not just filtered white noise. For instance, the construction can enable a multiresolution analysis of 'regularity' (or 'smoothness') that easily distinguishes between low-pass filtered i.i.d. Gaussian noise and a pulse train or sinusoid with a random phase offset (for example, X k = 1 + sin( π ( k + J ) / 1000) for any integer k , where J is an integer drawn uniformly at random from 1, 2, . . . , 2000). More generally, the construction should enable discriminating between many interesting classes of stochastic processes, commensurate with the ability of multiwavelet-based multiresolution analysis to measure 'regularity,' 'intermittency,' distributional characteristics (say, Gaussian versus Poisson), etc. Any globally stationary stochastic process - with or without intermittent fluctuations - can be modeled as above as a locally stationary stochastic process (of course, Bruna et al. (2015) treat the former directly, to great advantage in the analysis of homogeneous turbulence and other phenomena from statistical physics). Every model in the parametric family constituting the complex-valued convnet calculates relevant features, windowed spectra of the form in equation 6 and equation 7. The absolute values in equation 6 and equation 7 are the key nonlinearity, a reflection of the local stationarity - the local translation-invariance - of the process and its relevant features.
Acknowledgments
We would like to thank Keith Adams, Lubomir Bourdev, Rob Fergus, Armand Joulin, Manohar Paluri, Christian Puhrsch, Marc'Aurelio Ranzato, Ben Recht, and Rachel Ward.
/negationslash

Figure 1: A flow chart for the 'wavelet transform' of an input vector: each box ' ω =0' corresponds to equation 7 with ω =0 or (equivalently) to equation 9; each box ' ω =0' corresponds to equation 7 - convolution followed by taking the absolute value of every entry followed by local averaging; each circle ' ↓ ' corresponds to subsampling (say, retaining only every other entry)


$$ \label{colored} X_j = \sum_{k=-\infty}^{\infty} f_{j-k} , Z_k $$ \tag{colored}
$$ \label{absspec} \tilde{X}(\omega) = \lim_{n \to \infty} \E\left| \frac{1}{\sqrt{2n+1}} \sum_{k=-n}^n e^{-i k \omega } X_k \right| $$ \tag{absspec}
$$ g_k(\omega) = e^{i k \omega} g_k $$
$$ h_l = \frac{1}{2n+1} \sum_{j=-n+l}^{n+l} g_j $$
Remark. The absolute spectrum can be more robust than the power spectrum, in the same sense that the mean absolute deviation can be more robust than the variance or standard deviation. The power spectrum is more fundamental in a certain sense, yet the absolute spectrum may be preferable for applications to machine learning. We conjecture that both can work about the same. We focus on the absolute spectrum to simplify the exposition.
Remark. In practice, decimation or subsampling is important to avoid overfitting in the data-driven approach discussed below, by limiting the number of degrees of freedom appropriately. Even when the signal is not a strictly stationary stochastic process, the averaging in equationconv --- the leftmost summation --- performs the ``cycle spinning'' ofcoifman-donoho to avoid artifacts that would otherwise arise due to windows' partitioning after subsampling. The averaging reduces the variance; wider averaging would further reduce the variance.
Remark. Sequences that are finite rather than doubly infinite provide only enough information for estimating a smoothed version of the spectrum. Alternatively, a finite amount of data provides information for estimating multiscale windowed spectra yielding time-frequency (or space-Fourier) resolution similar to the multiresolution analysis of wavelets.
Remark. SIFT, HOG, SURF, etc.\ oflowe1999, lowe2004, dalal-triggs, bay-ess-tuytelaars-van_gool, and others are more analogous to the multiwavelet architecture of Figuremultifig than to the more general wavelet-packet architecture of Figure~packetfig.
Remark. In consonance with the ``best-basis'' approach ofcoifman-meyer-quake-wickerhauser andsaito-coifman, a potentially more efficient possibility is to restrict the convolutional filters in equation~conv to be windowed exponentials that are designed completely a priori, aside from one overall scaling factor per filter, fitting/learning/training only the scaling factors. How best to effect this approach is an open question.
References
A Mathematical Motivation for
Complex-valued Convolutional Networks
Joan Bruna, Soumith Chintala, Yann LeCun, Serkan Piantino, Arthur Szlam, Mark Tygert Facebook Artificial Intelligence Research, 1 Facebook Way, Menlo Park, California 94025
Keywords: deep learning, neural networks, harmonic analysis
Abstract: A complex-valued convolutional network (convnet) implements the repeated application of the following composition of three operations, recursively applying the composition to an input vector of nonnegative real numbers: (1) convolution with complex-valued vectors followed by (2) taking the absolute value of every entry of the resulting vectors followed by (3) local averaging. For processing real-valued random vectors, complex-valued convnets can be viewed as “data-driven multiscale windowed power spectra,” “data-driven multiscale windowed absolute spectra,” “data-driven multiwavelet absolute values,” or (in their most general configuration) “data-driven nonlinear multiwavelet packets.” Indeed, complex-valued convnets can calculate multiscale windowed spectra when the convnet filters are windowed complex-valued exponentials. Standard real-valued convnets, using rectified linear units (ReLUs), sigmoidal (for example, logistic or tanh) nonlinearities, max. pooling, etc., do not obviously exhibit the same exact correspondence with data-driven wavelets (whereas for complex-valued convnets, the correspondence is much more than just a vague analogy). Courtesy of the exact correspondence, the remarkably rich and rigorous body of mathematical analysis for wavelets applies directly to (complex-valued) convnets.
Convolutional networks (convnets) have become increasingly important to artificial intelligence in recent years, as reviewed by LeCun et al., (2015). The present paper presents a theoretical argument for complex-valued convnets and their remarkable performance; complex-valued convnets turn out to calculate “data-driven multiscale windowed spectra” characterizing certain stochastic processes common in the modeling of time series (such as audio) and natural images (including patterns and textures). We motivate the construction of such multiscale spectra via “local averages of multiwavelet absolute values” or, more generally, “nonlinear multiwavelet packets.”
A textbook treatment of all concepts and terms used above and below is given by Mallat, (2008). Further information is available in the original work of Daubechies, (1992), Meyer, (1993), Coifman et al., (1994), Coifman and Donoho, (1995), Simoncelli and Freeman, (1995), Meyer and Coifman, (1997), LeCun et al., (1998), Donoho et al., (2003), Srivastava et al., (2003), Rabiner and Schafer, (2007), and Mallat, (2008), for example. The work of Haensch and Hellwich, (2010), Mallat, (2010), Poggio et al., (2012), Bruna and Mallat, (2013), Bruna et al., (2015), and Chintala et al., (2015) also develops complex-valued convnets, providing copious applications and numerical experiments. A related, more sophisticated connection (to renormalization group theory) is given by Mehta and Schwab, (2014). Our exposition relies on nothing but the basic signal processing treated by Mallat, (2008). Via the connections discussed below, the rich, rigorous mathematical analysis surveyed by Daubechies, (1992), Meyer, (1993), Mallat, (2008), and others applies directly to complex-valued convnets.
Citing such connections, the present paper’s anonymous reviews suggested viewing complex-valued convnets as a kind of baseline architecture for much of the deep learning reviewed by LeCun et al., (2015). Section 6 presents numerical analyses corroborating this viewpoint. Having such a theoretical basis for deep learning could help in paring down the combinatorial explosion of possibilities for future developments, while probably also illuminating further possibilities.
The present paper proceeds as follows: Section 2 reviews stationary stochastic processes and their spectra. Section 3 reviews locally stationary stochastic processes and the connection of their spectra to stages in a complex-valued convnet. Section 4 introduces multiscale (multiple stages in a convnet). Section 5 describes the fitting/learning/training that the connection to convnets facilitates. Section 6 briefly compares on a common benchmark the accuracies for the complex-valued convnets of Chintala et al., (2015) to those for the scattering transforms of Mallat, (2010) and for the standard real-valued convnets of Krizhevsky et al., (2012). Section 7 generalizes and summarizes the aforementioned sections.
For simplicity, we first limit consideration to the special case of a doubly infinite sequence of nonnegative random variables Xksubscript𝑋𝑘X_{k}, where k𝑘k ranges over the integers. This input data will be the result of convolving an unmeasured independent and identically distributed (i.i.d.) sequence Zksubscript𝑍𝑘Z_{k}, where k𝑘k ranges over the integers, with an unknown sequence of real numbers fksubscript𝑓𝑘f_{k}, where k𝑘k ranges over the integers (this latter sequence is known as a “filter,” whereas the i.i.d. sequence is known as “white noise”):
for any integer j𝑗j. Such a sequence Xksubscript𝑋𝑘X_{k}, with k𝑘k ranging over the integers, is a (strictly) “stationary stochastic process.” The terminology “strictly stationary” refers to the fact that lagging or shifting the process preserves the probability distribution of the process: indeed, for any integer l𝑙l, the shift Yk=Xk−lsubscript𝑌𝑘subscript𝑋𝑘𝑙Y_{k}=X_{k-l}, where k𝑘k ranges over the integers, satisfies
for any integer j𝑗j, where Zk′=Zk−lsubscriptsuperscript𝑍′𝑘subscript𝑍𝑘𝑙Z^{\prime}{k}=Z{k-l}; the sequence Zk′subscriptsuperscript𝑍′𝑘Z^{\prime}{k}, with k𝑘k ranging over the integers, is i.i.d. with the same distribution as Zksubscript𝑍𝑘Z{k}, where k𝑘k ranges over the integers.
The associated “absolute spectrum” is
for any real number ω𝜔\omega (usually we consider not just any, but instead restrict consideration to a sequence running from 00 to about 2π2𝜋2\pi). Please note that lagging or shifting the process changes neither the probability distribution of the process (since the process is stationary) nor the absolute spectrum: for any integer l𝑙l, the shift Yk=Xk−lsubscript𝑌𝑘subscript𝑋𝑘𝑙Y_{k}=X_{k-l} yields Y(ω)=X(ω)𝑌𝜔𝑋𝜔\tilde{Y}(\omega)=\tilde{X}(\omega) for any real number ω𝜔\omega, due to the absolute value in equation 3.
for any real number ω𝜔\omega; there is an extra squaring under the expectation in equation 4 compared to equation 3. Again, lagging or shifting the process changes neither the probability distribution of the process nor the power spectrum: for any integer l𝑙l, the shift Yk=Xk−lsubscript𝑌𝑘subscript𝑋𝑘𝑙Y_{k}=X_{k-l} yields Y(ω)=X(ω)𝑌𝜔𝑋𝜔{\tilde{\vphantom{\rule{1.0pt}{6.53334pt}}\smash{\tilde{Y}}}}(\omega)={\tilde{\vphantom{\rule{1.0pt}{6.53334pt}}\smash{\tilde{X}}}}(\omega) for any real number ω𝜔\omega, due to the absolute value in equation 4. The remainder of the present paper focuses on the absolute spectrum; most of the discussion applies to the power spectrum, too.
The absolute spectrum can be more robust than the power spectrum, in the same sense that the mean absolute deviation can be more robust than the variance or standard deviation. The power spectrum is more fundamental in a certain sense, yet the absolute spectrum may be preferable for applications to machine learning. We conjecture that both can work about the same. We focus on the absolute spectrum to simplify the exposition.
In practice, the input data is seldom strictly stationary, but usually only locally stationary, that is, equation 1 becomes
for any integer j𝑗j, where fk(j)subscriptsuperscript𝑓𝑗𝑘f^{(j)}{k} changes much more slowly when changing j𝑗j than when changing k𝑘k. To accommodate such data, we introduce windowed spectra; for any even nonnegative-valued sequence gksubscript𝑔𝑘g{k}, with k𝑘k ranging through the integers — this sequence could be samples of a Gaussian or any other window suitable for Gabor analysis (the data itself will determine g𝑔g during training) — we consider
for any integer l𝑙l, with some positive integer n𝑛n. The extra summation in equation 6 averages away noise and is a kind of approximation to the expected value in equation 3. Usually gksubscript𝑔𝑘g_{k} is fairly close to 111 for k=−n𝑘𝑛k=-n, −n+1𝑛1-n+1, …, n−1𝑛1n-1, n𝑛n, and gksubscript𝑔𝑘g_{k} is fairly close to 00 for |k|>n𝑘𝑛|k|>n, making a reasonably smooth transition between 0 and 1. The most important difference between equation 3 and equation 6 is the absence of a limit in the latter (hence the terminology, “local” spectrum).
Due to the absolute value, equation 6 is equivalent to
for any even nonnegative-valued sequence gksubscript𝑔𝑘g_{k}, with k𝑘k ranging through the integers, where
for any integer k𝑘k (“even” means that g−k=gksubscript𝑔𝑘subscript𝑔𝑘g_{-k}=g_{k} for every integer k𝑘k). Please note that the right-hand side of equation 7 is just a convolution followed by the absolute value followed by local averaging; this will facilitate fitting/learning/training using data — enabling a “data-driven” approach — in Section 5.
In most cases, the ideal choices of n𝑛n and width of the window in equation 7, that is, the ideal number of indices for which gksubscript𝑔𝑘g_{k} is substantially nonzero, are far from obvious. Often, in fact, multiple widths are relevant (say, wider for lower-frequency variations than for higher frequency). Not knowing the ideal a priori, we use multiple windows on multiple scales. An especially efficient multiscale implementation processes the results of the lowest-frequency channels recursively. For the lowest frequency, ω=0𝜔0\omega=0, and when Xksubscript𝑋𝑘X_{k} is nonnegative for every integer k𝑘k (for example, the input Xksubscript𝑋𝑘X_{k} could be the Xksubscript𝑋𝑘\tilde{X}_{k} arising from previous processing), equation 7 simplifies to
for any integer l𝑙l, where
for any integer l𝑙l, and again gjsubscript𝑔𝑗g_{j}, with j𝑗j ranging through the integers, is an even sequence of nonnegative real numbers (“even” means that g−j=gjsubscript𝑔𝑗subscript𝑔𝑗g_{-j}=g_{j} for every integer j𝑗j). The result of equation 9 is simply a convolution with the input sequence, and further convolutions — say via recursive processing of the form in equation 7 — can undo this convolution and set the effective window however desired in later stages. The deconvolution and subsequent convolution with the windowed exponential of a later stage is numerically stable if the later window is wider than the preceding. In particular, recursively processing the zero-frequency channels in this way can implement a “wavelet transform” (if each recursive stage considers only two values for ω𝜔\omega, one zero and one nonzero — see Figure 1) or a “multiwavelet transform” (if each recursive stage considers multiple values for ω𝜔\omega, with one of the values being zero — see Figure 2). For multidimensional signals, multiwavelets detect local directionality beyond what wavelets provide. If we recursively process the higher-frequency channels, too, then we obtain a “nonlinear wavelet packet transform” or a “nonlinear multiwavelet packet transform” — a kind of nonlinear iterated filter bank — see Figure 3. Linearly recombining the different frequency channels may help realize local rotation-invariance and other potentially desirable properties (indeed, Mallat, (2010) did this for rotations and other transformations) — including generating harmonics when processing audio signals. The transforms just discussed are undecimated, but interleaving appropriate decimation or subsampling applied to the sequences yields the usual decimated transforms.
In practice, decimation or subsampling is important to avoid overfitting in the data-driven approach discussed below, by limiting the number of degrees of freedom appropriately. Even when the signal is not a strictly stationary stochastic process, the averaging in equation 7 — the leftmost summation — performs the “cycle spinning” of Coifman and Donoho, (1995) to avoid artifacts that would otherwise arise due to windows’ partitioning after subsampling. The averaging reduces the variance; wider averaging would further reduce the variance.
Sequences that are finite rather than doubly infinite provide only enough information for estimating a smoothed version of the spectrum. Alternatively, a finite amount of data provides information for estimating multiscale windowed spectra yielding time-frequency (or space-Fourier) resolution similar to the multiresolution analysis of wavelets.
SIFT, HOG, SURF, etc. of Lowe, (1999), Lowe, (2004), Dalal and Triggs, (2005), Bay et al., (2008), and others are more analogous to the multiwavelet architecture of Figure 2 than to the more general wavelet-packet architecture of Figure 3.
The “multiwavelet transform” constitutes a desirable baseline model. We can easily adapt to the data the choices of windows and indeed the whole recursive structure of the processing (whether restricting the recursion to the zero-frequency channels, or also allowing the recursive processing of higher-frequency channels). Viewing the convolutional filters in equation 7 that serve as windowed exponentials as parameters, the desirable baseline is just one member of a parametric family of models. This parametric family is known as a “complex-valued convolutional network.” We can fit (that is, learn or train) the parameters to the data via optimization procedures such as stochastic gradient descent in conjunction with “backpropagation” (backpropagation is the chain rule of Calculus applied to calculate gradients of our recursively composed operations). For “supervised learning,” we optimize according to a specified objective, usually using the multiscale spectra as inputs to a scheme for classification or regression, as detailed by LeCun et al., (1998), for example.
In consonance with the “best-basis” approach of Coifman et al., (1994) and Saito and Coifman, (1995), a potentially more efficient possibility is to restrict the convolutional filters in equation 7 to be windowed exponentials that are designed completely a priori, aside from one overall scaling factor per filter, fitting/learning/training only the scaling factors. How best to effect this approach is an open question.
The following reports the classification accuracies for the complex-valued convnets of Chintala et al., (2015), the standard real-valued convnets of Krizhevsky et al., (2012), and the scattering transforms of Oyallon and Mallat, (2015), on a benchmark data set, “CIFAR-10,” from Krizhevsky, (2009) (CIFAR-10 contains 50,000 images in its training set and 10,000 images in its testing set; each image falls into one of ten classes, is full-color, and consists of a 32×32323232\times 32 grid of pixels): According to Table 4 of Oyallon and Mallat, (2015), the scattering transforms attain an error rate of 18%percent1818% on the test set, after training their classifiers on the training set. According to Section 3.3 of Krizhevsky et al., (2012), a standard real-valued convnet attains an error rate of 13%percent1313% on the test set without the “local response normalization” of that Section 3.3, and attains 11%percent1111% with the local response normalization. The complex-valued convnets detailed in Chintala et al., (2015) attain an error rate of 12%percent1212% on the test set, at least when using a larger net and training with enough iterations for the test error to settle down and converge (for complex-valued convnets, accuracy seems to improve as the net becomes larger — for the error rate of 12%percent1212%, a net eight times the size of that reported in Table 1 of Chintala et al., (2015) was sufficient, using the same kernel sizes and other parameter settings as for Table 1). Augmenting the training images with their mirror images improved convergence to the reported accuracies. All in all, the extensively trained real- and complex-valued convnets yielded similar error rates, which are about a third less than those which scattering transforms attained. Of course, the fitting/learning/training involved for classification with the scattering transforms is much less extensive.
While the above concerns Xksubscript𝑋𝑘X_{k}, where k𝑘k ranges over the integers, extension to analyzing Xj,ksubscript𝑋𝑗𝑘X_{j,k}, where j𝑗j and k𝑘k range over the integers, is straightforward — the latter could be a “locally homogeneous random field.” Also, the infinite range of the integers is far from essential; implementations on computers obviously use only finite sequences. Moreover, the above construction is appropriate for processing any locally stationary stochastic process, not just filtered white noise. For instance, the construction can enable a multiresolution analysis of “regularity” (or “smoothness”) that easily distinguishes between low-pass filtered i.i.d. Gaussian noise and a pulse train or sinusoid with a random phase offset (for example, Xk=1+sin(π(k+J)/1000)subscript𝑋𝑘1𝜋𝑘𝐽1000X_{k}=1+\sin(\pi(k+J)/1000) for any integer k𝑘k, where J𝐽J is an integer drawn uniformly at random from 1, 2, …, 2000). More generally, the construction should enable discriminating between many interesting classes of stochastic processes, commensurate with the ability of multiwavelet-based multiresolution analysis to measure “regularity,” “intermittency,” distributional characteristics (say, Gaussian versus Poisson), etc. Any globally stationary stochastic process — with or without intermittent fluctuations — can be modeled as above as a locally stationary stochastic process (of course, Bruna et al., (2015) treat the former directly, to great advantage in the analysis of homogeneous turbulence and other phenomena from statistical physics). Every model in the parametric family constituting the complex-valued convnet calculates relevant features, windowed spectra of the form in equation 6 and equation 7. The absolute values in equation 6 and equation 7 are the key nonlinearity, a reflection of the local stationarity — the local translation-invariance — of the process and its relevant features.
We would like to thank Keith Adams, Lubomir Bourdev, Rob Fergus, Armand Joulin, Manohar Paluri, Christian Puhrsch, Marc’Aurelio Ranzato, Ben Recht, and Rachel Ward.
A flow chart for the “wavelet transform” of an input vector: each box “ω𝜔\omega==00” corresponds to equation 7 with ω𝜔\omega==00 or (equivalently) to equation 9; each box “ω𝜔\omega≠\neq00” corresponds to equation 7 — convolution followed by taking the absolute value of every entry followed by local averaging; each circle “↓↓\downarrow” corresponds to subsampling (say, retaining only every other entry)
$$ X_{j}=\sum_{k=-\infty}^{\infty}f_{j-k},Z_{k} $$ \tag{S2.E1}
$$ \tilde{X}(\omega)=\lim_{n\to\infty}{\bf E}\left|\frac{1}{\sqrt{2n+1}}\sum_{k=-n}^{n}e^{-ik\omega}X_{k}\right| $$ \tag{S2.E3}
$$ g_{k}(\omega)=e^{ik\omega}g_{k} $$ \tag{S3.E8}
Remark1. Remark 1. The absolute spectrum can be more robust than the power spectrum, in the same sense that the mean absolute deviation can be more robust than the variance or standard deviation. The power spectrum is more fundamental in a certain sense, yet the absolute spectrum may be preferable for applications to machine learning. We conjecture that both can work about the same. We focus on the absolute spectrum to simplify the exposition.
Remark1. Remark 2. In practice, decimation or subsampling is important to avoid overfitting in the data-driven approach discussed below, by limiting the number of degrees of freedom appropriately. Even when the signal is not a strictly stationary stochastic process, the averaging in equation 7 — the leftmost summation — performs the “cycle spinning” of Coifman and Donoho, (1995) to avoid artifacts that would otherwise arise due to windows’ partitioning after subsampling. The averaging reduces the variance; wider averaging would further reduce the variance.
Remark1. Remark 3. Sequences that are finite rather than doubly infinite provide only enough information for estimating a smoothed version of the spectrum. Alternatively, a finite amount of data provides information for estimating multiscale windowed spectra yielding time-frequency (or space-Fourier) resolution similar to the multiresolution analysis of wavelets.
Remark1. Remark 4. SIFT, HOG, SURF, etc. of Lowe, (1999), Lowe, (2004), Dalal and Triggs, (2005), Bay et al., (2008), and others are more analogous to the multiwavelet architecture of Figure 2 than to the more general wavelet-packet architecture of Figure 3.
Remark1. Remark 5. In consonance with the “best-basis” approach of Coifman et al., (1994) and Saito and Coifman, (1995), a potentially more efficient possibility is to restrict the convolutional filters in equation 7 to be windowed exponentials that are designed completely a priori, aside from one overall scaling factor per filter, fitting/learning/training only the scaling factors. How best to effect this approach is an open question.


References
[bay-ess-tuytelaars-van_gool] Bay, H., Ess, A., Tuytelaars, T., and Gool, L.~V. (2008). \newblock Speeded-up robust features ({SURF}). \newblock {\em Computer Vision Image Understanding}, 110(3):346--359.
[bruna-mallat] Bruna, J. and Mallat, S. (2013). \newblock Invariant scattering convolutional networks. \newblock {\em IEEE Trans.\ Pattern Analysis Machine Intel.}, 35(8):1872--1886.
[bruna-mallat-bacry-muzy] Bruna, J., Mallat, S., Bacry, E., and Muzy, J.-F. (2015). \newblock Intermittent process analysis with scattering moments. \newblock {\em Ann.\ Statist.}, 43(1):323--351.
[chintala-ranzato-szlam-tian-tygert-zaremba] Chintala, S., Ranzato, M., Szlam, A., Tian, Y., Tygert, M., and Zaremba, W. (2015). \newblock Scale-invariant learning and convolutional networks. \newblock Technical Report 1506.08230, arXiv.
[coifman-donoho] Coifman, R.~R. and Donoho, D. (1995). \newblock Translation-invariant denoising. \newblock In Antoniadis, A. and Oppenheim, G., editors, {\em Wavelets and Statistics}, volume 103 of {\em Lecture Notes in Statistics}, pages 125--150. Springer.
[coifman-meyer-quake-wickerhauser] Coifman, R.~R., Meyer, Y., Quake, S., and Wickerhauser, M.~V. (1994). \newblock Signal processing and compression with wavelet packets. \newblock In Byrnes, J.~S., Byrnes, J.~L., Hargreaves, K.~A., and Berry, K., editors, {\em Wavelets and Their Applications}, volume 442 of {\em NATO ASI Series C: Mathematical and Physical Sciences}, pages 363--379. Springer.
[dalal-triggs] Dalal, N. and Triggs, B. (2005). \newblock Histograms of oriented gradients for human detection. \newblock In {\em IEEE Computer Society Conf.\ Computer Vision and Pattern Recognition 2005}, volume~1, pages 886--893. IEEE.
[daubechies] Daubechies, I. (1992). \newblock {\em Ten Lectures on Wavelets}. \newblock CBMS-NSF Regional Conference Series in Applied Mathematics. SIAM.
[donoho-mallat-von_sachs-samuelides] Donoho, D., Mallat, S., {von Sachs}, R., and Samuelides, Y. (2003). \newblock Locally stationary covariance and signal estimation with macrotiles. \newblock {\em IEEE Trans.\ Signal Processing}, 51(3):614--627.
[haensch-hellwich] Haensch, R. and Hellwich, O. (2010). \newblock Complex-valued convolutional neural networks for object detection in {PolSAR} data. \newblock In {\em 8th European Conf.\ EUSAR}, pages 1--4. IEEE.
[krizhevsky] Krizhevsky, A. (2009). \newblock Learning multiple layers of features from tiny images. \newblock Technical Report Master's Thesis, University of Toronto Department of Computer Science.
[krizhevsky-sutskever-hinton] Krizhevsky, A., Sutskever, I., and Hinton, G. (2012). \newblock Image{N}et classification with deep convolutional neural networks. \newblock In {\em Advances in Neural Information Processing Systems}, pages 1097--1105. Curran Associates.
[lecun-bengio-hinton] Le{C}un, Y., Bengio, Y., and Hinton, G. (2015). \newblock Deep learning. \newblock {\em Nature}, 521(7553):436--444.
[lecun-bottou-bengio-haffner] {LeC}un, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998). \newblock Gradient-based learning applied to document recognition. \newblock {\em Proc. IEEE}, 86(11):2278--2324.
[lowe1999] Lowe, D.G. (1999). \newblock Object recognition from local scale-invariant features. \newblock In {\em Proc.\ 7th IEEE Internat.\ Conf.\ Computer Vision}, volume2, pages 1150--1157. IEEE.
[lowe2004] Lowe, D.~G. (2004). \newblock Distinctive image features from scale-invariant keypoints. \newblock {\em Internat.\ J.\ Computer Vision}, 60(2):91--110.
[mallat2008] Mallat, S. (2008). \newblock {\em A Wavelet Tour of Signal Processing: the Sparse Way}. \newblock Academic Press, 3rd edition.
[mallat2010] Mallat, S. (2010). \newblock Recursive interferometric representations. \newblock In {\em Proc.\ EUSIPCO Conf.\ 2010}, pages 716--720. EURASIP.
[mehta-schwab] Mehta, P. and Schwab, D.~J. (2014). \newblock An exact mapping between the variational renormalization group and deep learning. \newblock Technical Report 1410.3831, arXiv.
[meyer] Meyer, Y. (1993). \newblock {\em Wavelets and Operators}, volume~37 of {\em Cambridge Studies in Advanced Mathematics}. \newblock Cambridge University Press.
[meyer-coifman] Meyer, Y. and Coifman, R.R. (1997). \newblock {\em Wavelets: Calder'on-Zygmund and Multilinear Operators}, volume48 of {\em Cambridge Studies in Advanced Mathematics}. \newblock Cambridge University Press.
[oyallon-mallat] Oyallon, E. and Mallat, S. (2015). \newblock Deep roto-translation scattering for object classification. \newblock In {\em IEEE Computer Society Conf.\ Computer Vision and Pattern Recognition 2015}, volume~1, pages 2865--2873. IEEE.
[poggio-mutch-leibo-rosasco-tacchetti] Poggio, T., Mutch, J., Leibo, J., Rosasco, L., and Tacchetti, A. (2012). \newblock The computational magic of the ventral stream: sketch of a theory (and why some deep architectures work). \newblock Technical Report MIT-CSAIL-TR-2012-035, MIT CSAIL, Cambridge, MA.
[rabiner-schafer] Rabiner, L.~R. and Schafer, R.W. (2007). \newblock {\em Introduction to Digital Speech Processing}, volume1 of {\em Foundations and Trends in Signal Processing}. \newblock NOW.
[saito-coifman] Saito, N. and Coifman, R.~R. (1995). \newblock Local discriminant bases and their applications. \newblock {\em J. Math. Imaging Vision}, 5(4):337--358.
[simoncelli-freeman] Simoncelli, E.~P. and Freeman, W.T. (1995). \newblock The steerable pyramid: a flexible architecture for multi-scale derivative computation. \newblock In {\em Proc.\ Internat.\ Conf.\ Image Processing 1995}, volume3, pages 444--447. IEEE.
[srivastava-lee-simoncelli-zhu] Srivastava, A., Lee, A.~B., Simoncelli, E.~P., and Zhu, S. (2003). \newblock On advances in statistical modeling of natural images. \newblock {\em J. Math. Imaging Vision}, 18(1):17--33.
[bib5] Coifman and Donoho, (1995) Coifman, R. R. and Donoho, D. (1995). Translation-invariant denoising. In Antoniadis, A. and Oppenheim, G., editors, Wavelets and Statistics, volume 103 of Lecture Notes in Statistics, pages 125–150. Springer.
[bib6] Coifman et al., (1994) Coifman, R. R., Meyer, Y., Quake, S., and Wickerhauser, M. V. (1994). Signal processing and compression with wavelet packets. In Byrnes, J. S., Byrnes, J. L., Hargreaves, K. A., and Berry, K., editors, Wavelets and Their Applications, volume 442 of NATO ASI Series C: Mathematical and Physical Sciences, pages 363–379. Springer.
[bib7] Dalal and Triggs, (2005) Dalal, N. and Triggs, B. (2005). Histograms of oriented gradients for human detection. In IEEE Computer Society Conf. Computer Vision and Pattern Recognition 2005, volume 1, pages 886–893. IEEE.
[bib9] Donoho et al., (2003) Donoho, D., Mallat, S., von Sachs, R., and Samuelides, Y. (2003). Locally stationary covariance and signal estimation with macrotiles. IEEE Trans. Signal Processing, 51(3):614–627.
[bib12] Krizhevsky et al., (2012) Krizhevsky, A., Sutskever, I., and Hinton, G. (2012). ImageNet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, pages 1097–1105. Curran Associates.
[bib22] Oyallon and Mallat, (2015) Oyallon, E. and Mallat, S. (2015). Deep roto-translation scattering for object classification. In IEEE Computer Society Conf. Computer Vision and Pattern Recognition 2015, volume 1, pages 2865–2873. IEEE.
[bib23] Poggio et al., (2012) Poggio, T., Mutch, J., Leibo, J., Rosasco, L., and Tacchetti, A. (2012). The computational magic of the ventral stream: sketch of a theory (and why some deep architectures work). Technical Report MIT-CSAIL-TR-2012-035, MIT CSAIL, Cambridge, MA.