DSPRelated.com
Forums

Most commonly used window functions

Started by Max November 21, 2015
On Wednesday, November 25, 2015 at 12:34:55 AM UTC-5, robert bristow-johnson wrote:
> On Sunday, November 22, 2015 at 11:15:34 PM UTC-5, Max wrote: > > On Sun, 22 Nov 2015 16:01:48 -0800 (PST), radams2000@gmail.com wrote: > > > > >One thing that no one mentioned yet is that if you are reconstructing a time-domain signal using the ifft (let's assume an overlap of 50%), then the window needs to have the property that when you shift the window by half the frame length and add it to itself, the sum must be 1.0 in the overlapped region. Otherwise alias cancellation doesn't work. This assumes the only window is on the input side; it's also possible to window during the overlap-add operation and then the product of the input and output windows has to meet this condition. > > > > > >Anyway this rules out the majority of possible windows, unless you don't care about reconstruction. > > > > > >Bob > > > > Very good point, Bob! Which windows still qualify? > > windows that are "complementary". with the proper scaling (either time or amplitude or maybe both), these are complementary: > > Rectangular > Hann > Hamming (times 0.92) > triangular (is that the Blackman)?
no, it's "Bartlett". my memory for these names is mush. r b-j
On Wednesday, November 25, 2015 at 3:55:54 PM UTC-8, robert bristow-johnson wrote:
> On Wednesday, November 25, 2015 at 6:11:57 PM UTC-5, dbd wrote:
...
>.> >.> That's exactly what Princen and Bradley designed to do. >.> >.> J. P. Princen and A. B. Bradley "Analysis/synthesis filter bank design >.> based on time domain aliasing cancellation", IEEE Trans. Acoust., Speech, >.> Signal Processing, vol. 34, pp.1153 -1161 1986 >.> >.> J. P. Princen , A. W. Johnson and A. B. Bradley "Subband/transform coding using filter bank designs based on time domain aliasing cancellation", Proc. ICAASP, pp.2161 -2164 1987 > > > > well, i don't have access to that. but as i look around, it seems to have application to MDCT used in a codec like Dolby AAC.
If you haven't found free access to one of these you are either too incompetent or too lazy to use google. But then, being to lazy to learn the context in a technical discussion -is- incompetent. For the similarly limited: /www.ece.lsu.edu/pub/aravena/CoGS_Literature/FilterBanks/SubbandCodinfg-Princen1987.pdf
> > now we know that complementary power crossfades (that's the term i am using, i am viewing a window as the concatenation of a fade up function followed by a fade down function) is what you use for incoherent phase sinusoids. but if the sinusoids are coherent (and all of the phase vocoders i have worked on made sure that every sinusoid had its phase adjusted so that they were coherent across adjacent frames) the crossfades must be complementary **voltage**. they must cross-over when they're both at 0.5 (whereas the complementary power crossfades cross over when both are at 0.707). >
The context is not crossfades. We are discussing the OP's question, not "all of the phase vocoders" that you have worked on. Princen and Bradley refer to windows applied in the different transform domains before t/f and f/t transforms. For those who read Max's second paragraph in the original post and follow that I am responding about Max's "STFT and subsequent IFFT (B-H used for both)", some actual examples of complementary windows can be found in Table I of: Audio Engineering Society Convention Paper Presented at the 119th Convention 2005 October 7-10 New York, New York USA A New Class of Smooth Power Complementary Windows and their Application to Audio Signal Processing Deepen Sinha Anibal J. S. Ferreira at: http://www.atc-labs.com/technology/misc/windows/docs/aes119_218_ds.pdf Please note: I did not "have access" to this, but I do know how to use Google. Some of you can too. Dale B. Dalrymple
On Wednesday, November 25, 2015 at 8:50:53 PM UTC-5, dbd wrote:
> On Wednesday, November 25, 2015 at 3:55:54 PM UTC-8, robert bristow-johnson wrote: > > On Wednesday, November 25, 2015 at 6:11:57 PM UTC-5, dbd wrote: > ... > >.> > >.> That's exactly what Princen and Bradley designed to do. > >.> > >.> J. P. Princen and A. B. Bradley "Analysis/synthesis filter bank design > >.> based on time domain aliasing cancellation", IEEE Trans. Acoust., Speech, > >.> Signal Processing, vol. 34, pp.1153 -1161 1986 > >.> > >.> J. P. Princen , A. W. Johnson and A. B. Bradley "Subband/transform coding using filter bank designs based on time domain aliasing cancellation", Proc. ICAASP, pp.2161 -2164 1987 > > > > > > > well, i don't have access to that. but as i look around, it seems to have application to MDCT used in a codec like Dolby AAC. > > If you haven't found free access to one of these you are either too incompetent or too lazy to use google. But then, being to lazy to learn the context in a technical discussion -is- incompetent. > > For the similarly limited: > /www.ece.lsu.edu/pub/aravena/CoGS_Literature/FilterBanks/SubbandCodinfg-Princen1987.pdf >
well, Dale, looks like you're back to grinding axes again. i almost never have luck getting an on-line copy of an IEEE paper (which was the first one you sited) from behind a pay wall. and my lazy habit is, that when i want to point someone to a reference on the web, i give them a URL. thank you for that one.
> > > > now we know that complementary power crossfades (that's the term i am using, i am viewing a window as the concatenation of a fade up function followed by a fade down function) is what you use for incoherent phase sinusoids. but if the sinusoids are coherent (and all of the phase vocoders i have worked on made sure that every sinusoid had its phase adjusted so that they were coherent across adjacent frames) the crossfades must be complementary **voltage**. they must cross-over when they're both at 0.5 (whereas the complementary power crossfades cross over when both are at 0.707). > > > > The context is not crossfades.
absolutely wrong. where the context is "perfect reconstruction" and "frames", it's about crossfades, whether you like that semantics or not. in fact i noticed Bob Adams using a similar terminology in another thread describing a fade out function as a half window. call it what you want, but these windows are a concatenation of a fade up function adjacent to a fade down function. and they are complementary if the fade up half and fade down half add, in some sense, to 1. now, in what sense might that be?? so the OP is quoting from a bunch of audio and music authors (that i recognize, dunno if you did) and Bob Adams picks it up a little - remember this: On Sunday, November 22, 2015 at 11:15:34 PM UTC-5, Max wrote:
> On Sun, 22 Nov 2015 16:01:48 -0800 (PST), radams2000@gmail.com wrote: > > >One thing that no one mentioned yet is that if you are reconstructing a time-domain signal using the ifft (let's assume an overlap of 50%), then the window needs to have the property that when you shift the window by half the frame length and add it to itself, the sum must be 1.0 in the overlapped region. Otherwise alias cancellation doesn't work. This assumes the only window is on the input side; it's also possible to window during the overlap-add operation and then the product of the input and output windows has to meet this condition. > > > >Anyway this rules out the majority of possible windows, unless you don't care about reconstruction. > > > >Bob > > Very good point, Bob! Which windows still qualify?
and that is the context where i picked it up. my response to that was technically spot on. point 1: a complementary power window will not add to 1 at 50% overlap, but i know that the Hann^2 window will add to 1 with 75% overlap. point 2: when doing perfect reconstruction using "STFT with subsequent IFFT" (hmmmm, where did those words come from?), there is no mention of DCT nor MDCT nor any of those other contexts where complementary power windows are used. do you know why? (think "perfectly correlated" vs. "uncorrelated".) if you use windows that satisfy the "Princen-Bradley condition" (which is a term i never ran across before, but it appears to simply mean "complementary power" windows), your "STFT with subsequent IFFT" operation will fail "perfect reconstruction" if the process in the frequency domain is the null process. (except, i seem to remember that Hann^2 and 75% overlap provides for perfect reconstruction.) the reason for that is math. not what you or i or Princen or Bradley say.
> > For those who read Max's second paragraph in the original post and follow that I am responding about Max's "STFT and subsequent IFFT (B-H used for both)", some actual examples of complementary windows can be found in Table I of: > > Audio Engineering Society > Convention Paper > Presented at the 119th Convention > 2005 October 7-10 New York, New York USA > > A New Class of Smooth Power > Complementary Windows and their > Application to Audio Signal Processing > Deepen Sinha > Anibal J. S. Ferreira > > at: > http://www.atc-labs.com/technology/misc/windows/docs/aes119_218_ds.pdf >
the paper is about lossy coding (hence references to MDCT, Vorbis, Dolby AAC) of audio. it's not about perfect reconstruction. and there is no indication from the OP that this person is dealing with lossy coding (where phase is tossed to the wind, audio at frame boundaries are uncorrelated, and complementary power windows are better than complementary amplitude). that is not in general what the "STFT and subsequent IFFT" is about in audio analysis and processing. sometimes people want in the output exactly what they put in the input in the case where the processing algorithm (in the STFT domain) is null. and if that is the case, complementary power windows along with 50% overlap will fail "perfect reconstruction". and that is because 1/sqrt(2) + 1/sqrt(2) does not add to 1. (but 1/2 + 1/2 does add to 1.) maybe the OP wants to do lossy coding with a DCT. i dunno. (and i dunno why he or she is referencing Xavier Serra, if that's the case.) r b-j
On Wednesday, November 25, 2015 at 6:47:49 PM UTC-8, robert bristow-johnson wrote:
> On Wednesday, November 25, 2015 at 8:50:53 PM UTC-5, dbd wrote: > > On Wednesday, November 25, 2015 at 3:55:54 PM UTC-8, robert bristow-johnson wrote: > > > On Wednesday, November 25, 2015 at 6:11:57 PM UTC-5, dbd wrote: > > ...
> > > well, i don't have access to that. but as i look around, it seems to have application to MDCT used in a codec like Dolby AAC. > > > > If you haven't found free access to one of these you are either too incompetent or too lazy to use google. But then, being to lazy to learn the context in a technical discussion -is- incompetent. > > > > For the similarly limited: > > /www.ece.lsu.edu/pub/aravena/CoGS_Literature/FilterBanks/SubbandCodinfg-Princen1987.pdf > > > > well, Dale, looks like you're back to grinding axes again. i almost never have luck getting an on-line copy of an IEEE paper (which was the first one you sited) from behind a pay wall. and my lazy habit is, that when i want to point someone to a reference on the web, i give them a URL. >
Luck has little to do with it. All of the papers I've mentioned in this thread except the 2010 comp.dsp paper are available behind pay walls. They are also all available for free on the web. Unfortunately, non-paywall urls can be transient and the urls alone don't provide enough information to allow identifying the papers. I think it is important for other readers of comp.dsp to understand that their access to information isn't subject to the limitations of your disabilities or personal choices. ...
> > The context is not crossfades. > > absolutely wrong. > > where the context is "perfect reconstruction" and "frames", it's about crossfades, whether you like that semantics or not. in fact i noticed Bob Adams using a similar terminology in another thread describing a fade out function as a half window. >
I have no problem with the semantics or terminology, but perfect reconstruction in frame based processing includes the effects of all processing, not just the input windowing.
> call it what you want, but these windows are a concatenation of a fade up function adjacent to a fade down function. and they are complementary if the fade up half and fade down half add, in some sense, to 1. > > now, in what sense might that be??
Yes, and there are at least 2 senses. Bob mentions one in the paragraph you quoted, the one with windowing only in the time domain where perfect reconstruction requires windows to be arithmetically complementary. There is another with two windows, one in time and one in frequency, which is the one the Princen-Bradley criteria refer to. Their frame-based perfect reconstruction algorithm is described in the first paper I cited. I'll repeat the reference below.
> > so the OP is quoting from a bunch of audio and music authors (that i recognize, dunno if you did) and Bob Adams picks it up a little - remember this: >
...
> > my response to that was technically spot on. > > point 1: a complementary power window will not add to 1 at 50% overlap, but i know that the Hann^2 window will add to 1 with 75% overlap.
Power complementary windows are not applied in a way that allows their effects to be determined by simple addition. You need to look at the frame processing algorithm they are used in and how they are applied in it.
> > point 2: when doing perfect reconstruction using "STFT with subsequent IFFT" (hmmmm, where did those words come from?), there is no mention of DCT nor MDCT nor any of those other contexts where complementary power windows are used. do you know why? (think "perfectly correlated" vs. "uncorrelated".) if you use windows that satisfy the "Princen-Bradley condition" (which is a term i never ran across before, but it appears to simply mean "complementary power" windows), your "STFT with subsequent IFFT" operation will fail "perfect reconstruction" if the process in the frequency domain is the null process. (except, i seem to remember that Hann^2 and 75% overlap provides for perfect reconstruction.) >
I'm glad you have some recollections and can identify some things that you don't remember, but you aren't doing well at inventing interpretations of them. You should try reading some of the citations for comprehension so we can discuss the same signal processes. You keep demonstrating that you haven't looked at the manner in which the power complementary windows are applied. Also, DCTs and DSTs -are- discrete Fourier transforms (DFT), have the short time property and have fast implementations, forward and reverse. They are often chosen as faster DFTs when signal processing goals don't require a CXin-CXout form of DFT or IDFT.
> the reason for that is math. not what you or i or Princen or Bradley say. > > > > > > For those who read Max's second paragraph in the original post and follow that I am responding about Max's "STFT and subsequent IFFT (B-H used for both)", some actual examples of complementary windows can be found in Table I of: > > > > Audio Engineering Society > > Convention Paper > > Presented at the 119th Convention > > 2005 October 7-10 New York, New York USA > > > > A New Class of Smooth Power > > Complementary Windows and their > > Application to Audio Signal Processing > > Deepen Sinha > > Anibal J. S. Ferreira > > > > at: > > http://www.atc-labs.com/technology/misc/windows/docs/aes119_218_ds.pdf > > > > the paper is about lossy coding (hence references to MDCT, Vorbis, Dolby AAC) of audio. it's not about perfect reconstruction. and there is no indication from the OP that this person is dealing with lossy coding (where phase is tossed to the wind, audio at frame boundaries are uncorrelated, and complementary power windows are better than complementary amplitude). >
I referred readers to Table I for examples of power complementary windows. The table reference is appropriate regardless of what types of processing the paper may describe.
> that is not in general what the "STFT and subsequent IFFT" is about in audio analysis and processing. sometimes people want in the output exactly what they put in the input in the case where the processing algorithm (in the STFT domain) is null. and if that is the case, complementary power windows along with 50% overlap will fail "perfect reconstruction". and that is because 1/sqrt(2) + 1/sqrt(2) does not add to 1. (but 1/2 + 1/2 does add to 1.)
"In General" includes many approaches to using FFT => IFFT to produce perfect reconstruction. You should read and learn about them. I've already provided some references. Here is a url for the Princen Bradley paper under discussion. http://www.sm.luth.se/csee/courses/sms/047/2004/lectures/MDCT_Princen96.pdf (accessed 20151128, if it fails go to google: J. P. Princen and A. B. Bradley "Analysis/synthesis filter bank design based on time domain aliasing cancellation", IEEE Trans. Acoust., Speech, Signal Processing, vol. 34, pp.1153 -1161 1986)
> > maybe the OP wants to do lossy coding with a DCT. i dunno. (and i dunno why he or she is referencing Xavier Serra, if that's the case.)
... The OP is free to chose what to do and learn about, including the choice of goals and selection of transforms and interconnections to achieve those goals. I've offered the OP access to a different signal processing architecture than Bob has. You don't seem to be able to recognize the difference. Dale B. Dalrymple
On Saturday, November 28, 2015 at 8:11:08 PM UTC-5, dbd wrote:
> On Wednesday, November 25, 2015 at 6:47:49 PM UTC-8, robert bristow-johnson wrote: > > On Wednesday, November 25, 2015 at 8:50:53 PM UTC-5, dbd wrote:
...
> > > > > > If you haven't found free access to one of these you are either too incompetent or too lazy to use google. But then, being to lazy to learn the context in a technical discussion -is- incompetent. > > > > > > For the similarly limited: > > > /www.ece.lsu.edu/pub/aravena/CoGS_Literature/FilterBanks/SubbandCodinfg-Princen1987.pdf > > > > > > > well, Dale, looks like you're back to grinding axes again. i almost never have luck getting an on-line copy of an IEEE paper (which was the first one you sited) from behind a pay wall. and my lazy habit is, that when i want to point someone to a reference on the web, i give them a URL. > > > > Luck has little to do with it. All of the papers I've mentioned in this thread except the 2010 comp.dsp paper are available behind pay walls. They are also all available for free on the web. Unfortunately, non-paywall urls can be transient and the urls alone don't provide enough information to allow identifying the papers.
and luck has "little" to do with that.
> ... > > > The context is not crossfades. > > > > absolutely wrong. > > > > where the context is "perfect reconstruction" and "frames", it's about crossfades, whether you like that semantics or not. in fact i noticed Bob Adams using a similar terminology in another thread describing a fade out function as a half window. > > > > I have no problem with the semantics or terminology, but perfect reconstruction in frame based processing includes the effects of all processing, not just the input windowing. >
absolutely correct. now, if there is other processing (in the frequency domain), not just the "null process", the all bets are off. you're not getting "perfect reconstruction". within that semantic (that is "perfect reconstruction"), i have been careful to always say with it "null process".
> > call it what you want, but these windows are a concatenation of a fade up function adjacent to a fade down function. and they are complementary if the fade up half and fade down half add, in some sense, to 1. > > > > now, in what sense might that be?? > > Yes, and there are at least 2 senses. Bob mentions one in the paragraph you quoted, the one with windowing only in the time domain where perfect reconstruction requires windows to be arithmetically complementary. There is another with two windows, one in time and one in frequency,
when applied to the frequency-domain data, that sure-as-hell ain't a "null process'. what i thought Bob Adams meant was applying the window twice to time-domain data. first on the outset (the "analysis window") and again at reconstruction (the "synthesis window").
> > Power complementary windows are not applied in a way that allows their effects to be determined by simple addition. You need to look at the frame processing algorithm they are used in and how they are applied in it. >
i am considering *only* the null processing algorithm and then i am testing for "perfect reconstruction", which i take to mean that the signal coming out of the whole damn thing is exactly the signal going in, with the possible exception of a *known* delay.
> > > > point 2: when doing perfect reconstruction using "STFT with subsequent IFFT" (hmmmm, where did those words come from?), there is no mention of DCT nor MDCT nor any of those other contexts where complementary power windows are used. do you know why? (think "perfectly correlated" vs. "uncorrelated".) if you use windows that satisfy the "Princen-Bradley condition" (which is a term i never ran across before, but it appears to simply mean "complementary power" windows), your "STFT with subsequent IFFT" operation will fail "perfect reconstruction" if the process in the frequency domain is the null process. (except, i seem to remember that Hann^2 and 75% overlap provides for perfect reconstruction.) > > > > I'm glad you have some recollections and can identify some things that you don't remember, but you aren't doing well at inventing interpretations of them. You should try reading some of the citations for comprehension so we can discuss the same signal processes.
> You keep demonstrating that you haven't looked at the manner in which the power complementary windows are applied. >
> Also, DCTs and DSTs -are- discrete Fourier transforms (DFT), have the short time property and have fast implementations, forward and reverse. They are often chosen as faster DFTs when signal processing goals don't require a CXin-CXout form of DFT or IDFT. >
but i think that DCT (or the DST which doesn't seem to be used much) tosses phase information to the wind. maybe i'm incorrect, but when i look it up: DCT: N-1 X[k] = SUM{ x[n] cos( (pi/2N) (2n+1) k ) } n=0 can't find the inverse DCT definition for the moment, which causes me to ask "is this perfectly invertable? does x[n] come back from X[k] with all frequency components lined up in phase as they were going in?" i have to confess that i don't know the answer. i had been thinking that the answer is "no". i had been thinking that the DCT sorta destroys phase information.
> > I referred readers to Table I for examples of power complementary windows. The table reference is appropriate regardless of what types of processing the paper may describe. >
no. *some* processing messes up the signal. the only exception i am aware of is the "null process" that does nothing to the signal.
> > that is not in general what the "STFT and subsequent IFFT" is about in audio analysis and processing. sometimes people want in the output exactly what they put in the input in the case where the processing algorithm (in the STFT domain) is null. and if that is the case, complementary power windows along with 50% overlap will fail "perfect reconstruction". and that is because 1/sqrt(2) + 1/sqrt(2) does not add to 1. (but 1/2 + 1/2 does add to 1.) > > "In General" includes many approaches to using FFT => IFFT to produce perfect reconstruction.
**if** the process in *null*.
> You should read and learn about them.
Dale, i confess i have never ever ever used the DCT. but 15 years ago and prior, i have done soooo many frame-based, windowed, "FFT => IFFT" apps that i have lost count. i have published (my only publication for an IEEE conference, feel free to look it up) one of those applications. i actually know something about this frame-by-frame frequency-domain processing stuff. usually i boil it down to "phase-vocoder" or "sinusoidal modeling". i have never done a lossy data-compression app like AAC or Vorbis. but, i am not relying on that to make the case. i *only* rely on the math to make a claim. one thing that i had not been addressing is, again, what Bob Adams brought up: yes, the fact is that if a complementary-power window is applied *twice* to the same data, that is effectively the same as a single application of a complementary-voltage window (which is the square of the complementary-power window) and that *will* result in perfect reconstruction, with 50% overlap, if the process in-between is the null process. maybe this is where we are talking past each other, Dale. but if one window is applied to frequency-domain data while the other is applied to time domain data, then all bets are off. *unless* the window is Gaussian (see my IEEE paper). but Gaussian windows are not complementary in any sense that i am aware of. so, in *my* paper, we applied the Gaussian window as the "analysis window", did our processing in the frequency domain in a manner which preserves the Gaussian window property and which adjusts phase in such a manner that the phases of each frequency component is aligned at the frame boundaries. after transforming back to the time domain, i multiply by a Hann (or "flattened-Hann") window and divide by the Gaussian window (i team that up into a single synthesis window definition), which leaves the data in a complementary-add state. then lastly, overlap-add with 50% overlap. it is also possible to multiply the input data with a Hann window (which is *not* complementary-power), transform into frequency domain, do phase-preserving processing, transform back, apply *another* Hann window and overlap-add with 75% overlap and get perfect reconstruction. Miller Puckette and Jean Laroche did that, but i never duplicated the work. however i once did the overlap and summation of the Hann^2 window to see that it did indeed add to a constant. try that out, Dale. i also know that when splicing uncorrelated data together, to avoid a little "dip" in energy at the middle of the splice, that complementary-power splice functions are needed. but when the data is perfectly correlated (which means that all frequency components are phase aligned), a complementary-voltage splice function should be used. in the music-dsp mailing list, i wrote about what to do if the (normalized) correlation is somewhere in-between zero and 1. you get a window that is something in-between a complementary-power and complementary-voltage window. another guy, Olli Niemitalo (from Norway, i think), picked up on that because he did similar research. r b-j
On Monday, November 30, 2015 at 2:23:34 PM UTC-8, robert bristow-johnson wrote:
> On Saturday, November 28, 2015 at 8:11:08 PM UTC-5, dbd wrote: > > On Wednesday, November 25, 2015 at 6:47:49 PM UTC-8, robert bristow-johnson
...
> > I'm glad you have some recollections and can identify some things that you don't remember, but you aren't doing well at inventing interpretations of them. You should try reading some of the citations for comprehension so we can discuss the same signal processes. > > > You keep demonstrating that you haven't looked at the manner in which the power complementary windows are applied. > > > > > Also, DCTs and DSTs -are- discrete Fourier transforms (DFT), have the short time property and have fast implementations, forward and reverse. They are often chosen as faster DFTs when signal processing goals don't require a CXin-CXout form of DFT or IDFT. > > > > but i think that DCT (or the DST which doesn't seem to be used much) tosses phase information to the wind. maybe i'm incorrect, but when i look it up: > DCT: > N-1 > X[k] = SUM{ x[n] cos( (pi/2N) (2n+1) k ) } > n=0 > can't find the inverse DCT definition for the moment, which causes me to ask "is this perfectly invertable? does x[n] come back from X[k] with all frequency components lined up in phase as they were going in?"
The last time we discussed the DCT I think you knew the answer to this. The proper choices of DCT-IDCT pairs give perfect reconstruction.
> > i have to confess that i don't know the answer. i had been thinking that the answer is "no". i had been thinking that the DCT sorta destroys phase information.
In the past you would have found the answer before posting. Your former capacity for informed commentary is sorely missed.
>
...
> > "In General" includes many approaches to using FFT => IFFT to produce perfect reconstruction. > > **if** the process in *null*.
That's what perfect reconstruction means.
> > You should read and learn about them. > > Dale, i confess i have never ever ever used the DCT. but 15 years ago and prior, i have done soooo many frame-based, windowed, "FFT => IFFT" apps that i have lost count. i have published (my only publication for an IEEE conference, feel free to look it up) one of those applications. i actually know something about this frame-by-frame frequency-domain processing stuff.
We have both used these processes for many years. For three decades I've used them to generate precise test signals in dynamic signal analysers and to wrap prefect reconstruction t>f and f>t wrappers around frequency domain beamformers.
> usually i boil it down to "phase-vocoder" or "sinusoidal modeling". i have never done a lossy data-compression app like AAC or Vorbis. but, i am not relying on that to make the case. i *only* rely on the math to make a claim. >
The general field of LOTs, lapped orthogonal transforms, now includes many arctitectures for overlap-add processing using DFTs, DCTs and/or DSTs and their inverses that have the perfect reconstruction property besides the form you have used in the "phase-vocoder". The maths have been done and are available but only to those who can and do choose to read them. ...
> maybe this is where we are talking past each other, Dale. >
Where we talk past each other is that you are casting your interpretation and remarks onto the overlap add structure you used long ago in "phase-vocoders". That structure is not the context in which power complementary windows are used. As long as you fail to deal with the appropriate architechtures, the only resort for everyone discussing power-complementary windows is to talk and listen past the noises you are making in terms of your memories of "phase-vocoders". Dale B. Dalrymple
On Mon, 30 Nov 2015 14:23:29 -0800 (PST), robert bristow-johnson
<rbj@audioimagination.com> wrote:


Hey guys.  I wanted to check back to say thanks for the info, even
though I'll need some study time to follow the full extent of the
conversation.  I didn't mean to cause any dissent here, but I've also
picked up a couple things from the exchange, so there.

My initial question was probably relatively simple. I wanted to find
out which window types were suited to breaking an audio stream into
short groups of spectral factors via STFT, but with the option to
reassemble the stream after subsequent IFFT. So 'overlap add' is the
usual term, I believe.  Not sure about the difference between the two
approaches that have been mentioned here, so I'll need to read up on
that.

Some of Xavier Serra's approach has commonality with Julius O Smith,
of course. They've coauthored some papers.  So I've found some more
info in JOS's "Spectral" book, which I'll be going through.

Anyway, thanks. I didn't want you  to think I had ignored the
additional followups in the thread. Just trying to catch up is all.