[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: psychoacoustically driven temporal approximation

Hello Alberto,

I'm not entirely sure, but this sounds a bit like my thoughts when I
started my Ph.D. many years ago: can I synthesize an audio waveform
from a perceptual representation of another audio signal?  (and what
does that imply about that particular perceptual representation?)

My answer was to so by iterative resynthesis: make a first good guess
of the inverse perceptual transform, then correct for the error.  Key
point is that the correction needs to go in the right direction.  My
perceptual transform was a set of sparsely sampled hilbert envelopes
of the outputs of a gammatone filterbank.

If you want you can have a look at my thesis "A Sparse Auditory
Envelope Representation with Iterative Reconstruction for Audio
Coding", linked from my homepage
(http://jthiem.bitbucket.org/research.html), you can find the MATLAB
code on that page too. Of course, in my thesis I refer to work that
others have done in a similar vein.


On 4 March 2014 12:05, JesterN Alberto Novello <jestern77@xxxxxxxx> wrote:
> Hi all,
> i'm trying to find a way to approximate the sample values of an audio
> waveform in time domain.
> I want a method that takes care of approximating perceptually-relevant audio
> bands better than others.
> Basically a spectral-weighted temporal approximation method.
> In my head it's not clear how to connect frequency components to specific
> samples in the time domain.
> Any DSP wizard out there with a good idea/papers ?
> Best regards
> Alberto

Joachim Thiemann :: http://jthiem.bitbucket.org ::