Hacker News new | comments | show | ask | jobs | submit login

Well, and in fact the wavelet-based JPEG-2000 standard does produce better visual quality at smaller file size than the DFT-based JPEG standard. JPEG-2000 compressors haven't been as broadly adopted, probably because they're slower. I'm not sure if that's because of wavelets, because of unoptimized implementations of JPEG-2000, or because some other aspect of the JPEG-2000 standard makes it more computationally intensive. There's further discussion at https://lobste.rs/s/oxuu1k/jpeg_2000_the_better_alternative_....



It's probably slower because of the wavelets.

You'll often hear that wavelets are faster than Fourier based methods, because the DWT is O(n) while FFT is O(n log(n)), but that doesn't hold up when the FFT window is a fixed size (as in the case of JPEG). Since JPEG uses a window size of 8px, the "log n" is 3, whereas JPEG 2000 uses a wavelet with 8 coefficients. So it's basically O(n) in both cases, with a much larger constant factor on JPEG 2000's DWT.


Doesn't JPEG use a window size of 8×8 = 64px, with a sort of boustrophedon diagonal path through those 64 pixels?


Yes to the first part, but both the DWT and DCT are linearly separable, so you can compare their performance by just looking at one dimension.

As for the zigzag path, it doesn't run through the 64 pixels, but rather through the 64 coefficients that come out of the DCT. That's part of the coding stage, which is a separate animal. That said, JPEG's coding scheme is way simpler, and probably way faster than JPEG 2000's, so that might also have something to do with the perf difference.

My overall impression of JPEG 2000 is that it seems like they started with "wavelets can surely be used to achieve superior image compression to JPEG", and then made whatever sacrifices were necessary, in terms of implementation complexity and computational resources, to bear out that premise. You could make similar sacrifices and beat the hell out of JPEG with a DCT-based codec too (e.g. with larger windows, overlapping transforms, better psychovisual models etc.)


I see! That makes more sense. Thank you for explaining?

A linearly-separated 2D 8×8 DFT should involve 6 butterflies per pixel, no? And I guess a 2D DWT with 8 coefficients results in 16 multiply-accumulates per pixel? Does it depend on the order of the DWT?


> Does it depend on the order of the DWT?

Not sure what you mean by the order of the DWT. Do you mean how deep you transform (how many levels of the transform you apply), or the number of taps on the filter?

It definitely depends on the number of taps on the filter.

IIRC, as for the depth DWT down to one scale coefficient is still O(N) (although there's another constant factor multiplied on beyond the number of filter coefficients). Even if you only do two levels, I believe you only cut the time in half, and you lose a lot of opportunity for compression if you do. I'm not sure what depth JPEG 2000 goes to, but it's probably more than three, so I think you can largely discount any perf gains there.


I meant the depth. Thank you!




Guidelines | FAQ | Support | API | Security | Lists | Bookmarklet | Legal | Apply to YC | Contact

Search: