On Wed, 17 Aug 2022, Ronald S. Bultje wrote:
On Wed, Aug 17, 2022 at 4:32 PM Martin Storsjö <mar...@martin.st> wrote:
This avoids overflows on some inputs in the x86 case, where the
assembly version would clip/overflow differently from the
C reference function.
This doesn't seem to be a real issue with actual input data, but
only with the previous fully random input data.
I'm a bit scared of this change... If we can trigger overflows with specific
pixel patterns, doesn't that make FFmpeg input-data exploitable? Think of
how that would go with corporate users with user-provided input data.
No, this most probably isn't a real issue with actual filters - it's only
that the current checkasm test was overly simplistic.
The input to this DSP function isn't raw user-provided input pixels, but
16 bit integers produced as the output of the first (horizontal) scaling
step. Yesterday when I wrote the patch, I hadn't checked exactly what the
range of those values were, and I assumed it wasn't the whole int16_t
range - but apparently they can range at least up to max int16_t. (They
most probably can't range down to the minimum negative int16_t though -
simulating that aspect would be nice too.)
The filter coefficients should add up to 4096. The input sample range is
15 bits (plus sign), and filter coefficients add a total magnitude of 12
bits, giving a total range of 27 bits (plus sign). After shifting down by
19 bits at the end, this produces 8 bits output (which is clipped).
The critical stage here is the 27 bits, where there's still plenty of
headroom (4 bits) for filter overshoot - with a real filter.
However in the current test, the filter coefficients are just plain random
int16_t values in the whole range - and that can easily cause overflows in
the intermediates.
So I guess we shouldn't scale down the input "pixels" here as they
actually can use the whole range up to max int16_t (but ideally, they
wouldn't range further down below zero than what you'd get from the
maximal negative filter overshoot either), but we should scale down the
fully random filter coefficients, so that they can't overflow even if they
all happen to align in the worst way.
Alternatively we could construct a more realistic test filter, e.g.
something like what's used in the hscale test. There, if the filter should
add up to 1<<F, and we have N filter coefficients, we have all of them but
one be set to -((1<<F)/(N-1)) and one set to ((1<<(F+1)) - 1). It doesn't
look much like a real actual filter, but keeps most properties - it adds
up to the right sum and doesn't trigger unreal overflows and it produces
both positive and negative values.
Anyway, I'll update the patch and make a clearer comment for it.
// Martin
_______________________________________________
ffmpeg-devel mailing list
ffmpeg-devel@ffmpeg.org
https://ffmpeg.org/mailman/listinfo/ffmpeg-devel
To unsubscribe, visit link above, or email
ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe".