| // Copyright 2014 Google Inc. All Rights Reserved. |
| // |
| // Use of this source code is governed by a BSD-style license |
| // that can be found in the COPYING file in the root of the source |
| // tree. An additional intellectual property rights grant can be found |
| // in the file PATENTS. All contributing project authors may |
| // be found in the AUTHORS file in the root of the source tree. |
| // ----------------------------------------------------------------------------- |
| // |
| // SSE2 variant of methods for lossless decoder |
| // |
| // Author: Skal (pascal.massimino@gmail.com) |
| |
| #include "third_party/chromium_headless/libwebp/dsp/dsp.h" |
| |
| #if defined(WEBP_USE_SSE2) |
| #include <assert.h> |
| #include <emmintrin.h> |
| #include "third_party/chromium_headless/libwebp/dsp/lossless.h" |
| |
| //------------------------------------------------------------------------------ |
| // Predictor Transform |
| |
| static WEBP_INLINE uint32_t ClampedAddSubtractFull(uint32_t c0, uint32_t c1, |
| uint32_t c2) { |
| const __m128i zero = _mm_setzero_si128(); |
| const __m128i C0 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(c0), zero); |
| const __m128i C1 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(c1), zero); |
| const __m128i C2 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(c2), zero); |
| const __m128i V1 = _mm_add_epi16(C0, C1); |
| const __m128i V2 = _mm_sub_epi16(V1, C2); |
| const __m128i b = _mm_packus_epi16(V2, V2); |
| const uint32_t output = _mm_cvtsi128_si32(b); |
| return output; |
| } |
| |
| static WEBP_INLINE uint32_t ClampedAddSubtractHalf(uint32_t c0, uint32_t c1, |
| uint32_t c2) { |
| const __m128i zero = _mm_setzero_si128(); |
| const __m128i C0 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(c0), zero); |
| const __m128i C1 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(c1), zero); |
| const __m128i B0 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(c2), zero); |
| const __m128i avg = _mm_add_epi16(C1, C0); |
| const __m128i A0 = _mm_srli_epi16(avg, 1); |
| const __m128i A1 = _mm_sub_epi16(A0, B0); |
| const __m128i BgtA = _mm_cmpgt_epi16(B0, A0); |
| const __m128i A2 = _mm_sub_epi16(A1, BgtA); |
| const __m128i A3 = _mm_srai_epi16(A2, 1); |
| const __m128i A4 = _mm_add_epi16(A0, A3); |
| const __m128i A5 = _mm_packus_epi16(A4, A4); |
| const uint32_t output = _mm_cvtsi128_si32(A5); |
| return output; |
| } |
| |
| static WEBP_INLINE uint32_t Select(uint32_t a, uint32_t b, uint32_t c) { |
| int pa_minus_pb; |
| const __m128i zero = _mm_setzero_si128(); |
| const __m128i A0 = _mm_cvtsi32_si128(a); |
| const __m128i B0 = _mm_cvtsi32_si128(b); |
| const __m128i C0 = _mm_cvtsi32_si128(c); |
| const __m128i AC0 = _mm_subs_epu8(A0, C0); |
| const __m128i CA0 = _mm_subs_epu8(C0, A0); |
| const __m128i BC0 = _mm_subs_epu8(B0, C0); |
| const __m128i CB0 = _mm_subs_epu8(C0, B0); |
| const __m128i AC = _mm_or_si128(AC0, CA0); |
| const __m128i BC = _mm_or_si128(BC0, CB0); |
| const __m128i pa = _mm_unpacklo_epi8(AC, zero); // |a - c| |
| const __m128i pb = _mm_unpacklo_epi8(BC, zero); // |b - c| |
| const __m128i diff = _mm_sub_epi16(pb, pa); |
| { |
| int16_t out[8]; |
| _mm_storeu_si128((__m128i*)out, diff); |
| pa_minus_pb = out[0] + out[1] + out[2] + out[3]; |
| } |
| return (pa_minus_pb <= 0) ? a : b; |
| } |
| |
| static WEBP_INLINE __m128i Average2_128i(uint32_t a0, uint32_t a1) { |
| const __m128i zero = _mm_setzero_si128(); |
| const __m128i A0 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(a0), zero); |
| const __m128i A1 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(a1), zero); |
| const __m128i sum = _mm_add_epi16(A1, A0); |
| const __m128i avg = _mm_srli_epi16(sum, 1); |
| return avg; |
| } |
| |
| static WEBP_INLINE uint32_t Average2(uint32_t a0, uint32_t a1) { |
| const __m128i avg = Average2_128i(a0, a1); |
| const __m128i A2 = _mm_packus_epi16(avg, avg); |
| const uint32_t output = _mm_cvtsi128_si32(A2); |
| return output; |
| } |
| |
| static WEBP_INLINE uint32_t Average3(uint32_t a0, uint32_t a1, uint32_t a2) { |
| const __m128i zero = _mm_setzero_si128(); |
| const __m128i avg1 = Average2_128i(a0, a2); |
| const __m128i A1 = _mm_unpacklo_epi8(_mm_cvtsi32_si128(a1), zero); |
| const __m128i sum = _mm_add_epi16(avg1, A1); |
| const __m128i avg2 = _mm_srli_epi16(sum, 1); |
| const __m128i A2 = _mm_packus_epi16(avg2, avg2); |
| const uint32_t output = _mm_cvtsi128_si32(A2); |
| return output; |
| } |
| |
| static WEBP_INLINE uint32_t Average4(uint32_t a0, uint32_t a1, |
| uint32_t a2, uint32_t a3) { |
| const __m128i avg1 = Average2_128i(a0, a1); |
| const __m128i avg2 = Average2_128i(a2, a3); |
| const __m128i sum = _mm_add_epi16(avg2, avg1); |
| const __m128i avg3 = _mm_srli_epi16(sum, 1); |
| const __m128i A0 = _mm_packus_epi16(avg3, avg3); |
| const uint32_t output = _mm_cvtsi128_si32(A0); |
| return output; |
| } |
| |
| static uint32_t Predictor5(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = Average3(left, top[0], top[1]); |
| return pred; |
| } |
| static uint32_t Predictor6(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = Average2(left, top[-1]); |
| return pred; |
| } |
| static uint32_t Predictor7(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = Average2(left, top[0]); |
| return pred; |
| } |
| static uint32_t Predictor8(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = Average2(top[-1], top[0]); |
| (void)left; |
| return pred; |
| } |
| static uint32_t Predictor9(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = Average2(top[0], top[1]); |
| (void)left; |
| return pred; |
| } |
| static uint32_t Predictor10(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = Average4(left, top[-1], top[0], top[1]); |
| return pred; |
| } |
| static uint32_t Predictor11(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = Select(top[0], left, top[-1]); |
| return pred; |
| } |
| static uint32_t Predictor12(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = ClampedAddSubtractFull(left, top[0], top[-1]); |
| return pred; |
| } |
| static uint32_t Predictor13(uint32_t left, const uint32_t* const top) { |
| const uint32_t pred = ClampedAddSubtractHalf(left, top[0], top[-1]); |
| return pred; |
| } |
| |
| //------------------------------------------------------------------------------ |
| // Subtract-Green Transform |
| |
| static void AddGreenToBlueAndRed(uint32_t* argb_data, int num_pixels) { |
| int i; |
| for (i = 0; i + 4 <= num_pixels; i += 4) { |
| const __m128i in = _mm_loadu_si128((__m128i*)&argb_data[i]); // argb |
| const __m128i A = _mm_srli_epi16(in, 8); // 0 a 0 g |
| const __m128i B = _mm_shufflelo_epi16(A, _MM_SHUFFLE(2, 2, 0, 0)); |
| const __m128i C = _mm_shufflehi_epi16(B, _MM_SHUFFLE(2, 2, 0, 0)); // 0g0g |
| const __m128i out = _mm_add_epi8(in, C); |
| _mm_storeu_si128((__m128i*)&argb_data[i], out); |
| } |
| // fallthrough and finish off with plain-C |
| VP8LAddGreenToBlueAndRed_C(argb_data + i, num_pixels - i); |
| } |
| |
| //------------------------------------------------------------------------------ |
| // Color Transform |
| |
| static void TransformColorInverse(const VP8LMultipliers* const m, |
| uint32_t* argb_data, int num_pixels) { |
| // sign-extended multiplying constants, pre-shifted by 5. |
| #define CST(X) (((int16_t)(m->X << 8)) >> 5) // sign-extend |
| const __m128i mults_rb = _mm_set_epi16( |
| CST(green_to_red_), CST(green_to_blue_), |
| CST(green_to_red_), CST(green_to_blue_), |
| CST(green_to_red_), CST(green_to_blue_), |
| CST(green_to_red_), CST(green_to_blue_)); |
| const __m128i mults_b2 = _mm_set_epi16( |
| CST(red_to_blue_), 0, CST(red_to_blue_), 0, |
| CST(red_to_blue_), 0, CST(red_to_blue_), 0); |
| #undef CST |
| const __m128i mask_ag = _mm_set1_epi32(0xff00ff00); // alpha-green masks |
| int i; |
| for (i = 0; i + 4 <= num_pixels; i += 4) { |
| const __m128i in = _mm_loadu_si128((__m128i*)&argb_data[i]); // argb |
| const __m128i A = _mm_and_si128(in, mask_ag); // a 0 g 0 |
| const __m128i B = _mm_shufflelo_epi16(A, _MM_SHUFFLE(2, 2, 0, 0)); |
| const __m128i C = _mm_shufflehi_epi16(B, _MM_SHUFFLE(2, 2, 0, 0)); // g0g0 |
| const __m128i D = _mm_mulhi_epi16(C, mults_rb); // x dr x db1 |
| const __m128i E = _mm_add_epi8(in, D); // x r' x b' |
| const __m128i F = _mm_slli_epi16(E, 8); // r' 0 b' 0 |
| const __m128i G = _mm_mulhi_epi16(F, mults_b2); // x db2 0 0 |
| const __m128i H = _mm_srli_epi32(G, 8); // 0 x db2 0 |
| const __m128i I = _mm_add_epi8(H, F); // r' x b'' 0 |
| const __m128i J = _mm_srli_epi16(I, 8); // 0 r' 0 b'' |
| const __m128i out = _mm_or_si128(J, A); |
| _mm_storeu_si128((__m128i*)&argb_data[i], out); |
| } |
| // Fall-back to C-version for left-overs. |
| VP8LTransformColorInverse_C(m, argb_data + i, num_pixels - i); |
| } |
| |
| //------------------------------------------------------------------------------ |
| // Color-space conversion functions |
| |
| static void ConvertBGRAToRGBA(const uint32_t* src, |
| int num_pixels, uint8_t* dst) { |
| const __m128i* in = (const __m128i*)src; |
| __m128i* out = (__m128i*)dst; |
| while (num_pixels >= 8) { |
| const __m128i bgra0 = _mm_loadu_si128(in++); // bgra0|bgra1|bgra2|bgra3 |
| const __m128i bgra4 = _mm_loadu_si128(in++); // bgra4|bgra5|bgra6|bgra7 |
| const __m128i v0l = _mm_unpacklo_epi8(bgra0, bgra4); // b0b4g0g4r0r4a0a4... |
| const __m128i v0h = _mm_unpackhi_epi8(bgra0, bgra4); // b2b6g2g6r2r6a2a6... |
| const __m128i v1l = _mm_unpacklo_epi8(v0l, v0h); // b0b2b4b6g0g2g4g6... |
| const __m128i v1h = _mm_unpackhi_epi8(v0l, v0h); // b1b3b5b7g1g3g5g7... |
| const __m128i v2l = _mm_unpacklo_epi8(v1l, v1h); // b0...b7 | g0...g7 |
| const __m128i v2h = _mm_unpackhi_epi8(v1l, v1h); // r0...r7 | a0...a7 |
| const __m128i ga0 = _mm_unpackhi_epi64(v2l, v2h); // g0...g7 | a0...a7 |
| const __m128i rb0 = _mm_unpacklo_epi64(v2h, v2l); // r0...r7 | b0...b7 |
| const __m128i rg0 = _mm_unpacklo_epi8(rb0, ga0); // r0g0r1g1 ... r6g6r7g7 |
| const __m128i ba0 = _mm_unpackhi_epi8(rb0, ga0); // b0a0b1a1 ... b6a6b7a7 |
| const __m128i rgba0 = _mm_unpacklo_epi16(rg0, ba0); // rgba0|rgba1... |
| const __m128i rgba4 = _mm_unpackhi_epi16(rg0, ba0); // rgba4|rgba5... |
| _mm_storeu_si128(out++, rgba0); |
| _mm_storeu_si128(out++, rgba4); |
| num_pixels -= 8; |
| } |
| // left-overs |
| VP8LConvertBGRAToRGBA_C((const uint32_t*)in, num_pixels, (uint8_t*)out); |
| } |
| |
| static void ConvertBGRAToRGBA4444(const uint32_t* src, |
| int num_pixels, uint8_t* dst) { |
| const __m128i mask_0x0f = _mm_set1_epi8(0x0f); |
| const __m128i mask_0xf0 = _mm_set1_epi8(0xf0); |
| const __m128i* in = (const __m128i*)src; |
| __m128i* out = (__m128i*)dst; |
| while (num_pixels >= 8) { |
| const __m128i bgra0 = _mm_loadu_si128(in++); // bgra0|bgra1|bgra2|bgra3 |
| const __m128i bgra4 = _mm_loadu_si128(in++); // bgra4|bgra5|bgra6|bgra7 |
| const __m128i v0l = _mm_unpacklo_epi8(bgra0, bgra4); // b0b4g0g4r0r4a0a4... |
| const __m128i v0h = _mm_unpackhi_epi8(bgra0, bgra4); // b2b6g2g6r2r6a2a6... |
| const __m128i v1l = _mm_unpacklo_epi8(v0l, v0h); // b0b2b4b6g0g2g4g6... |
| const __m128i v1h = _mm_unpackhi_epi8(v0l, v0h); // b1b3b5b7g1g3g5g7... |
| const __m128i v2l = _mm_unpacklo_epi8(v1l, v1h); // b0...b7 | g0...g7 |
| const __m128i v2h = _mm_unpackhi_epi8(v1l, v1h); // r0...r7 | a0...a7 |
| const __m128i ga0 = _mm_unpackhi_epi64(v2l, v2h); // g0...g7 | a0...a7 |
| const __m128i rb0 = _mm_unpacklo_epi64(v2h, v2l); // r0...r7 | b0...b7 |
| const __m128i ga1 = _mm_srli_epi16(ga0, 4); // g0-|g1-|...|a6-|a7- |
| const __m128i rb1 = _mm_and_si128(rb0, mask_0xf0); // -r0|-r1|...|-b6|-a7 |
| const __m128i ga2 = _mm_and_si128(ga1, mask_0x0f); // g0-|g1-|...|a6-|a7- |
| const __m128i rgba0 = _mm_or_si128(ga2, rb1); // rg0..rg7 | ba0..ba7 |
| const __m128i rgba1 = _mm_srli_si128(rgba0, 8); // ba0..ba7 | 0 |
| #ifdef WEBP_SWAP_16BIT_CSP |
| const __m128i rgba = _mm_unpacklo_epi8(rgba1, rgba0); // barg0...barg7 |
| #else |
| const __m128i rgba = _mm_unpacklo_epi8(rgba0, rgba1); // rgba0...rgba7 |
| #endif |
| _mm_storeu_si128(out++, rgba); |
| num_pixels -= 8; |
| } |
| // left-overs |
| VP8LConvertBGRAToRGBA4444_C((const uint32_t*)in, num_pixels, (uint8_t*)out); |
| } |
| |
| static void ConvertBGRAToRGB565(const uint32_t* src, |
| int num_pixels, uint8_t* dst) { |
| const __m128i mask_0xe0 = _mm_set1_epi8(0xe0); |
| const __m128i mask_0xf8 = _mm_set1_epi8(0xf8); |
| const __m128i mask_0x07 = _mm_set1_epi8(0x07); |
| const __m128i* in = (const __m128i*)src; |
| __m128i* out = (__m128i*)dst; |
| while (num_pixels >= 8) { |
| const __m128i bgra0 = _mm_loadu_si128(in++); // bgra0|bgra1|bgra2|bgra3 |
| const __m128i bgra4 = _mm_loadu_si128(in++); // bgra4|bgra5|bgra6|bgra7 |
| const __m128i v0l = _mm_unpacklo_epi8(bgra0, bgra4); // b0b4g0g4r0r4a0a4... |
| const __m128i v0h = _mm_unpackhi_epi8(bgra0, bgra4); // b2b6g2g6r2r6a2a6... |
| const __m128i v1l = _mm_unpacklo_epi8(v0l, v0h); // b0b2b4b6g0g2g4g6... |
| const __m128i v1h = _mm_unpackhi_epi8(v0l, v0h); // b1b3b5b7g1g3g5g7... |
| const __m128i v2l = _mm_unpacklo_epi8(v1l, v1h); // b0...b7 | g0...g7 |
| const __m128i v2h = _mm_unpackhi_epi8(v1l, v1h); // r0...r7 | a0...a7 |
| const __m128i ga0 = _mm_unpackhi_epi64(v2l, v2h); // g0...g7 | a0...a7 |
| const __m128i rb0 = _mm_unpacklo_epi64(v2h, v2l); // r0...r7 | b0...b7 |
| const __m128i rb1 = _mm_and_si128(rb0, mask_0xf8); // -r0..-r7|-b0..-b7 |
| const __m128i g_lo1 = _mm_srli_epi16(ga0, 5); |
| const __m128i g_lo2 = _mm_and_si128(g_lo1, mask_0x07); // g0-...g7-|xx (3b) |
| const __m128i g_hi1 = _mm_slli_epi16(ga0, 3); |
| const __m128i g_hi2 = _mm_and_si128(g_hi1, mask_0xe0); // -g0...-g7|xx (3b) |
| const __m128i b0 = _mm_srli_si128(rb1, 8); // -b0...-b7|0 |
| const __m128i rg1 = _mm_or_si128(rb1, g_lo2); // gr0...gr7|xx |
| const __m128i b1 = _mm_srli_epi16(b0, 3); |
| const __m128i gb1 = _mm_or_si128(b1, g_hi2); // bg0...bg7|xx |
| #ifdef WEBP_SWAP_16BIT_CSP |
| const __m128i rgba = _mm_unpacklo_epi8(gb1, rg1); // rggb0...rggb7 |
| #else |
| const __m128i rgba = _mm_unpacklo_epi8(rg1, gb1); // bgrb0...bgrb7 |
| #endif |
| _mm_storeu_si128(out++, rgba); |
| num_pixels -= 8; |
| } |
| // left-overs |
| VP8LConvertBGRAToRGB565_C((const uint32_t*)in, num_pixels, (uint8_t*)out); |
| } |
| |
| static void ConvertBGRAToBGR(const uint32_t* src, |
| int num_pixels, uint8_t* dst) { |
| const __m128i mask_l = _mm_set_epi32(0, 0x00ffffff, 0, 0x00ffffff); |
| const __m128i mask_h = _mm_set_epi32(0x00ffffff, 0, 0x00ffffff, 0); |
| const __m128i* in = (const __m128i*)src; |
| const uint8_t* const end = dst + num_pixels * 3; |
| // the last storel_epi64 below writes 8 bytes starting at offset 18 |
| while (dst + 26 <= end) { |
| const __m128i bgra0 = _mm_loadu_si128(in++); // bgra0|bgra1|bgra2|bgra3 |
| const __m128i bgra4 = _mm_loadu_si128(in++); // bgra4|bgra5|bgra6|bgra7 |
| const __m128i a0l = _mm_and_si128(bgra0, mask_l); // bgr0|0|bgr0|0 |
| const __m128i a4l = _mm_and_si128(bgra4, mask_l); // bgr0|0|bgr0|0 |
| const __m128i a0h = _mm_and_si128(bgra0, mask_h); // 0|bgr0|0|bgr0 |
| const __m128i a4h = _mm_and_si128(bgra4, mask_h); // 0|bgr0|0|bgr0 |
| const __m128i b0h = _mm_srli_epi64(a0h, 8); // 000b|gr00|000b|gr00 |
| const __m128i b4h = _mm_srli_epi64(a4h, 8); // 000b|gr00|000b|gr00 |
| const __m128i c0 = _mm_or_si128(a0l, b0h); // rgbrgb00|rgbrgb00 |
| const __m128i c4 = _mm_or_si128(a4l, b4h); // rgbrgb00|rgbrgb00 |
| const __m128i c2 = _mm_srli_si128(c0, 8); |
| const __m128i c6 = _mm_srli_si128(c4, 8); |
| _mm_storel_epi64((__m128i*)(dst + 0), c0); |
| _mm_storel_epi64((__m128i*)(dst + 6), c2); |
| _mm_storel_epi64((__m128i*)(dst + 12), c4); |
| _mm_storel_epi64((__m128i*)(dst + 18), c6); |
| dst += 24; |
| num_pixels -= 8; |
| } |
| // left-overs |
| VP8LConvertBGRAToBGR_C((const uint32_t*)in, num_pixels, dst); |
| } |
| |
| //------------------------------------------------------------------------------ |
| // Entry point |
| |
| extern void VP8LDspInitSSE2(void); |
| |
| WEBP_TSAN_IGNORE_FUNCTION void VP8LDspInitSSE2(void) { |
| VP8LPredictors[5] = Predictor5; |
| VP8LPredictors[6] = Predictor6; |
| VP8LPredictors[7] = Predictor7; |
| VP8LPredictors[8] = Predictor8; |
| VP8LPredictors[9] = Predictor9; |
| VP8LPredictors[10] = Predictor10; |
| VP8LPredictors[11] = Predictor11; |
| VP8LPredictors[12] = Predictor12; |
| VP8LPredictors[13] = Predictor13; |
| |
| VP8LAddGreenToBlueAndRed = AddGreenToBlueAndRed; |
| VP8LTransformColorInverse = TransformColorInverse; |
| |
| VP8LConvertBGRAToRGBA = ConvertBGRAToRGBA; |
| VP8LConvertBGRAToRGBA4444 = ConvertBGRAToRGBA4444; |
| VP8LConvertBGRAToRGB565 = ConvertBGRAToRGB565; |
| VP8LConvertBGRAToBGR = ConvertBGRAToBGR; |
| } |
| |
| #else // !WEBP_USE_SSE2 |
| |
| WEBP_DSP_INIT_STUB(VP8LDspInitSSE2) |
| |
| #endif // WEBP_USE_SSE2 |