mirror of
				https://github.com/webmproject/libwebp.git
				synced 2025-10-31 10:25:46 +01:00 
			
		
		
		
	Merge "Implement lossless transforms in NEON."
This commit is contained in:
		| @@ -139,6 +139,249 @@ static void ConvertBGRAToRGB(const uint32_t* src, | ||||
|  | ||||
| #endif   // !WORK_AROUND_GCC | ||||
|  | ||||
|  | ||||
| //------------------------------------------------------------------------------ | ||||
| // Predictor Transform | ||||
|  | ||||
| #define LOAD_U32_AS_U8(IN) vreinterpret_u8_u32(vdup_n_u32((IN))) | ||||
| #define LOAD_U32P_AS_U8(IN) vreinterpret_u8_u32(vld1_u32((IN))) | ||||
| #define LOADQ_U32_AS_U8(IN) vreinterpretq_u8_u32(vdupq_n_u32((IN))) | ||||
| #define LOADQ_U32P_AS_U8(IN) vreinterpretq_u8_u32(vld1q_u32((IN))) | ||||
| #define GET_U8_AS_U32(IN) vget_lane_u32(vreinterpret_u32_u8((IN)), 0); | ||||
| #define GETQ_U8_AS_U32(IN) vgetq_lane_u32(vreinterpretq_u32_u8((IN)), 0); | ||||
| #define STOREQ_U8_AS_U32P(OUT, IN) vst1q_u32((OUT), vreinterpretq_u32_u8((IN))); | ||||
|  | ||||
| static WEBP_INLINE uint8x8_t Average2_u8_NEON(uint32_t a0, uint32_t a1) { | ||||
|   const uint8x8_t A0 = LOAD_U32_AS_U8(a0); | ||||
|   const uint8x8_t A1 = LOAD_U32_AS_U8(a1); | ||||
|   return vhadd_u8(A0, A1); | ||||
| } | ||||
|  | ||||
| static WEBP_INLINE uint32_t ClampedAddSubtractHalf_NEON(uint32_t c0, | ||||
|                                                         uint32_t c1, | ||||
|                                                         uint32_t c2) { | ||||
|   const uint8x8_t avg = Average2_u8_NEON(c0, c1); | ||||
|   // Remove one to c2 when bigger than avg. | ||||
|   const uint8x8_t C2 = LOAD_U32_AS_U8(c2); | ||||
|   const uint8x8_t cmp = vcgt_u8(C2, avg); | ||||
|   const uint8x8_t C2_1 = vadd_u8(C2, cmp); | ||||
|   // Compute half of the difference between avg and c2. | ||||
|   const int8x8_t diff_avg = vreinterpret_s8_u8(vhsub_u8(avg, C2_1)); | ||||
|   // Compute the sum with avg and saturate. | ||||
|   const int16x8_t avg_16 = vreinterpretq_s16_u16(vmovl_u8(avg)); | ||||
|   const uint8x8_t res = vqmovun_s16(vaddw_s8(avg_16, diff_avg)); | ||||
|   const uint32_t output = GET_U8_AS_U32(res); | ||||
|   return output; | ||||
| } | ||||
|  | ||||
| static WEBP_INLINE uint32_t Average2_NEON(uint32_t a0, uint32_t a1) { | ||||
|   const uint8x8_t avg_u8x8 = Average2_u8_NEON(a0, a1); | ||||
|   const uint32_t avg = GET_U8_AS_U32(avg_u8x8); | ||||
|   return avg; | ||||
| } | ||||
|  | ||||
| static WEBP_INLINE uint32_t Average3_NEON(uint32_t a0, uint32_t a1, | ||||
|                                           uint32_t a2) { | ||||
|   const uint8x8_t avg0 = Average2_u8_NEON(a0, a2); | ||||
|   const uint8x8_t A1 = LOAD_U32_AS_U8(a1); | ||||
|   const uint32_t avg = GET_U8_AS_U32(vhadd_u8(avg0, A1)); | ||||
|   return avg; | ||||
| } | ||||
|  | ||||
| static uint32_t Predictor5_NEON(uint32_t left, const uint32_t* const top) { | ||||
|   return Average3_NEON(left, top[0], top[1]); | ||||
| } | ||||
| static uint32_t Predictor6_NEON(uint32_t left, const uint32_t* const top) { | ||||
|   return Average2_NEON(left, top[-1]); | ||||
| } | ||||
| static uint32_t Predictor7_NEON(uint32_t left, const uint32_t* const top) { | ||||
|   return Average2_NEON(left, top[0]); | ||||
| } | ||||
| static uint32_t Predictor13_NEON(uint32_t left, const uint32_t* const top) { | ||||
|   return ClampedAddSubtractHalf_NEON(left, top[0], top[-1]); | ||||
| } | ||||
|  | ||||
| // Batch versions of those functions. | ||||
|  | ||||
| // Predictor0: ARGB_BLACK. | ||||
| static void PredictorAdd0_NEON(const uint32_t* in, const uint32_t* upper, | ||||
|                                int num_pixels, uint32_t* out) { | ||||
|   int i; | ||||
|   const uint8x16_t black = vreinterpretq_u8_u32(vdupq_n_u32(ARGB_BLACK)); | ||||
|   for (i = 0; i + 4 <= num_pixels; i += 4) { | ||||
|     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]); | ||||
|     const uint8x16_t res = vaddq_u8(src, black); | ||||
|     STOREQ_U8_AS_U32P(&out[i], res); | ||||
|   } | ||||
|   VP8LPredictorsAdd_C[0](in + i, upper + i, num_pixels - i, out + i); | ||||
| } | ||||
|  | ||||
| // Predictor1: left. | ||||
| static void PredictorAdd1_NEON(const uint32_t* in, const uint32_t* upper, | ||||
|                                int num_pixels, uint32_t* out) { | ||||
|   int i; | ||||
|   const uint8x16_t zero = LOADQ_U32_AS_U8(0); | ||||
|   for (i = 0; i + 4 <= num_pixels; i += 4) { | ||||
|     // a | b | c | d | ||||
|     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]); | ||||
|     // 0 | a | b | c | ||||
|     const uint8x16_t shift0 = vextq_u8(zero, src, 12); | ||||
|     // a | a + b | b + c | c + d | ||||
|     const uint8x16_t sum0 = vaddq_u8(src, shift0); | ||||
|     // 0 | 0 | a | a + b | ||||
|     const uint8x16_t shift1 = vextq_u8(zero, sum0, 8); | ||||
|     // a | a + b | a + b + c | a + b + c + d | ||||
|     const uint8x16_t sum1 = vaddq_u8(sum0, shift1); | ||||
|     const uint8x16_t prev = LOADQ_U32_AS_U8(out[i - 1]); | ||||
|     const uint8x16_t res = vaddq_u8(sum1, prev); | ||||
|     STOREQ_U8_AS_U32P(&out[i], res); | ||||
|   } | ||||
|   VP8LPredictorsAdd_C[1](in + i, upper + i, num_pixels - i, out + i); | ||||
| } | ||||
|  | ||||
| // Macro that adds 32-bit integers from IN using mod 256 arithmetic | ||||
| // per 8 bit channel. | ||||
| #define GENERATE_PREDICTOR_1(X, IN)                                       \ | ||||
| static void PredictorAdd##X##_NEON(const uint32_t* in,                    \ | ||||
|                                    const uint32_t* upper, int num_pixels, \ | ||||
|                                    uint32_t* out) {                       \ | ||||
|   int i;                                                                  \ | ||||
|   for (i = 0; i + 4 <= num_pixels; i += 4) {                              \ | ||||
|     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);                      \ | ||||
|     const uint8x16_t other = LOADQ_U32P_AS_U8(&(IN));                     \ | ||||
|     const uint8x16_t res = vaddq_u8(src, other);                          \ | ||||
|     STOREQ_U8_AS_U32P(&out[i], res);                                      \ | ||||
|   }                                                                       \ | ||||
|   VP8LPredictorsAdd_C[(X)](in + i, upper + i, num_pixels - i, out + i);   \ | ||||
| } | ||||
| // Predictor2: Top. | ||||
| GENERATE_PREDICTOR_1(2, upper[i]) | ||||
| // Predictor3: Top-right. | ||||
| GENERATE_PREDICTOR_1(3, upper[i + 1]) | ||||
| // Predictor4: Top-left. | ||||
| GENERATE_PREDICTOR_1(4, upper[i - 1]) | ||||
| #undef GENERATE_PREDICTOR_1 | ||||
|  | ||||
| #define GENERATE_PREDICTOR_2(X, IN)                                       \ | ||||
| static void PredictorAdd##X##_NEON(const uint32_t* in,                    \ | ||||
|                                    const uint32_t* upper, int num_pixels, \ | ||||
|                                    uint32_t* out) {                       \ | ||||
|   int i;                                                                  \ | ||||
|   for (i = 0; i + 4 <= num_pixels; i += 4) {                              \ | ||||
|     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);                      \ | ||||
|     const uint8x16_t Tother = LOADQ_U32P_AS_U8(&(IN));                    \ | ||||
|     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]);                     \ | ||||
|     const uint8x16_t avg = vhaddq_u8(T, Tother);                          \ | ||||
|     const uint8x16_t res = vaddq_u8(avg, src);                            \ | ||||
|     STOREQ_U8_AS_U32P(&out[i], res);                                      \ | ||||
|   }                                                                       \ | ||||
|   VP8LPredictorsAdd_C[(X)](in + i, upper + i, num_pixels - i, out + i);   \ | ||||
| } | ||||
| // Predictor8: average TL T. | ||||
| GENERATE_PREDICTOR_2(8, upper[i - 1]) | ||||
| // Predictor9: average T TR. | ||||
| GENERATE_PREDICTOR_2(9, upper[i + 1]) | ||||
| #undef GENERATE_PREDICTOR_2 | ||||
|  | ||||
| // Predictor10: average of (average of (L,TL), average of (T, TR)). | ||||
| static void PredictorAdd10_NEON(const uint32_t* in, const uint32_t* upper, | ||||
|                                 int num_pixels, uint32_t* out) { | ||||
|   int i, j; | ||||
|   for (i = 0; i + 4 <= num_pixels; i += 4) { | ||||
|     uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]); | ||||
|     uint8x16_t TL = LOADQ_U32P_AS_U8(&upper[i - 1]); | ||||
|     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]); | ||||
|     const uint8x16_t TR = LOADQ_U32P_AS_U8(&upper[i + 1]); | ||||
|     uint8x16_t avgTTR = vhaddq_u8(T, TR); | ||||
|     for (j = 0; j < 4; ++j) { | ||||
|       const uint8x16_t L = LOADQ_U32_AS_U8(out[i + j - 1]); | ||||
|       const uint8x16_t avgLTL = vhaddq_u8(L, TL); | ||||
|       const uint8x16_t avg = vhaddq_u8(avgTTR, avgLTL); | ||||
|       const uint8x16_t res = vaddq_u8(avg, src); | ||||
|       out[i + j] = GETQ_U8_AS_U32(res); | ||||
|       // Rotate the pre-computed values for the next iteration. | ||||
|       avgTTR = vextq_u8(avgTTR, avgTTR, 4); | ||||
|       TL = vextq_u8(TL, TL, 4); | ||||
|       src = vextq_u8(src, src, 4); | ||||
|     } | ||||
|   } | ||||
|   VP8LPredictorsAdd_C[10](in + i, upper + i, num_pixels - i, out + i); | ||||
| } | ||||
|  | ||||
| // Predictor11: select. | ||||
| static void PredictorAdd11_NEON(const uint32_t* in, const uint32_t* upper, | ||||
|                                 int num_pixels, uint32_t* out) { | ||||
|   int i, j; | ||||
|   const int32x2_t zero_s32 = vdup_n_s32(0); | ||||
|   for (i = 0; i + 4 <= num_pixels; i += 4) { | ||||
|     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]); | ||||
|     uint8x16_t TL = LOADQ_U32P_AS_U8(&upper[i - 1]); | ||||
|     // |T - TL| | ||||
|     uint8x16_t pTTL = vabdq_u8(T, TL); | ||||
|     // T + in | ||||
|     uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]); | ||||
|     uint32x4_t sumTin = vreinterpretq_u32_u8(vaddq_u8(T, src)); | ||||
|     for (j = 0; j < 4; ++j) { | ||||
|       const uint8x8_t L = LOAD_U32_AS_U8(out[i + j - 1]); | ||||
|       const uint8x8_t pLTL = vabd_u8(L, vget_low_u8(TL));  // |L - TL| | ||||
|       const int16x4_t diff = vget_low_s16( | ||||
|           vreinterpretq_s16_u16(vsubl_u8(pLTL, vget_low_u8(pTTL)))); | ||||
|       // Horizontal add the adjacent pairs twice to get the sum of the first | ||||
|       // four signed 16-bit integers. | ||||
|       // The first add cannot be vpaddl_s16 as it would return a int32x2_t | ||||
|       // which would lead to a int64x1_t for the second one (which would be | ||||
|       // hard to deal with). | ||||
|       const int16x4_t sum = vpadd_s16(diff, diff); | ||||
|       const int32x2_t pa_minus_pb = vpaddl_s16(sum); | ||||
|       const uint32x2_t cmp = vcle_s32(pa_minus_pb, zero_s32); | ||||
|       // L + in | ||||
|       const uint32x2_t sumLin = | ||||
|           vreinterpret_u32_u8(vadd_u8(L, vget_low_u8(src))); | ||||
|       // Add to top (pre-computed) or left. | ||||
|       const uint32x2_t output = vbsl_u32(cmp, vget_low_u32(sumTin), sumLin); | ||||
|       out[i + j] = vget_lane_u32(output, 0); | ||||
|       // Rotate the pre-computed values for the next iteration. | ||||
|       pTTL = vextq_u8(pTTL, pTTL, 4); | ||||
|       TL = vextq_u8(TL, TL, 4); | ||||
|       src = vextq_u8(src, src, 4); | ||||
|       sumTin = vextq_u32(sumTin, sumTin, 1); | ||||
|     } | ||||
|   } | ||||
|   VP8LPredictorsAdd_C[11](in + i, upper + i, num_pixels - i, out + i); | ||||
| } | ||||
|  | ||||
| // Predictor12: ClampedAddSubtractFull. | ||||
| static void PredictorAdd12_NEON(const uint32_t* in, const uint32_t* upper, | ||||
|                                 int num_pixels, uint32_t* out) { | ||||
|   int i, j; | ||||
|   // +4 to not read outside of memory. | ||||
|   for (i = 0; i + 4 <= num_pixels; i += 2) { | ||||
|     uint8x8_t src = LOAD_U32P_AS_U8(&in[i]); | ||||
|     const uint8x8_t TL = LOAD_U32P_AS_U8(&upper[i - 1]); | ||||
|     const uint8x8_t T = LOAD_U32P_AS_U8(&upper[i]); | ||||
|     int16x8_t diff = vreinterpretq_s16_u16(vsubl_u8(T, TL)); | ||||
|     for (j = 0; j < 2; ++j) { | ||||
|       const uint8x8_t L8 = LOAD_U32_AS_U8(out[i + j - 1]); | ||||
|       const int16x8_t L = vreinterpretq_s16_u16(vmovl_u8(L8)); | ||||
|       const int16x8_t sum = vaddq_s16(L, diff); | ||||
|       const uint8x8_t res = vadd_u8(vqmovun_s16(sum), src); | ||||
|       out[i + j] = vget_lane_u32(vreinterpret_u32_u8(res), 0); | ||||
|       // Shift the pre-computed value for the next iteration. | ||||
|       diff = vextq_s16(diff, diff, 4); | ||||
|       src = vext_u8(src, src, 4); | ||||
|     } | ||||
|   } | ||||
|   VP8LPredictorsAdd_C[12](in + i, upper + i, num_pixels - i, out + i); | ||||
| } | ||||
|  | ||||
| #undef LOAD_U32_AS_U8 | ||||
| #undef LOAD_U32P_AS_U8 | ||||
| #undef LOADQ_U32_AS_U8 | ||||
| #undef LOADQ_U32P_AS_U8 | ||||
| #undef GET_U8_AS_U32 | ||||
| #undef GETQ_U8_AS_U32 | ||||
| #undef STOREQ_U8_AS_U32P | ||||
|  | ||||
| //------------------------------------------------------------------------------ | ||||
| // Subtract-Green Transform | ||||
|  | ||||
| @@ -256,6 +499,22 @@ static void TransformColorInverse(const VP8LMultipliers* const m, | ||||
| extern void VP8LDspInitNEON(void); | ||||
|  | ||||
| WEBP_TSAN_IGNORE_FUNCTION void VP8LDspInitNEON(void) { | ||||
|   VP8LPredictors[5] = Predictor5_NEON; | ||||
|   VP8LPredictors[6] = Predictor6_NEON; | ||||
|   VP8LPredictors[7] = Predictor7_NEON; | ||||
|   VP8LPredictors[13] = Predictor13_NEON; | ||||
|  | ||||
|   VP8LPredictorsAdd[0] = PredictorAdd0_NEON; | ||||
|   VP8LPredictorsAdd[1] = PredictorAdd1_NEON; | ||||
|   VP8LPredictorsAdd[2] = PredictorAdd2_NEON; | ||||
|   VP8LPredictorsAdd[3] = PredictorAdd3_NEON; | ||||
|   VP8LPredictorsAdd[4] = PredictorAdd4_NEON; | ||||
|   VP8LPredictorsAdd[8] = PredictorAdd8_NEON; | ||||
|   VP8LPredictorsAdd[9] = PredictorAdd9_NEON; | ||||
|   VP8LPredictorsAdd[10] = PredictorAdd10_NEON; | ||||
|   VP8LPredictorsAdd[11] = PredictorAdd11_NEON; | ||||
|   VP8LPredictorsAdd[12] = PredictorAdd12_NEON; | ||||
|  | ||||
|   VP8LConvertBGRAToRGBA = ConvertBGRAToRGBA; | ||||
|   VP8LConvertBGRAToBGR = ConvertBGRAToBGR; | ||||
|   VP8LConvertBGRAToRGB = ConvertBGRAToRGB; | ||||
|   | ||||
		Reference in New Issue
	
	Block a user