OLD | NEW |
1 /* | 1 /* |
2 * Copyright (c) 2014 The WebM project authors. All Rights Reserved. | 2 * Copyright (c) 2014 The WebM project authors. All Rights Reserved. |
3 * | 3 * |
4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
9 */ | 9 */ |
10 | 10 |
(...skipping 220 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
231 *q9s16 = vaddq_s16(q1s16, q6s16); | 231 *q9s16 = vaddq_s16(q1s16, q6s16); |
232 *q10s16 = vaddq_s16(q2s16, q5s16); | 232 *q10s16 = vaddq_s16(q2s16, q5s16); |
233 *q11s16 = vaddq_s16(q3s16, q4s16); | 233 *q11s16 = vaddq_s16(q3s16, q4s16); |
234 *q12s16 = vsubq_s16(q3s16, q4s16); | 234 *q12s16 = vsubq_s16(q3s16, q4s16); |
235 *q13s16 = vsubq_s16(q2s16, q5s16); | 235 *q13s16 = vsubq_s16(q2s16, q5s16); |
236 *q14s16 = vsubq_s16(q1s16, q6s16); | 236 *q14s16 = vsubq_s16(q1s16, q6s16); |
237 *q15s16 = vsubq_s16(q0s16, q7s16); | 237 *q15s16 = vsubq_s16(q0s16, q7s16); |
238 return; | 238 return; |
239 } | 239 } |
240 | 240 |
241 void vp9_idct8x8_64_add_neon( | 241 void vpx_idct8x8_64_add_neon( |
242 int16_t *input, | 242 int16_t *input, |
243 uint8_t *dest, | 243 uint8_t *dest, |
244 int dest_stride) { | 244 int dest_stride) { |
245 uint8_t *d1, *d2; | 245 uint8_t *d1, *d2; |
246 uint8x8_t d0u8, d1u8, d2u8, d3u8; | 246 uint8x8_t d0u8, d1u8, d2u8, d3u8; |
247 uint64x1_t d0u64, d1u64, d2u64, d3u64; | 247 uint64x1_t d0u64, d1u64, d2u64, d3u64; |
248 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; | 248 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; |
249 uint16x8_t q8u16, q9u16, q10u16, q11u16; | 249 uint16x8_t q8u16, q9u16, q10u16, q11u16; |
250 | 250 |
251 q8s16 = vld1q_s16(input); | 251 q8s16 = vld1q_s16(input); |
(...skipping 92 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
344 d2 += dest_stride; | 344 d2 += dest_stride; |
345 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); | 345 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); |
346 d2 += dest_stride; | 346 d2 += dest_stride; |
347 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); | 347 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); |
348 d2 += dest_stride; | 348 d2 += dest_stride; |
349 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); | 349 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); |
350 d2 += dest_stride; | 350 d2 += dest_stride; |
351 return; | 351 return; |
352 } | 352 } |
353 | 353 |
354 void vp9_idct8x8_12_add_neon( | 354 void vpx_idct8x8_12_add_neon( |
355 int16_t *input, | 355 int16_t *input, |
356 uint8_t *dest, | 356 uint8_t *dest, |
357 int dest_stride) { | 357 int dest_stride) { |
358 uint8_t *d1, *d2; | 358 uint8_t *d1, *d2; |
359 uint8x8_t d0u8, d1u8, d2u8, d3u8; | 359 uint8x8_t d0u8, d1u8, d2u8, d3u8; |
360 int16x4_t d10s16, d11s16, d12s16, d13s16, d16s16; | 360 int16x4_t d10s16, d11s16, d12s16, d13s16, d16s16; |
361 int16x4_t d26s16, d27s16, d28s16, d29s16; | 361 int16x4_t d26s16, d27s16, d28s16, d29s16; |
362 uint64x1_t d0u64, d1u64, d2u64, d3u64; | 362 uint64x1_t d0u64, d1u64, d2u64, d3u64; |
363 int16x8_t q0s16, q1s16, q2s16, q3s16, q4s16, q5s16, q6s16, q7s16; | 363 int16x8_t q0s16, q1s16, q2s16, q3s16, q4s16, q5s16, q6s16, q7s16; |
364 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; | 364 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; |
(...skipping 166 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
531 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d0u8)); | 531 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d0u8)); |
532 d2 += dest_stride; | 532 d2 += dest_stride; |
533 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); | 533 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); |
534 d2 += dest_stride; | 534 d2 += dest_stride; |
535 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); | 535 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); |
536 d2 += dest_stride; | 536 d2 += dest_stride; |
537 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); | 537 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); |
538 d2 += dest_stride; | 538 d2 += dest_stride; |
539 return; | 539 return; |
540 } | 540 } |
OLD | NEW |