| OLD | NEW |
| 1 /* | 1 /* |
| 2 * Copyright (c) 2014 The WebM project authors. All Rights Reserved. | 2 * Copyright (c) 2014 The WebM project authors. All Rights Reserved. |
| 3 * | 3 * |
| 4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
| 5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
| 6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
| 7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
| 8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
| 9 */ | 9 */ |
| 10 | 10 |
| (...skipping 220 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 231 *q9s16 = vaddq_s16(q1s16, q6s16); | 231 *q9s16 = vaddq_s16(q1s16, q6s16); |
| 232 *q10s16 = vaddq_s16(q2s16, q5s16); | 232 *q10s16 = vaddq_s16(q2s16, q5s16); |
| 233 *q11s16 = vaddq_s16(q3s16, q4s16); | 233 *q11s16 = vaddq_s16(q3s16, q4s16); |
| 234 *q12s16 = vsubq_s16(q3s16, q4s16); | 234 *q12s16 = vsubq_s16(q3s16, q4s16); |
| 235 *q13s16 = vsubq_s16(q2s16, q5s16); | 235 *q13s16 = vsubq_s16(q2s16, q5s16); |
| 236 *q14s16 = vsubq_s16(q1s16, q6s16); | 236 *q14s16 = vsubq_s16(q1s16, q6s16); |
| 237 *q15s16 = vsubq_s16(q0s16, q7s16); | 237 *q15s16 = vsubq_s16(q0s16, q7s16); |
| 238 return; | 238 return; |
| 239 } | 239 } |
| 240 | 240 |
| 241 void vp9_idct8x8_64_add_neon( | 241 void vpx_idct8x8_64_add_neon( |
| 242 int16_t *input, | 242 int16_t *input, |
| 243 uint8_t *dest, | 243 uint8_t *dest, |
| 244 int dest_stride) { | 244 int dest_stride) { |
| 245 uint8_t *d1, *d2; | 245 uint8_t *d1, *d2; |
| 246 uint8x8_t d0u8, d1u8, d2u8, d3u8; | 246 uint8x8_t d0u8, d1u8, d2u8, d3u8; |
| 247 uint64x1_t d0u64, d1u64, d2u64, d3u64; | 247 uint64x1_t d0u64, d1u64, d2u64, d3u64; |
| 248 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; | 248 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; |
| 249 uint16x8_t q8u16, q9u16, q10u16, q11u16; | 249 uint16x8_t q8u16, q9u16, q10u16, q11u16; |
| 250 | 250 |
| 251 q8s16 = vld1q_s16(input); | 251 q8s16 = vld1q_s16(input); |
| (...skipping 92 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 344 d2 += dest_stride; | 344 d2 += dest_stride; |
| 345 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); | 345 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); |
| 346 d2 += dest_stride; | 346 d2 += dest_stride; |
| 347 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); | 347 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); |
| 348 d2 += dest_stride; | 348 d2 += dest_stride; |
| 349 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); | 349 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); |
| 350 d2 += dest_stride; | 350 d2 += dest_stride; |
| 351 return; | 351 return; |
| 352 } | 352 } |
| 353 | 353 |
| 354 void vp9_idct8x8_12_add_neon( | 354 void vpx_idct8x8_12_add_neon( |
| 355 int16_t *input, | 355 int16_t *input, |
| 356 uint8_t *dest, | 356 uint8_t *dest, |
| 357 int dest_stride) { | 357 int dest_stride) { |
| 358 uint8_t *d1, *d2; | 358 uint8_t *d1, *d2; |
| 359 uint8x8_t d0u8, d1u8, d2u8, d3u8; | 359 uint8x8_t d0u8, d1u8, d2u8, d3u8; |
| 360 int16x4_t d10s16, d11s16, d12s16, d13s16, d16s16; | 360 int16x4_t d10s16, d11s16, d12s16, d13s16, d16s16; |
| 361 int16x4_t d26s16, d27s16, d28s16, d29s16; | 361 int16x4_t d26s16, d27s16, d28s16, d29s16; |
| 362 uint64x1_t d0u64, d1u64, d2u64, d3u64; | 362 uint64x1_t d0u64, d1u64, d2u64, d3u64; |
| 363 int16x8_t q0s16, q1s16, q2s16, q3s16, q4s16, q5s16, q6s16, q7s16; | 363 int16x8_t q0s16, q1s16, q2s16, q3s16, q4s16, q5s16, q6s16, q7s16; |
| 364 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; | 364 int16x8_t q8s16, q9s16, q10s16, q11s16, q12s16, q13s16, q14s16, q15s16; |
| (...skipping 166 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 531 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d0u8)); | 531 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d0u8)); |
| 532 d2 += dest_stride; | 532 d2 += dest_stride; |
| 533 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); | 533 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d1u8)); |
| 534 d2 += dest_stride; | 534 d2 += dest_stride; |
| 535 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); | 535 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d2u8)); |
| 536 d2 += dest_stride; | 536 d2 += dest_stride; |
| 537 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); | 537 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); |
| 538 d2 += dest_stride; | 538 d2 += dest_stride; |
| 539 return; | 539 return; |
| 540 } | 540 } |
| OLD | NEW |