| OLD | NEW |
| (Empty) |
| 1 /* | |
| 2 * Copyright (c) 2013 The WebM project authors. All Rights Reserved. | |
| 3 * | |
| 4 * Use of this source code is governed by a BSD-style license | |
| 5 * that can be found in the LICENSE file in the root of the source | |
| 6 * tree. An additional intellectual property rights grant can be found | |
| 7 * in the file PATENTS. All contributing project authors may | |
| 8 * be found in the AUTHORS file in the root of the source tree. | |
| 9 */ | |
| 10 | |
| 11 #include <assert.h> | |
| 12 #include <stdio.h> | |
| 13 | |
| 14 #include "./vpx_dsp_rtcd.h" | |
| 15 #include "vpx_dsp/mips/vpx_common_dspr2.h" | |
| 16 #include "vpx_dsp/vpx_convolve.h" | |
| 17 #include "vpx_dsp/vpx_dsp_common.h" | |
| 18 #include "vpx_ports/mem.h" | |
| 19 | |
| 20 #if HAVE_DSPR2 | |
| 21 static void convolve_bi_avg_horiz_4_dspr2(const uint8_t *src, | |
| 22 int32_t src_stride, | |
| 23 uint8_t *dst, | |
| 24 int32_t dst_stride, | |
| 25 const int16_t *filter_x0, | |
| 26 int32_t h) { | |
| 27 int32_t y; | |
| 28 uint8_t *cm = vpx_ff_cropTbl; | |
| 29 int32_t Temp1, Temp2, Temp3, Temp4; | |
| 30 uint32_t vector4a = 64; | |
| 31 uint32_t tp1, tp2; | |
| 32 uint32_t p1, p2, p3; | |
| 33 uint32_t tn1, tn2; | |
| 34 const int16_t *filter = &filter_x0[3]; | |
| 35 uint32_t filter45; | |
| 36 | |
| 37 filter45 = ((const int32_t *)filter)[0]; | |
| 38 | |
| 39 for (y = h; y--;) { | |
| 40 /* prefetch data to cache memory */ | |
| 41 prefetch_load(src + src_stride); | |
| 42 prefetch_load(src + src_stride + 32); | |
| 43 prefetch_store(dst + dst_stride); | |
| 44 | |
| 45 __asm__ __volatile__ ( | |
| 46 "ulw %[tp1], 0(%[src]) \n\t" | |
| 47 "ulw %[tp2], 4(%[src]) \n\t" | |
| 48 | |
| 49 /* even 1. pixel */ | |
| 50 "mtlo %[vector4a], $ac3 \n\t" | |
| 51 "mthi $zero, $ac3 \n\t" | |
| 52 "preceu.ph.qbr %[p1], %[tp1] \n\t" | |
| 53 "preceu.ph.qbl %[p2], %[tp1] \n\t" | |
| 54 "dpa.w.ph $ac3, %[p1], %[filter45] \n\t" | |
| 55 "extp %[Temp1], $ac3, 31 \n\t" | |
| 56 | |
| 57 /* even 2. pixel */ | |
| 58 "mtlo %[vector4a], $ac2 \n\t" | |
| 59 "mthi $zero, $ac2 \n\t" | |
| 60 "balign %[tp2], %[tp1], 3 \n\t" | |
| 61 "dpa.w.ph $ac2, %[p2], %[filter45] \n\t" | |
| 62 "extp %[Temp3], $ac2, 31 \n\t" | |
| 63 | |
| 64 "lbu %[p2], 3(%[dst]) \n\t"
/* load odd 2 */ | |
| 65 | |
| 66 /* odd 1. pixel */ | |
| 67 "lbux %[tp1], %[Temp1](%[cm]) \n\t"
/* even 1 */ | |
| 68 "mtlo %[vector4a], $ac3 \n\t" | |
| 69 "mthi $zero, $ac3 \n\t" | |
| 70 "lbu %[Temp1], 1(%[dst]) \n\t"
/* load odd 1 */ | |
| 71 "preceu.ph.qbr %[p1], %[tp2] \n\t" | |
| 72 "preceu.ph.qbl %[p3], %[tp2] \n\t" | |
| 73 "dpa.w.ph $ac3, %[p1], %[filter45] \n\t" | |
| 74 "extp %[Temp2], $ac3, 31 \n\t" | |
| 75 | |
| 76 "lbu %[tn2], 0(%[dst]) \n\t"
/* load even 1 */ | |
| 77 | |
| 78 /* odd 2. pixel */ | |
| 79 "lbux %[tp2], %[Temp3](%[cm]) \n\t"
/* even 2 */ | |
| 80 "mtlo %[vector4a], $ac2 \n\t" | |
| 81 "mthi $zero, $ac2 \n\t" | |
| 82 "lbux %[tn1], %[Temp2](%[cm]) \n\t"
/* odd 1 */ | |
| 83 "addqh_r.w %[tn2], %[tn2], %[tp1] \n\t"
/* average even 1 */ | |
| 84 "dpa.w.ph $ac2, %[p3], %[filter45] \n\t" | |
| 85 "extp %[Temp4], $ac2, 31 \n\t" | |
| 86 | |
| 87 "lbu %[tp1], 2(%[dst]) \n\t"
/* load even 2 */ | |
| 88 "sb %[tn2], 0(%[dst]) \n\t"
/* store even 1 */ | |
| 89 | |
| 90 /* clamp */ | |
| 91 "addqh_r.w %[Temp1], %[Temp1], %[tn1] \n\t"
/* average odd 1 */ | |
| 92 "lbux %[p3], %[Temp4](%[cm]) \n\t"
/* odd 2 */ | |
| 93 "sb %[Temp1], 1(%[dst]) \n\t"
/* store odd 1 */ | |
| 94 | |
| 95 "addqh_r.w %[tp1], %[tp1], %[tp2] \n\t"
/* average even 2 */ | |
| 96 "sb %[tp1], 2(%[dst]) \n\t"
/* store even 2 */ | |
| 97 | |
| 98 "addqh_r.w %[p2], %[p2], %[p3] \n\t"
/* average odd 2 */ | |
| 99 "sb %[p2], 3(%[dst]) \n\t"
/* store odd 2 */ | |
| 100 | |
| 101 : [tp1] "=&r" (tp1), [tp2] "=&r" (tp2), | |
| 102 [tn1] "=&r" (tn1), [tn2] "=&r" (tn2), | |
| 103 [p1] "=&r" (p1), [p2] "=&r" (p2), [p3] "=&r" (p3), | |
| 104 [Temp1] "=&r" (Temp1), [Temp2] "=&r" (Temp2), | |
| 105 [Temp3] "=&r" (Temp3), [Temp4] "=&r" (Temp4) | |
| 106 : [filter45] "r" (filter45), [vector4a] "r" (vector4a), | |
| 107 [cm] "r" (cm), [dst] "r" (dst), [src] "r" (src) | |
| 108 ); | |
| 109 | |
| 110 /* Next row... */ | |
| 111 src += src_stride; | |
| 112 dst += dst_stride; | |
| 113 } | |
| 114 } | |
| 115 | |
| 116 static void convolve_bi_avg_horiz_8_dspr2(const uint8_t *src, | |
| 117 int32_t src_stride, | |
| 118 uint8_t *dst, | |
| 119 int32_t dst_stride, | |
| 120 const int16_t *filter_x0, | |
| 121 int32_t h) { | |
| 122 int32_t y; | |
| 123 uint8_t *cm = vpx_ff_cropTbl; | |
| 124 uint32_t vector4a = 64; | |
| 125 int32_t Temp1, Temp2, Temp3; | |
| 126 uint32_t tp1, tp2, tp3, tp4; | |
| 127 uint32_t p1, p2, p3, p4, n1; | |
| 128 uint32_t st0, st1; | |
| 129 const int16_t *filter = &filter_x0[3]; | |
| 130 uint32_t filter45;; | |
| 131 | |
| 132 filter45 = ((const int32_t *)filter)[0]; | |
| 133 | |
| 134 for (y = h; y--;) { | |
| 135 /* prefetch data to cache memory */ | |
| 136 prefetch_load(src + src_stride); | |
| 137 prefetch_load(src + src_stride + 32); | |
| 138 prefetch_store(dst + dst_stride); | |
| 139 | |
| 140 __asm__ __volatile__ ( | |
| 141 "ulw %[tp1], 0(%[src]) \n\t" | |
| 142 "ulw %[tp2], 4(%[src]) \n\t" | |
| 143 | |
| 144 /* even 1. pixel */ | |
| 145 "mtlo %[vector4a], $ac3 \n\t" | |
| 146 "mthi $zero, $ac3 \n\t" | |
| 147 "mtlo %[vector4a], $ac2 \n\t" | |
| 148 "mthi $zero, $ac2 \n\t" | |
| 149 "preceu.ph.qbr %[p1], %[tp1] \n\t" | |
| 150 "preceu.ph.qbl %[p2], %[tp1] \n\t" | |
| 151 "preceu.ph.qbr %[p3], %[tp2] \n\t" | |
| 152 "preceu.ph.qbl %[p4], %[tp2] \n\t" | |
| 153 "ulw %[tp3], 8(%[src]) \n\t" | |
| 154 "dpa.w.ph $ac3, %[p1], %[filter45] \n\t" | |
| 155 "extp %[Temp1], $ac3, 31 \n\t" | |
| 156 "lbu %[Temp2], 0(%[dst]) \n\t" | |
| 157 "lbu %[tp4], 2(%[dst]) \n\t" | |
| 158 | |
| 159 /* even 2. pixel */ | |
| 160 "dpa.w.ph $ac2, %[p2], %[filter45] \n\t" | |
| 161 "extp %[Temp3], $ac2, 31 \n\t" | |
| 162 | |
| 163 /* even 3. pixel */ | |
| 164 "lbux %[st0], %[Temp1](%[cm]) \n\t" | |
| 165 "mtlo %[vector4a], $ac1 \n\t" | |
| 166 "mthi $zero, $ac1 \n\t" | |
| 167 "lbux %[st1], %[Temp3](%[cm]) \n\t" | |
| 168 "dpa.w.ph $ac1, %[p3], %[filter45] \n\t" | |
| 169 "extp %[Temp1], $ac1, 31 \n\t" | |
| 170 | |
| 171 "addqh_r.w %[Temp2], %[Temp2], %[st0] \n\t" | |
| 172 "addqh_r.w %[tp4], %[tp4], %[st1] \n\t" | |
| 173 "sb %[Temp2], 0(%[dst]) \n\t" | |
| 174 "sb %[tp4], 2(%[dst]) \n\t" | |
| 175 | |
| 176 /* even 4. pixel */ | |
| 177 "mtlo %[vector4a], $ac2 \n\t" | |
| 178 "mthi $zero, $ac2 \n\t" | |
| 179 "mtlo %[vector4a], $ac3 \n\t" | |
| 180 "mthi $zero, $ac3 \n\t" | |
| 181 | |
| 182 "balign %[tp3], %[tp2], 3 \n\t" | |
| 183 "balign %[tp2], %[tp1], 3 \n\t" | |
| 184 | |
| 185 "lbux %[st0], %[Temp1](%[cm]) \n\t" | |
| 186 "lbu %[Temp2], 4(%[dst]) \n\t" | |
| 187 "addqh_r.w %[Temp2], %[Temp2], %[st0] \n\t" | |
| 188 | |
| 189 "dpa.w.ph $ac2, %[p4], %[filter45] \n\t" | |
| 190 "extp %[Temp3], $ac2, 31 \n\t" | |
| 191 | |
| 192 /* odd 1. pixel */ | |
| 193 "mtlo %[vector4a], $ac1 \n\t" | |
| 194 "mthi $zero, $ac1 \n\t" | |
| 195 "sb %[Temp2], 4(%[dst]) \n\t" | |
| 196 "preceu.ph.qbr %[p1], %[tp2] \n\t" | |
| 197 "preceu.ph.qbl %[p2], %[tp2] \n\t" | |
| 198 "preceu.ph.qbr %[p3], %[tp3] \n\t" | |
| 199 "preceu.ph.qbl %[p4], %[tp3] \n\t" | |
| 200 "dpa.w.ph $ac3, %[p1], %[filter45] \n\t" | |
| 201 "extp %[Temp2], $ac3, 31 \n\t" | |
| 202 | |
| 203 "lbu %[tp1], 6(%[dst]) \n\t" | |
| 204 | |
| 205 /* odd 2. pixel */ | |
| 206 "mtlo %[vector4a], $ac3 \n\t" | |
| 207 "mthi $zero, $ac3 \n\t" | |
| 208 "mtlo %[vector4a], $ac2 \n\t" | |
| 209 "mthi $zero, $ac2 \n\t" | |
| 210 "lbux %[st0], %[Temp3](%[cm]) \n\t" | |
| 211 "dpa.w.ph $ac1, %[p2], %[filter45] \n\t" | |
| 212 "extp %[Temp3], $ac1, 31 \n\t" | |
| 213 | |
| 214 "lbu %[tp2], 1(%[dst]) \n\t" | |
| 215 "lbu %[tp3], 3(%[dst]) \n\t" | |
| 216 "addqh_r.w %[tp1], %[tp1], %[st0] \n\t" | |
| 217 | |
| 218 /* odd 3. pixel */ | |
| 219 "lbux %[st1], %[Temp2](%[cm]) \n\t" | |
| 220 "dpa.w.ph $ac3, %[p3], %[filter45] \n\t" | |
| 221 "addqh_r.w %[tp2], %[tp2], %[st1] \n\t" | |
| 222 "extp %[Temp2], $ac3, 31 \n\t" | |
| 223 | |
| 224 "lbu %[tp4], 5(%[dst]) \n\t" | |
| 225 | |
| 226 /* odd 4. pixel */ | |
| 227 "sb %[tp2], 1(%[dst]) \n\t" | |
| 228 "sb %[tp1], 6(%[dst]) \n\t" | |
| 229 "dpa.w.ph $ac2, %[p4], %[filter45] \n\t" | |
| 230 "extp %[Temp1], $ac2, 31 \n\t" | |
| 231 | |
| 232 "lbu %[tp1], 7(%[dst]) \n\t" | |
| 233 | |
| 234 /* clamp */ | |
| 235 "lbux %[p4], %[Temp3](%[cm]) \n\t" | |
| 236 "addqh_r.w %[tp3], %[tp3], %[p4] \n\t" | |
| 237 | |
| 238 "lbux %[p2], %[Temp2](%[cm]) \n\t" | |
| 239 "addqh_r.w %[tp4], %[tp4], %[p2] \n\t" | |
| 240 | |
| 241 "lbux %[p1], %[Temp1](%[cm]) \n\t" | |
| 242 "addqh_r.w %[tp1], %[tp1], %[p1] \n\t" | |
| 243 | |
| 244 /* store bytes */ | |
| 245 "sb %[tp3], 3(%[dst]) \n\t" | |
| 246 "sb %[tp4], 5(%[dst]) \n\t" | |
| 247 "sb %[tp1], 7(%[dst]) \n\t" | |
| 248 | |
| 249 : [tp1] "=&r" (tp1), [tp2] "=&r" (tp2), | |
| 250 [tp3] "=&r" (tp3), [tp4] "=&r" (tp4), | |
| 251 [st0] "=&r" (st0), [st1] "=&r" (st1), | |
| 252 [p1] "=&r" (p1), [p2] "=&r" (p2), [p3] "=&r" (p3), [p4] "=&r" (p4), | |
| 253 [n1] "=&r" (n1), | |
| 254 [Temp1] "=&r" (Temp1), [Temp2] "=&r" (Temp2), [Temp3] "=&r" (Temp3) | |
| 255 : [filter45] "r" (filter45), [vector4a] "r" (vector4a), | |
| 256 [cm] "r" (cm), [dst] "r" (dst), [src] "r" (src) | |
| 257 ); | |
| 258 | |
| 259 /* Next row... */ | |
| 260 src += src_stride; | |
| 261 dst += dst_stride; | |
| 262 } | |
| 263 } | |
| 264 | |
| 265 static void convolve_bi_avg_horiz_16_dspr2(const uint8_t *src_ptr, | |
| 266 int32_t src_stride, | |
| 267 uint8_t *dst_ptr, | |
| 268 int32_t dst_stride, | |
| 269 const int16_t *filter_x0, | |
| 270 int32_t h, | |
| 271 int32_t count) { | |
| 272 int32_t y, c; | |
| 273 const uint8_t *src; | |
| 274 uint8_t *dst; | |
| 275 uint8_t *cm = vpx_ff_cropTbl; | |
| 276 uint32_t vector_64 = 64; | |
| 277 int32_t Temp1, Temp2, Temp3; | |
| 278 uint32_t qload1, qload2, qload3; | |
| 279 uint32_t p1, p2, p3, p4, p5; | |
| 280 uint32_t st1, st2, st3; | |
| 281 const int16_t *filter = &filter_x0[3]; | |
| 282 uint32_t filter45;; | |
| 283 | |
| 284 filter45 = ((const int32_t *)filter)[0]; | |
| 285 | |
| 286 for (y = h; y--;) { | |
| 287 src = src_ptr; | |
| 288 dst = dst_ptr; | |
| 289 | |
| 290 /* prefetch data to cache memory */ | |
| 291 prefetch_load(src_ptr + src_stride); | |
| 292 prefetch_load(src_ptr + src_stride + 32); | |
| 293 prefetch_store(dst_ptr + dst_stride); | |
| 294 | |
| 295 for (c = 0; c < count; c++) { | |
| 296 __asm__ __volatile__ ( | |
| 297 "ulw %[qload1], 0(%[src]) \n\t" | |
| 298 "ulw %[qload2], 4(%[src]) \n\t" | |
| 299 | |
| 300 /* even 1. pixel */ | |
| 301 "mtlo %[vector_64], $ac1 \n\t" /*
even 1 */ | |
| 302 "mthi $zero, $ac1 \n\t" | |
| 303 "mtlo %[vector_64], $ac2 \n\t" /*
even 2 */ | |
| 304 "mthi $zero, $ac2 \n\t" | |
| 305 "preceu.ph.qbr %[p1], %[qload1] \n\t" | |
| 306 "preceu.ph.qbl %[p2], %[qload1] \n\t" | |
| 307 "preceu.ph.qbr %[p3], %[qload2] \n\t" | |
| 308 "preceu.ph.qbl %[p4], %[qload2] \n\t" | |
| 309 "ulw %[qload3], 8(%[src]) \n\t" | |
| 310 "dpa.w.ph $ac1, %[p1], %[filter45] \n\t" /*
even 1 */ | |
| 311 "extp %[Temp1], $ac1, 31 \n\t" /*
even 1 */ | |
| 312 "lbu %[st2], 0(%[dst]) \n\t" /*
load even 1 from dst */ | |
| 313 | |
| 314 /* even 2. pixel */ | |
| 315 "mtlo %[vector_64], $ac3 \n\t" /*
even 3 */ | |
| 316 "mthi $zero, $ac3 \n\t" | |
| 317 "preceu.ph.qbr %[p1], %[qload3] \n\t" | |
| 318 "preceu.ph.qbl %[p5], %[qload3] \n\t" | |
| 319 "ulw %[qload1], 12(%[src]) \n\t" | |
| 320 "dpa.w.ph $ac2, %[p2], %[filter45] \n\t" /*
even 1 */ | |
| 321 "extp %[Temp2], $ac2, 31 \n\t" /*
even 1 */ | |
| 322 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
even 1 */ | |
| 323 | |
| 324 "lbu %[qload3], 2(%[dst]) \n\t" /*
load even 2 from dst */ | |
| 325 | |
| 326 /* even 3. pixel */ | |
| 327 "mtlo %[vector_64], $ac1 \n\t" /*
even 4 */ | |
| 328 "mthi $zero, $ac1 \n\t" | |
| 329 "addqh_r.w %[st2], %[st2], %[st1] \n\t" /*
average even 1 */ | |
| 330 "preceu.ph.qbr %[p2], %[qload1] \n\t" | |
| 331 "sb %[st2], 0(%[dst]) \n\t" /*
store even 1 to dst */ | |
| 332 "dpa.w.ph $ac3, %[p3], %[filter45] \n\t" /*
even 3 */ | |
| 333 "extp %[Temp3], $ac3, 31 \n\t" /*
even 3 */ | |
| 334 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
even 1 */ | |
| 335 | |
| 336 /* even 4. pixel */ | |
| 337 "mtlo %[vector_64], $ac2 \n\t" /*
even 5 */ | |
| 338 "mthi $zero, $ac2 \n\t" | |
| 339 "addqh_r.w %[qload3], %[qload3], %[st2] \n\t" /*
average even 2 */ | |
| 340 "preceu.ph.qbl %[p3], %[qload1] \n\t" | |
| 341 "sb %[qload3], 2(%[dst]) \n\t" /*
store even 2 to dst */ | |
| 342 "lbu %[qload3], 4(%[dst]) \n\t" /*
load even 3 from dst */ | |
| 343 "lbu %[qload1], 6(%[dst]) \n\t" /*
load even 4 from dst */ | |
| 344 "dpa.w.ph $ac1, %[p4], %[filter45] \n\t" /*
even 4 */ | |
| 345 "extp %[Temp1], $ac1, 31 \n\t" /*
even 4 */ | |
| 346 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
even 3 */ | |
| 347 | |
| 348 /* even 5. pixel */ | |
| 349 "mtlo %[vector_64], $ac3 \n\t" /*
even 6 */ | |
| 350 "mthi $zero, $ac3 \n\t" | |
| 351 "addqh_r.w %[qload3], %[qload3], %[st3] \n\t" /*
average even 3 */ | |
| 352 "sb %[qload3], 4(%[dst]) \n\t" /*
store even 3 to dst */ | |
| 353 "dpa.w.ph $ac2, %[p1], %[filter45] \n\t" /*
even 5 */ | |
| 354 "extp %[Temp2], $ac2, 31 \n\t" /*
even 5 */ | |
| 355 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
even 4 */ | |
| 356 | |
| 357 /* even 6. pixel */ | |
| 358 "mtlo %[vector_64], $ac1 \n\t" /*
even 7 */ | |
| 359 "mthi $zero, $ac1 \n\t" | |
| 360 "addqh_r.w %[qload1], %[qload1], %[st1] \n\t" /*
average even 4 */ | |
| 361 "sb %[qload1], 6(%[dst]) \n\t" /*
store even 4 to dst */ | |
| 362 "dpa.w.ph $ac3, %[p5], %[filter45] \n\t" /*
even 6 */ | |
| 363 "lbu %[qload2], 8(%[dst]) \n\t" /*
load even 5 from dst */ | |
| 364 "extp %[Temp3], $ac3, 31 \n\t" /*
even 6 */ | |
| 365 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
even 5 */ | |
| 366 | |
| 367 /* even 7. pixel */ | |
| 368 "mtlo %[vector_64], $ac2 \n\t" /*
even 8 */ | |
| 369 "mthi $zero, $ac2 \n\t" | |
| 370 "addqh_r.w %[qload2], %[qload2], %[st2] \n\t" /*
average even 5 */ | |
| 371 "sb %[qload2], 8(%[dst]) \n\t" /*
store even 5 to dst */ | |
| 372 "dpa.w.ph $ac1, %[p2], %[filter45] \n\t" /*
even 7 */ | |
| 373 "lbu %[qload3], 10(%[dst]) \n\t" /*
load even 6 from dst */ | |
| 374 "extp %[Temp1], $ac1, 31 \n\t" /*
even 7 */ | |
| 375 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
even 6 */ | |
| 376 | |
| 377 "lbu %[st2], 12(%[dst]) \n\t" /*
load even 7 from dst */ | |
| 378 | |
| 379 /* even 8. pixel */ | |
| 380 "mtlo %[vector_64], $ac3 \n\t" /*
odd 1 */ | |
| 381 "mthi $zero, $ac3 \n\t" | |
| 382 "addqh_r.w %[qload3], %[qload3], %[st3] \n\t" /*
average even 6 */ | |
| 383 "dpa.w.ph $ac2, %[p3], %[filter45] \n\t" /*
even 8 */ | |
| 384 "sb %[qload3], 10(%[dst]) \n\t" /*
store even 6 to dst */ | |
| 385 "extp %[Temp2], $ac2, 31 \n\t" /*
even 8 */ | |
| 386 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
even 7 */ | |
| 387 | |
| 388 /* ODD pixels */ | |
| 389 "ulw %[qload1], 1(%[src]) \n\t" | |
| 390 "ulw %[qload2], 5(%[src]) \n\t" | |
| 391 | |
| 392 "addqh_r.w %[st2], %[st2], %[st1] \n\t" /*
average even 7 */ | |
| 393 | |
| 394 /* odd 1. pixel */ | |
| 395 "mtlo %[vector_64], $ac1 \n\t" /*
odd 2 */ | |
| 396 "mthi $zero, $ac1 \n\t" | |
| 397 "preceu.ph.qbr %[p1], %[qload1] \n\t" | |
| 398 "preceu.ph.qbl %[p2], %[qload1] \n\t" | |
| 399 "preceu.ph.qbr %[p3], %[qload2] \n\t" | |
| 400 "preceu.ph.qbl %[p4], %[qload2] \n\t" | |
| 401 "sb %[st2], 12(%[dst]) \n\t" /*
store even 7 to dst */ | |
| 402 "ulw %[qload3], 9(%[src]) \n\t" | |
| 403 "dpa.w.ph $ac3, %[p1], %[filter45] \n\t" /*
odd 1 */ | |
| 404 "lbu %[qload2], 14(%[dst]) \n\t" /*
load even 8 from dst */ | |
| 405 "extp %[Temp3], $ac3, 31 \n\t" /*
odd 1 */ | |
| 406 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
even 8 */ | |
| 407 | |
| 408 "lbu %[st1], 1(%[dst]) \n\t" /*
load odd 1 from dst */ | |
| 409 | |
| 410 /* odd 2. pixel */ | |
| 411 "mtlo %[vector_64], $ac2 \n\t" /*
odd 3 */ | |
| 412 "mthi $zero, $ac2 \n\t" | |
| 413 "addqh_r.w %[qload2], %[qload2], %[st2] \n\t" /*
average even 8 */ | |
| 414 "preceu.ph.qbr %[p1], %[qload3] \n\t" | |
| 415 "preceu.ph.qbl %[p5], %[qload3] \n\t" | |
| 416 "sb %[qload2], 14(%[dst]) \n\t" /*
store even 8 to dst */ | |
| 417 "ulw %[qload1], 13(%[src]) \n\t" | |
| 418 "dpa.w.ph $ac1, %[p2], %[filter45] \n\t" /*
odd 2 */ | |
| 419 "lbu %[qload3], 3(%[dst]) \n\t" /*
load odd 2 from dst */ | |
| 420 "extp %[Temp1], $ac1, 31 \n\t" /*
odd 2 */ | |
| 421 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
odd 1 */ | |
| 422 | |
| 423 /* odd 3. pixel */ | |
| 424 "mtlo %[vector_64], $ac3 \n\t" /*
odd 4 */ | |
| 425 "mthi $zero, $ac3 \n\t" | |
| 426 "addqh_r.w %[st3], %[st3], %[st1] \n\t" /*
average odd 1 */ | |
| 427 "preceu.ph.qbr %[p2], %[qload1] \n\t" | |
| 428 "dpa.w.ph $ac2, %[p3], %[filter45] \n\t" /*
odd 3 */ | |
| 429 "sb %[st3], 1(%[dst]) \n\t" /*
store odd 1 to dst */ | |
| 430 "extp %[Temp2], $ac2, 31 \n\t" /*
odd 3 */ | |
| 431 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
odd 2 */ | |
| 432 | |
| 433 /* odd 4. pixel */ | |
| 434 "mtlo %[vector_64], $ac1 \n\t" /*
odd 5 */ | |
| 435 "mthi $zero, $ac1 \n\t" | |
| 436 "addqh_r.w %[qload3], %[qload3], %[st1] \n\t" /*
average odd 2 */ | |
| 437 "preceu.ph.qbl %[p3], %[qload1] \n\t" | |
| 438 "sb %[qload3], 3(%[dst]) \n\t" /*
store odd 2 to dst */ | |
| 439 "lbu %[qload1], 5(%[dst]) \n\t" /*
load odd 3 from dst */ | |
| 440 "dpa.w.ph $ac3, %[p4], %[filter45] \n\t" /*
odd 4 */ | |
| 441 "extp %[Temp3], $ac3, 31 \n\t" /*
odd 4 */ | |
| 442 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
odd 3 */ | |
| 443 | |
| 444 "lbu %[st1], 7(%[dst]) \n\t" /*
load odd 4 from dst */ | |
| 445 | |
| 446 /* odd 5. pixel */ | |
| 447 "mtlo %[vector_64], $ac2 \n\t" /*
odd 6 */ | |
| 448 "mthi $zero, $ac2 \n\t" | |
| 449 "addqh_r.w %[qload1], %[qload1], %[st2] \n\t" /*
average odd 3 */ | |
| 450 "sb %[qload1], 5(%[dst]) \n\t" /*
store odd 3 to dst */ | |
| 451 "dpa.w.ph $ac1, %[p1], %[filter45] \n\t" /*
odd 5 */ | |
| 452 "extp %[Temp1], $ac1, 31 \n\t" /*
odd 5 */ | |
| 453 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
odd 4 */ | |
| 454 | |
| 455 "lbu %[qload1], 9(%[dst]) \n\t" /*
load odd 5 from dst */ | |
| 456 | |
| 457 /* odd 6. pixel */ | |
| 458 "mtlo %[vector_64], $ac3 \n\t" /*
odd 7 */ | |
| 459 "mthi $zero, $ac3 \n\t" | |
| 460 "addqh_r.w %[st1], %[st1], %[st3] \n\t" /*
average odd 4 */ | |
| 461 "sb %[st1], 7(%[dst]) \n\t" /*
store odd 4 to dst */ | |
| 462 "dpa.w.ph $ac2, %[p5], %[filter45] \n\t" /*
odd 6 */ | |
| 463 "extp %[Temp2], $ac2, 31 \n\t" /*
odd 6 */ | |
| 464 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
odd 5 */ | |
| 465 | |
| 466 /* odd 7. pixel */ | |
| 467 "mtlo %[vector_64], $ac1 \n\t" /*
odd 8 */ | |
| 468 "mthi $zero, $ac1 \n\t" | |
| 469 "addqh_r.w %[qload1], %[qload1], %[st1] \n\t" /*
average odd 5 */ | |
| 470 "sb %[qload1], 9(%[dst]) \n\t" /*
store odd 5 to dst */ | |
| 471 "lbu %[qload2], 11(%[dst]) \n\t" /*
load odd 6 from dst */ | |
| 472 "dpa.w.ph $ac3, %[p2], %[filter45] \n\t" /*
odd 7 */ | |
| 473 "extp %[Temp3], $ac3, 31 \n\t" /*
odd 7 */ | |
| 474 | |
| 475 "lbu %[qload3], 13(%[dst]) \n\t" /*
load odd 7 from dst */ | |
| 476 | |
| 477 /* odd 8. pixel */ | |
| 478 "dpa.w.ph $ac1, %[p3], %[filter45] \n\t" /*
odd 8 */ | |
| 479 "extp %[Temp1], $ac1, 31 \n\t" /*
odd 8 */ | |
| 480 | |
| 481 "lbu %[qload1], 15(%[dst]) \n\t" /*
load odd 8 from dst */ | |
| 482 | |
| 483 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
odd 6 */ | |
| 484 "addqh_r.w %[qload2], %[qload2], %[st2] \n\t" /*
average odd 6 */ | |
| 485 | |
| 486 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
odd 7 */ | |
| 487 "addqh_r.w %[qload3], %[qload3], %[st3] \n\t" /*
average odd 7 */ | |
| 488 | |
| 489 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
odd 8 */ | |
| 490 "addqh_r.w %[qload1], %[qload1], %[st1] \n\t" /*
average odd 8 */ | |
| 491 | |
| 492 "sb %[qload2], 11(%[dst]) \n\t" /*
store odd 6 to dst */ | |
| 493 "sb %[qload3], 13(%[dst]) \n\t" /*
store odd 7 to dst */ | |
| 494 "sb %[qload1], 15(%[dst]) \n\t" /*
store odd 8 to dst */ | |
| 495 | |
| 496 : [qload1] "=&r" (qload1), [qload2] "=&r" (qload2), | |
| 497 [st1] "=&r" (st1), [st2] "=&r" (st2), [st3] "=&r" (st3), | |
| 498 [p1] "=&r" (p1), [p2] "=&r" (p2), [p3] "=&r" (p3), [p4] "=&r" (p4), | |
| 499 [qload3] "=&r" (qload3), [p5] "=&r" (p5), | |
| 500 [Temp1] "=&r" (Temp1), [Temp2] "=&r" (Temp2), [Temp3] "=&r" (Temp3) | |
| 501 : [filter45] "r" (filter45), [vector_64] "r" (vector_64), | |
| 502 [cm] "r" (cm), [dst] "r" (dst), [src] "r" (src) | |
| 503 ); | |
| 504 | |
| 505 src += 16; | |
| 506 dst += 16; | |
| 507 } | |
| 508 | |
| 509 /* Next row... */ | |
| 510 src_ptr += src_stride; | |
| 511 dst_ptr += dst_stride; | |
| 512 } | |
| 513 } | |
| 514 | |
| 515 static void convolve_bi_avg_horiz_64_dspr2(const uint8_t *src_ptr, | |
| 516 int32_t src_stride, | |
| 517 uint8_t *dst_ptr, | |
| 518 int32_t dst_stride, | |
| 519 const int16_t *filter_x0, | |
| 520 int32_t h) { | |
| 521 int32_t y, c; | |
| 522 const uint8_t *src; | |
| 523 uint8_t *dst; | |
| 524 uint8_t *cm = vpx_ff_cropTbl; | |
| 525 uint32_t vector_64 = 64; | |
| 526 int32_t Temp1, Temp2, Temp3; | |
| 527 uint32_t qload1, qload2, qload3; | |
| 528 uint32_t p1, p2, p3, p4, p5; | |
| 529 uint32_t st1, st2, st3; | |
| 530 const int16_t *filter = &filter_x0[3]; | |
| 531 uint32_t filter45;; | |
| 532 | |
| 533 filter45 = ((const int32_t *)filter)[0]; | |
| 534 | |
| 535 for (y = h; y--;) { | |
| 536 src = src_ptr; | |
| 537 dst = dst_ptr; | |
| 538 | |
| 539 /* prefetch data to cache memory */ | |
| 540 prefetch_load(src_ptr + src_stride); | |
| 541 prefetch_load(src_ptr + src_stride + 32); | |
| 542 prefetch_load(src_ptr + src_stride + 64); | |
| 543 prefetch_store(dst_ptr + dst_stride); | |
| 544 prefetch_store(dst_ptr + dst_stride + 32); | |
| 545 | |
| 546 for (c = 0; c < 4; c++) { | |
| 547 __asm__ __volatile__ ( | |
| 548 "ulw %[qload1], 0(%[src]) \n\t" | |
| 549 "ulw %[qload2], 4(%[src]) \n\t" | |
| 550 | |
| 551 /* even 1. pixel */ | |
| 552 "mtlo %[vector_64], $ac1 \n\t" /*
even 1 */ | |
| 553 "mthi $zero, $ac1 \n\t" | |
| 554 "mtlo %[vector_64], $ac2 \n\t" /*
even 2 */ | |
| 555 "mthi $zero, $ac2 \n\t" | |
| 556 "preceu.ph.qbr %[p1], %[qload1] \n\t" | |
| 557 "preceu.ph.qbl %[p2], %[qload1] \n\t" | |
| 558 "preceu.ph.qbr %[p3], %[qload2] \n\t" | |
| 559 "preceu.ph.qbl %[p4], %[qload2] \n\t" | |
| 560 "ulw %[qload3], 8(%[src]) \n\t" | |
| 561 "dpa.w.ph $ac1, %[p1], %[filter45] \n\t" /*
even 1 */ | |
| 562 "extp %[Temp1], $ac1, 31 \n\t" /*
even 1 */ | |
| 563 "lbu %[st2], 0(%[dst]) \n\t" /*
load even 1 from dst */ | |
| 564 | |
| 565 /* even 2. pixel */ | |
| 566 "mtlo %[vector_64], $ac3 \n\t" /*
even 3 */ | |
| 567 "mthi $zero, $ac3 \n\t" | |
| 568 "preceu.ph.qbr %[p1], %[qload3] \n\t" | |
| 569 "preceu.ph.qbl %[p5], %[qload3] \n\t" | |
| 570 "ulw %[qload1], 12(%[src]) \n\t" | |
| 571 "dpa.w.ph $ac2, %[p2], %[filter45] \n\t" /*
even 1 */ | |
| 572 "extp %[Temp2], $ac2, 31 \n\t" /*
even 1 */ | |
| 573 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
even 1 */ | |
| 574 | |
| 575 "lbu %[qload3], 2(%[dst]) \n\t" /*
load even 2 from dst */ | |
| 576 | |
| 577 /* even 3. pixel */ | |
| 578 "mtlo %[vector_64], $ac1 \n\t" /*
even 4 */ | |
| 579 "mthi $zero, $ac1 \n\t" | |
| 580 "addqh_r.w %[st2], %[st2], %[st1] \n\t" /*
average even 1 */ | |
| 581 "preceu.ph.qbr %[p2], %[qload1] \n\t" | |
| 582 "sb %[st2], 0(%[dst]) \n\t" /*
store even 1 to dst */ | |
| 583 "dpa.w.ph $ac3, %[p3], %[filter45] \n\t" /*
even 3 */ | |
| 584 "extp %[Temp3], $ac3, 31 \n\t" /*
even 3 */ | |
| 585 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
even 1 */ | |
| 586 | |
| 587 /* even 4. pixel */ | |
| 588 "mtlo %[vector_64], $ac2 \n\t" /*
even 5 */ | |
| 589 "mthi $zero, $ac2 \n\t" | |
| 590 "addqh_r.w %[qload3], %[qload3], %[st2] \n\t" /*
average even 2 */ | |
| 591 "preceu.ph.qbl %[p3], %[qload1] \n\t" | |
| 592 "sb %[qload3], 2(%[dst]) \n\t" /*
store even 2 to dst */ | |
| 593 "lbu %[qload3], 4(%[dst]) \n\t" /*
load even 3 from dst */ | |
| 594 "lbu %[qload1], 6(%[dst]) \n\t" /*
load even 4 from dst */ | |
| 595 "dpa.w.ph $ac1, %[p4], %[filter45] \n\t" /*
even 4 */ | |
| 596 "extp %[Temp1], $ac1, 31 \n\t" /*
even 4 */ | |
| 597 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
even 3 */ | |
| 598 | |
| 599 /* even 5. pixel */ | |
| 600 "mtlo %[vector_64], $ac3 \n\t" /*
even 6 */ | |
| 601 "mthi $zero, $ac3 \n\t" | |
| 602 "addqh_r.w %[qload3], %[qload3], %[st3] \n\t" /*
average even 3 */ | |
| 603 "sb %[qload3], 4(%[dst]) \n\t" /*
store even 3 to dst */ | |
| 604 "dpa.w.ph $ac2, %[p1], %[filter45] \n\t" /*
even 5 */ | |
| 605 "extp %[Temp2], $ac2, 31 \n\t" /*
even 5 */ | |
| 606 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
even 4 */ | |
| 607 | |
| 608 /* even 6. pixel */ | |
| 609 "mtlo %[vector_64], $ac1 \n\t" /*
even 7 */ | |
| 610 "mthi $zero, $ac1 \n\t" | |
| 611 "addqh_r.w %[qload1], %[qload1], %[st1] \n\t" /*
average even 4 */ | |
| 612 "sb %[qload1], 6(%[dst]) \n\t" /*
store even 4 to dst */ | |
| 613 "dpa.w.ph $ac3, %[p5], %[filter45] \n\t" /*
even 6 */ | |
| 614 "lbu %[qload2], 8(%[dst]) \n\t" /*
load even 5 from dst */ | |
| 615 "extp %[Temp3], $ac3, 31 \n\t" /*
even 6 */ | |
| 616 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
even 5 */ | |
| 617 | |
| 618 /* even 7. pixel */ | |
| 619 "mtlo %[vector_64], $ac2 \n\t" /*
even 8 */ | |
| 620 "mthi $zero, $ac2 \n\t" | |
| 621 "addqh_r.w %[qload2], %[qload2], %[st2] \n\t" /*
average even 5 */ | |
| 622 "sb %[qload2], 8(%[dst]) \n\t" /*
store even 5 to dst */ | |
| 623 "dpa.w.ph $ac1, %[p2], %[filter45] \n\t" /*
even 7 */ | |
| 624 "lbu %[qload3], 10(%[dst]) \n\t" /*
load even 6 from dst */ | |
| 625 "extp %[Temp1], $ac1, 31 \n\t" /*
even 7 */ | |
| 626 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
even 6 */ | |
| 627 | |
| 628 "lbu %[st2], 12(%[dst]) \n\t" /*
load even 7 from dst */ | |
| 629 | |
| 630 /* even 8. pixel */ | |
| 631 "mtlo %[vector_64], $ac3 \n\t" /*
odd 1 */ | |
| 632 "mthi $zero, $ac3 \n\t" | |
| 633 "addqh_r.w %[qload3], %[qload3], %[st3] \n\t" /*
average even 6 */ | |
| 634 "dpa.w.ph $ac2, %[p3], %[filter45] \n\t" /*
even 8 */ | |
| 635 "sb %[qload3], 10(%[dst]) \n\t" /*
store even 6 to dst */ | |
| 636 "extp %[Temp2], $ac2, 31 \n\t" /*
even 8 */ | |
| 637 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
even 7 */ | |
| 638 | |
| 639 /* ODD pixels */ | |
| 640 "ulw %[qload1], 1(%[src]) \n\t" | |
| 641 "ulw %[qload2], 5(%[src]) \n\t" | |
| 642 | |
| 643 "addqh_r.w %[st2], %[st2], %[st1] \n\t" /*
average even 7 */ | |
| 644 | |
| 645 /* odd 1. pixel */ | |
| 646 "mtlo %[vector_64], $ac1 \n\t" /*
odd 2 */ | |
| 647 "mthi $zero, $ac1 \n\t" | |
| 648 "preceu.ph.qbr %[p1], %[qload1] \n\t" | |
| 649 "preceu.ph.qbl %[p2], %[qload1] \n\t" | |
| 650 "preceu.ph.qbr %[p3], %[qload2] \n\t" | |
| 651 "preceu.ph.qbl %[p4], %[qload2] \n\t" | |
| 652 "sb %[st2], 12(%[dst]) \n\t" /*
store even 7 to dst */ | |
| 653 "ulw %[qload3], 9(%[src]) \n\t" | |
| 654 "dpa.w.ph $ac3, %[p1], %[filter45] \n\t" /*
odd 1 */ | |
| 655 "lbu %[qload2], 14(%[dst]) \n\t" /*
load even 8 from dst */ | |
| 656 "extp %[Temp3], $ac3, 31 \n\t" /*
odd 1 */ | |
| 657 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
even 8 */ | |
| 658 | |
| 659 "lbu %[st1], 1(%[dst]) \n\t" /*
load odd 1 from dst */ | |
| 660 | |
| 661 /* odd 2. pixel */ | |
| 662 "mtlo %[vector_64], $ac2 \n\t" /*
odd 3 */ | |
| 663 "mthi $zero, $ac2 \n\t" | |
| 664 "addqh_r.w %[qload2], %[qload2], %[st2] \n\t" /*
average even 8 */ | |
| 665 "preceu.ph.qbr %[p1], %[qload3] \n\t" | |
| 666 "preceu.ph.qbl %[p5], %[qload3] \n\t" | |
| 667 "sb %[qload2], 14(%[dst]) \n\t" /*
store even 8 to dst */ | |
| 668 "ulw %[qload1], 13(%[src]) \n\t" | |
| 669 "dpa.w.ph $ac1, %[p2], %[filter45] \n\t" /*
odd 2 */ | |
| 670 "lbu %[qload3], 3(%[dst]) \n\t" /*
load odd 2 from dst */ | |
| 671 "extp %[Temp1], $ac1, 31 \n\t" /*
odd 2 */ | |
| 672 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
odd 1 */ | |
| 673 | |
| 674 /* odd 3. pixel */ | |
| 675 "mtlo %[vector_64], $ac3 \n\t" /*
odd 4 */ | |
| 676 "mthi $zero, $ac3 \n\t" | |
| 677 "addqh_r.w %[st3], %[st3], %[st1] \n\t" /*
average odd 1 */ | |
| 678 "preceu.ph.qbr %[p2], %[qload1] \n\t" | |
| 679 "dpa.w.ph $ac2, %[p3], %[filter45] \n\t" /*
odd 3 */ | |
| 680 "sb %[st3], 1(%[dst]) \n\t" /*
store odd 1 to dst */ | |
| 681 "extp %[Temp2], $ac2, 31 \n\t" /*
odd 3 */ | |
| 682 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
odd 2 */ | |
| 683 | |
| 684 /* odd 4. pixel */ | |
| 685 "mtlo %[vector_64], $ac1 \n\t" /*
odd 5 */ | |
| 686 "mthi $zero, $ac1 \n\t" | |
| 687 "addqh_r.w %[qload3], %[qload3], %[st1] \n\t" /*
average odd 2 */ | |
| 688 "preceu.ph.qbl %[p3], %[qload1] \n\t" | |
| 689 "sb %[qload3], 3(%[dst]) \n\t" /*
store odd 2 to dst */ | |
| 690 "lbu %[qload1], 5(%[dst]) \n\t" /*
load odd 3 from dst */ | |
| 691 "dpa.w.ph $ac3, %[p4], %[filter45] \n\t" /*
odd 4 */ | |
| 692 "extp %[Temp3], $ac3, 31 \n\t" /*
odd 4 */ | |
| 693 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
odd 3 */ | |
| 694 | |
| 695 "lbu %[st1], 7(%[dst]) \n\t" /*
load odd 4 from dst */ | |
| 696 | |
| 697 /* odd 5. pixel */ | |
| 698 "mtlo %[vector_64], $ac2 \n\t" /*
odd 6 */ | |
| 699 "mthi $zero, $ac2 \n\t" | |
| 700 "addqh_r.w %[qload1], %[qload1], %[st2] \n\t" /*
average odd 3 */ | |
| 701 "sb %[qload1], 5(%[dst]) \n\t" /*
store odd 3 to dst */ | |
| 702 "dpa.w.ph $ac1, %[p1], %[filter45] \n\t" /*
odd 5 */ | |
| 703 "extp %[Temp1], $ac1, 31 \n\t" /*
odd 5 */ | |
| 704 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
odd 4 */ | |
| 705 | |
| 706 "lbu %[qload1], 9(%[dst]) \n\t" /*
load odd 5 from dst */ | |
| 707 | |
| 708 /* odd 6. pixel */ | |
| 709 "mtlo %[vector_64], $ac3 \n\t" /*
odd 7 */ | |
| 710 "mthi $zero, $ac3 \n\t" | |
| 711 "addqh_r.w %[st1], %[st1], %[st3] \n\t" /*
average odd 4 */ | |
| 712 "sb %[st1], 7(%[dst]) \n\t" /*
store odd 4 to dst */ | |
| 713 "dpa.w.ph $ac2, %[p5], %[filter45] \n\t" /*
odd 6 */ | |
| 714 "extp %[Temp2], $ac2, 31 \n\t" /*
odd 6 */ | |
| 715 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
odd 5 */ | |
| 716 | |
| 717 /* odd 7. pixel */ | |
| 718 "mtlo %[vector_64], $ac1 \n\t" /*
odd 8 */ | |
| 719 "mthi $zero, $ac1 \n\t" | |
| 720 "addqh_r.w %[qload1], %[qload1], %[st1] \n\t" /*
average odd 5 */ | |
| 721 "sb %[qload1], 9(%[dst]) \n\t" /*
store odd 5 to dst */ | |
| 722 "lbu %[qload2], 11(%[dst]) \n\t" /*
load odd 6 from dst */ | |
| 723 "dpa.w.ph $ac3, %[p2], %[filter45] \n\t" /*
odd 7 */ | |
| 724 "extp %[Temp3], $ac3, 31 \n\t" /*
odd 7 */ | |
| 725 | |
| 726 "lbu %[qload3], 13(%[dst]) \n\t" /*
load odd 7 from dst */ | |
| 727 | |
| 728 /* odd 8. pixel */ | |
| 729 "dpa.w.ph $ac1, %[p3], %[filter45] \n\t" /*
odd 8 */ | |
| 730 "extp %[Temp1], $ac1, 31 \n\t" /*
odd 8 */ | |
| 731 | |
| 732 "lbu %[qload1], 15(%[dst]) \n\t" /*
load odd 8 from dst */ | |
| 733 | |
| 734 "lbux %[st2], %[Temp2](%[cm]) \n\t" /*
odd 6 */ | |
| 735 "addqh_r.w %[qload2], %[qload2], %[st2] \n\t" /*
average odd 6 */ | |
| 736 | |
| 737 "lbux %[st3], %[Temp3](%[cm]) \n\t" /*
odd 7 */ | |
| 738 "addqh_r.w %[qload3], %[qload3], %[st3] \n\t" /*
average odd 7 */ | |
| 739 | |
| 740 "lbux %[st1], %[Temp1](%[cm]) \n\t" /*
odd 8 */ | |
| 741 "addqh_r.w %[qload1], %[qload1], %[st1] \n\t" /*
average odd 8 */ | |
| 742 | |
| 743 "sb %[qload2], 11(%[dst]) \n\t" /*
store odd 6 to dst */ | |
| 744 "sb %[qload3], 13(%[dst]) \n\t" /*
store odd 7 to dst */ | |
| 745 "sb %[qload1], 15(%[dst]) \n\t" /*
store odd 8 to dst */ | |
| 746 | |
| 747 : [qload1] "=&r" (qload1), [qload2] "=&r" (qload2), | |
| 748 [st1] "=&r" (st1), [st2] "=&r" (st2), [st3] "=&r" (st3), | |
| 749 [p1] "=&r" (p1), [p2] "=&r" (p2), [p3] "=&r" (p3), [p4] "=&r" (p4), | |
| 750 [qload3] "=&r" (qload3), [p5] "=&r" (p5), | |
| 751 [Temp1] "=&r" (Temp1), [Temp2] "=&r" (Temp2), [Temp3] "=&r" (Temp3) | |
| 752 : [filter45] "r" (filter45), [vector_64] "r" (vector_64), | |
| 753 [cm] "r" (cm), [dst] "r" (dst), [src] "r" (src) | |
| 754 ); | |
| 755 | |
| 756 src += 16; | |
| 757 dst += 16; | |
| 758 } | |
| 759 | |
| 760 /* Next row... */ | |
| 761 src_ptr += src_stride; | |
| 762 dst_ptr += dst_stride; | |
| 763 } | |
| 764 } | |
| 765 | |
| 766 void vpx_convolve2_avg_horiz_dspr2(const uint8_t *src, ptrdiff_t src_stride, | |
| 767 uint8_t *dst, ptrdiff_t dst_stride, | |
| 768 const int16_t *filter_x, int x_step_q4, | |
| 769 const int16_t *filter_y, int y_step_q4, | |
| 770 int w, int h) { | |
| 771 if (16 == x_step_q4) { | |
| 772 uint32_t pos = 38; | |
| 773 | |
| 774 /* bit positon for extract from acc */ | |
| 775 __asm__ __volatile__ ( | |
| 776 "wrdsp %[pos], 1 \n\t" | |
| 777 : | |
| 778 : [pos] "r" (pos) | |
| 779 ); | |
| 780 | |
| 781 /* prefetch data to cache memory */ | |
| 782 prefetch_load(src); | |
| 783 prefetch_load(src + 32); | |
| 784 prefetch_store(dst); | |
| 785 | |
| 786 switch (w) { | |
| 787 case 4: | |
| 788 convolve_bi_avg_horiz_4_dspr2(src, src_stride, | |
| 789 dst, dst_stride, | |
| 790 filter_x, h); | |
| 791 break; | |
| 792 case 8: | |
| 793 convolve_bi_avg_horiz_8_dspr2(src, src_stride, | |
| 794 dst, dst_stride, | |
| 795 filter_x, h); | |
| 796 break; | |
| 797 case 16: | |
| 798 convolve_bi_avg_horiz_16_dspr2(src, src_stride, | |
| 799 dst, dst_stride, | |
| 800 filter_x, h, 1); | |
| 801 break; | |
| 802 case 32: | |
| 803 convolve_bi_avg_horiz_16_dspr2(src, src_stride, | |
| 804 dst, dst_stride, | |
| 805 filter_x, h, 2); | |
| 806 break; | |
| 807 case 64: | |
| 808 prefetch_load(src + 64); | |
| 809 prefetch_store(dst + 32); | |
| 810 | |
| 811 convolve_bi_avg_horiz_64_dspr2(src, src_stride, | |
| 812 dst, dst_stride, | |
| 813 filter_x, h); | |
| 814 break; | |
| 815 default: | |
| 816 vpx_convolve8_avg_horiz_c(src, src_stride, | |
| 817 dst, dst_stride, | |
| 818 filter_x, x_step_q4, | |
| 819 filter_y, y_step_q4, | |
| 820 w, h); | |
| 821 break; | |
| 822 } | |
| 823 } else { | |
| 824 vpx_convolve8_avg_horiz_c(src, src_stride, | |
| 825 dst, dst_stride, | |
| 826 filter_x, x_step_q4, | |
| 827 filter_y, y_step_q4, | |
| 828 w, h); | |
| 829 } | |
| 830 } | |
| 831 #endif | |
| OLD | NEW |