OLD | NEW |
---|---|
1 /* | 1 /* |
2 * Copyright 2016 The LibYuv Project Authors. All rights reserved. | 2 * Copyright 2016 The LibYuv Project Authors. All rights reserved. |
3 * | 3 * |
4 * Use of this source code is governed by a BSD-style license | 4 * Use of this source code is governed by a BSD-style license |
5 * that can be found in the LICENSE file in the root of the source | 5 * that can be found in the LICENSE file in the root of the source |
6 * tree. An additional intellectual property rights grant can be found | 6 * tree. An additional intellectual property rights grant can be found |
7 * in the file PATENTS. All contributing project authors may | 7 * in the file PATENTS. All contributing project authors may |
8 * be found in the AUTHORS file in the root of the source tree. | 8 * be found in the AUTHORS file in the root of the source tree. |
9 */ | 9 */ |
10 | 10 |
(...skipping 1287 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
1298 dst0 = (v16u8)__msa_ilvr_b((v16i8)vec2, (v16i8)vec0); | 1298 dst0 = (v16u8)__msa_ilvr_b((v16i8)vec2, (v16i8)vec0); |
1299 dst1 = (v16u8)__msa_ilvl_b((v16i8)vec2, (v16i8)vec0); | 1299 dst1 = (v16u8)__msa_ilvl_b((v16i8)vec2, (v16i8)vec0); |
1300 dst2 = (v16u8)__msa_ilvr_b((v16i8)vec3, (v16i8)vec1); | 1300 dst2 = (v16u8)__msa_ilvr_b((v16i8)vec3, (v16i8)vec1); |
1301 dst3 = (v16u8)__msa_ilvl_b((v16i8)vec3, (v16i8)vec1); | 1301 dst3 = (v16u8)__msa_ilvl_b((v16i8)vec3, (v16i8)vec1); |
1302 ST_UB4(dst0, dst1, dst2, dst3, dst_argb, 16); | 1302 ST_UB4(dst0, dst1, dst2, dst3, dst_argb, 16); |
1303 src_argb4444 += 32; | 1303 src_argb4444 += 32; |
1304 dst_argb += 64; | 1304 dst_argb += 64; |
1305 } | 1305 } |
1306 } | 1306 } |
1307 | 1307 |
1308 void ARGB1555ToARGBRow_MSA(const uint8* src_argb1555, | |
1309 uint8* dst_argb, | |
1310 int width) { | |
1311 int x; | |
1312 v8u16 src0, src1; | |
1313 v8u16 vec0, vec1, vec2, vec3, vec4, vec5; | |
1314 v16u8 reg0, reg1, reg2, reg3, reg4, reg5, reg6; | |
1315 v16u8 dst0, dst1, dst2, dst3; | |
1316 v8u16 const_0x1F = (v8u16)__msa_ldi_h(0x1F); | |
1317 | |
1318 for (x = 0; x < width; x += 16) { | |
1319 src0 = (v8u16)__msa_ld_h((v8u16*)src_argb1555, 0); | |
1320 src1 = (v8u16)__msa_ld_h((v8u16*)src_argb1555, 16); | |
1321 vec0 = src0 & const_0x1F; | |
1322 vec1 = src1 & const_0x1F; | |
1323 src0 = (v8u16)__msa_srli_h((v8i16)src0, 5); | |
1324 src1 = (v8u16)__msa_srli_h((v8i16)src1, 5); | |
1325 vec2 = src0 & const_0x1F; | |
1326 vec3 = src1 & const_0x1F; | |
1327 src0 = (v8u16)__msa_srli_h((v8i16)src0, 5); | |
1328 src1 = (v8u16)__msa_srli_h((v8i16)src1, 5); | |
1329 vec4 = src0 & const_0x1F; | |
1330 vec5 = src1 & const_0x1F; | |
1331 src0 = (v8u16)__msa_srli_h((v8i16)src0, 5); | |
1332 src1 = (v8u16)__msa_srli_h((v8i16)src1, 5); | |
1333 reg0 = (v16u8)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); | |
1334 reg1 = (v16u8)__msa_pckev_b((v16i8)vec3, (v16i8)vec2); | |
1335 reg2 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec4); | |
1336 reg3 = (v16u8)__msa_pckev_b((v16i8)src1, (v16i8)src0); | |
1337 reg4 = (v16u8)__msa_slli_b((v16i8)reg0, 3); | |
1338 reg5 = (v16u8)__msa_slli_b((v16i8)reg1, 3); | |
1339 reg6 = (v16u8)__msa_slli_b((v16i8)reg2, 3); | |
1340 reg4 |= (v16u8)__msa_srai_b((v16i8)reg0, 2); | |
1341 reg5 |= (v16u8)__msa_srai_b((v16i8)reg1, 2); | |
1342 reg6 |= (v16u8)__msa_srai_b((v16i8)reg2, 2); | |
1343 reg3 = -reg3; | |
1344 reg0 = (v16u8)__msa_ilvr_b((v16i8)reg6, (v16i8)reg4); | |
1345 reg1 = (v16u8)__msa_ilvl_b((v16i8)reg6, (v16i8)reg4); | |
1346 reg2 = (v16u8)__msa_ilvr_b((v16i8)reg3, (v16i8)reg5); | |
1347 reg3 = (v16u8)__msa_ilvl_b((v16i8)reg3, (v16i8)reg5); | |
1348 dst0 = (v16u8)__msa_ilvr_b((v16i8)reg2, (v16i8)reg0); | |
1349 dst1 = (v16u8)__msa_ilvl_b((v16i8)reg2, (v16i8)reg0); | |
1350 dst2 = (v16u8)__msa_ilvr_b((v16i8)reg3, (v16i8)reg1); | |
1351 dst3 = (v16u8)__msa_ilvl_b((v16i8)reg3, (v16i8)reg1); | |
1352 ST_UB4(dst0, dst1, dst2, dst3, dst_argb, 16); | |
1353 src_argb1555 += 32; | |
1354 dst_argb += 64; | |
1355 } | |
1356 } | |
1357 | |
1358 void RGB565ToARGBRow_MSA(const uint8* src_rgb565, uint8* dst_argb, int width) { | |
1359 int x; | |
1360 v8u16 src0, src1, vec0, vec1, vec2, vec3, vec4, vec5; | |
1361 v8u16 reg0, reg1, reg2, reg3, reg4, reg5; | |
1362 v16u8 res0, res1, res2, res3, dst0, dst1, dst2, dst3; | |
1363 v16u8 const_0xFF = (v16u8)__msa_ldi_b(0xFF); | |
1364 v8u16 const_0x1F = (v8u16)__msa_ldi_h(0x1F); | |
1365 v8u16 const_0x7E0 = (v8u16)__msa_fill_h(0x7E0); | |
1366 v8u16 const_0xF800 = (v8u16)__msa_fill_h(0xF800); | |
1367 | |
1368 for (x = 0; x < width; x += 16) { | |
1369 src0 = (v8u16)__msa_ld_h((v8u16*)src_rgb565, 0); | |
1370 src1 = (v8u16)__msa_ld_h((v8u16*)src_rgb565, 16); | |
1371 vec0 = src0 & const_0x1F; | |
1372 vec1 = src0 & const_0x7E0; | |
1373 vec2 = src0 & const_0xF800; | |
1374 vec3 = src1 & const_0x1F; | |
1375 vec4 = src1 & const_0x7E0; | |
1376 vec5 = src1 & const_0xF800; | |
1377 reg0 = (v8u16)__msa_slli_h((v8i16)vec0, 3); | |
1378 reg1 = (v8u16)__msa_srli_h((v8i16)vec1, 3); | |
1379 reg2 = (v8u16)__msa_srli_h((v8i16)vec2, 8); | |
1380 reg3 = (v8u16)__msa_slli_h((v8i16)vec3, 3); | |
1381 reg4 = (v8u16)__msa_srli_h((v8i16)vec4, 3); | |
1382 reg5 = (v8u16)__msa_srli_h((v8i16)vec5, 8); | |
1383 reg0 |= (v8u16)__msa_srli_h((v8i16)vec0, 2); | |
1384 reg1 |= (v8u16)__msa_srli_h((v8i16)vec1, 9); | |
1385 reg2 |= (v8u16)__msa_srli_h((v8i16)vec2, 13); | |
1386 reg3 |= (v8u16)__msa_srli_h((v8i16)vec3, 2); | |
1387 reg4 |= (v8u16)__msa_srli_h((v8i16)vec4, 9); | |
1388 reg5 |= (v8u16)__msa_srli_h((v8i16)vec5, 13); | |
1389 res0 = (v16u8)__msa_ilvev_b((v16i8)reg2, (v16i8)reg0); | |
1390 res1 = (v16u8)__msa_ilvev_b((v16i8)const_0xFF, (v16i8)reg1); | |
1391 res2 = (v16u8)__msa_ilvev_b((v16i8)reg5, (v16i8)reg3); | |
1392 res3 = (v16u8)__msa_ilvev_b((v16i8)const_0xFF, (v16i8)reg4); | |
1393 dst0 = (v16u8)__msa_ilvr_b((v16i8)res1, (v16i8)res0); | |
1394 dst1 = (v16u8)__msa_ilvl_b((v16i8)res1, (v16i8)res0); | |
1395 dst2 = (v16u8)__msa_ilvr_b((v16i8)res3, (v16i8)res2); | |
1396 dst3 = (v16u8)__msa_ilvl_b((v16i8)res3, (v16i8)res2); | |
1397 ST_UB4(dst0, dst1, dst2, dst3, dst_argb, 16); | |
1398 src_rgb565 += 32; | |
1399 dst_argb += 64; | |
1400 } | |
1401 } | |
1402 | |
1403 void RGB24ToARGBRow_MSA(const uint8* src_rgb24, uint8* dst_argb, int width) { | |
1404 int x; | |
1405 v16u8 src0, src1, src2; | |
1406 v16u8 vec0, vec1, vec2; | |
1407 v16u8 dst0, dst1, dst2, dst3; | |
1408 v16u8 const_0xFF = (v16u8)__msa_ldi_b(0xFF); | |
1409 v16i8 shuffler = {0, 1, 2, 16, 3, 4, 5, 17, 6, 7, 8, 18, 9, 10, 11, 19}; | |
1410 | |
1411 for (x = 0; x < width; x += 16) { | |
1412 src0 = (v16u8)__msa_ld_b((v16i8*)src_rgb24, 0); | |
1413 src1 = (v16u8)__msa_ld_b((v16i8*)src_rgb24, 16); | |
1414 src2 = (v16u8)__msa_ld_b((v16i8*)src_rgb24, 32); | |
1415 vec0 = (v16u8)__msa_sldi_b((v16i8)src1, (v16i8)src0, 12); | |
1416 vec1 = (v16u8)__msa_sldi_b((v16i8)src2, (v16i8)src1, 8); | |
1417 vec2 = (v16u8)__msa_sldi_b((v16i8)src2, (v16i8)src2, 4); | |
1418 dst0 = (v16u8)__msa_vshf_b(shuffler, (v16i8)const_0xFF, (v16i8)src0); | |
1419 dst1 = (v16u8)__msa_vshf_b(shuffler, (v16i8)const_0xFF, (v16i8)vec0); | |
1420 dst2 = (v16u8)__msa_vshf_b(shuffler, (v16i8)const_0xFF, (v16i8)vec1); | |
1421 dst3 = (v16u8)__msa_vshf_b(shuffler, (v16i8)const_0xFF, (v16i8)vec2); | |
1422 ST_UB4(dst0, dst1, dst2, dst3, dst_argb, 16); | |
1423 src_rgb24 += 48; | |
1424 dst_argb += 64; | |
1425 } | |
1426 } | |
1427 | |
1428 void RAWToARGBRow_MSA(const uint8* src_raw, uint8* dst_argb, int width) { | |
1429 int x; | |
1430 v16u8 src0, src1, src2; | |
1431 v16u8 vec0, vec1, vec2; | |
1432 v16u8 dst0, dst1, dst2, dst3; | |
1433 v16u8 const_0xFF = (v16u8)__msa_ldi_b(0xFF); | |
1434 v16i8 mask = {2, 1, 0, 16, 5, 4, 3, 17, 8, 7, 6, 18, 11, 10, 9, 19}; | |
1435 | |
1436 for (x = 0; x < width; x += 16) { | |
1437 src0 = (v16u8)__msa_ld_b((v16i8*)src_raw, 0); | |
1438 src1 = (v16u8)__msa_ld_b((v16i8*)src_raw, 16); | |
1439 src2 = (v16u8)__msa_ld_b((v16i8*)src_raw, 32); | |
1440 vec0 = (v16u8)__msa_sldi_b((v16i8)src1, (v16i8)src0, 12); | |
1441 vec1 = (v16u8)__msa_sldi_b((v16i8)src2, (v16i8)src1, 8); | |
1442 vec2 = (v16u8)__msa_sldi_b((v16i8)src2, (v16i8)src2, 4); | |
1443 dst0 = (v16u8)__msa_vshf_b(mask, (v16i8)const_0xFF, (v16i8)src0); | |
1444 dst1 = (v16u8)__msa_vshf_b(mask, (v16i8)const_0xFF, (v16i8)vec0); | |
1445 dst2 = (v16u8)__msa_vshf_b(mask, (v16i8)const_0xFF, (v16i8)vec1); | |
1446 dst3 = (v16u8)__msa_vshf_b(mask, (v16i8)const_0xFF, (v16i8)vec2); | |
1447 ST_UB4(dst0, dst1, dst2, dst3, dst_argb, 16); | |
1448 src_raw += 48; | |
1449 dst_argb += 64; | |
1450 } | |
1451 } | |
1452 | |
1453 void ARGB1555ToYRow_MSA(const uint8* src_argb1555, uint8* dst_y, int width) { | |
1454 int x; | |
1455 v8u16 src0, src1, vec0, vec1, vec2, vec3, vec4, vec5; | |
1456 v8u16 reg0, reg1, reg2, reg3, reg4, reg5; | |
1457 v16u8 dst0; | |
1458 v8u16 const_0x19 = (v8u16)__msa_ldi_h(0x19); | |
1459 v8u16 const_0x81 = (v8u16)__msa_ldi_h(0x81); | |
1460 v8u16 const_0x42 = (v8u16)__msa_ldi_h(0x42); | |
1461 v8u16 const_0x1F = (v8u16)__msa_ldi_h(0x1F); | |
1462 v8u16 const_0x1080 = (v8u16)__msa_fill_h(0x1080); | |
1463 | |
1464 for (x = 0; x < width; x += 16) { | |
1465 src0 = (v8u16)__msa_ld_b((v8i16*)src_argb1555, 0); | |
1466 src1 = (v8u16)__msa_ld_b((v8i16*)src_argb1555, 16); | |
1467 vec0 = src0 & const_0x1F; | |
1468 vec1 = src1 & const_0x1F; | |
1469 src0 = (v8u16)__msa_srai_h((v8i16)src0, 5); | |
1470 src1 = (v8u16)__msa_srai_h((v8i16)src1, 5); | |
1471 vec2 = src0 & const_0x1F; | |
1472 vec3 = src1 & const_0x1F; | |
1473 src0 = (v8u16)__msa_srai_h((v8i16)src0, 5); | |
1474 src1 = (v8u16)__msa_srai_h((v8i16)src1, 5); | |
1475 vec4 = src0 & const_0x1F; | |
1476 vec5 = src1 & const_0x1F; | |
1477 reg0 = (v8u16)__msa_slli_h((v8i16)vec0, 3); | |
1478 reg1 = (v8u16)__msa_slli_h((v8i16)vec1, 3); | |
1479 reg0 |= (v8u16)__msa_srai_h((v8i16)vec0, 2); | |
1480 reg1 |= (v8u16)__msa_srai_h((v8i16)vec1, 2); | |
1481 reg2 = (v8u16)__msa_slli_h((v8i16)vec2, 3); | |
1482 reg3 = (v8u16)__msa_slli_h((v8i16)vec3, 3); | |
1483 reg2 |= (v8u16)__msa_srai_h((v8i16)vec2, 2); | |
1484 reg3 |= (v8u16)__msa_srai_h((v8i16)vec3, 2); | |
1485 reg4 = (v8u16)__msa_slli_h((v8i16)vec4, 3); | |
1486 reg5 = (v8u16)__msa_slli_h((v8i16)vec5, 3); | |
1487 reg4 |= (v8u16)__msa_srai_h((v8i16)vec4, 2); | |
1488 reg5 |= (v8u16)__msa_srai_h((v8i16)vec5, 2); | |
1489 reg0 *= const_0x19; | |
1490 reg1 *= const_0x19; | |
1491 reg2 *= const_0x81; | |
1492 reg3 *= const_0x81; | |
1493 reg4 *= const_0x42; | |
1494 reg5 *= const_0x42; | |
1495 reg0 += reg2; | |
1496 reg1 += reg3; | |
1497 reg0 += reg4; | |
1498 reg1 += reg5; | |
1499 reg0 += const_0x1080; | |
1500 reg1 += const_0x1080; | |
1501 reg0 = (v8u16)__msa_srai_h((v8i16)reg0, 8); | |
1502 reg1 = (v8u16)__msa_srai_h((v8i16)reg1, 8); | |
1503 dst0 = (v16u8)__msa_pckev_b((v16i8)reg1, (v16i8)reg0); | |
1504 ST_UB(dst0, dst_y); | |
1505 src_argb1555 += 32; | |
1506 dst_y += 16; | |
1507 } | |
1508 } | |
1509 | |
1510 void RGB565ToYRow_MSA(const uint8* src_rgb565, uint8* dst_y, int width) { | |
1511 int x; | |
1512 v8u16 src0, src1, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; | |
1513 v8u16 reg0, reg1, reg2, reg3, reg4, reg5; | |
1514 v4u32 res0, res1, res2, res3; | |
1515 v16u8 dst0; | |
1516 v4u32 const_0x810019 = (v4u32)__msa_fill_w(0x810019); | |
fbarchard1
2017/01/11 20:26:31
caveat these constants subject to change.
manojkumar.bhosale
2017/01/13 08:42:35
Acknowledged.
| |
1517 v4u32 const_0x010042 = (v4u32)__msa_fill_w(0x010042); | |
1518 v8i16 const_0x1080 = __msa_fill_h(0x1080); | |
1519 v8u16 const_0x1F = (v8u16)__msa_ldi_h(0x1F); | |
1520 v8u16 const_0x7E0 = (v8u16)__msa_fill_h(0x7E0); | |
1521 v8u16 const_0xF800 = (v8u16)__msa_fill_h(0xF800); | |
1522 | |
1523 for (x = 0; x < width; x += 16) { | |
1524 src0 = (v8u16)__msa_ld_b((v8i16*)src_rgb565, 0); | |
1525 src1 = (v8u16)__msa_ld_b((v8i16*)src_rgb565, 16); | |
1526 vec0 = src0 & const_0x1F; | |
1527 vec1 = src0 & const_0x7E0; | |
1528 vec2 = src0 & const_0xF800; | |
1529 vec3 = src1 & const_0x1F; | |
1530 vec4 = src1 & const_0x7E0; | |
1531 vec5 = src1 & const_0xF800; | |
1532 reg0 = (v8u16)__msa_slli_h((v8i16)vec0, 3); | |
1533 reg1 = (v8u16)__msa_srli_h((v8i16)vec1, 3); | |
1534 reg2 = (v8u16)__msa_srli_h((v8i16)vec2, 8); | |
1535 reg3 = (v8u16)__msa_slli_h((v8i16)vec3, 3); | |
1536 reg4 = (v8u16)__msa_srli_h((v8i16)vec4, 3); | |
1537 reg5 = (v8u16)__msa_srli_h((v8i16)vec5, 8); | |
1538 reg0 |= (v8u16)__msa_srli_h((v8i16)vec0, 2); | |
1539 reg1 |= (v8u16)__msa_srli_h((v8i16)vec1, 9); | |
1540 reg2 |= (v8u16)__msa_srli_h((v8i16)vec2, 13); | |
1541 reg3 |= (v8u16)__msa_srli_h((v8i16)vec3, 2); | |
1542 reg4 |= (v8u16)__msa_srli_h((v8i16)vec4, 9); | |
1543 reg5 |= (v8u16)__msa_srli_h((v8i16)vec5, 13); | |
1544 vec0 = (v8u16)__msa_ilvr_h((v8i16)reg1, (v8i16)reg0); | |
1545 vec1 = (v8u16)__msa_ilvl_h((v8i16)reg1, (v8i16)reg0); | |
1546 vec2 = (v8u16)__msa_ilvr_h((v8i16)reg4, (v8i16)reg3); | |
1547 vec3 = (v8u16)__msa_ilvl_h((v8i16)reg4, (v8i16)reg3); | |
1548 vec4 = (v8u16)__msa_ilvr_h(const_0x1080, (v8i16)reg2); | |
1549 vec5 = (v8u16)__msa_ilvl_h(const_0x1080, (v8i16)reg2); | |
1550 vec6 = (v8u16)__msa_ilvr_h(const_0x1080, (v8i16)reg5); | |
1551 vec7 = (v8u16)__msa_ilvl_h(const_0x1080, (v8i16)reg5); | |
1552 res0 = __msa_dotp_u_w(vec0, (v8u16)const_0x810019); | |
1553 res1 = __msa_dotp_u_w(vec1, (v8u16)const_0x810019); | |
1554 res2 = __msa_dotp_u_w(vec2, (v8u16)const_0x810019); | |
1555 res3 = __msa_dotp_u_w(vec3, (v8u16)const_0x810019); | |
1556 res0 = __msa_dpadd_u_w(res0, vec4, (v8u16)const_0x010042); | |
1557 res1 = __msa_dpadd_u_w(res1, vec5, (v8u16)const_0x010042); | |
1558 res2 = __msa_dpadd_u_w(res2, vec6, (v8u16)const_0x010042); | |
1559 res3 = __msa_dpadd_u_w(res3, vec7, (v8u16)const_0x010042); | |
1560 res0 = (v4u32)__msa_srai_w((v4i32)res0, 8); | |
1561 res1 = (v4u32)__msa_srai_w((v4i32)res1, 8); | |
1562 res2 = (v4u32)__msa_srai_w((v4i32)res2, 8); | |
1563 res3 = (v4u32)__msa_srai_w((v4i32)res3, 8); | |
1564 vec0 = (v8u16)__msa_pckev_h((v8i16)res1, (v8i16)res0); | |
1565 vec1 = (v8u16)__msa_pckev_h((v8i16)res3, (v8i16)res2); | |
1566 dst0 = (v16u8)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); | |
1567 ST_UB(dst0, dst_y); | |
1568 src_rgb565 += 32; | |
1569 dst_y += 16; | |
1570 } | |
1571 } | |
1572 | |
1573 void RGB24ToYRow_MSA(const uint8* src_argb0, uint8* dst_y, int width) { | |
1574 int x; | |
1575 v16u8 src0, src1, src2, reg0, reg1, reg2, reg3, dst0; | |
1576 v8u16 vec0, vec1, vec2, vec3; | |
1577 v8u16 const_0x8119 = (v8u16)__msa_fill_h(0x8119); | |
1578 v8u16 const_0x42 = (v8u16)__msa_fill_h(0x42); | |
1579 v8u16 const_0x1080 = (v8u16)__msa_fill_h(0x1080); | |
1580 v16i8 mask0 = {0, 1, 2, 3, 3, 4, 5, 6, 6, 7, 8, 9, 9, 10, 11, 12}; | |
1581 v16i8 mask1 = {12, 13, 14, 15, 15, 16, 17, 18, | |
1582 18, 19, 20, 21, 21, 22, 23, 24}; | |
1583 v16i8 mask2 = {8, 9, 10, 11, 11, 12, 13, 14, 14, 15, 16, 17, 17, 18, 19, 20}; | |
1584 v16i8 mask3 = {4, 5, 6, 7, 7, 8, 9, 10, 10, 11, 12, 13, 13, 14, 15, 16}; | |
1585 v16i8 zero = {0}; | |
1586 | |
1587 for (x = 0; x < width; x += 16) { | |
1588 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb0, 0); | |
1589 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb0, 16); | |
1590 src2 = (v16u8)__msa_ld_b((v16i8*)src_argb0, 32); | |
1591 reg0 = (v16u8)__msa_vshf_b(mask0, zero, (v16i8)src0); | |
1592 reg1 = (v16u8)__msa_vshf_b(mask1, (v16i8)src1, (v16i8)src0); | |
1593 reg2 = (v16u8)__msa_vshf_b(mask2, (v16i8)src2, (v16i8)src1); | |
1594 reg3 = (v16u8)__msa_vshf_b(mask3, zero, (v16i8)src2); | |
1595 vec0 = (v8u16)__msa_pckev_h((v8i16)reg1, (v8i16)reg0); | |
1596 vec1 = (v8u16)__msa_pckev_h((v8i16)reg3, (v8i16)reg2); | |
1597 vec2 = (v8u16)__msa_pckod_h((v8i16)reg1, (v8i16)reg0); | |
1598 vec3 = (v8u16)__msa_pckod_h((v8i16)reg3, (v8i16)reg2); | |
1599 vec0 = __msa_dotp_u_h((v16u8)vec0, (v16u8)const_0x8119); | |
1600 vec1 = __msa_dotp_u_h((v16u8)vec1, (v16u8)const_0x8119); | |
1601 vec0 = __msa_dpadd_u_h(vec0, (v16u8)vec2, (v16u8)const_0x42); | |
1602 vec1 = __msa_dpadd_u_h(vec1, (v16u8)vec3, (v16u8)const_0x42); | |
1603 vec0 += const_0x1080; | |
1604 vec1 += const_0x1080; | |
1605 vec0 = (v8u16)__msa_srai_h((v8i16)vec0, 8); | |
1606 vec1 = (v8u16)__msa_srai_h((v8i16)vec1, 8); | |
1607 dst0 = (v16u8)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); | |
1608 ST_UB(dst0, dst_y); | |
1609 src_argb0 += 48; | |
1610 dst_y += 16; | |
1611 } | |
1612 } | |
1613 | |
1614 void RAWToYRow_MSA(const uint8* src_argb0, uint8* dst_y, int width) { | |
1615 int x; | |
1616 v16u8 src0, src1, src2, reg0, reg1, reg2, reg3, dst0; | |
1617 v8u16 vec0, vec1, vec2, vec3; | |
1618 v8u16 const_0x8142 = (v8u16)__msa_fill_h(0x8142); | |
1619 v8u16 const_0x19 = (v8u16)__msa_fill_h(0x19); | |
1620 v8u16 const_0x1080 = (v8u16)__msa_fill_h(0x1080); | |
1621 v16i8 mask0 = {0, 1, 2, 3, 3, 4, 5, 6, 6, 7, 8, 9, 9, 10, 11, 12}; | |
1622 v16i8 mask1 = {12, 13, 14, 15, 15, 16, 17, 18, | |
1623 18, 19, 20, 21, 21, 22, 23, 24}; | |
1624 v16i8 mask2 = {8, 9, 10, 11, 11, 12, 13, 14, 14, 15, 16, 17, 17, 18, 19, 20}; | |
1625 v16i8 mask3 = {4, 5, 6, 7, 7, 8, 9, 10, 10, 11, 12, 13, 13, 14, 15, 16}; | |
1626 v16i8 zero = {0}; | |
1627 | |
1628 for (x = 0; x < width; x += 16) { | |
1629 src0 = (v16u8)__msa_ld_b((v16i8*)src_argb0, 0); | |
1630 src1 = (v16u8)__msa_ld_b((v16i8*)src_argb0, 16); | |
1631 src2 = (v16u8)__msa_ld_b((v16i8*)src_argb0, 32); | |
1632 reg0 = (v16u8)__msa_vshf_b(mask0, zero, (v16i8)src0); | |
1633 reg1 = (v16u8)__msa_vshf_b(mask1, (v16i8)src1, (v16i8)src0); | |
1634 reg2 = (v16u8)__msa_vshf_b(mask2, (v16i8)src2, (v16i8)src1); | |
1635 reg3 = (v16u8)__msa_vshf_b(mask3, zero, (v16i8)src2); | |
1636 vec0 = (v8u16)__msa_pckev_h((v8i16)reg1, (v8i16)reg0); | |
1637 vec1 = (v8u16)__msa_pckev_h((v8i16)reg3, (v8i16)reg2); | |
1638 vec2 = (v8u16)__msa_pckod_h((v8i16)reg1, (v8i16)reg0); | |
1639 vec3 = (v8u16)__msa_pckod_h((v8i16)reg3, (v8i16)reg2); | |
1640 vec0 = __msa_dotp_u_h((v16u8)vec0, (v16u8)const_0x8142); | |
1641 vec1 = __msa_dotp_u_h((v16u8)vec1, (v16u8)const_0x8142); | |
1642 vec0 = __msa_dpadd_u_h(vec0, (v16u8)vec2, (v16u8)const_0x19); | |
1643 vec1 = __msa_dpadd_u_h(vec1, (v16u8)vec3, (v16u8)const_0x19); | |
1644 vec0 += const_0x1080; | |
1645 vec1 += const_0x1080; | |
1646 vec0 = (v8u16)__msa_srai_h((v8i16)vec0, 8); | |
1647 vec1 = (v8u16)__msa_srai_h((v8i16)vec1, 8); | |
1648 dst0 = (v16u8)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); | |
1649 ST_UB(dst0, dst_y); | |
1650 src_argb0 += 48; | |
1651 dst_y += 16; | |
1652 } | |
1653 } | |
1654 | |
1655 void ARGB1555ToUVRow_MSA(const uint8* src_argb1555, | |
1656 int src_stride_argb1555, | |
1657 uint8* dst_u, | |
1658 uint8* dst_v, | |
1659 int width) { | |
1660 int x; | |
1661 const uint16* s = (const uint16*)src_argb1555; | |
1662 const uint16* t = (const uint16*)(src_argb1555 + src_stride_argb1555); | |
1663 int64_t res0, res1; | |
1664 v8u16 src0, src1, src2, src3, reg0, reg1, reg2, reg3; | |
1665 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6; | |
1666 v16u8 dst0; | |
1667 v8u16 const_0x70 = (v8u16)__msa_ldi_h(0x70); | |
1668 v8u16 const_0x4A = (v8u16)__msa_ldi_h(0x4A); | |
1669 v8u16 const_0x26 = (v8u16)__msa_ldi_h(0x26); | |
1670 v8u16 const_0x5E = (v8u16)__msa_ldi_h(0x5E); | |
1671 v8u16 const_0x12 = (v8u16)__msa_ldi_h(0x12); | |
1672 v8u16 const_0x8080 = (v8u16)__msa_fill_h(0x8080); | |
1673 v8u16 const_0x1F = (v8u16)__msa_ldi_h(0x1F); | |
1674 | |
1675 for (x = 0; x < width; x += 16) { | |
1676 src0 = (v8u16)__msa_ld_b((v8i16*)s, 0); | |
1677 src1 = (v8u16)__msa_ld_b((v8i16*)s, 16); | |
1678 src2 = (v8u16)__msa_ld_b((v8i16*)t, 0); | |
1679 src3 = (v8u16)__msa_ld_b((v8i16*)t, 16); | |
1680 vec0 = src0 & const_0x1F; | |
1681 vec1 = src1 & const_0x1F; | |
1682 vec0 += src2 & const_0x1F; | |
1683 vec1 += src3 & const_0x1F; | |
1684 vec0 = (v8u16)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); | |
1685 src0 = (v8u16)__msa_srai_h((v8i16)src0, 5); | |
1686 src1 = (v8u16)__msa_srai_h((v8i16)src1, 5); | |
1687 src2 = (v8u16)__msa_srai_h((v8i16)src2, 5); | |
1688 src3 = (v8u16)__msa_srai_h((v8i16)src3, 5); | |
1689 vec2 = src0 & const_0x1F; | |
1690 vec3 = src1 & const_0x1F; | |
1691 vec2 += src2 & const_0x1F; | |
1692 vec3 += src3 & const_0x1F; | |
1693 vec2 = (v8u16)__msa_pckev_b((v16i8)vec3, (v16i8)vec2); | |
1694 src0 = (v8u16)__msa_srai_h((v8i16)src0, 5); | |
1695 src1 = (v8u16)__msa_srai_h((v8i16)src1, 5); | |
1696 src2 = (v8u16)__msa_srai_h((v8i16)src2, 5); | |
1697 src3 = (v8u16)__msa_srai_h((v8i16)src3, 5); | |
1698 vec4 = src0 & const_0x1F; | |
1699 vec5 = src1 & const_0x1F; | |
1700 vec4 += src2 & const_0x1F; | |
1701 vec5 += src3 & const_0x1F; | |
1702 vec4 = (v8u16)__msa_pckev_b((v16i8)vec5, (v16i8)vec4); | |
1703 vec0 = __msa_hadd_u_h((v16u8)vec0, (v16u8)vec0); | |
1704 vec2 = __msa_hadd_u_h((v16u8)vec2, (v16u8)vec2); | |
1705 vec4 = __msa_hadd_u_h((v16u8)vec4, (v16u8)vec4); | |
1706 vec6 = (v8u16)__msa_slli_h((v8i16)vec0, 1); | |
1707 vec6 |= (v8u16)__msa_srai_h((v8i16)vec0, 6); | |
1708 vec0 = (v8u16)__msa_slli_h((v8i16)vec2, 1); | |
1709 vec0 |= (v8u16)__msa_srai_h((v8i16)vec2, 6); | |
1710 vec2 = (v8u16)__msa_slli_h((v8i16)vec4, 1); | |
1711 vec2 |= (v8u16)__msa_srai_h((v8i16)vec4, 6); | |
1712 reg0 = vec6 * const_0x70; | |
1713 reg1 = vec0 * const_0x4A; | |
1714 reg2 = vec2 * const_0x70; | |
1715 reg3 = vec0 * const_0x5E; | |
1716 reg0 += const_0x8080; | |
1717 reg1 += vec2 * const_0x26; | |
1718 reg2 += const_0x8080; | |
1719 reg3 += vec6 * const_0x12; | |
1720 reg0 -= reg1; | |
1721 reg2 -= reg3; | |
1722 reg0 = (v8u16)__msa_srai_h((v8i16)reg0, 8); | |
1723 reg2 = (v8u16)__msa_srai_h((v8i16)reg2, 8); | |
1724 dst0 = (v16u8)__msa_pckev_b((v16i8)reg2, (v16i8)reg0); | |
1725 res0 = __msa_copy_u_d((v2i64)dst0, 0); | |
1726 res1 = __msa_copy_u_d((v2i64)dst0, 1); | |
1727 SD(res0, dst_u); | |
1728 SD(res1, dst_v); | |
1729 s += 16; | |
1730 t += 16; | |
1731 dst_u += 8; | |
1732 dst_v += 8; | |
1733 } | |
1734 } | |
1735 | |
1736 void RGB565ToUVRow_MSA(const uint8* src_rgb565, | |
fbarchard1
2017/01/11 20:26:31
note these conversions from various RGB formats to
manojkumar.bhosale
2017/01/13 08:42:35
Ok. Keeping them for now as there would be perform
| |
1737 int src_stride_rgb565, | |
1738 uint8* dst_u, | |
1739 uint8* dst_v, | |
1740 int width) { | |
1741 int x; | |
1742 const uint16* s = (const uint16*)src_rgb565; | |
1743 const uint16* t = (const uint16*)(src_rgb565 + src_stride_rgb565); | |
1744 int64_t res0, res1; | |
1745 v8u16 src0, src1, src2, src3, reg0, reg1, reg2, reg3; | |
1746 v8u16 vec0, vec1, vec2, vec3, vec4, vec5; | |
1747 v16u8 dst0; | |
1748 v8u16 const_0x70 = (v8u16)__msa_ldi_h(0x70); | |
1749 v8u16 const_0x4A = (v8u16)__msa_ldi_h(0x4A); | |
1750 v8u16 const_0x26 = (v8u16)__msa_ldi_h(0x26); | |
1751 v8u16 const_0x5E = (v8u16)__msa_ldi_h(0x5E); | |
1752 v8u16 const_0x12 = (v8u16)__msa_ldi_h(0x12); | |
1753 v8u16 const_32896 = (v8u16)__msa_fill_h(0x8080); | |
1754 v8u16 const_0x1F = (v8u16)__msa_ldi_h(0x1F); | |
1755 v8u16 const_0x3F = (v8u16)__msa_fill_h(0x3F); | |
1756 | |
1757 for (x = 0; x < width; x += 16) { | |
1758 src0 = (v8u16)__msa_ld_b((v8i16*)s, 0); | |
1759 src1 = (v8u16)__msa_ld_b((v8i16*)s, 16); | |
1760 src2 = (v8u16)__msa_ld_b((v8i16*)t, 0); | |
1761 src3 = (v8u16)__msa_ld_b((v8i16*)t, 16); | |
1762 vec0 = src0 & const_0x1F; | |
1763 vec1 = src1 & const_0x1F; | |
1764 vec0 += src2 & const_0x1F; | |
1765 vec1 += src3 & const_0x1F; | |
1766 vec0 = (v8u16)__msa_pckev_b((v16i8)vec1, (v16i8)vec0); | |
1767 src0 = (v8u16)__msa_srai_h((v8i16)src0, 5); | |
1768 src1 = (v8u16)__msa_srai_h((v8i16)src1, 5); | |
1769 src2 = (v8u16)__msa_srai_h((v8i16)src2, 5); | |
1770 src3 = (v8u16)__msa_srai_h((v8i16)src3, 5); | |
1771 vec2 = src0 & const_0x3F; | |
1772 vec3 = src1 & const_0x3F; | |
1773 vec2 += src2 & const_0x3F; | |
1774 vec3 += src3 & const_0x3F; | |
1775 vec1 = (v8u16)__msa_pckev_b((v16i8)vec3, (v16i8)vec2); | |
1776 src0 = (v8u16)__msa_srai_h((v8i16)src0, 6); | |
1777 src1 = (v8u16)__msa_srai_h((v8i16)src1, 6); | |
1778 src2 = (v8u16)__msa_srai_h((v8i16)src2, 6); | |
1779 src3 = (v8u16)__msa_srai_h((v8i16)src3, 6); | |
1780 vec4 = src0 & const_0x1F; | |
1781 vec5 = src1 & const_0x1F; | |
1782 vec4 += src2 & const_0x1F; | |
1783 vec5 += src3 & const_0x1F; | |
1784 vec2 = (v8u16)__msa_pckev_b((v16i8)vec5, (v16i8)vec4); | |
1785 vec0 = __msa_hadd_u_h((v16u8)vec0, (v16u8)vec0); | |
1786 vec1 = __msa_hadd_u_h((v16u8)vec1, (v16u8)vec1); | |
1787 vec2 = __msa_hadd_u_h((v16u8)vec2, (v16u8)vec2); | |
1788 vec3 = (v8u16)__msa_slli_h((v8i16)vec0, 1); | |
1789 vec3 |= (v8u16)__msa_srai_h((v8i16)vec0, 6); | |
1790 vec4 = (v8u16)__msa_slli_h((v8i16)vec2, 1); | |
1791 vec4 |= (v8u16)__msa_srai_h((v8i16)vec2, 6); | |
1792 reg0 = vec3 * const_0x70; | |
1793 reg1 = vec1 * const_0x4A; | |
1794 reg2 = vec4 * const_0x70; | |
1795 reg3 = vec1 * const_0x5E; | |
1796 reg0 += const_32896; | |
1797 reg1 += vec4 * const_0x26; | |
1798 reg2 += const_32896; | |
1799 reg3 += vec3 * const_0x12; | |
1800 reg0 -= reg1; | |
1801 reg2 -= reg3; | |
1802 reg0 = (v8u16)__msa_srai_h((v8i16)reg0, 8); | |
1803 reg2 = (v8u16)__msa_srai_h((v8i16)reg2, 8); | |
1804 dst0 = (v16u8)__msa_pckev_b((v16i8)reg2, (v16i8)reg0); | |
1805 res0 = __msa_copy_u_d((v2i64)dst0, 0); | |
1806 res1 = __msa_copy_u_d((v2i64)dst0, 1); | |
1807 SD(res0, dst_u); | |
1808 SD(res1, dst_v); | |
1809 s += 16; | |
1810 t += 16; | |
1811 dst_u += 8; | |
1812 dst_v += 8; | |
1813 } | |
1814 } | |
1815 | |
1816 void RGB24ToUVRow_MSA(const uint8* src_rgb0, | |
1817 int src_stride_rgb, | |
1818 uint8* dst_u, | |
1819 uint8* dst_v, | |
1820 int width) { | |
1821 int x; | |
1822 const uint8* s = src_rgb0; | |
1823 const uint8* t = src_rgb0 + src_stride_rgb; | |
1824 int64 res0, res1; | |
1825 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; | |
1826 v16u8 inp0, inp1, inp2, inp3, inp4, inp5; | |
1827 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; | |
1828 v8i16 reg0, reg1, reg2, reg3; | |
1829 v16u8 dst0; | |
1830 v8u16 const_0x70 = (v8u16)__msa_fill_h(0x70); | |
1831 v8u16 const_0x4A = (v8u16)__msa_fill_h(0x4A); | |
1832 v8u16 const_0x26 = (v8u16)__msa_fill_h(0x26); | |
1833 v8u16 const_0x5E = (v8u16)__msa_fill_h(0x5E); | |
1834 v8u16 const_0x12 = (v8u16)__msa_fill_h(0x12); | |
1835 v8u16 const_0x8080 = (v8u16)__msa_fill_h(0x8080); | |
1836 v16i8 mask = {0, 1, 2, 16, 3, 4, 5, 17, 6, 7, 8, 18, 9, 10, 11, 19}; | |
1837 v16i8 zero = {0}; | |
1838 | |
1839 for (x = 0; x < width; x += 16) { | |
1840 inp0 = (v16u8)__msa_ld_b((v16i8*)s, 0); | |
1841 inp1 = (v16u8)__msa_ld_b((v16i8*)s, 16); | |
1842 inp2 = (v16u8)__msa_ld_b((v16i8*)s, 32); | |
1843 inp3 = (v16u8)__msa_ld_b((v16i8*)t, 0); | |
1844 inp4 = (v16u8)__msa_ld_b((v16i8*)t, 16); | |
1845 inp5 = (v16u8)__msa_ld_b((v16i8*)t, 32); | |
1846 src1 = (v16u8)__msa_sldi_b((v16i8)inp1, (v16i8)inp0, 12); | |
1847 src5 = (v16u8)__msa_sldi_b((v16i8)inp4, (v16i8)inp3, 12); | |
1848 src2 = (v16u8)__msa_sldi_b((v16i8)inp2, (v16i8)inp1, 8); | |
1849 src6 = (v16u8)__msa_sldi_b((v16i8)inp5, (v16i8)inp4, 8); | |
1850 src3 = (v16u8)__msa_sldi_b((v16i8)inp2, (v16i8)inp2, 4); | |
1851 src7 = (v16u8)__msa_sldi_b((v16i8)inp5, (v16i8)inp5, 4); | |
1852 src0 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)inp0); | |
1853 src1 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src1); | |
1854 src2 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src2); | |
1855 src3 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src3); | |
1856 src4 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)inp3); | |
1857 src5 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src5); | |
1858 src6 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src6); | |
1859 src7 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src7); | |
1860 vec0 = (v8u16)__msa_ilvr_b((v16i8)src4, (v16i8)src0); | |
1861 vec1 = (v8u16)__msa_ilvl_b((v16i8)src4, (v16i8)src0); | |
1862 vec2 = (v8u16)__msa_ilvr_b((v16i8)src5, (v16i8)src1); | |
1863 vec3 = (v8u16)__msa_ilvl_b((v16i8)src5, (v16i8)src1); | |
1864 vec4 = (v8u16)__msa_ilvr_b((v16i8)src6, (v16i8)src2); | |
1865 vec5 = (v8u16)__msa_ilvl_b((v16i8)src6, (v16i8)src2); | |
1866 vec6 = (v8u16)__msa_ilvr_b((v16i8)src7, (v16i8)src3); | |
1867 vec7 = (v8u16)__msa_ilvl_b((v16i8)src7, (v16i8)src3); | |
1868 vec0 = (v8u16)__msa_hadd_u_h((v16u8)vec0, (v16u8)vec0); | |
1869 vec1 = (v8u16)__msa_hadd_u_h((v16u8)vec1, (v16u8)vec1); | |
1870 vec2 = (v8u16)__msa_hadd_u_h((v16u8)vec2, (v16u8)vec2); | |
1871 vec3 = (v8u16)__msa_hadd_u_h((v16u8)vec3, (v16u8)vec3); | |
1872 vec4 = (v8u16)__msa_hadd_u_h((v16u8)vec4, (v16u8)vec4); | |
1873 vec5 = (v8u16)__msa_hadd_u_h((v16u8)vec5, (v16u8)vec5); | |
1874 vec6 = (v8u16)__msa_hadd_u_h((v16u8)vec6, (v16u8)vec6); | |
1875 vec7 = (v8u16)__msa_hadd_u_h((v16u8)vec7, (v16u8)vec7); | |
1876 reg0 = (v8i16)__msa_pckev_d((v2i64)vec1, (v2i64)vec0); | |
1877 reg1 = (v8i16)__msa_pckev_d((v2i64)vec3, (v2i64)vec2); | |
1878 reg2 = (v8i16)__msa_pckev_d((v2i64)vec5, (v2i64)vec4); | |
1879 reg3 = (v8i16)__msa_pckev_d((v2i64)vec7, (v2i64)vec6); | |
1880 reg0 += (v8i16)__msa_pckod_d((v2i64)vec1, (v2i64)vec0); | |
1881 reg1 += (v8i16)__msa_pckod_d((v2i64)vec3, (v2i64)vec2); | |
1882 reg2 += (v8i16)__msa_pckod_d((v2i64)vec5, (v2i64)vec4); | |
1883 reg3 += (v8i16)__msa_pckod_d((v2i64)vec7, (v2i64)vec6); | |
1884 reg0 = __msa_srai_h((v8i16)reg0, 2); | |
1885 reg1 = __msa_srai_h((v8i16)reg1, 2); | |
1886 reg2 = __msa_srai_h((v8i16)reg2, 2); | |
1887 reg3 = __msa_srai_h((v8i16)reg3, 2); | |
1888 vec4 = (v8u16)__msa_pckev_h(reg1, reg0); | |
1889 vec5 = (v8u16)__msa_pckev_h(reg3, reg2); | |
1890 vec6 = (v8u16)__msa_pckod_h(reg1, reg0); | |
1891 vec7 = (v8u16)__msa_pckod_h(reg3, reg2); | |
1892 vec0 = (v8u16)__msa_pckev_h((v8i16)vec5, (v8i16)vec4); | |
1893 vec1 = (v8u16)__msa_pckev_h((v8i16)vec7, (v8i16)vec6); | |
1894 vec2 = (v8u16)__msa_pckod_h((v8i16)vec5, (v8i16)vec4); | |
1895 vec3 = vec0 * const_0x70; | |
1896 vec4 = vec1 * const_0x4A; | |
1897 vec5 = vec2 * const_0x26; | |
1898 vec2 *= const_0x70; | |
1899 vec1 *= const_0x5E; | |
1900 vec0 *= const_0x12; | |
1901 reg0 = __msa_subv_h((v8i16)vec3, (v8i16)vec4); | |
1902 reg1 = __msa_subv_h((v8i16)const_0x8080, (v8i16)vec5); | |
1903 reg2 = __msa_subv_h((v8i16)vec2, (v8i16)vec1); | |
1904 reg3 = __msa_subv_h((v8i16)const_0x8080, (v8i16)vec0); | |
1905 reg0 += reg1; | |
1906 reg2 += reg3; | |
1907 reg0 = __msa_srai_h(reg0, 8); | |
1908 reg2 = __msa_srai_h(reg2, 8); | |
1909 dst0 = (v16u8)__msa_pckev_b((v16i8)reg2, (v16i8)reg0); | |
1910 res0 = __msa_copy_u_d((v2i64)dst0, 0); | |
1911 res1 = __msa_copy_u_d((v2i64)dst0, 1); | |
1912 SD(res0, dst_u); | |
1913 SD(res1, dst_v); | |
1914 t += 48; | |
1915 s += 48; | |
1916 dst_u += 8; | |
1917 dst_v += 8; | |
1918 } | |
1919 } | |
1920 | |
1921 void RAWToUVRow_MSA(const uint8* src_rgb0, | |
1922 int src_stride_rgb, | |
1923 uint8* dst_u, | |
1924 uint8* dst_v, | |
1925 int width) { | |
1926 int x; | |
1927 const uint8* s = src_rgb0; | |
1928 const uint8* t = src_rgb0 + src_stride_rgb; | |
1929 int64 res0, res1; | |
1930 v16u8 inp0, inp1, inp2, inp3, inp4, inp5; | |
1931 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; | |
1932 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; | |
1933 v8i16 reg0, reg1, reg2, reg3; | |
1934 v16u8 dst0; | |
1935 v8u16 const_0x70 = (v8u16)__msa_fill_h(0x70); | |
1936 v8u16 const_0x4A = (v8u16)__msa_fill_h(0x4A); | |
1937 v8u16 const_0x26 = (v8u16)__msa_fill_h(0x26); | |
1938 v8u16 const_0x5E = (v8u16)__msa_fill_h(0x5E); | |
1939 v8u16 const_0x12 = (v8u16)__msa_fill_h(0x12); | |
1940 v8u16 const_0x8080 = (v8u16)__msa_fill_h(0x8080); | |
1941 v16i8 mask = {0, 1, 2, 16, 3, 4, 5, 17, 6, 7, 8, 18, 9, 10, 11, 19}; | |
1942 v16i8 zero = {0}; | |
1943 | |
1944 for (x = 0; x < width; x += 16) { | |
1945 inp0 = (v16u8)__msa_ld_b((v16i8*)s, 0); | |
1946 inp1 = (v16u8)__msa_ld_b((v16i8*)s, 16); | |
1947 inp2 = (v16u8)__msa_ld_b((v16i8*)s, 32); | |
1948 inp3 = (v16u8)__msa_ld_b((v16i8*)t, 0); | |
1949 inp4 = (v16u8)__msa_ld_b((v16i8*)t, 16); | |
1950 inp5 = (v16u8)__msa_ld_b((v16i8*)t, 32); | |
1951 src1 = (v16u8)__msa_sldi_b((v16i8)inp1, (v16i8)inp0, 12); | |
1952 src5 = (v16u8)__msa_sldi_b((v16i8)inp4, (v16i8)inp3, 12); | |
1953 src2 = (v16u8)__msa_sldi_b((v16i8)inp2, (v16i8)inp1, 8); | |
1954 src6 = (v16u8)__msa_sldi_b((v16i8)inp5, (v16i8)inp4, 8); | |
1955 src3 = (v16u8)__msa_sldi_b((v16i8)inp2, (v16i8)inp2, 4); | |
1956 src7 = (v16u8)__msa_sldi_b((v16i8)inp5, (v16i8)inp5, 4); | |
1957 src0 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)inp0); | |
1958 src1 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src1); | |
1959 src2 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src2); | |
1960 src3 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src3); | |
1961 src4 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)inp3); | |
1962 src5 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src5); | |
1963 src6 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src6); | |
1964 src7 = (v16u8)__msa_vshf_b(mask, (v16i8)zero, (v16i8)src7); | |
1965 vec0 = (v8u16)__msa_ilvr_b((v16i8)src4, (v16i8)src0); | |
1966 vec1 = (v8u16)__msa_ilvl_b((v16i8)src4, (v16i8)src0); | |
1967 vec2 = (v8u16)__msa_ilvr_b((v16i8)src5, (v16i8)src1); | |
1968 vec3 = (v8u16)__msa_ilvl_b((v16i8)src5, (v16i8)src1); | |
1969 vec4 = (v8u16)__msa_ilvr_b((v16i8)src6, (v16i8)src2); | |
1970 vec5 = (v8u16)__msa_ilvl_b((v16i8)src6, (v16i8)src2); | |
1971 vec6 = (v8u16)__msa_ilvr_b((v16i8)src7, (v16i8)src3); | |
1972 vec7 = (v8u16)__msa_ilvl_b((v16i8)src7, (v16i8)src3); | |
1973 vec0 = (v8u16)__msa_hadd_u_h((v16u8)vec0, (v16u8)vec0); | |
1974 vec1 = (v8u16)__msa_hadd_u_h((v16u8)vec1, (v16u8)vec1); | |
1975 vec2 = (v8u16)__msa_hadd_u_h((v16u8)vec2, (v16u8)vec2); | |
1976 vec3 = (v8u16)__msa_hadd_u_h((v16u8)vec3, (v16u8)vec3); | |
1977 vec4 = (v8u16)__msa_hadd_u_h((v16u8)vec4, (v16u8)vec4); | |
1978 vec5 = (v8u16)__msa_hadd_u_h((v16u8)vec5, (v16u8)vec5); | |
1979 vec6 = (v8u16)__msa_hadd_u_h((v16u8)vec6, (v16u8)vec6); | |
1980 vec7 = (v8u16)__msa_hadd_u_h((v16u8)vec7, (v16u8)vec7); | |
1981 reg0 = (v8i16)__msa_pckev_d((v2i64)vec1, (v2i64)vec0); | |
1982 reg1 = (v8i16)__msa_pckev_d((v2i64)vec3, (v2i64)vec2); | |
1983 reg2 = (v8i16)__msa_pckev_d((v2i64)vec5, (v2i64)vec4); | |
1984 reg3 = (v8i16)__msa_pckev_d((v2i64)vec7, (v2i64)vec6); | |
1985 reg0 += (v8i16)__msa_pckod_d((v2i64)vec1, (v2i64)vec0); | |
1986 reg1 += (v8i16)__msa_pckod_d((v2i64)vec3, (v2i64)vec2); | |
1987 reg2 += (v8i16)__msa_pckod_d((v2i64)vec5, (v2i64)vec4); | |
1988 reg3 += (v8i16)__msa_pckod_d((v2i64)vec7, (v2i64)vec6); | |
1989 reg0 = __msa_srai_h(reg0, 2); | |
1990 reg1 = __msa_srai_h(reg1, 2); | |
1991 reg2 = __msa_srai_h(reg2, 2); | |
1992 reg3 = __msa_srai_h(reg3, 2); | |
1993 vec4 = (v8u16)__msa_pckev_h((v8i16)reg1, (v8i16)reg0); | |
1994 vec5 = (v8u16)__msa_pckev_h((v8i16)reg3, (v8i16)reg2); | |
1995 vec6 = (v8u16)__msa_pckod_h((v8i16)reg1, (v8i16)reg0); | |
1996 vec7 = (v8u16)__msa_pckod_h((v8i16)reg3, (v8i16)reg2); | |
1997 vec0 = (v8u16)__msa_pckod_h((v8i16)vec5, (v8i16)vec4); | |
1998 vec1 = (v8u16)__msa_pckev_h((v8i16)vec7, (v8i16)vec6); | |
1999 vec2 = (v8u16)__msa_pckev_h((v8i16)vec5, (v8i16)vec4); | |
2000 vec3 = vec0 * const_0x70; | |
2001 vec4 = vec1 * const_0x4A; | |
2002 vec5 = vec2 * const_0x26; | |
2003 vec2 *= const_0x70; | |
2004 vec1 *= const_0x5E; | |
2005 vec0 *= const_0x12; | |
2006 reg0 = __msa_subv_h((v8i16)vec3, (v8i16)vec4); | |
2007 reg1 = __msa_subv_h((v8i16)const_0x8080, (v8i16)vec5); | |
2008 reg2 = __msa_subv_h((v8i16)vec2, (v8i16)vec1); | |
2009 reg3 = __msa_subv_h((v8i16)const_0x8080, (v8i16)vec0); | |
2010 reg0 += reg1; | |
2011 reg2 += reg3; | |
2012 reg0 = __msa_srai_h(reg0, 8); | |
2013 reg2 = __msa_srai_h(reg2, 8); | |
2014 dst0 = (v16u8)__msa_pckev_b((v16i8)reg2, (v16i8)reg0); | |
2015 res0 = __msa_copy_u_d((v2i64)dst0, 0); | |
2016 res1 = __msa_copy_u_d((v2i64)dst0, 1); | |
2017 SD(res0, dst_u); | |
2018 SD(res1, dst_v); | |
2019 t += 48; | |
2020 s += 48; | |
2021 dst_u += 8; | |
2022 dst_v += 8; | |
2023 } | |
2024 } | |
2025 | |
1308 #ifdef __cplusplus | 2026 #ifdef __cplusplus |
1309 } // extern "C" | 2027 } // extern "C" |
1310 } // namespace libyuv | 2028 } // namespace libyuv |
1311 #endif | 2029 #endif |
1312 | 2030 |
1313 #endif // !defined(LIBYUV_DISABLE_MSA) && defined(__mips_msa) | 2031 #endif // !defined(LIBYUV_DISABLE_MSA) && defined(__mips_msa) |
OLD | NEW |