OLD | NEW |
1 // Copyright 2012 the V8 project authors. All rights reserved. | 1 // Copyright 2012 the V8 project authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 #include <limits.h> // For LONG_MIN, LONG_MAX. | 5 #include <limits.h> // For LONG_MIN, LONG_MAX. |
6 | 6 |
7 #if V8_TARGET_ARCH_MIPS64 | 7 #if V8_TARGET_ARCH_MIPS64 |
8 | 8 |
9 #include "src/base/division-by-constant.h" | 9 #include "src/base/division-by-constant.h" |
10 #include "src/bootstrapper.h" | 10 #include "src/bootstrapper.h" |
(...skipping 1302 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1313 daddiu(rd, zero_reg, (j.imm64_ & kImm16Mask)); | 1313 daddiu(rd, zero_reg, (j.imm64_ & kImm16Mask)); |
1314 } else if (!(j.imm64_ & kHiMask)) { | 1314 } else if (!(j.imm64_ & kHiMask)) { |
1315 ori(rd, zero_reg, (j.imm64_ & kImm16Mask)); | 1315 ori(rd, zero_reg, (j.imm64_ & kImm16Mask)); |
1316 } else if (!(j.imm64_ & kImm16Mask)) { | 1316 } else if (!(j.imm64_ & kImm16Mask)) { |
1317 lui(rd, (j.imm64_ >> kLuiShift) & kImm16Mask); | 1317 lui(rd, (j.imm64_ >> kLuiShift) & kImm16Mask); |
1318 } else { | 1318 } else { |
1319 lui(rd, (j.imm64_ >> kLuiShift) & kImm16Mask); | 1319 lui(rd, (j.imm64_ >> kLuiShift) & kImm16Mask); |
1320 ori(rd, rd, (j.imm64_ & kImm16Mask)); | 1320 ori(rd, rd, (j.imm64_ & kImm16Mask)); |
1321 } | 1321 } |
1322 } else { | 1322 } else { |
1323 if (is_int48(j.imm64_)) { | 1323 if (kArchVariant == kMips64r6) { |
1324 if ((j.imm64_ >> 32) & kImm16Mask) { | 1324 int64_t imm = j.imm64_; |
1325 lui(rd, (j.imm64_ >> 32) & kImm16Mask); | 1325 bool lui_emited = false; |
1326 if ((j.imm64_ >> 16) & kImm16Mask) { | 1326 if (((imm >> kLuiShift) & kImm16Mask) != 0) { |
1327 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); | 1327 lui(rd, (imm >> kLuiShift) & kImm16Mask); |
| 1328 lui_emited = true; |
| 1329 } |
| 1330 if ((imm & kImm16Mask) != 0) { |
| 1331 ori(rd, rd, (imm & kImm16Mask)); |
| 1332 } else if (!lui_emited) { |
| 1333 or_(rd, zero_reg, zero_reg); |
| 1334 } |
| 1335 if ((imm >> 31) & 0x1) { |
| 1336 imm = (imm >> 32) + 1; |
| 1337 } else { |
| 1338 imm = imm >> 32; |
| 1339 } |
| 1340 if (imm & kImm16Mask) { |
| 1341 dahi(rd, imm & kImm16Mask); |
| 1342 } |
| 1343 if (!is_int48(j.imm64_)) { |
| 1344 if ((imm >> 15) & 0x1) { |
| 1345 imm = (imm >> 16) + 1; |
| 1346 } else { |
| 1347 imm = imm >> 16; |
| 1348 } |
| 1349 if (imm & kImm16Mask) { |
| 1350 dati(rd, imm & kImm16Mask); |
| 1351 } |
| 1352 } |
| 1353 } else { |
| 1354 if (is_int48(j.imm64_)) { |
| 1355 if ((j.imm64_ >> 32) & kImm16Mask) { |
| 1356 lui(rd, (j.imm64_ >> 32) & kImm16Mask); |
| 1357 if ((j.imm64_ >> 16) & kImm16Mask) { |
| 1358 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); |
| 1359 } |
| 1360 } else { |
| 1361 ori(rd, zero_reg, (j.imm64_ >> 16) & kImm16Mask); |
| 1362 } |
| 1363 dsll(rd, rd, 16); |
| 1364 if (j.imm64_ & kImm16Mask) { |
| 1365 ori(rd, rd, j.imm64_ & kImm16Mask); |
1328 } | 1366 } |
1329 } else { | 1367 } else { |
1330 ori(rd, zero_reg, (j.imm64_ >> 16) & kImm16Mask); | 1368 lui(rd, (j.imm64_ >> 48) & kImm16Mask); |
1331 } | 1369 if ((j.imm64_ >> 32) & kImm16Mask) { |
1332 dsll(rd, rd, 16); | 1370 ori(rd, rd, (j.imm64_ >> 32) & kImm16Mask); |
1333 if (j.imm64_ & kImm16Mask) { | 1371 } |
1334 ori(rd, rd, j.imm64_ & kImm16Mask); | 1372 if ((j.imm64_ >> 16) & kImm16Mask) { |
1335 } | |
1336 } else { | |
1337 lui(rd, (j.imm64_ >> 48) & kImm16Mask); | |
1338 if ((j.imm64_ >> 32) & kImm16Mask) { | |
1339 ori(rd, rd, (j.imm64_ >> 32) & kImm16Mask); | |
1340 } | |
1341 if ((j.imm64_ >> 16) & kImm16Mask) { | |
1342 dsll(rd, rd, 16); | |
1343 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); | |
1344 if (j.imm64_ & kImm16Mask) { | |
1345 dsll(rd, rd, 16); | 1373 dsll(rd, rd, 16); |
1346 ori(rd, rd, j.imm64_ & kImm16Mask); | 1374 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); |
| 1375 if (j.imm64_ & kImm16Mask) { |
| 1376 dsll(rd, rd, 16); |
| 1377 ori(rd, rd, j.imm64_ & kImm16Mask); |
| 1378 } else { |
| 1379 dsll(rd, rd, 16); |
| 1380 } |
1347 } else { | 1381 } else { |
1348 dsll(rd, rd, 16); | 1382 if (j.imm64_ & kImm16Mask) { |
1349 } | 1383 dsll32(rd, rd, 0); |
1350 } else { | 1384 ori(rd, rd, j.imm64_ & kImm16Mask); |
1351 if (j.imm64_ & kImm16Mask) { | 1385 } else { |
1352 dsll32(rd, rd, 0); | 1386 dsll32(rd, rd, 0); |
1353 ori(rd, rd, j.imm64_ & kImm16Mask); | 1387 } |
1354 } else { | |
1355 dsll32(rd, rd, 0); | |
1356 } | 1388 } |
1357 } | 1389 } |
1358 } | 1390 } |
1359 } | 1391 } |
1360 } else if (MustUseReg(j.rmode_)) { | 1392 } else if (MustUseReg(j.rmode_)) { |
1361 RecordRelocInfo(j.rmode_, j.imm64_); | 1393 RecordRelocInfo(j.rmode_, j.imm64_); |
1362 lui(rd, (j.imm64_ >> 32) & kImm16Mask); | 1394 lui(rd, (j.imm64_ >> 32) & kImm16Mask); |
1363 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); | 1395 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); |
1364 dsll(rd, rd, 16); | 1396 dsll(rd, rd, 16); |
1365 ori(rd, rd, j.imm64_ & kImm16Mask); | 1397 ori(rd, rd, j.imm64_ & kImm16Mask); |
1366 } else if (mode == ADDRESS_LOAD) { | 1398 } else if (mode == ADDRESS_LOAD) { |
1367 // We always need the same number of instructions as we may need to patch | 1399 // We always need the same number of instructions as we may need to patch |
1368 // this code to load another value which may need all 4 instructions. | 1400 // this code to load another value which may need all 4 instructions. |
1369 lui(rd, (j.imm64_ >> 32) & kImm16Mask); | 1401 lui(rd, (j.imm64_ >> 32) & kImm16Mask); |
1370 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); | 1402 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); |
1371 dsll(rd, rd, 16); | 1403 dsll(rd, rd, 16); |
1372 ori(rd, rd, j.imm64_ & kImm16Mask); | 1404 ori(rd, rd, j.imm64_ & kImm16Mask); |
1373 } else { | 1405 } else { |
1374 lui(rd, (j.imm64_ >> 48) & kImm16Mask); | 1406 if (kArchVariant == kMips64r6) { |
1375 ori(rd, rd, (j.imm64_ >> 32) & kImm16Mask); | 1407 int64_t imm = j.imm64_; |
1376 dsll(rd, rd, 16); | 1408 lui(rd, (imm >> kLuiShift) & kImm16Mask); |
1377 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); | 1409 if (imm & kImm16Mask) { |
1378 dsll(rd, rd, 16); | 1410 ori(rd, rd, (imm & kImm16Mask)); |
1379 ori(rd, rd, j.imm64_ & kImm16Mask); | 1411 } |
| 1412 if ((imm >> 31) & 0x1) { |
| 1413 imm = (imm >> 32) + 1; |
| 1414 } else { |
| 1415 imm = imm >> 32; |
| 1416 } |
| 1417 dahi(rd, imm & kImm16Mask); |
| 1418 if ((imm >> 15) & 0x1) { |
| 1419 imm = (imm >> 16) + 1; |
| 1420 } else { |
| 1421 imm = imm >> 16; |
| 1422 } |
| 1423 dati(rd, imm & kImm16Mask); |
| 1424 } else { |
| 1425 lui(rd, (j.imm64_ >> 48) & kImm16Mask); |
| 1426 ori(rd, rd, (j.imm64_ >> 32) & kImm16Mask); |
| 1427 dsll(rd, rd, 16); |
| 1428 ori(rd, rd, (j.imm64_ >> 16) & kImm16Mask); |
| 1429 dsll(rd, rd, 16); |
| 1430 ori(rd, rd, j.imm64_ & kImm16Mask); |
| 1431 } |
1380 } | 1432 } |
1381 } | 1433 } |
1382 | 1434 |
1383 | 1435 |
1384 void MacroAssembler::MultiPush(RegList regs) { | 1436 void MacroAssembler::MultiPush(RegList regs) { |
1385 int16_t num_to_push = NumberOfBitsSet(regs); | 1437 int16_t num_to_push = NumberOfBitsSet(regs); |
1386 int16_t stack_offset = num_to_push * kPointerSize; | 1438 int16_t stack_offset = num_to_push * kPointerSize; |
1387 | 1439 |
1388 Dsubu(sp, sp, Operand(stack_offset)); | 1440 Dsubu(sp, sp, Operand(stack_offset)); |
1389 for (int16_t i = kNumRegisters - 1; i >= 0; i--) { | 1441 for (int16_t i = kNumRegisters - 1; i >= 0; i--) { |
(...skipping 5115 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6505 if (mag.shift > 0) sra(result, result, mag.shift); | 6557 if (mag.shift > 0) sra(result, result, mag.shift); |
6506 srl(at, dividend, 31); | 6558 srl(at, dividend, 31); |
6507 Addu(result, result, Operand(at)); | 6559 Addu(result, result, Operand(at)); |
6508 } | 6560 } |
6509 | 6561 |
6510 | 6562 |
6511 } // namespace internal | 6563 } // namespace internal |
6512 } // namespace v8 | 6564 } // namespace v8 |
6513 | 6565 |
6514 #endif // V8_TARGET_ARCH_MIPS64 | 6566 #endif // V8_TARGET_ARCH_MIPS64 |
OLD | NEW |