| Index: src/ia32/assembler-ia32.cc
|
| diff --git a/src/ia32/assembler-ia32.cc b/src/ia32/assembler-ia32.cc
|
| index 35eb6ac80a80510ed24b7b7c1a570e0909559cd0..98caaeb897133bd73352fef683a155c284debd69 100644
|
| --- a/src/ia32/assembler-ia32.cc
|
| +++ b/src/ia32/assembler-ia32.cc
|
| @@ -122,6 +122,10 @@ void CpuFeatures::ProbeImpl(bool cross_compile) {
|
| OSHasAVXSupport()) {
|
| supported_ |= 1u << FMA3;
|
| }
|
| + if (cpu.has_bmi1() && FLAG_enable_bmi1) supported_ |= 1u << BMI1;
|
| + if (cpu.has_bmi2() && FLAG_enable_bmi2) supported_ |= 1u << BMI2;
|
| + if (cpu.has_lzcnt() && FLAG_enable_lzcnt) supported_ |= 1u << LZCNT;
|
| + if (cpu.has_popcnt() && FLAG_enable_popcnt) supported_ |= 1u << POPCNT;
|
| if (strcmp(FLAG_mcpu, "auto") == 0) {
|
| if (cpu.is_atom()) supported_ |= 1u << ATOM;
|
| } else if (strcmp(FLAG_mcpu, "atom") == 0) {
|
| @@ -132,10 +136,14 @@ void CpuFeatures::ProbeImpl(bool cross_compile) {
|
|
|
| void CpuFeatures::PrintTarget() { }
|
| void CpuFeatures::PrintFeatures() {
|
| - printf("SSE3=%d SSE4_1=%d AVX=%d FMA3=%d ATOM=%d\n",
|
| - CpuFeatures::IsSupported(SSE3), CpuFeatures::IsSupported(SSE4_1),
|
| - CpuFeatures::IsSupported(AVX), CpuFeatures::IsSupported(FMA3),
|
| - CpuFeatures::IsSupported(ATOM));
|
| + printf(
|
| + "SSE3=%d SSE4_1=%d AVX=%d FMA3=%d BMI1=%d BMI2=%d LZCNT=%d POPCNT=%d "
|
| + "ATOM=%d\n",
|
| + CpuFeatures::IsSupported(SSE3), CpuFeatures::IsSupported(SSE4_1),
|
| + CpuFeatures::IsSupported(AVX), CpuFeatures::IsSupported(FMA3),
|
| + CpuFeatures::IsSupported(BMI1), CpuFeatures::IsSupported(BMI2),
|
| + CpuFeatures::IsSupported(LZCNT), CpuFeatures::IsSupported(POPCNT),
|
| + CpuFeatures::IsSupported(ATOM));
|
| }
|
|
|
|
|
| @@ -2657,6 +2665,67 @@ void Assembler::vss(byte op, XMMRegister dst, XMMRegister src1,
|
| }
|
|
|
|
|
| +void Assembler::bmi1(byte op, Register reg, Register vreg, const Operand& rm) {
|
| + DCHECK(IsEnabled(BMI1));
|
| + EnsureSpace ensure_space(this);
|
| + emit_vex_prefix(vreg, kLZ, kNone, k0F38, kW0);
|
| + EMIT(op);
|
| + emit_operand(reg, rm);
|
| +}
|
| +
|
| +
|
| +void Assembler::tzcnt(Register dst, const Operand& src) {
|
| + DCHECK(IsEnabled(BMI1));
|
| + EnsureSpace ensure_space(this);
|
| + EMIT(0xF3);
|
| + EMIT(0x0F);
|
| + EMIT(0xBC);
|
| + emit_operand(dst, src);
|
| +}
|
| +
|
| +
|
| +void Assembler::lzcnt(Register dst, const Operand& src) {
|
| + DCHECK(IsEnabled(LZCNT));
|
| + EnsureSpace ensure_space(this);
|
| + EMIT(0xF3);
|
| + EMIT(0x0F);
|
| + EMIT(0xBD);
|
| + emit_operand(dst, src);
|
| +}
|
| +
|
| +
|
| +void Assembler::popcnt(Register dst, const Operand& src) {
|
| + DCHECK(IsEnabled(POPCNT));
|
| + EnsureSpace ensure_space(this);
|
| + EMIT(0xF3);
|
| + EMIT(0x0F);
|
| + EMIT(0xB8);
|
| + emit_operand(dst, src);
|
| +}
|
| +
|
| +
|
| +void Assembler::bmi2(SIMDPrefix pp, byte op, Register reg, Register vreg,
|
| + const Operand& rm) {
|
| + DCHECK(IsEnabled(BMI2));
|
| + EnsureSpace ensure_space(this);
|
| + emit_vex_prefix(vreg, kLZ, pp, k0F38, kW0);
|
| + EMIT(op);
|
| + emit_operand(reg, rm);
|
| +}
|
| +
|
| +
|
| +void Assembler::rorx(Register dst, const Operand& src, byte imm8) {
|
| + DCHECK(IsEnabled(BMI2));
|
| + DCHECK(is_uint8(imm8));
|
| + Register vreg = {0}; // VEX.vvvv unused
|
| + EnsureSpace ensure_space(this);
|
| + emit_vex_prefix(vreg, kLZ, kF2, k0F3A, kW0);
|
| + EMIT(0xF0);
|
| + emit_operand(dst, src);
|
| + EMIT(imm8);
|
| +}
|
| +
|
| +
|
| void Assembler::emit_sse_operand(XMMRegister reg, const Operand& adr) {
|
| Register ireg = { reg.code() };
|
| emit_operand(ireg, adr);
|
| @@ -2682,7 +2751,8 @@ void Assembler::emit_vex_prefix(XMMRegister vreg, VectorLength l, SIMDPrefix pp,
|
| LeadingOpcode mm, VexW w) {
|
| if (mm != k0F || w != kW0) {
|
| EMIT(0xc4);
|
| - EMIT(0xc0 | mm);
|
| + // Change RXB from "110" to "111" to align with gdb disassembler.
|
| + EMIT(0xe0 | mm);
|
| EMIT(w | ((~vreg.code() & 0xf) << 3) | l | pp);
|
| } else {
|
| EMIT(0xc5);
|
| @@ -2691,6 +2761,13 @@ void Assembler::emit_vex_prefix(XMMRegister vreg, VectorLength l, SIMDPrefix pp,
|
| }
|
|
|
|
|
| +void Assembler::emit_vex_prefix(Register vreg, VectorLength l, SIMDPrefix pp,
|
| + LeadingOpcode mm, VexW w) {
|
| + XMMRegister ivreg = {vreg.code()};
|
| + emit_vex_prefix(ivreg, l, pp, mm, w);
|
| +}
|
| +
|
| +
|
| void Assembler::GrowBuffer() {
|
| DCHECK(buffer_overflow());
|
| if (!own_buffer_) FATAL("external code buffer is too small");
|
|
|