mirror of
https://git.suyu.dev/suyu/dynarmic.git
synced 2026-03-10 06:46:27 +00:00
IR: Vector instructions now take esize argument in emitter
This commit is contained in:
@@ -18,18 +18,7 @@ bool TranslatorVisitor::DUP_gen(bool Q, Imm<5> imm5, Reg Rn, Vec Vd) {
|
||||
|
||||
const IR::UAny element = X(esize, Rn);
|
||||
|
||||
const IR::U128 result = [&]{
|
||||
switch (esize) {
|
||||
case 8:
|
||||
return Q ? ir.VectorBroadcast8(element) : ir.VectorBroadcastLower8(element);
|
||||
case 16:
|
||||
return Q ? ir.VectorBroadcast16(element) : ir.VectorBroadcastLower16(element);
|
||||
case 32:
|
||||
return Q ? ir.VectorBroadcast32(element) : ir.VectorBroadcastLower32(element);
|
||||
default:
|
||||
return ir.VectorBroadcast64(element);
|
||||
}
|
||||
}();
|
||||
const IR::U128 result = Q ? ir.VectorBroadcast(esize, element) : ir.VectorBroadcastLower(esize, element);
|
||||
|
||||
V(datasize, Vd, result);
|
||||
|
||||
|
||||
@@ -16,7 +16,7 @@ bool TranslatorVisitor::MOVI(bool Q, bool op, Imm<1> a, Imm<1> b, Imm<1> c, Imm<
|
||||
// also FMOV (vector, immediate) when cmode == 0b1111
|
||||
const auto movi = [&]{
|
||||
u64 imm64 = AdvSIMDExpandImm(op, cmode, concatenate(a, b, c, d, e, f, g, h));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast64(ir.Imm64(imm64));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast(64, ir.Imm64(imm64));
|
||||
V(128, Vd, imm);
|
||||
return true;
|
||||
};
|
||||
@@ -24,7 +24,7 @@ bool TranslatorVisitor::MOVI(bool Q, bool op, Imm<1> a, Imm<1> b, Imm<1> c, Imm<
|
||||
// MVNI
|
||||
const auto mvni = [&]{
|
||||
u64 imm64 = ~AdvSIMDExpandImm(op, cmode, concatenate(a, b, c, d, e, f, g, h));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast64(ir.Imm64(imm64));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast(64, ir.Imm64(imm64));
|
||||
V(128, Vd, imm);
|
||||
return true;
|
||||
};
|
||||
@@ -32,7 +32,7 @@ bool TranslatorVisitor::MOVI(bool Q, bool op, Imm<1> a, Imm<1> b, Imm<1> c, Imm<
|
||||
// ORR (vector, immediate)
|
||||
const auto orr = [&]{
|
||||
u64 imm64 = AdvSIMDExpandImm(op, cmode, concatenate(a, b, c, d, e, f, g, h));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast64(ir.Imm64(imm64));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast(64, ir.Imm64(imm64));
|
||||
const IR::U128 operand = V(datasize, Vd);
|
||||
const IR::U128 result = ir.VectorOr(operand, imm);
|
||||
V(datasize, Vd, result);
|
||||
@@ -42,7 +42,7 @@ bool TranslatorVisitor::MOVI(bool Q, bool op, Imm<1> a, Imm<1> b, Imm<1> c, Imm<
|
||||
// BIC (vector, immediate)
|
||||
const auto bic = [&]{
|
||||
u64 imm64 = ~AdvSIMDExpandImm(op, cmode, concatenate(a, b, c, d, e, f, g, h));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast64(ir.Imm64(imm64));
|
||||
const IR::U128 imm = datasize == 64 ? ir.ZeroExtendToQuad(ir.Imm64(imm64)) : ir.VectorBroadcast(64, ir.Imm64(imm64));
|
||||
const IR::U128 operand = V(datasize, Vd);
|
||||
const IR::U128 result = ir.VectorAnd(operand, imm);
|
||||
V(datasize, Vd, result);
|
||||
|
||||
@@ -13,24 +13,12 @@ bool TranslatorVisitor::ZIP1(bool Q, Imm<2> size, Vec Vm, Vec Vn, Vec Vd) {
|
||||
return ReservedValue();
|
||||
}
|
||||
|
||||
const size_t esize = 8 << size.ZeroExtend<size_t>();
|
||||
const size_t datasize = Q ? 128 : 64;
|
||||
|
||||
const IR::U128 result = [&] {
|
||||
const IR::U128 operand1 = V(datasize, Vn);
|
||||
const IR::U128 operand2 = V(datasize, Vm);
|
||||
|
||||
switch (size.ZeroExtend()) {
|
||||
case 0b00:
|
||||
return ir.VectorInterleaveLower8(operand1, operand2);
|
||||
case 0b01:
|
||||
return ir.VectorInterleaveLower16(operand1, operand2);
|
||||
case 0b10:
|
||||
return ir.VectorInterleaveLower32(operand1, operand2);
|
||||
case 0b11:
|
||||
default:
|
||||
return ir.VectorInterleaveLower64(operand1, operand2);
|
||||
}
|
||||
}();
|
||||
const IR::U128 operand1 = V(datasize, Vn);
|
||||
const IR::U128 operand2 = V(datasize, Vm);
|
||||
const IR::U128 result = ir.VectorInterleaveLower(esize, operand1, operand2);
|
||||
|
||||
V(datasize, Vd, result);
|
||||
return true;
|
||||
|
||||
@@ -22,19 +22,7 @@ bool TranslatorVisitor::SHL_2(bool Q, Imm<4> immh, Imm<3> immb, Vec Vn, Vec Vd)
|
||||
const u8 shift_amount = concatenate(immh, immb).ZeroExtend<u8>() - static_cast<u8>(esize);
|
||||
|
||||
const IR::U128 operand = V(datasize, Vn);
|
||||
const IR::U128 result = [&]{
|
||||
switch (esize) {
|
||||
case 8:
|
||||
return ir.VectorLogicalShiftLeft8(operand, shift_amount);
|
||||
case 16:
|
||||
return ir.VectorLogicalShiftLeft16(operand, shift_amount);
|
||||
case 32:
|
||||
return ir.VectorLogicalShiftLeft32(operand, shift_amount);
|
||||
case 64:
|
||||
default:
|
||||
return ir.VectorLogicalShiftLeft64(operand, shift_amount);
|
||||
}
|
||||
}();
|
||||
const IR::U128 result = ir.VectorLogicalShiftLeft(esize, operand, shift_amount);
|
||||
|
||||
V(datasize, Vd, result);
|
||||
return true;
|
||||
|
||||
@@ -16,18 +16,7 @@ bool TranslatorVisitor::ADD_vector(bool Q, Imm<2> size, Vec Vm, Vec Vn, Vec Vd)
|
||||
auto operand1 = V(datasize, Vn);
|
||||
auto operand2 = V(datasize, Vm);
|
||||
|
||||
auto result = [&]{
|
||||
switch (esize) {
|
||||
case 8:
|
||||
return ir.VectorAdd8(operand1, operand2);
|
||||
case 16:
|
||||
return ir.VectorAdd16(operand1, operand2);
|
||||
case 32:
|
||||
return ir.VectorAdd32(operand1, operand2);
|
||||
default:
|
||||
return ir.VectorAdd64(operand1, operand2);
|
||||
}
|
||||
}();
|
||||
auto result = ir.VectorAdd(esize, operand1, operand2);
|
||||
|
||||
V(datasize, Vd, result);
|
||||
|
||||
@@ -42,18 +31,7 @@ bool TranslatorVisitor::ADDP_vec(bool Q, Imm<2> size, Vec Vm, Vec Vn, Vec Vd) {
|
||||
const IR::U128 operand1 = V(datasize, Vn);
|
||||
const IR::U128 operand2 = V(datasize, Vm);
|
||||
|
||||
const IR::U128 result = [&]{
|
||||
switch (esize) {
|
||||
case 8:
|
||||
return Q ? ir.VectorPairedAdd8(operand1, operand2) : ir.VectorPairedAddLower8(operand1, operand2);
|
||||
case 16:
|
||||
return Q ? ir.VectorPairedAdd16(operand1, operand2) : ir.VectorPairedAddLower16(operand1, operand2);
|
||||
case 32:
|
||||
return Q ? ir.VectorPairedAdd32(operand1, operand2) : ir.VectorPairedAddLower32(operand1, operand2);
|
||||
default:
|
||||
return ir.VectorPairedAdd64(operand1, operand2);
|
||||
}
|
||||
}();
|
||||
const IR::U128 result = Q ? ir.VectorPairedAdd(esize, operand1, operand2) : ir.VectorPairedAddLower(esize, operand1, operand2);
|
||||
|
||||
V(datasize, Vd, result);
|
||||
|
||||
@@ -128,18 +106,7 @@ bool TranslatorVisitor::CMEQ_reg_2(bool Q, Imm<2> size, Vec Vm, Vec Vn, Vec Vd)
|
||||
const IR::U128 operand1 = V(datasize, Vn);
|
||||
const IR::U128 operand2 = V(datasize, Vm);
|
||||
|
||||
IR::U128 result = [&]{
|
||||
switch (esize) {
|
||||
case 8:
|
||||
return ir.VectorEqual8(operand1, operand2);
|
||||
case 16:
|
||||
return ir.VectorEqual16(operand1, operand2);
|
||||
case 32:
|
||||
return ir.VectorEqual32(operand1, operand2);
|
||||
default:
|
||||
return ir.VectorEqual64(operand1, operand2);
|
||||
}
|
||||
}();
|
||||
IR::U128 result = ir.VectorEqual(esize, operand1, operand2);
|
||||
|
||||
if (datasize == 64) {
|
||||
result = ir.VectorZeroUpper(result);
|
||||
@@ -170,8 +137,7 @@ bool TranslatorVisitor::BIF(bool Q, Vec Vm, Vec Vn, Vec Vd) {
|
||||
auto operand4 = V(datasize, Vn);
|
||||
auto operand3 = ir.VectorNot(V(datasize, Vm));
|
||||
|
||||
auto result = ir.VectorEor(operand1,
|
||||
ir.VectorAnd(ir.VectorEor(operand1, operand4), operand3));
|
||||
auto result = ir.VectorEor(operand1, ir.VectorAnd(ir.VectorEor(operand1, operand4), operand3));
|
||||
|
||||
V(datasize, Vd, result);
|
||||
|
||||
@@ -185,8 +151,7 @@ bool TranslatorVisitor::BIT(bool Q, Vec Vm, Vec Vn, Vec Vd) {
|
||||
auto operand4 = V(datasize, Vn);
|
||||
auto operand3 = V(datasize, Vm);
|
||||
|
||||
auto result = ir.VectorEor(operand1,
|
||||
ir.VectorAnd(ir.VectorEor(operand1, operand4), operand3));
|
||||
auto result = ir.VectorEor(operand1, ir.VectorAnd(ir.VectorEor(operand1, operand4), operand3));
|
||||
|
||||
V(datasize, Vd, result);
|
||||
|
||||
@@ -200,8 +165,7 @@ bool TranslatorVisitor::BSL(bool Q, Vec Vm, Vec Vn, Vec Vd) {
|
||||
auto operand1 = V(datasize, Vm);
|
||||
auto operand3 = V(datasize, Vd);
|
||||
|
||||
auto result = ir.VectorEor(operand1,
|
||||
ir.VectorAnd(ir.VectorEor(operand1, operand4), operand3));
|
||||
auto result = ir.VectorEor(operand1, ir.VectorAnd(ir.VectorEor(operand1, operand4), operand3));
|
||||
|
||||
V(datasize, Vd, result);
|
||||
|
||||
|
||||
Reference in New Issue
Block a user