@@ -2968,7 +2968,7 @@ void Assembler::movb(Register dst, Address src) {
2968
2968
}
2969
2969
2970
2970
void Assembler::movddup(XMMRegister dst, XMMRegister src) {
2971
- NOT_LP64( assert(VM_Version::supports_sse3(), "") );
2971
+ assert(VM_Version::supports_sse3(), "");
2972
2972
int vector_len = VM_Version::supports_avx512novl() ? AVX_512bit : AVX_128bit;
2973
2973
InstructionAttr attributes(vector_len, /* rex_w */ VM_Version::supports_evex(), /* legacy_mode */ false, /* no_mask_reg */ true, /* uses_vl */ true);
2974
2974
attributes.set_rex_vex_w_reverted();
@@ -2977,7 +2977,7 @@ void Assembler::movddup(XMMRegister dst, XMMRegister src) {
2977
2977
}
2978
2978
2979
2979
void Assembler::movddup(XMMRegister dst, Address src) {
2980
- NOT_LP64( assert(VM_Version::supports_sse3(), "") );
2980
+ assert(VM_Version::supports_sse3(), "");
2981
2981
InstructionMark im(this);
2982
2982
InstructionAttr attributes(AVX_128bit, /* rex_w */ VM_Version::supports_evex(), /* legacy_mode */ false, /* no_mask_reg */ true, /* uses_vl */ true);
2983
2983
attributes.set_address_attributes(/* tuple_type */ EVEX_DUP, /* input_size_in_bits */ EVEX_64bit);
@@ -4614,7 +4614,7 @@ void Assembler::vpacksswb(XMMRegister dst, XMMRegister nds, XMMRegister src, int
4614
4614
}
4615
4615
4616
4616
void Assembler::packssdw(XMMRegister dst, XMMRegister src) {
4617
- assert(VM_Version::supports_sse2(), "");
4617
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
4618
4618
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ _legacy_mode_bw, /* no_mask_reg */ true, /* uses_vl */ true);
4619
4619
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
4620
4620
emit_int16(0x6B, (0xC0 | encode));
@@ -4820,7 +4820,7 @@ void Assembler::pcmpestri(XMMRegister dst, XMMRegister src, int imm8) {
4820
4820
4821
4821
// In this context, the dst vector contains the components that are equal, non equal components are zeroed in dst
4822
4822
void Assembler::pcmpeqb(XMMRegister dst, XMMRegister src) {
4823
- assert(VM_Version::supports_sse2(), "");
4823
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
4824
4824
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ true, /* no_mask_reg */ true, /* uses_vl */ false);
4825
4825
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
4826
4826
emit_int16(0x74, (0xC0 | encode));
@@ -4968,7 +4968,7 @@ void Assembler::evpcmpeqb(KRegister kdst, KRegister mask, XMMRegister nds, Addre
4968
4968
4969
4969
// In this context, the dst vector contains the components that are equal, non equal components are zeroed in dst
4970
4970
void Assembler::pcmpeqw(XMMRegister dst, XMMRegister src) {
4971
- assert(VM_Version::supports_sse2(), "");
4971
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
4972
4972
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ true, /* no_mask_reg */ true, /* uses_vl */ false);
4973
4973
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
4974
4974
emit_int16(0x75, (0xC0 | encode));
@@ -5017,7 +5017,7 @@ void Assembler::evpcmpeqw(KRegister kdst, XMMRegister nds, Address src, int vect
5017
5017
5018
5018
// In this context, the dst vector contains the components that are equal, non equal components are zeroed in dst
5019
5019
void Assembler::pcmpeqd(XMMRegister dst, XMMRegister src) {
5020
- assert(VM_Version::supports_sse2(), "");
5020
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
5021
5021
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ true, /* no_mask_reg */ true, /* uses_vl */ false);
5022
5022
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
5023
5023
emit_int16(0x76, (0xC0 | encode));
@@ -5122,7 +5122,7 @@ void Assembler::pcmpgtq(XMMRegister dst, XMMRegister src) {
5122
5122
}
5123
5123
5124
5124
void Assembler::pmovmskb(Register dst, XMMRegister src) {
5125
- assert(VM_Version::supports_sse2(), "");
5125
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
5126
5126
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ true, /* no_mask_reg */ true, /* uses_vl */ false);
5127
5127
int encode = simd_prefix_and_encode(as_XMMRegister(dst->encoding()), xnoreg, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
5128
5128
emit_int16((unsigned char)0xD7, (0xC0 | encode));
@@ -5188,7 +5188,7 @@ void Assembler::pextrq(Address dst, XMMRegister src, int imm8) {
5188
5188
}
5189
5189
5190
5190
void Assembler::pextrw(Register dst, XMMRegister src, int imm8) {
5191
- assert(VM_Version::supports_sse2(), "");
5191
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
5192
5192
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ _legacy_mode_bw, /* no_mask_reg */ true, /* uses_vl */ false);
5193
5193
int encode = simd_prefix_and_encode(as_XMMRegister(dst->encoding()), xnoreg, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
5194
5194
emit_int24((unsigned char)0xC5, (0xC0 | encode), imm8);
@@ -5274,14 +5274,14 @@ void Assembler::vpinsrq(XMMRegister dst, XMMRegister nds, Register src, int imm8
5274
5274
}
5275
5275
5276
5276
void Assembler::pinsrw(XMMRegister dst, Register src, int imm8) {
5277
- assert(VM_Version::supports_sse2(), "");
5277
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
5278
5278
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ _legacy_mode_bw, /* no_mask_reg */ true, /* uses_vl */ false);
5279
5279
int encode = simd_prefix_and_encode(dst, dst, as_XMMRegister(src->encoding()), VEX_SIMD_66, VEX_OPCODE_0F, &attributes, true);
5280
5280
emit_int24((unsigned char)0xC4, (0xC0 | encode), imm8);
5281
5281
}
5282
5282
5283
5283
void Assembler::pinsrw(XMMRegister dst, Address src, int imm8) {
5284
- assert(VM_Version::supports_sse2(), "");
5284
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
5285
5285
InstructionMark im(this);
5286
5286
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ _legacy_mode_bw, /* no_mask_reg */ true, /* uses_vl */ false);
5287
5287
attributes.set_address_attributes(/* tuple_type */ EVEX_T1S, /* input_size_in_bits */ EVEX_16bit);
@@ -8712,7 +8712,7 @@ void Assembler::pmulld(XMMRegister dst, XMMRegister src) {
8712
8712
}
8713
8713
8714
8714
void Assembler::pmuludq(XMMRegister dst, XMMRegister src) {
8715
- assert(VM_Version::supports_sse2(), "");
8715
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
8716
8716
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ false, /* no_mask_reg */ true, /* uses_vl */ true);
8717
8717
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
8718
8718
emit_int16((unsigned char)0xF4, (0xC0 | encode));
@@ -8813,7 +8813,7 @@ void Assembler::vpminsb(XMMRegister dst, XMMRegister nds, XMMRegister src, int v
8813
8813
}
8814
8814
8815
8815
void Assembler::pminsw(XMMRegister dst, XMMRegister src) {
8816
- assert(VM_Version::supports_sse2(), "");
8816
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
8817
8817
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ _legacy_mode_bw, /* no_mask_reg */ true, /* uses_vl */ true);
8818
8818
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
8819
8819
emit_int16((unsigned char)0xEA, (0xC0 | encode));
@@ -8892,7 +8892,7 @@ void Assembler::vpmaxsb(XMMRegister dst, XMMRegister nds, XMMRegister src, int v
8892
8892
}
8893
8893
8894
8894
void Assembler::pmaxsw(XMMRegister dst, XMMRegister src) {
8895
- assert(VM_Version::supports_sse2(), "");
8895
+ NOT_LP64( assert(VM_Version::supports_sse2(), "");)
8896
8896
InstructionAttr attributes(AVX_128bit, /* rex_w */ false, /* legacy_mode */ _legacy_mode_bw, /* no_mask_reg */ true, /* uses_vl */ true);
8897
8897
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F, &attributes);
8898
8898
emit_int16((unsigned char)0xEE, (0xC0 | encode));
@@ -12401,15 +12401,15 @@ void Assembler::evpternlogq(XMMRegister dst, int imm8, KRegister mask, XMMRegist
12401
12401
12402
12402
void Assembler::gf2p8affineqb(XMMRegister dst, XMMRegister src, int imm8) {
12403
12403
assert(VM_Version::supports_gfni(), "");
12404
- assert(VM_Version::supports_sse(), "");
12404
+ NOT_LP64( assert(VM_Version::supports_sse(), "");)
12405
12405
InstructionAttr attributes(AVX_128bit, /* rex_w */ true, /* legacy_mode */ false, /* no_mask_reg */ true, /* uses_vl */ false);
12406
12406
int encode = simd_prefix_and_encode(dst, dst, src, VEX_SIMD_66, VEX_OPCODE_0F_3A, &attributes);
12407
12407
emit_int24((unsigned char)0xCE, (unsigned char)(0xC0 | encode), imm8);
12408
12408
}
12409
12409
12410
12410
void Assembler::vgf2p8affineqb(XMMRegister dst, XMMRegister src2, XMMRegister src3, int imm8, int vector_len) {
12411
12411
assert(VM_Version::supports_gfni(), "requires GFNI support");
12412
- assert(VM_Version::supports_sse(), "");
12412
+ NOT_LP64( assert(VM_Version::supports_sse(), "");)
12413
12413
InstructionAttr attributes(vector_len, /* vex_w */ true, /* legacy_mode */ false, /* no_mask_reg */ true, /* uses_vl */ true);
12414
12414
int encode = vex_prefix_and_encode(dst->encoding(), src2->encoding(), src3->encoding(), VEX_SIMD_66, VEX_OPCODE_0F_3A, &attributes);
12415
12415
emit_int24((unsigned char)0xCE, (unsigned char)(0xC0 | encode), imm8);
0 commit comments