모두의 코드
VPMASKMOV (Intel x86/64 assembly instruction)

작성일 : 2020-09-01 이 글은 593 번 읽혔습니다.

VPMASKMOV

Conditional SIMD Integer Packed Loads and Stores

참고 사항

아래 표를 해석하는 방법은 x86-64 명령어 레퍼런스 읽는 법 글을 참조하시기 바랍니다.

Opcode/
Instruction

Op/
En

64/32
-bit
Mode

CPUID
Feature
Flag

Description

VEX.NDS.128.66.0F38.W0 8C /rVPMASK-MOVD xmm1, xmm2, m128

RVM

V/V

AVX2

Conditionally load dword values from m128 using mask in xmm2 and store in xmm1.

VEX.NDS.256.66.0F38.W0 8C /r
VPMASKMOVD ymm1 ymm2 m256

RVM

V/V

AVX2

Conditionally load dword values from m256 using mask in ymm2 and store in ymm1.

VEX.NDS.128.66.0F38.W1 8C /r
VPMASKMOVQ xmm1 xmm2 m128

RVM

V/V

AVX2

Conditionally load qword values from m128 using mask in xmm2 and store in xmm1.

VEX.NDS.256.66.0F38.W1 8C /r
VPMASKMOVQ ymm1 ymm2 m256

RVM

V/V

AVX2

Conditionally load qword values from m256 using mask in ymm2 and store in ymm1.

VEX.NDS.128.66.0F38.W0 8E /r
VPMASKMOVD m128 xmm1 xmm2

MVR

V/V

AVX2

Conditionally store dword values from xmm2 using mask in xmm1.

VEX.NDS.256.66.0F38.W0 8E /r
VPMASKMOVD m256 ymm1 ymm2

MVR

V/V

AVX2

Conditionally store dword values from ymm2 using mask in ymm1.

VEX.NDS.128.66.0F38.W1 8E /r
VPMASKMOVQ m128 xmm1 xmm2

MVR

V/V

AVX2

Conditionally store qword values from xmm2 using mask in xmm1.

VEX.NDS.256.66.0F38.W1 8E /r
VPMASKMOVQ m256 ymm1 ymm2

MVR

V/V

AVX2

Conditionally store qword values from ymm2 using mask in ymm1.

Instruction Operand Encoding

Op/En

Operand 1

Operand 2

Operand 3

Operand 4

RVM

ModRM:reg (w)

VEX.vvvv

ModRM:r/m (r)

NA

MVR

ModRM:r/m (w)

VEX.vvvv

ModRM:reg (r)

NA

Description

Conditionally moves packed data elements from the second source operand into the corresponding data element of the destination operand, depending on the mask bits associated with each data element. The mask bits are specified in the first source operand.

The mask bit for each data element is the most significant bit of that element in the first source operand. If a mask is 1, the corresponding data element is copied from the second source operand to the destination operand. If the mask is 0, the corresponding data element is set to zero in the load form of these instructions, and unmodified in the store form.

The second source operand is a memory address for the load form of these instructions. The destination operand is a memory address for the store form of these instructions. The other operands are either XMM registers (for VEX.128 version) or YMM registers (for VEX.256 version).

Faults occur only due to mask-bit required memory accesses that caused the faults. Faults will not occur due to referencing any memory location if the corresponding mask bit for that memory location is 0. For example, no faults will be detected if the mask bits are all zero.

Unlike previous MASKMOV instructions (MASKMOVQ and MASKMOVDQU), a nontemporal hint is not applied to these instructions.

Instruction behavior on alignment check reporting with mask bits of less than all 1s are the same as with mask bits of all 1s.

VMASKMOV should not be used to access memory mapped I/O as the ordering of the individual loads or stores it does is implementation specific.

In cases where mask bits indicate data should not be loaded or stored paging A and D bits will be set in an imple-mentation dependent way. However, A and D bits are always set for pages where data is actually loaded/stored.

Note: for load forms, the first source (the mask) is encoded in VEX.vvvv; the second source is encoded in rmfield, and the destination register is encoded in regfield.

Note: for store forms, the first source (the mask) is encoded in VEX.vvvv; the second source register is encoded in regfield, and the destination memory location is encoded in rmfield.

Operation

VPMASKMOVD - 256-bit load

DEST[31:0] <-  IF (SRC1[31]) Load_32(mem) ELSE 0 
DEST[63:32] <-  IF (SRC1[63]) Load_32(mem + 4) ELSE 0 
DEST[95:64] <-  IF (SRC1[95]) Load_32(mem + 8) ELSE 0 
DEST[127:96] <-  IF (SRC1[127]) Load_32(mem + 12) ELSE 0 
DEST[159:128] <-  IF (SRC1[159]) Load_32(mem + 16) ELSE 0 
DEST[191:160] <-  IF (SRC1[191]) Load_32(mem + 20) ELSE 0 
DEST[223:192] <-  IF (SRC1[223]) Load_32(mem + 24) ELSE 0 
DEST[255:224] <-  IF (SRC1[255]) Load_32(mem + 28) ELSE 0 

VPMASKMOVD -128-bit load

DEST[31:0] <-  IF (SRC1[31]) Load_32(mem) ELSE 0 
DEST[63:32] <-  IF (SRC1[63]) Load_32(mem + 4) ELSE 0 
DEST[95:64] <-  IF (SRC1[95]) Load_32(mem + 8) ELSE 0 
DEST[127:97] <-  IF (SRC1[127]) Load_32(mem + 12) ELSE 0 
DEST[VLMAX-1:128] <-  0

VPMASKMOVQ - 256-bit load

DEST[63:0] <-  IF (SRC1[63]) Load_64(mem) ELSE 0 
DEST[127:64] <-  IF (SRC1[127]) Load_64(mem + 8) ELSE 0 
DEST[195:128] <-  IF (SRC1[191]) Load_64(mem + 16) ELSE 0 
DEST[255:196] <-  IF (SRC1[255]) Load_64(mem + 24) ELSE 0 

VPMASKMOVQ - 128-bit load

DEST[63:0] <-  IF (SRC1[63]) Load_64(mem) ELSE 0 
DEST[127:64] <-  IF (SRC1[127]) Load_64(mem + 16) ELSE 0
DEST[VLMAX-1:128] <-  0

VPMASKMOVD - 256-bit store

IF (SRC1[31]) DEST[31:0] <-  SRC2[31:0] 
IF (SRC1[63]) DEST[63:32] <-  SRC2[63:32] 
IF (SRC1[95]) DEST[95:64] <-  SRC2[95:64] 
IF (SRC1[127]) DEST[127:96] <-  SRC2[127:96] 
IF (SRC1[159]) DEST[159:128] <- SRC2[159:128] 
IF (SRC1[191]) DEST[191:160] <-  SRC2[191:160] 
IF (SRC1[223]) DEST[223:192] <-  SRC2[223:192] 
IF (SRC1[255]) DEST[255:224] <-  SRC2[255:224] 

VPMASKMOVD - 128-bit store

IF (SRC1[31]) DEST[31:0] <-  SRC2[31:0] 
IF (SRC1[63]) DEST[63:32] <-  SRC2[63:32] 
IF (SRC1[95]) DEST[95:64] <-  SRC2[95:64] 
IF (SRC1[127]) DEST[127:96] <-  SRC2[127:96] 

VPMASKMOVQ - 256-bit store

IF (SRC1[63]) DEST[63:0] <-  SRC2[63:0] 
IF (SRC1[127]) DEST[127:64] <- SRC2[127:64] 
IF (SRC1[191]) DEST[191:128] <-  SRC2[191:128] 
IF (SRC1[255]) DEST[255:192] <-  SRC2[255:192] 

VPMASKMOVQ - 128-bit store

IF (SRC1[63]) DEST[63:0] <-  SRC2[63:0] 
IF (SRC1[127]) DEST[127:64] <- SRC2[127:64] 

Intel C/C++ Compiler Intrinsic Equivalent

VPMASKMOVD : __m256i _mm256_maskload_epi32(int const *a,
                                           __m256i mask) VPMASKMOVD
    : void _mm256_maskstore_epi32(int *a, __m256i mask, __m256i b) VPMASKMOVQ
    : __m256i _mm256_maskload_epi64(__int64 const *a, __m256i mask);
VPMASKMOVQ : void _mm256_maskstore_epi64(__int64 *a, __m256i mask, __m256d b);
VPMASKMOVD : __m128i _mm_maskload_epi32(int const *a, __m128i mask) VPMASKMOVD
    : void _mm_maskstore_epi32(int *a, __m128i mask, __m128 b) VPMASKMOVQ
    : __m128i _mm_maskload_epi64(__int cont *a, __m128i mask);
VPMASKMOVQ : void _mm_maskstore_epi64(__int64 *a, __m128i mask, __m128i b);

SIMD Floating-Point Exceptions

None

Other Exceptions

See Exceptions Type 6 (No AC# reported for any mask bit combinations).

첫 댓글을 달아주세요!
프로필 사진 없음
강좌에 관련 없이 궁금한 내용은 여기를 사용해주세요

    댓글을 불러오는 중입니다..