image/svg+xmlVSCATTERPF1DPS/VSCATTERPF1QPS/VSCATTERPF1DPD/VSCATTERPF1QPD—Sparse Prefetch Packed SP/DP Data Values with Signed Dword, Signed Qword Indices Using T1 Hint with Intent to WriteInstruction Operand EncodingDescriptionThe instruction conditionally prefetches up to sixteen 32-bit or eight 64-bit integer byte data elements. The elements are specified via the VSIB (i.e., the index register is an zmm, holding packed indices). Elements will only be prefetched if their corresponding mask bit is one. cache lines will be brought into exclusive state (RFO) specified by a locality hint (T1):• T1 (temporal data)—prefetch data into the second level cache.[PS data] For dword indices, the instruction will prefetch sixteen memory locations. For qword indices, the instruc-tion will prefetch eight values.[PD data] For dword and qword indices, the instruction will prefetch eight memory locations. Note that:(1) The prefetches may happen in any order (or not at all). The instruction is a hint.(2) The mask is left unchanged.(3) Not valid with 16-bit effective addresses. Will deliver a #UD fault.(4) No FP nor memory faults may be produced by this instruction.(5) Prefetches do not handle cache line splits(6) A #UD is signaled if the memory operand is encoded without the SIB byte.OperationBASE_ADDR stands for the memory operand base address (a GPR); may not existVINDEX stands for the memory operand vector of indices (a vector register)SCALE stands for the memory operand scalar (1, 2, 4 or 8)DISP is the optional 1, 2 or 4 byte displacementPREFETCH(mem, Level, State) Prefetches a byte memory location pointed by ‘mem’ into the cache level specified by ‘Level’; a request for exclusive/ownership is done if ‘State’ is 1. Note that the memory location ignore cache line splits. This operation is considered a hint for the processor and may be skipped depending on implementation.Opcode/InstructionOp/En64/32 bit Mode SupportCPUID Feature FlagDescriptionEVEX.512.66.0F38.W0 C6 /6 /vsib VSCATTERPF1DPS vm32z {k1}AV/VAVX512PFUsing signed dword indices, prefetch sparse byte memory locations containing single-precision data using writemask k1 and T1 hint with intent to write.EVEX.512.66.0F38.W0 C7 /6 /vsibVSCATTERPF1QPS vm64z {k1}AV/VAVX512PFUsing signed qword indices, prefetch sparse byte memory locations containing single-precision data using writemask k1 and T1 hint with intent to write.EVEX.512.66.0F38.W1 C6 /6 /vsib VSCATTERPF1DPD vm32y {k1}AV/VAVX512PFUsing signed dword indices, prefetch sparse byte memory locations containing double-precision data using writemask k1 and T1 hint with intent to write.EVEX.512.66.0F38.W1 C7 /6 /vsibVSCATTERPF1QPD vm64z {k1}AV/VAVX512PFUsing signed qword indices, prefetch sparse byte memory locations containing double-precision data using writemask k1 and T1 hint with intent to write.Op/EnTuple TypeOperand 1Operand 2Operand 3Operand 4ATuple1 ScalarBaseReg (R): VSIB:base,VectorReg(R): VSIB:indexNANANA

image/svg+xmlVSCATTERPF1DPS (EVEX encoded version)(KL, VL) = (16, 512)FOR j := 0 TO KL-1i := j * 32IF k1[j] Prefetch( [BASE_ADDR + SignExtend(VINDEX[i+31:i]) * SCALE + DISP], Level=1, RFO = 1)FI;ENDFORVSCATTERPF1DPD (EVEX encoded version)(KL, VL) = (8, 512)FOR j := 0 TO KL-1i := j * 64k := j * 32IF k1[j] Prefetch( [BASE_ADDR + SignExtend(VINDEX[k+31:k]) * SCALE + DISP], Level=1, RFO = 1)FI;ENDFORVSCATTERPF1QPS (EVEX encoded version)(KL, VL) = (8, 512)FOR j := 0 TO KL-1i := j * 64IF k1[j] Prefetch( [BASE_ADDR + SignExtend(VINDEX[i+63:i]) * SCALE + DISP], Level=1, RFO = 1)FI;ENDFORVSCATTERPF1QPD (EVEX encoded version)(KL, VL) = (8, 512)FOR j := 0 TO KL-1i := j * 64k := j * 64IF k1[j] Prefetch( [BASE_ADDR + SignExtend(VINDEX[k+63:k]) * SCALE + DISP], Level=1, RFO = 1)FI;ENDFORIntel C/C++ Compiler Intrinsic EquivalentVSCATTERPF1DPD void _mm512_prefetch_i32scatter_pd(void *base, __m256i vdx, int scale, int hint);VSCATTERPF1DPD void _mm512_mask_prefetch_i32scatter_pd(void *base, __mmask8 m, __m256i vdx, int scale, int hint);VSCATTERPF1DPS void _mm512_prefetch_i32scatter_ps(void *base, __m512i vdx, int scale, int hint);VSCATTERPF1DPS void _mm512_mask_prefetch_i32scatter_ps(void *base, __mmask16 m, __m512i vdx, int scale, int hint);VSCATTERPF1QPD void _mm512_prefetch_i64scatter_pd(void * base, __m512i vdx, int scale, int hint);VSCATTERPF1QPD void _mm512_mask_prefetch_i64scatter_pd(void * base, __mmask8 m, __m512i vdx, int scale, int hint);VSCATTERPF1QPS void _mm512_prefetch_i64scatter_ps(void *base, __m512i vdx, int scale, int hint);VSCATTERPF1QPS void _mm512_mask_prefetch_i64scatter_ps(void *base, __mmask8 m, __m512i vdx, int scale, int hint);SIMD Floating-Point ExceptionsNoneOther ExceptionsSee Table2-62, “Type E12NP Class Exception Conditions”.

image/svg+xml

This UNOFFICIAL reference was generated from the official Intel® 64 and IA-32 Architectures Software Developer’s Manual by a dumb script. There is no guarantee that some parts aren't mangled or broken and is distributed WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.