[llvm] r214719 - [SKX] Enabling load/store instructions: encoding

Robert Khasanov rob.khasanov at gmail.com
Tue Aug 5 02:46:24 PDT 2014


Hi Adam,


2014-08-05 0:47 GMT+04:00 Adam Nemet <anemet at apple.com>:

> Hi Robert,
>
> On Aug 4, 2014, at 7:35 AM, Robert Khasanov <rob.khasanov at gmail.com>
> wrote:
>
> > Author: rkhasanov
> > Date: Mon Aug  4 09:35:15 2014
> > New Revision: 214719
> >
> > URL: http://llvm.org/viewvc/llvm-project?rev=214719&view=rev
> > Log:
> > [SKX] Enabling load/store instructions: encoding
> > Instructions: VMOVAPD, VMOVAPS, VMOVDQA8, VMOVDQA16,
> VMOVDQA32,VMOVDQA64, VMOVDQU8, VMOVDQU16, VMOVDQU32,VMOVDQU64, VMOVUPD,
> VMOVUPS,
> >
> > Reviewed by Elena Demikhovsky <elena.demikhovsky at intel.com>
> >
>
> Next time, it would be good if you could describe the approach/design in
> the commit log.  I think I figured it out this time and like the extra
> layer of multiclasses (_vl) you’re adding!
>
>
Ok. Next time I will provide more details about patch in commit log.


> I am also staring to gravitate toward adding some helper definitions that
> would map things like the number of vector elements to the corresponding
> ValueTypes and RegisterClasses, etc — I think that otherwise we will have
> things like !cast<RegisterClass>("VK"##vsz256##"WM”) all over the place.
>  I’ll let you know once I have something more concrete.
>
> I have one more question below:
>
>
> > Added:
> >    llvm/trunk/test/MC/X86/x86-64-avx512bw_vl.s
> >    llvm/trunk/test/MC/X86/x86-64-avx512f_vl.s
> > Modified:
> >    llvm/trunk/lib/Target/X86/X86InstrAVX512.td
> >    llvm/trunk/lib/Target/X86/X86InstrFragmentsSIMD.td
> >    llvm/trunk/lib/Target/X86/X86InstrInfo.cpp
> >    llvm/trunk/test/CodeGen/X86/avx512-vec-cmp.ll
> >    llvm/trunk/test/MC/X86/avx512-encodings.s
> >    llvm/trunk/test/MC/X86/x86-64-avx512bw.s
> >    llvm/trunk/utils/TableGen/X86DisassemblerTables.cpp
> >
> > Modified: llvm/trunk/lib/Target/X86/X86InstrAVX512.td
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86InstrAVX512.td?rev=214719&r1=214718&r2=214719&view=diff
> >
> ==============================================================================
> > --- llvm/trunk/lib/Target/X86/X86InstrAVX512.td (original)
> > +++ llvm/trunk/lib/Target/X86/X86InstrAVX512.td Mon Aug  4 09:35:15 2014
> > @@ -1433,104 +1433,176 @@ def : Pat<(v8i1 (X86vsrli VK8:$src, (i8
> > // AVX-512 - Aligned and unaligned load and store
> > //
> >
> > -multiclass avx512_load<bits<8> opc, RegisterClass RC, RegisterClass KRC,
> > -                            X86MemOperand x86memop, PatFrag ld_frag,
> > -                            string asm, Domain d,
> > -                            ValueType vt, bit IsReMaterializable = 1> {
> > +multiclass avx512_load<bits<8> opc, string OpcodeStr, PatFrag ld_frag,
> > +                       RegisterClass KRC, RegisterClass RC,
> > +                       ValueType vt, ValueType zvt, X86MemOperand memop,
> > +                       Domain d, bit IsReMaterializable = 1> {
> > let hasSideEffects = 0 in {
> >   def rr : AVX512PI<opc, MRMSrcReg, (outs RC:$dst), (ins RC:$src),
> > -              !strconcat(asm, " \t{$src, $dst|$dst, $src}"), [], d>,
> > -              EVEX;
> > +                    !strconcat(OpcodeStr, "\t{$src, $dst|$dst, $src}"),
> [],
> > +                    d>, EVEX;
> >   def rrkz : AVX512PI<opc, MRMSrcReg, (outs RC:$dst), (ins KRC:$mask,
> RC:$src),
> > -               !strconcat(asm,
> > -               " \t{$src, ${dst} {${mask}} {z}|${dst} {${mask}} {z},
> $src}"),
> > -               [], d>, EVEX, EVEX_KZ;
> > -  }
> > -  let canFoldAsLoad = 1, isReMaterializable = IsReMaterializable in
> > -  def rm : AVX512PI<opc, MRMSrcMem, (outs RC:$dst), (ins x86memop:$src),
> > -              !strconcat(asm, " \t{$src, $dst|$dst, $src}"),
> > -               [(set (vt RC:$dst), (ld_frag addr:$src))], d>, EVEX;
> > -  let Constraints = "$src1 = $dst",  hasSideEffects = 0 in {
> > -  def rrk : AVX512PI<opc, MRMSrcReg, (outs RC:$dst),
> > -                                     (ins RC:$src1, KRC:$mask,
> RC:$src2),
> > -              !strconcat(asm,
> > -              " \t{$src2, ${dst} {${mask}}|${dst} {${mask}}, $src2}"),
> [], d>,
> > -              EVEX, EVEX_K;
> > -  let mayLoad = 1 in
> > -  def rmk : AVX512PI<opc, MRMSrcMem, (outs RC:$dst),
> > -                                (ins RC:$src1, KRC:$mask,
> x86memop:$src2),
> > -              !strconcat(asm,
> > -              " \t{$src2, ${dst} {${mask}}|${dst} {${mask}}, $src2}"),
> > -               [], d>, EVEX, EVEX_K;
> > +                      !strconcat(OpcodeStr, "\t{$src, ${dst} {${mask}}
> {z}|",
> > +                       "${dst} {${mask}} {z}, $src}"), [], d>, EVEX,
> EVEX_KZ;
> > +  }
> > +  let canFoldAsLoad = 1, isReMaterializable = IsReMaterializable,
> > +      SchedRW = [WriteLoad] in
> > +  def rm : AVX512PI<opc, MRMSrcMem, (outs RC:$dst), (ins memop:$src),
> > +                    !strconcat(OpcodeStr, "\t{$src, $dst|$dst, $src}"),
> > +                    [(set RC:$dst, (vt (bitconvert (ld_frag
> addr:$src))))],
> > +                    d>, EVEX;
> > +
> > +  let AddedComplexity = 20 in {
> > +  let Constraints = "$src0 = $dst",  hasSideEffects = 0 in {
> > +  let hasSideEffects = 0 in
> > +    def rrk : AVX512PI<opc, MRMSrcReg, (outs RC:$dst),
> > +                     (ins RC:$src0, KRC:$mask, RC:$src1),
> > +                     !strconcat(OpcodeStr, "\t{$src1, ${dst}
> {${mask}}|",
> > +                      "${dst} {${mask}}, $src1}"),
> > +                     [(set RC:$dst, (vt (vselect KRC:$mask,
> > +                                          (vt RC:$src1),
> > +                                          (vt RC:$src0))))],
> > +                     d>, EVEX, EVEX_K;
> > +  let mayLoad = 1, SchedRW = [WriteLoad] in
> > +    def rmk : AVX512PI<opc, MRMSrcMem, (outs RC:$dst),
> > +                     (ins RC:$src0, KRC:$mask, memop:$src1),
> > +                     !strconcat(OpcodeStr, "\t{$src1, ${dst}
> {${mask}}|",
> > +                      "${dst} {${mask}}, $src1}"),
> > +                     [(set RC:$dst, (vt
> > +                         (vselect KRC:$mask,
> > +                                 (vt (bitconvert (ld_frag addr:$src1))),
> > +                                 (vt RC:$src0))))],
> > +                     d>, EVEX, EVEX_K;
> > +  }
> > +  let mayLoad = 1, SchedRW = [WriteLoad] in
> > +    def rmkz : AVX512PI<opc, MRMSrcMem, (outs RC:$dst),
> > +                      (ins KRC:$mask, memop:$src),
> > +                      !strconcat(OpcodeStr, "\t{$src, ${dst} {${mask}}
> {z}|",
> > +                       "${dst} {${mask}} {z}, $src}"),
> > +                      [(set RC:$dst, (vt
> > +                           (vselect KRC:$mask,
> > +                                     (vt (bitconvert (ld_frag
> addr:$src))),
> > +                                     (vt (bitconvert (zvt
> immAllZerosV))))))],
> > +                      d>, EVEX, EVEX_KZ;
> >   }
> > -  let mayLoad = 1 in
> > -  def rmkz : AVX512PI<opc, MRMSrcMem, (outs RC:$dst),
> > -                      (ins KRC:$mask, x86memop:$src2),
> > -              !strconcat(asm,
> > -              " \t{$src2, ${dst} {${mask}} {z}|${dst} {${mask}} {z},
> $src2}"),
> > -               [], d>, EVEX, EVEX_KZ;
> > }
> >
> > -multiclass avx512_store<bits<8> opc, RegisterClass RC, RegisterClass
> KRC,
> > -                            X86MemOperand x86memop, PatFrag store_frag,
> > -                            string asm, Domain d, ValueType vt> {
> > +multiclass avx512_load_vl<bits<8> opc, string OpcodeStr, string ld_pat,
> > +                          string elty, string elsz, string vsz512,
> > +                          string vsz256, string vsz128, Domain d,
> > +                          Predicate prd, bit IsReMaterializable = 1> {
> > +  let Predicates = [prd] in
> > +  defm Z : avx512_load<opc, OpcodeStr,
> > +                       !cast<PatFrag>(ld_pat##"v"##vsz512##elty##elsz),
> > +                       !cast<RegisterClass>("VK"##vsz512##"WM"), VR512,
> > +                       !cast<ValueType>("v"##vsz512##elty##elsz),
> v16i32,
> > +                       !cast<X86MemOperand>(elty##"512mem"), d,
> > +                       IsReMaterializable>, EVEX_V512;
> > +
> > +  let Predicates = [prd, HasVLX] in {
> > +    defm Z256 : avx512_load<opc, OpcodeStr,
> > +                       !cast<PatFrag>(ld_pat##!if(!eq(elty,"f"),
> > +                             "v"##vsz256##elty##elsz, "v4i64")),
> > +                       !cast<RegisterClass>("VK"##vsz256##"WM"), VR256X,
> > +                       !cast<ValueType>("v"##vsz256##elty##elsz), v8i32,
> > +                       !cast<X86MemOperand>(elty##"256mem"), d,
> > +                       IsReMaterializable>, EVEX_V256;
> > +
> > +    defm Z128 : avx512_load<opc, OpcodeStr,
> > +                       !cast<PatFrag>(ld_pat##!if(!eq(elty,"f"),
> > +                             "v"##vsz128##elty##elsz, "v2i64")),
> > +                       !cast<RegisterClass>("VK"##vsz128##"WM"), VR128X,
> > +                       !cast<ValueType>("v"##vsz128##elty##elsz), v4i32,
> > +                       !cast<X86MemOperand>(elty##"128mem"), d,
> > +                       IsReMaterializable>, EVEX_V128;
> > +  }
> > +}
> > +
> > +
> > +multiclass avx512_store<bits<8> opc, string OpcodeStr, PatFrag st_frag,
> > +                        ValueType OpVT, RegisterClass KRC,
> RegisterClass RC,
> > +                        X86MemOperand memop, Domain d> {
> >   let isAsmParserOnly = 1, hasSideEffects = 0 in {
> >   def rr_alt : AVX512PI<opc, MRMDestReg, (outs RC:$dst), (ins RC:$src),
> > -              !strconcat(asm, " \t{$src, $dst|$dst, $src}"), [], d>,
> > +              !strconcat(OpcodeStr, "\t{$src, $dst|$dst, $src}"), [],
> d>,
> >               EVEX;
> >   let Constraints = "$src1 = $dst" in
> > -  def alt_rrk : AVX512PI<opc, MRMDestReg, (outs  RC:$dst),
> > -                                     (ins RC:$src1, KRC:$mask,
> RC:$src2),
> > -              !strconcat(asm,
> > -              " \t{$src2, ${dst} {${mask}}|${dst} {${mask}}, $src2}"),
> [], d>,
> > +  def rrk_alt : AVX512PI<opc, MRMDestReg, (outs  RC:$dst),
> > +                                          (ins RC:$src1, KRC:$mask,
> RC:$src2),
> > +              !strconcat(OpcodeStr,
> > +              "\t{$src2, ${dst} {${mask}}|${dst} {${mask}}, $src2}"),
> [], d>,
> >               EVEX, EVEX_K;
> > -  def alt_rrkz : AVX512PI<opc, MRMDestReg, (outs  RC:$dst),
> > +  def rrkz_alt : AVX512PI<opc, MRMDestReg, (outs  RC:$dst),
> >                                            (ins KRC:$mask, RC:$src),
> > -              !strconcat(asm,
> > -              " \t{$src, ${dst} {${mask}} {z}|${dst} {${mask}} {z},
> $src}"),
> > +              !strconcat(OpcodeStr,
> > +              "\t{$src, ${dst} {${mask}} {z}|${dst} {${mask}} {z},
> $src}"),
> >               [], d>, EVEX, EVEX_KZ;
> >   }
> >   let mayStore = 1 in {
> > -  def mr : AVX512PI<opc, MRMDestMem, (outs), (ins x86memop:$dst,
> RC:$src),
> > -              !strconcat(asm, " \t{$src, $dst|$dst, $src}"),
> > -               [(store_frag (vt RC:$src), addr:$dst)], d>, EVEX;
> > +  def mr : AVX512PI<opc, MRMDestMem, (outs), (ins memop:$dst, RC:$src),
> > +                    !strconcat(OpcodeStr, "\t{$src, $dst|$dst, $src}"),
> > +                    [(st_frag (OpVT RC:$src), addr:$dst)], d>, EVEX;
> >   def mrk : AVX512PI<opc, MRMDestMem, (outs),
> > -                                (ins x86memop:$dst, KRC:$mask, RC:$src),
> > -              !strconcat(asm,
> > -              " \t{$src, ${dst} {${mask}}|${dst} {${mask}}, $src}"),
> > +                                      (ins memop:$dst, KRC:$mask,
> RC:$src),
> > +              !strconcat(OpcodeStr,
> > +              "\t{$src, ${dst} {${mask}}|${dst} {${mask}}, $src}"),
> >                [], d>, EVEX, EVEX_K;
> > -  def mrkz : AVX512PI<opc, MRMDestMem, (outs),
> > -                      (ins x86memop:$dst, KRC:$mask, RC:$src),
> > -              !strconcat(asm,
> > -              " \t{$src, ${dst} {${mask}} {z}|${dst} {${mask}} {z},
> $src}"),
> > -               [], d>, EVEX, EVEX_KZ;
> >   }
> > }
> >
> > -defm VMOVAPSZ : avx512_load<0x28, VR512, VK16WM, f512mem,
> alignedloadv16f32,
> > -                              "vmovaps", SSEPackedSingle, v16f32>,
> > -                avx512_store<0x29, VR512, VK16WM, f512mem,
> alignedstore512,
> > -                              "vmovaps", SSEPackedSingle, v16f32>,
> > -                               PS, EVEX_V512, EVEX_CD8<32, CD8VF>;
> > -defm VMOVAPDZ : avx512_load<0x28, VR512, VK8WM, f512mem,
> alignedloadv8f64,
> > -                              "vmovapd", SSEPackedDouble, v8f64>,
> > -                avx512_store<0x29, VR512, VK8WM, f512mem,
> alignedstore512,
> > -                              "vmovapd", SSEPackedDouble, v8f64>,
> > -                              PD, EVEX_V512, VEX_W,
> > -                              EVEX_CD8<64, CD8VF>;
> > -defm VMOVUPSZ : avx512_load<0x10, VR512, VK16WM, f512mem, loadv16f32,
> > -                              "vmovups", SSEPackedSingle, v16f32>,
> > -                avx512_store<0x11, VR512, VK16WM, f512mem, store,
> > -                              "vmovups", SSEPackedSingle, v16f32>,
> > -                              PS, EVEX_V512, EVEX_CD8<32, CD8VF>;
> > -defm VMOVUPDZ : avx512_load<0x10, VR512, VK8WM, f512mem, loadv8f64,
> > -                              "vmovupd", SSEPackedDouble, v8f64, 0>,
> > -                avx512_store<0x11, VR512, VK8WM, f512mem, store,
> > -                              "vmovupd", SSEPackedDouble, v8f64>,
> > -                               PD, EVEX_V512, VEX_W,
> > -                               EVEX_CD8<64, CD8VF>;
> > +
> > +multiclass avx512_store_vl<bits<8> opc, string OpcodeStr, string st_pat,
> > +                           string st_suff_512, string st_suff_256,
> > +                           string st_suff_128, string elty, string elsz,
> > +                           string vsz512, string vsz256, string vsz128,
> > +                           Domain d, Predicate prd> {
> > +  let Predicates = [prd] in
> > +  defm Z : avx512_store<opc, OpcodeStr,
> !cast<PatFrag>(st_pat##st_suff_512),
> > +                        !cast<ValueType>("v"##vsz512##elty##elsz),
> > +                        !cast<RegisterClass>("VK"##vsz512##"WM"), VR512,
> > +                        !cast<X86MemOperand>(elty##"512mem"), d>,
> EVEX_V512;
> > +
> > +  let Predicates = [prd, HasVLX] in {
> > +    defm Z256 : avx512_store<opc, OpcodeStr,
> !cast<PatFrag>(st_pat##st_suff_256),
> > +                             !cast<ValueType>("v"##vsz256##elty##elsz),
> > +                             !cast<RegisterClass>("VK"##vsz256##"WM"),
> VR256X,
> > +                             !cast<X86MemOperand>(elty##"256mem"), d>,
> EVEX_V256;
> > +
> > +    defm Z128 : avx512_store<opc, OpcodeStr,
> !cast<PatFrag>(st_pat##st_suff_128),
> > +                             !cast<ValueType>("v"##vsz128##elty##elsz),
> > +                             !cast<RegisterClass>("VK"##vsz128##"WM"),
> VR128X,
> > +                             !cast<X86MemOperand>(elty##"128mem"), d>,
> EVEX_V128;
> > +  }
> > +}
> > +
> > +defm VMOVAPS : avx512_load_vl<0x28, "vmovaps", "alignedload", "f", "32",
> > +                              "16", "8", "4", SSEPackedSingle,
> HasAVX512>,
> > +               avx512_store_vl<0x29, "vmovaps", "alignedstore",
> > +                               "512", "256", "", "f", "32", "16", "8",
> "4",
> > +                               SSEPackedSingle, HasAVX512>,
> > +                              PS, EVEX_CD8<32, CD8VF>;
> > +
> > +defm VMOVAPD : avx512_load_vl<0x28, "vmovapd", "alignedload", "f", "64",
> > +                              "8", "4", "2", SSEPackedDouble,
> HasAVX512>,
> > +               avx512_store_vl<0x29, "vmovapd", "alignedstore",
> > +                               "512", "256", "", "f", "64", "8", "4",
> "2",
> > +                               SSEPackedDouble, HasAVX512>,
> > +                              PD, VEX_W, EVEX_CD8<64, CD8VF>;
> > +
> > +defm VMOVUPS : avx512_load_vl<0x10, "vmovups", "load", "f", "32",
> > +                              "16", "8", "4", SSEPackedSingle,
> HasAVX512>,
> > +               avx512_store_vl<0x11, "vmovups", "store", "", "", "",
> "f", "32",
> > +                              "16", "8", "4", SSEPackedSingle,
> HasAVX512>,
> > +                              PS, EVEX_CD8<32, CD8VF>;
> > +
> > +defm VMOVUPD : avx512_load_vl<0x10, "vmovupd", "load", "f", "64",
> > +                              "8", "4", "2", SSEPackedDouble,
> HasAVX512, 0>,
> > +               avx512_store_vl<0x11, "vmovupd", "store", "", "", "",
> "f", "64",
> > +                              "8", "4", "2", SSEPackedDouble,
> HasAVX512>,
> > +                             PD, VEX_W, EVEX_CD8<64, CD8VF>;
> > +
> > def: Pat<(v8f64 (int_x86_avx512_mask_loadu_pd_512 addr:$ptr,
> > -                 (bc_v8f64 (v16i32 immAllZerosV)), GR8:$mask)),
> > +                (bc_v8f64 (v16i32 immAllZerosV)), GR8:$mask)),
> >        (VMOVUPDZrmkz (v8i1 (COPY_TO_REGCLASS GR8:$mask, VK8WM)),
> addr:$ptr)>;
> >
> > def: Pat<(v16f32 (int_x86_avx512_mask_loadu_ps_512 addr:$ptr,
> > @@ -1546,75 +1618,80 @@ def: Pat<(int_x86_avx512_mask_storeu_pd_
> >          (VMOVUPDZmrk addr:$ptr, (v8i1 (COPY_TO_REGCLASS GR8:$mask,
> VK8WM)),
> >             VR512:$src)>;
> >
> > -defm VMOVDQA32: avx512_load<0x6F, VR512, VK16WM, i512mem,
> alignedloadv16i32,
> > -                              "vmovdqa32", SSEPackedInt, v16i32>,
> > -                avx512_store<0x7F, VR512, VK16WM, i512mem,
> alignedstore512,
> > -                              "vmovdqa32", SSEPackedInt, v16i32>,
> > -                               PD, EVEX_V512, EVEX_CD8<32, CD8VF>;
> > -defm VMOVDQA64: avx512_load<0x6F, VR512, VK8WM, i512mem,
> alignedloadv8i64,
> > -                              "vmovdqa64", SSEPackedInt, v8i64>,
> > -                avx512_store<0x7F, VR512, VK8WM, i512mem,
> alignedstore512,
> > -                              "vmovdqa64", SSEPackedInt, v8i64>,
> > -                               PD, VEX_W, EVEX_V512, EVEX_CD8<64,
> CD8VF>;
> > -defm VMOVDQU32: avx512_load<0x6F, VR512, VK16WM, i512mem, load,
> > -                              "vmovdqu32", SSEPackedInt, v16i32>,
> > -                avx512_store<0x7F, VR512, VK16WM, i512mem, store,
> > -                              "vmovdqu32", SSEPackedInt, v16i32>,
> > -                               XS, EVEX_V512, EVEX_CD8<32, CD8VF>;
> > -defm VMOVDQU64: avx512_load<0x6F, VR512, VK8WM, i512mem, load,
> > -                              "vmovdqu64", SSEPackedInt, v8i64>,
> > -                avx512_store<0x7F, VR512, VK8WM, i512mem, store,
> > -                              "vmovdqu64", SSEPackedInt, v8i64>,
> > -                               XS, VEX_W, EVEX_V512, EVEX_CD8<64,
> CD8VF>;
> > +defm VMOVDQA32 : avx512_load_vl<0x6F, "vmovdqa32", "alignedload", "i",
> "32",
> > +                                "16", "8", "4", SSEPackedInt,
> HasAVX512>,
> > +                 avx512_store_vl<0x7F, "vmovdqa32", "alignedstore",
> > +                                 "512", "256", "", "i", "32", "16",
> "8", "4",
> > +                                 SSEPackedInt, HasAVX512>,
> > +                                PD, EVEX_CD8<32, CD8VF>;
> > +
> > +defm VMOVDQA64 : avx512_load_vl<0x6F, "vmovdqa64", "alignedload", "i",
> "64",
> > +                                "8", "4", "2", SSEPackedInt, HasAVX512>,
> > +                 avx512_store_vl<0x7F, "vmovdqa64", "alignedstore",
> > +                                 "512", "256", "", "i", "64", "8", "4",
> "2",
> > +                                 SSEPackedInt, HasAVX512>,
> > +                                PD, VEX_W, EVEX_CD8<64, CD8VF>;
> > +
> > +defm VMOVDQU8 : avx512_load_vl<0x6F, "vmovdqu8", "load", "i", "8",
> > +                               "64", "32", "16", SSEPackedInt, HasBWI>,
> > +                 avx512_store_vl<0x7F, "vmovdqu8", "store", "", "", "",
> > +                                 "i", "8", "64", "32", "16",
> SSEPackedInt,
> > +                                 HasBWI>, XD, EVEX_CD8<8, CD8VF>;
> > +
> > +defm VMOVDQU16 : avx512_load_vl<0x6F, "vmovdqu16", "load", "i", "16",
> > +                                "32", "16", "8", SSEPackedInt, HasBWI>,
> > +                 avx512_store_vl<0x7F, "vmovdqu16", "store", "", "", "",
> > +                                 "i", "16", "32", "16", "8",
> SSEPackedInt,
> > +                                 HasBWI>, XD, VEX_W, EVEX_CD8<16,
> CD8VF>;
> > +
> > +defm VMOVDQU32 : avx512_load_vl<0x6F, "vmovdqu32", "load", "i", "32",
> > +                                "16", "8", "4", SSEPackedInt,
> HasAVX512>,
> > +                 avx512_store_vl<0x7F, "vmovdqu32", "store", "", "", "",
> > +                                 "i", "32", "16", "8", "4",
> SSEPackedInt,
> > +                                 HasAVX512>, XS, EVEX_CD8<32, CD8VF>;
> > +
> > +defm VMOVDQU64 : avx512_load_vl<0x6F, "vmovdqu64", "load", "i", "64",
> > +                                "8", "4", "2", SSEPackedInt, HasAVX512>,
> > +                 avx512_store_vl<0x7F, "vmovdqu64", "store", "", "", "",
> > +                                 "i", "64", "8", "4", "2", SSEPackedInt,
> > +                                 HasAVX512>, XS, VEX_W, EVEX_CD8<64,
> CD8VF>;
> >
> > def: Pat<(v16i32 (int_x86_avx512_mask_loadu_d_512 addr:$ptr,
> >                  (v16i32 immAllZerosV), GR16:$mask)),
> > -       (VMOVDQU32rmkz (v16i1 (COPY_TO_REGCLASS GR16:$mask, VK16WM)),
> addr:$ptr)>;
> > +       (VMOVDQU32Zrmkz (v16i1 (COPY_TO_REGCLASS GR16:$mask, VK16WM)),
> addr:$ptr)>;
> >
> > def: Pat<(v8i64 (int_x86_avx512_mask_loadu_q_512 addr:$ptr,
> > -                 (bc_v8i64 (v16i32 immAllZerosV)), GR8:$mask)),
> > -       (VMOVDQU64rmkz (v8i1 (COPY_TO_REGCLASS GR8:$mask, VK8WM)),
> addr:$ptr)>;
> > +                (bc_v8i64 (v16i32 immAllZerosV)), GR8:$mask)),
> > +       (VMOVDQU64Zrmkz (v8i1 (COPY_TO_REGCLASS GR8:$mask, VK8WM)),
> addr:$ptr)>;
> >
> > def: Pat<(int_x86_avx512_mask_storeu_d_512 addr:$ptr, (v16i32
> VR512:$src),
> > -          GR16:$mask),
> > -         (VMOVDQU32mrk addr:$ptr, (v16i1 (COPY_TO_REGCLASS GR16:$mask,
> VK16WM)),
> > +            GR16:$mask),
> > +         (VMOVDQU32Zmrk addr:$ptr, (v16i1 (COPY_TO_REGCLASS GR16:$mask,
> VK16WM)),
> >             VR512:$src)>;
> > def: Pat<(int_x86_avx512_mask_storeu_q_512 addr:$ptr, (v8i64 VR512:$src),
> > -          GR8:$mask),
> > -         (VMOVDQU64mrk addr:$ptr, (v8i1 (COPY_TO_REGCLASS GR8:$mask,
> VK8WM)),
> > +            GR8:$mask),
> > +         (VMOVDQU64Zmrk addr:$ptr, (v8i1 (COPY_TO_REGCLASS GR8:$mask,
> VK8WM)),
> >             VR512:$src)>;
> >
> > let AddedComplexity = 20 in {
> > def : Pat<(v8i64 (vselect VK8WM:$mask, (v8i64 VR512:$src),
> > -                           (bc_v8i64 (v16i32 immAllZerosV)))),
> > -                  (VMOVDQU64rrkz VK8WM:$mask, VR512:$src)>;
> > +                          (bc_v8i64 (v16i32 immAllZerosV)))),
> > +                  (VMOVDQU64Zrrkz VK8WM:$mask, VR512:$src)>;
> >
> > def : Pat<(v8i64 (vselect VK8WM:$mask, (bc_v8i64 (v16i32 immAllZerosV)),
> > -                  (v8i64 VR512:$src))),
> > -   (VMOVDQU64rrkz (COPY_TO_REGCLASS (KNOTWrr (COPY_TO_REGCLASS
> VK8:$mask, VK16)),
> > +                          (v8i64 VR512:$src))),
> > +   (VMOVDQU64Zrrkz (COPY_TO_REGCLASS (KNOTWrr (COPY_TO_REGCLASS
> VK8:$mask, VK16)),
> >                                               VK8), VR512:$src)>;
> >
> > def : Pat<(v16i32 (vselect VK16WM:$mask, (v16i32 VR512:$src),
> >                            (v16i32 immAllZerosV))),
> > -                  (VMOVDQU32rrkz VK16WM:$mask, VR512:$src)>;
> > +                  (VMOVDQU32Zrrkz VK16WM:$mask, VR512:$src)>;
> >
> > def : Pat<(v16i32 (vselect VK16WM:$mask, (v16i32 immAllZerosV),
> > -                   (v16i32 VR512:$src))),
> > -   (VMOVDQU32rrkz (KNOTWrr VK16WM:$mask), VR512:$src)>;
> > -
> > -def : Pat<(v16f32 (vselect VK16WM:$mask, (v16f32 VR512:$src1),
> > -                           (v16f32 VR512:$src2))),
> > -                  (VMOVUPSZrrk VR512:$src2, VK16WM:$mask, VR512:$src1)>;
> > -def : Pat<(v8f64 (vselect VK8WM:$mask, (v8f64 VR512:$src1),
> > -                           (v8f64 VR512:$src2))),
> > -                  (VMOVUPDZrrk VR512:$src2, VK8WM:$mask, VR512:$src1)>;
> > -def : Pat<(v16i32 (vselect VK16WM:$mask, (v16i32 VR512:$src1),
> > -                           (v16i32 VR512:$src2))),
> > -                  (VMOVDQU32rrk VR512:$src2, VK16WM:$mask,
> VR512:$src1)>;
> > -def : Pat<(v8i64 (vselect VK8WM:$mask, (v8i64 VR512:$src1),
> > -                           (v8i64 VR512:$src2))),
> > -                  (VMOVDQU64rrk VR512:$src2, VK8WM:$mask, VR512:$src1)>;
> > +                           (v16i32 VR512:$src))),
> > +                  (VMOVDQU32Zrrkz (KNOTWrr VK16WM:$mask), VR512:$src)>;
> > }
> > +
> > // Move Int Doubleword to Packed Double Int
> > //
> > def VMOVDI2PDIZrr : AVX512BI<0x6E, MRMSrcReg, (outs VR128X:$dst), (ins
> GR32:$src),
> >
> > Modified: llvm/trunk/lib/Target/X86/X86InstrFragmentsSIMD.td
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86InstrFragmentsSIMD.td?rev=214719&r1=214718&r2=214719&view=diff
> >
> ==============================================================================
> > --- llvm/trunk/lib/Target/X86/X86InstrFragmentsSIMD.td (original)
> > +++ llvm/trunk/lib/Target/X86/X86InstrFragmentsSIMD.td Mon Aug  4
> 09:35:15 2014
> > @@ -311,6 +311,8 @@ def loadv4i64    : PatFrag<(ops node:$pt
> > // 512-bit load pattern fragments
> > def loadv16f32   : PatFrag<(ops node:$ptr), (v16f32 (load node:$ptr))>;
> > def loadv8f64    : PatFrag<(ops node:$ptr), (v8f64 (load node:$ptr))>;
> > +def loadv64i8    : PatFrag<(ops node:$ptr), (v64i8 (load node:$ptr))>;
> > +def loadv32i16   : PatFrag<(ops node:$ptr), (v32i16 (load node:$ptr))>;
> > def loadv16i32   : PatFrag<(ops node:$ptr), (v16i32 (load node:$ptr))>;
> > def loadv8i64    : PatFrag<(ops node:$ptr), (v8i64 (load node:$ptr))>;
> >
> >
> > Modified: llvm/trunk/lib/Target/X86/X86InstrInfo.cpp
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86InstrInfo.cpp?rev=214719&r1=214718&r2=214719&view=diff
> >
> ==============================================================================
> > --- llvm/trunk/lib/Target/X86/X86InstrInfo.cpp (original)
> > +++ llvm/trunk/lib/Target/X86/X86InstrInfo.cpp Mon Aug  4 09:35:15 2014
> > @@ -601,10 +601,10 @@ X86InstrInfo::X86InstrInfo(X86Subtarget
> >     // AVX-512 foldable instructions
> >     { X86::VMOV64toPQIZrr,  X86::VMOVQI2PQIZrm,       0 },
> >     { X86::VMOVDI2SSZrr,    X86::VMOVDI2SSZrm,        0 },
> > -    { X86::VMOVDQA32rr,     X86::VMOVDQA32rm,         TB_ALIGN_64 },
> > -    { X86::VMOVDQA64rr,     X86::VMOVDQA64rm,         TB_ALIGN_64 },
> > -    { X86::VMOVDQU32rr,     X86::VMOVDQU32rm,         0 },
> > -    { X86::VMOVDQU64rr,     X86::VMOVDQU64rm,         0 },
> > +    { X86::VMOVDQA32Zrr,    X86::VMOVDQA32Zrm,        TB_ALIGN_64 },
> > +    { X86::VMOVDQA64Zrr,    X86::VMOVDQA64Zrm,        TB_ALIGN_64 },
> > +    { X86::VMOVDQU32Zrr,    X86::VMOVDQU32Zrm,        0 },
> > +    { X86::VMOVDQU64Zrr,    X86::VMOVDQU64Zrm,        0 },
> >     { X86::VPABSDZrr,       X86::VPABSDZrm,           0 },
> >     { X86::VPABSQZrr,       X86::VPABSQZrm,           0 },
> >
> >
> > Modified: llvm/trunk/test/CodeGen/X86/avx512-vec-cmp.ll
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/avx512-vec-cmp.ll?rev=214719&r1=214718&r2=214719&view=diff
> >
> ==============================================================================
> > --- llvm/trunk/test/CodeGen/X86/avx512-vec-cmp.ll (original)
> > +++ llvm/trunk/test/CodeGen/X86/avx512-vec-cmp.ll Mon Aug  4 09:35:15
> 2014
> > @@ -2,7 +2,7 @@
> >
> > ; CHECK-LABEL: test1
> > ; CHECK: vcmpleps
> > -; CHECK: vmovups
> > +; CHECK: vmovaps
> > ; CHECK: ret
> > define <16 x float> @test1(<16 x float> %x, <16 x float> %y) nounwind {
> >   %mask = fcmp ole <16 x float> %x, %y
> > @@ -12,7 +12,7 @@ define <16 x float> @test1(<16 x float>
> >
> > ; CHECK-LABEL: test2
> > ; CHECK: vcmplepd
> > -; CHECK: vmovupd
> > +; CHECK: vmovapd
> > ; CHECK: ret
> > define <8 x double> @test2(<8 x double> %x, <8 x double> %y) nounwind {
> >   %mask = fcmp ole <8 x double> %x, %y
> > @@ -22,7 +22,7 @@ define <8 x double> @test2(<8 x double>
> >
> > ; CHECK-LABEL: test3
> > ; CHECK: vpcmpeqd  (%rdi)
> > -; CHECK: vmovdqu32
> > +; CHECK: vmovdqa32
> > ; CHECK: ret
> > define <16 x i32> @test3(<16 x i32> %x, <16 x i32> %x1, <16 x i32>* %yp)
> nounwind {
> >   %y = load <16 x i32>* %yp, align 4
> > @@ -33,7 +33,7 @@ define <16 x i32> @test3(<16 x i32> %x,
> >
> > ; CHECK-LABEL: @test4_unsigned
> > ; CHECK: vpcmpnltud
> > -; CHECK: vmovdqu32
> > +; CHECK: vmovdqa32
> > ; CHECK: ret
> > define <16 x i32> @test4_unsigned(<16 x i32> %x, <16 x i32> %y) nounwind
> {
> >   %mask = icmp uge <16 x i32> %x, %y
> > @@ -43,7 +43,7 @@ define <16 x i32> @test4_unsigned(<16 x
> >
> > ; CHECK-LABEL: test5
> > ; CHECK: vpcmpeqq {{.*}}%k1
> > -; CHECK: vmovdqu64 {{.*}}%k1
> > +; CHECK: vmovdqa64 {{.*}}%k1
> > ; CHECK: ret
> > define <8 x i64> @test5(<8 x i64> %x, <8 x i64> %y) nounwind {
> >   %mask = icmp eq <8 x i64> %x, %y
> > @@ -53,7 +53,7 @@ define <8 x i64> @test5(<8 x i64> %x, <8
> >
> > ; CHECK-LABEL: test6_unsigned
> > ; CHECK: vpcmpnleuq {{.*}}%k1
> > -; CHECK: vmovdqu64 {{.*}}%k1
> > +; CHECK: vmovdqa64 {{.*}}%k1
> > ; CHECK: ret
> > define <8 x i64> @test6_unsigned(<8 x i64> %x, <8 x i64> %y) nounwind {
> >   %mask = icmp ugt <8 x i64> %x, %y
> >
>
> Can you please explain these?
>
>
In this patch I added pattern for lowering for aligned loads (see rmk
versions in avx512_load multiclass). As aligned load is more preferable
than unaligned one they were chosen.
Next patch will include more lowering tests.

-- Robert


> Thanks,
> Adam
>
>
> > Modified: llvm/trunk/test/MC/X86/avx512-encodings.s
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/test/MC/X86/avx512-encodings.s?rev=214719&r1=214718&r2=214719&view=diff
> >
> ==============================================================================
> > --- llvm/trunk/test/MC/X86/avx512-encodings.s (original)
> > +++ llvm/trunk/test/MC/X86/avx512-encodings.s Mon Aug  4 09:35:15 2014
> > @@ -449,6 +449,294 @@
> > // CHECK:  encoding: [0x62,0xf1,0x64,0x58,0x5d,0x9a,0xfc,0xfd,0xff,0xff]
> >           vminps -516(%rdx){1to16}, %zmm3, %zmm3
> >
> > +// CHECK: vmovapd %zmm14, %zmm7
> > +// CHECK:  encoding: [0x62,0xd1,0xfd,0x48,0x28,0xfe]
> > +          vmovapd %zmm14, %zmm7
> > +
> > +// CHECK: vmovapd %zmm14, %zmm7 {%k5}
> > +// CHECK:  encoding: [0x62,0xd1,0xfd,0x4d,0x28,0xfe]
> > +          vmovapd %zmm14, %zmm7 {%k5}
> > +
> > +// CHECK: vmovapd %zmm14, %zmm7 {%k5} {z}
> > +// CHECK:  encoding: [0x62,0xd1,0xfd,0xcd,0x28,0xfe]
> > +          vmovapd %zmm14, %zmm7 {%k5} {z}
> > +
> > +// CHECK: vmovapd (%rcx), %zmm7
> > +// CHECK:  encoding: [0x62,0xf1,0xfd,0x48,0x28,0x39]
> > +          vmovapd (%rcx), %zmm7
> > +
> > +// CHECK: vmovapd 291(%rax,%r14,8), %zmm7
> > +// CHECK:  encoding:
> [0x62,0xb1,0xfd,0x48,0x28,0xbc,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovapd 291(%rax,%r14,8), %zmm7
> > +
> > +// CHECK: vmovapd 8128(%rdx), %zmm7
> > +// CHECK:  encoding: [0x62,0xf1,0xfd,0x48,0x28,0x7a,0x7f]
> > +          vmovapd 8128(%rdx), %zmm7
> > +
> > +// CHECK: vmovapd 8192(%rdx), %zmm7
> > +// CHECK:  encoding: [0x62,0xf1,0xfd,0x48,0x28,0xba,0x00,0x20,0x00,0x00]
> > +          vmovapd 8192(%rdx), %zmm7
> > +
> > +// CHECK: vmovapd -8192(%rdx), %zmm7
> > +// CHECK:  encoding: [0x62,0xf1,0xfd,0x48,0x28,0x7a,0x80]
> > +          vmovapd -8192(%rdx), %zmm7
> > +
> > +// CHECK: vmovapd -8256(%rdx), %zmm7
> > +// CHECK:  encoding: [0x62,0xf1,0xfd,0x48,0x28,0xba,0xc0,0xdf,0xff,0xff]
> > +          vmovapd -8256(%rdx), %zmm7
> > +
> > +// CHECK: vmovaps %zmm9, %zmm5
> > +// CHECK:  encoding: [0x62,0xd1,0x7c,0x48,0x28,0xe9]
> > +          vmovaps %zmm9, %zmm5
> > +
> > +// CHECK: vmovaps %zmm9, %zmm5 {%k1}
> > +// CHECK:  encoding: [0x62,0xd1,0x7c,0x49,0x28,0xe9]
> > +          vmovaps %zmm9, %zmm5 {%k1}
> > +
> > +// CHECK: vmovaps %zmm9, %zmm5 {%k1} {z}
> > +// CHECK:  encoding: [0x62,0xd1,0x7c,0xc9,0x28,0xe9]
> > +          vmovaps %zmm9, %zmm5 {%k1} {z}
> > +
> > +// CHECK: vmovaps (%rcx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7c,0x48,0x28,0x29]
> > +          vmovaps (%rcx), %zmm5
> > +
> > +// CHECK: vmovaps 291(%rax,%r14,8), %zmm5
> > +// CHECK:  encoding:
> [0x62,0xb1,0x7c,0x48,0x28,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovaps 291(%rax,%r14,8), %zmm5
> > +
> > +// CHECK: vmovaps 8128(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7c,0x48,0x28,0x6a,0x7f]
> > +          vmovaps 8128(%rdx), %zmm5
> > +
> > +// CHECK: vmovaps 8192(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7c,0x48,0x28,0xaa,0x00,0x20,0x00,0x00]
> > +          vmovaps 8192(%rdx), %zmm5
> > +
> > +// CHECK: vmovaps -8192(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7c,0x48,0x28,0x6a,0x80]
> > +          vmovaps -8192(%rdx), %zmm5
> > +
> > +// CHECK: vmovaps -8256(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7c,0x48,0x28,0xaa,0xc0,0xdf,0xff,0xff]
> > +          vmovaps -8256(%rdx), %zmm5
> > +
> > +// CHECK: vmovdqa32 %zmm18, %zmm22
> > +// CHECK:  encoding: [0x62,0xa1,0x7d,0x48,0x6f,0xf2]
> > +          vmovdqa32 %zmm18, %zmm22
> > +
> > +// CHECK: vmovdqa32 %zmm18, %zmm22 {%k6}
> > +// CHECK:  encoding: [0x62,0xa1,0x7d,0x4e,0x6f,0xf2]
> > +          vmovdqa32 %zmm18, %zmm22 {%k6}
> > +
> > +// CHECK: vmovdqa32 %zmm18, %zmm22 {%k6} {z}
> > +// CHECK:  encoding: [0x62,0xa1,0x7d,0xce,0x6f,0xf2]
> > +          vmovdqa32 %zmm18, %zmm22 {%k6} {z}
> > +
> > +// CHECK: vmovdqa32 (%rcx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x6f,0x31]
> > +          vmovdqa32 (%rcx), %zmm22
> > +
> > +// CHECK: vmovdqa32 291(%rax,%r14,8), %zmm22
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7d,0x48,0x6f,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa32 291(%rax,%r14,8), %zmm22
> > +
> > +// CHECK: vmovdqa32 8128(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x6f,0x72,0x7f]
> > +          vmovdqa32 8128(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqa32 8192(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x6f,0xb2,0x00,0x20,0x00,0x00]
> > +          vmovdqa32 8192(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqa32 -8192(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x6f,0x72,0x80]
> > +          vmovdqa32 -8192(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqa32 -8256(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x6f,0xb2,0xc0,0xdf,0xff,0xff]
> > +          vmovdqa32 -8256(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqa64 %zmm12, %zmm22
> > +// CHECK:  encoding: [0x62,0xc1,0xfd,0x48,0x6f,0xf4]
> > +          vmovdqa64 %zmm12, %zmm22
> > +
> > +// CHECK: vmovdqa64 %zmm12, %zmm22 {%k5}
> > +// CHECK:  encoding: [0x62,0xc1,0xfd,0x4d,0x6f,0xf4]
> > +          vmovdqa64 %zmm12, %zmm22 {%k5}
> > +
> > +// CHECK: vmovdqa64 %zmm12, %zmm22 {%k5} {z}
> > +// CHECK:  encoding: [0x62,0xc1,0xfd,0xcd,0x6f,0xf4]
> > +          vmovdqa64 %zmm12, %zmm22 {%k5} {z}
> > +
> > +// CHECK: vmovdqa64 (%rcx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x6f,0x31]
> > +          vmovdqa64 (%rcx), %zmm22
> > +
> > +// CHECK: vmovdqa64 291(%rax,%r14,8), %zmm22
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x48,0x6f,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa64 291(%rax,%r14,8), %zmm22
> > +
> > +// CHECK: vmovdqa64 8128(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x6f,0x72,0x7f]
> > +          vmovdqa64 8128(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqa64 8192(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x6f,0xb2,0x00,0x20,0x00,0x00]
> > +          vmovdqa64 8192(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqa64 -8192(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x6f,0x72,0x80]
> > +          vmovdqa64 -8192(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqa64 -8256(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x6f,0xb2,0xc0,0xdf,0xff,0xff]
> > +          vmovdqa64 -8256(%rdx), %zmm22
> > +
> > +// CHECK: vmovdqu32 %zmm24, %zmm5
> > +// CHECK:  encoding: [0x62,0x91,0x7e,0x48,0x6f,0xe8]
> > +          vmovdqu32 %zmm24, %zmm5
> > +
> > +// CHECK: vmovdqu32 %zmm24, %zmm5 {%k5}
> > +// CHECK:  encoding: [0x62,0x91,0x7e,0x4d,0x6f,0xe8]
> > +          vmovdqu32 %zmm24, %zmm5 {%k5}
> > +
> > +// CHECK: vmovdqu32 %zmm24, %zmm5 {%k5} {z}
> > +// CHECK:  encoding: [0x62,0x91,0x7e,0xcd,0x6f,0xe8]
> > +          vmovdqu32 %zmm24, %zmm5 {%k5} {z}
> > +
> > +// CHECK: vmovdqu32 (%rcx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7e,0x48,0x6f,0x29]
> > +          vmovdqu32 (%rcx), %zmm5
> > +
> > +// CHECK: vmovdqu32 291(%rax,%r14,8), %zmm5
> > +// CHECK:  encoding:
> [0x62,0xb1,0x7e,0x48,0x6f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu32 291(%rax,%r14,8), %zmm5
> > +
> > +// CHECK: vmovdqu32 8128(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7e,0x48,0x6f,0x6a,0x7f]
> > +          vmovdqu32 8128(%rdx), %zmm5
> > +
> > +// CHECK: vmovdqu32 8192(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7e,0x48,0x6f,0xaa,0x00,0x20,0x00,0x00]
> > +          vmovdqu32 8192(%rdx), %zmm5
> > +
> > +// CHECK: vmovdqu32 -8192(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7e,0x48,0x6f,0x6a,0x80]
> > +          vmovdqu32 -8192(%rdx), %zmm5
> > +
> > +// CHECK: vmovdqu32 -8256(%rdx), %zmm5
> > +// CHECK:  encoding: [0x62,0xf1,0x7e,0x48,0x6f,0xaa,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu32 -8256(%rdx), %zmm5
> > +
> > +// CHECK: vmovdqu64 %zmm15, %zmm6
> > +// CHECK:  encoding: [0x62,0xd1,0xfe,0x48,0x6f,0xf7]
> > +          vmovdqu64 %zmm15, %zmm6
> > +
> > +// CHECK: vmovdqu64 %zmm15, %zmm6 {%k3}
> > +// CHECK:  encoding: [0x62,0xd1,0xfe,0x4b,0x6f,0xf7]
> > +          vmovdqu64 %zmm15, %zmm6 {%k3}
> > +
> > +// CHECK: vmovdqu64 %zmm15, %zmm6 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0xd1,0xfe,0xcb,0x6f,0xf7]
> > +          vmovdqu64 %zmm15, %zmm6 {%k3} {z}
> > +
> > +// CHECK: vmovdqu64 (%rcx), %zmm6
> > +// CHECK:  encoding: [0x62,0xf1,0xfe,0x48,0x6f,0x31]
> > +          vmovdqu64 (%rcx), %zmm6
> > +
> > +// CHECK: vmovdqu64 291(%rax,%r14,8), %zmm6
> > +// CHECK:  encoding:
> [0x62,0xb1,0xfe,0x48,0x6f,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu64 291(%rax,%r14,8), %zmm6
> > +
> > +// CHECK: vmovdqu64 8128(%rdx), %zmm6
> > +// CHECK:  encoding: [0x62,0xf1,0xfe,0x48,0x6f,0x72,0x7f]
> > +          vmovdqu64 8128(%rdx), %zmm6
> > +
> > +// CHECK: vmovdqu64 8192(%rdx), %zmm6
> > +// CHECK:  encoding: [0x62,0xf1,0xfe,0x48,0x6f,0xb2,0x00,0x20,0x00,0x00]
> > +          vmovdqu64 8192(%rdx), %zmm6
> > +
> > +// CHECK: vmovdqu64 -8192(%rdx), %zmm6
> > +// CHECK:  encoding: [0x62,0xf1,0xfe,0x48,0x6f,0x72,0x80]
> > +          vmovdqu64 -8192(%rdx), %zmm6
> > +
> > +// CHECK: vmovdqu64 -8256(%rdx), %zmm6
> > +// CHECK:  encoding: [0x62,0xf1,0xfe,0x48,0x6f,0xb2,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu64 -8256(%rdx), %zmm6
> > +
> > +// CHECK: vmovupd %zmm9, %zmm27
> > +// CHECK:  encoding: [0x62,0x41,0xfd,0x48,0x10,0xd9]
> > +          vmovupd %zmm9, %zmm27
> > +
> > +// CHECK: vmovupd %zmm9, %zmm27 {%k2}
> > +// CHECK:  encoding: [0x62,0x41,0xfd,0x4a,0x10,0xd9]
> > +          vmovupd %zmm9, %zmm27 {%k2}
> > +
> > +// CHECK: vmovupd %zmm9, %zmm27 {%k2} {z}
> > +// CHECK:  encoding: [0x62,0x41,0xfd,0xca,0x10,0xd9]
> > +          vmovupd %zmm9, %zmm27 {%k2} {z}
> > +
> > +// CHECK: vmovupd (%rcx), %zmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x48,0x10,0x19]
> > +          vmovupd (%rcx), %zmm27
> > +
> > +// CHECK: vmovupd 291(%rax,%r14,8), %zmm27
> > +// CHECK:  encoding:
> [0x62,0x21,0xfd,0x48,0x10,0x9c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovupd 291(%rax,%r14,8), %zmm27
> > +
> > +// CHECK: vmovupd 8128(%rdx), %zmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x48,0x10,0x5a,0x7f]
> > +          vmovupd 8128(%rdx), %zmm27
> > +
> > +// CHECK: vmovupd 8192(%rdx), %zmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x48,0x10,0x9a,0x00,0x20,0x00,0x00]
> > +          vmovupd 8192(%rdx), %zmm27
> > +
> > +// CHECK: vmovupd -8192(%rdx), %zmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x48,0x10,0x5a,0x80]
> > +          vmovupd -8192(%rdx), %zmm27
> > +
> > +// CHECK: vmovupd -8256(%rdx), %zmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x48,0x10,0x9a,0xc0,0xdf,0xff,0xff]
> > +          vmovupd -8256(%rdx), %zmm27
> > +
> > +// CHECK: vmovups %zmm22, %zmm22
> > +// CHECK:  encoding: [0x62,0xa1,0x7c,0x48,0x10,0xf6]
> > +          vmovups %zmm22, %zmm22
> > +
> > +// CHECK: vmovups %zmm22, %zmm22 {%k3}
> > +// CHECK:  encoding: [0x62,0xa1,0x7c,0x4b,0x10,0xf6]
> > +          vmovups %zmm22, %zmm22 {%k3}
> > +
> > +// CHECK: vmovups %zmm22, %zmm22 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0xa1,0x7c,0xcb,0x10,0xf6]
> > +          vmovups %zmm22, %zmm22 {%k3} {z}
> > +
> > +// CHECK: vmovups (%rcx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x48,0x10,0x31]
> > +          vmovups (%rcx), %zmm22
> > +
> > +// CHECK: vmovups 291(%rax,%r14,8), %zmm22
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7c,0x48,0x10,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovups 291(%rax,%r14,8), %zmm22
> > +
> > +// CHECK: vmovups 8128(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x48,0x10,0x72,0x7f]
> > +          vmovups 8128(%rdx), %zmm22
> > +
> > +// CHECK: vmovups 8192(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x48,0x10,0xb2,0x00,0x20,0x00,0x00]
> > +          vmovups 8192(%rdx), %zmm22
> > +
> > +// CHECK: vmovups -8192(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x48,0x10,0x72,0x80]
> > +          vmovups -8192(%rdx), %zmm22
> > +
> > +// CHECK: vmovups -8256(%rdx), %zmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x48,0x10,0xb2,0xc0,0xdf,0xff,0xff]
> > +          vmovups -8256(%rdx), %zmm22
> > +
> > // CHECK: vmulpd %zmm23, %zmm4, %zmm24
> > // CHECK:  encoding: [0x62,0x21,0xdd,0x48,0x59,0xc7]
> >           vmulpd %zmm23, %zmm4, %zmm24
> > @@ -2557,6 +2845,230 @@
> > // CHECK:  encoding: [0xc5,0x78,0x93,0xea]
> >           kmovw  %k2, %r13d
> >
> > +// CHECK: vmovapd %zmm18, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x29,0x11]
> > +          vmovapd %zmm18, (%rcx)
> > +
> > +// CHECK: vmovapd %zmm18, (%rcx) {%k6}
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x4e,0x29,0x11]
> > +          vmovapd %zmm18, (%rcx) {%k6}
> > +
> > +// CHECK: vmovapd %zmm18, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x48,0x29,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovapd %zmm18, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovapd %zmm18, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x29,0x52,0x7f]
> > +          vmovapd %zmm18, 8128(%rdx)
> > +
> > +// CHECK: vmovapd %zmm18, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x29,0x92,0x00,0x20,0x00,0x00]
> > +          vmovapd %zmm18, 8192(%rdx)
> > +
> > +// CHECK: vmovapd %zmm18, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x29,0x52,0x80]
> > +          vmovapd %zmm18, -8192(%rdx)
> > +
> > +// CHECK: vmovapd %zmm18, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x29,0x92,0xc0,0xdf,0xff,0xff]
> > +          vmovapd %zmm18, -8256(%rdx)
> > +
> > +// CHECK: vmovaps %zmm9, (%rcx)
> > +// CHECK:  encoding: [0x62,0x71,0x7c,0x48,0x29,0x09]
> > +          vmovaps %zmm9, (%rcx)
> > +
> > +// CHECK: vmovaps %zmm9, (%rcx) {%k3}
> > +// CHECK:  encoding: [0x62,0x71,0x7c,0x4b,0x29,0x09]
> > +          vmovaps %zmm9, (%rcx) {%k3}
> > +
> > +// CHECK: vmovaps %zmm9, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x31,0x7c,0x48,0x29,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovaps %zmm9, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovaps %zmm9, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0x7c,0x48,0x29,0x4a,0x7f]
> > +          vmovaps %zmm9, 8128(%rdx)
> > +
> > +// CHECK: vmovaps %zmm9, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0x7c,0x48,0x29,0x8a,0x00,0x20,0x00,0x00]
> > +          vmovaps %zmm9, 8192(%rdx)
> > +
> > +// CHECK: vmovaps %zmm9, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0x7c,0x48,0x29,0x4a,0x80]
> > +          vmovaps %zmm9, -8192(%rdx)
> > +
> > +// CHECK: vmovaps %zmm9, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0x7c,0x48,0x29,0x8a,0xc0,0xdf,0xff,0xff]
> > +          vmovaps %zmm9, -8256(%rdx)
> > +
> > +// CHECK: vmovdqa32 %zmm18, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x7f,0x11]
> > +          vmovdqa32 %zmm18, (%rcx)
> > +
> > +// CHECK: vmovdqa32 %zmm18, (%rcx) {%k4}
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x4c,0x7f,0x11]
> > +          vmovdqa32 %zmm18, (%rcx) {%k4}
> > +
> > +// CHECK: vmovdqa32 %zmm18, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7d,0x48,0x7f,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa32 %zmm18, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqa32 %zmm18, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x7f,0x52,0x7f]
> > +          vmovdqa32 %zmm18, 8128(%rdx)
> > +
> > +// CHECK: vmovdqa32 %zmm18, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x7f,0x92,0x00,0x20,0x00,0x00]
> > +          vmovdqa32 %zmm18, 8192(%rdx)
> > +
> > +// CHECK: vmovdqa32 %zmm18, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x7f,0x52,0x80]
> > +          vmovdqa32 %zmm18, -8192(%rdx)
> > +
> > +// CHECK: vmovdqa32 %zmm18, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x48,0x7f,0x92,0xc0,0xdf,0xff,0xff]
> > +          vmovdqa32 %zmm18, -8256(%rdx)
> > +
> > +// CHECK: vmovdqa64 %zmm19, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x7f,0x19]
> > +          vmovdqa64 %zmm19, (%rcx)
> > +
> > +// CHECK: vmovdqa64 %zmm19, (%rcx) {%k5}
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x4d,0x7f,0x19]
> > +          vmovdqa64 %zmm19, (%rcx) {%k5}
> > +
> > +// CHECK: vmovdqa64 %zmm19, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x48,0x7f,0x9c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa64 %zmm19, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqa64 %zmm19, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x7f,0x5a,0x7f]
> > +          vmovdqa64 %zmm19, 8128(%rdx)
> > +
> > +// CHECK: vmovdqa64 %zmm19, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x7f,0x9a,0x00,0x20,0x00,0x00]
> > +          vmovdqa64 %zmm19, 8192(%rdx)
> > +
> > +// CHECK: vmovdqa64 %zmm19, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x7f,0x5a,0x80]
> > +          vmovdqa64 %zmm19, -8192(%rdx)
> > +
> > +// CHECK: vmovdqa64 %zmm19, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x48,0x7f,0x9a,0xc0,0xdf,0xff,0xff]
> > +          vmovdqa64 %zmm19, -8256(%rdx)
> > +
> > +// CHECK: vmovdqu32 %zmm22, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x48,0x7f,0x31]
> > +          vmovdqu32 %zmm22, (%rcx)
> > +
> > +// CHECK: vmovdqu32 %zmm22, (%rcx) {%k1}
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x49,0x7f,0x31]
> > +          vmovdqu32 %zmm22, (%rcx) {%k1}
> > +
> > +// CHECK: vmovdqu32 %zmm22, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7e,0x48,0x7f,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu32 %zmm22, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu32 %zmm22, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x48,0x7f,0x72,0x7f]
> > +          vmovdqu32 %zmm22, 8128(%rdx)
> > +
> > +// CHECK: vmovdqu32 %zmm22, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x48,0x7f,0xb2,0x00,0x20,0x00,0x00]
> > +          vmovdqu32 %zmm22, 8192(%rdx)
> > +
> > +// CHECK: vmovdqu32 %zmm22, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x48,0x7f,0x72,0x80]
> > +          vmovdqu32 %zmm22, -8192(%rdx)
> > +
> > +// CHECK: vmovdqu32 %zmm22, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x48,0x7f,0xb2,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu32 %zmm22, -8256(%rdx)
> > +
> > +// CHECK: vmovdqu64 %zmm24, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x48,0x7f,0x01]
> > +          vmovdqu64 %zmm24, (%rcx)
> > +
> > +// CHECK: vmovdqu64 %zmm24, (%rcx) {%k5}
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x4d,0x7f,0x01]
> > +          vmovdqu64 %zmm24, (%rcx) {%k5}
> > +
> > +// CHECK: vmovdqu64 %zmm24, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0xfe,0x48,0x7f,0x84,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu64 %zmm24, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu64 %zmm24, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x48,0x7f,0x42,0x7f]
> > +          vmovdqu64 %zmm24, 8128(%rdx)
> > +
> > +// CHECK: vmovdqu64 %zmm24, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x48,0x7f,0x82,0x00,0x20,0x00,0x00]
> > +          vmovdqu64 %zmm24, 8192(%rdx)
> > +
> > +// CHECK: vmovdqu64 %zmm24, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x48,0x7f,0x42,0x80]
> > +          vmovdqu64 %zmm24, -8192(%rdx)
> > +
> > +// CHECK: vmovdqu64 %zmm24, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x48,0x7f,0x82,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu64 %zmm24, -8256(%rdx)
> > +
> > +// CHECK: vmovupd %zmm10, (%rcx)
> > +// CHECK:  encoding: [0x62,0x71,0xfd,0x48,0x11,0x11]
> > +          vmovupd %zmm10, (%rcx)
> > +
> > +// CHECK: vmovupd %zmm10, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0x71,0xfd,0x4f,0x11,0x11]
> > +          vmovupd %zmm10, (%rcx) {%k7}
> > +
> > +// CHECK: vmovupd %zmm10, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x31,0xfd,0x48,0x11,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovupd %zmm10, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovupd %zmm10, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0xfd,0x48,0x11,0x52,0x7f]
> > +          vmovupd %zmm10, 8128(%rdx)
> > +
> > +// CHECK: vmovupd %zmm10, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0xfd,0x48,0x11,0x92,0x00,0x20,0x00,0x00]
> > +          vmovupd %zmm10, 8192(%rdx)
> > +
> > +// CHECK: vmovupd %zmm10, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0xfd,0x48,0x11,0x52,0x80]
> > +          vmovupd %zmm10, -8192(%rdx)
> > +
> > +// CHECK: vmovupd %zmm10, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0x71,0xfd,0x48,0x11,0x92,0xc0,0xdf,0xff,0xff]
> > +          vmovupd %zmm10, -8256(%rdx)
> > +
> > +// CHECK: vmovups %zmm24, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x48,0x11,0x01]
> > +          vmovups %zmm24, (%rcx)
> > +
> > +// CHECK: vmovups %zmm24, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x4f,0x11,0x01]
> > +          vmovups %zmm24, (%rcx) {%k7}
> > +
> > +// CHECK: vmovups %zmm24, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0x7c,0x48,0x11,0x84,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovups %zmm24, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovups %zmm24, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x48,0x11,0x42,0x7f]
> > +          vmovups %zmm24, 8128(%rdx)
> > +
> > +// CHECK: vmovups %zmm24, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x48,0x11,0x82,0x00,0x20,0x00,0x00]
> > +          vmovups %zmm24, 8192(%rdx)
> > +
> > +// CHECK: vmovups %zmm24, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x48,0x11,0x42,0x80]
> > +          vmovups %zmm24, -8192(%rdx)
> > +
> > +// CHECK: vmovups %zmm24, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x48,0x11,0x82,0xc0,0xdf,0xff,0xff]
> > +          vmovups %zmm24, -8256(%rdx)
> > +
> > // CHECK: vpmovqb %zmm2, %xmm3
> > // CHECK:  encoding: [0x62,0xf2,0x7e,0x48,0x32,0xd3]
> >           vpmovqb %zmm2, %xmm3
> >
> > Modified: llvm/trunk/test/MC/X86/x86-64-avx512bw.s
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/test/MC/X86/x86-64-avx512bw.s?rev=214719&r1=214718&r2=214719&view=diff
> >
> ==============================================================================
> > --- llvm/trunk/test/MC/X86/x86-64-avx512bw.s (original)
> > +++ llvm/trunk/test/MC/X86/x86-64-avx512bw.s Mon Aug  4 09:35:15 2014
> > @@ -1,5 +1,77 @@
> > // RUN: llvm-mc -triple x86_64-unknown-unknown -mcpu=skx
>  --show-encoding %s | FileCheck %s
> >
> > +// CHECK: vmovdqu8 %zmm19, %zmm29
> > +// CHECK:  encoding: [0x62,0x21,0x7f,0x48,0x6f,0xeb]
> > +          vmovdqu8 %zmm19, %zmm29
> > +
> > +// CHECK: vmovdqu8 %zmm19, %zmm29 {%k7}
> > +// CHECK:  encoding: [0x62,0x21,0x7f,0x4f,0x6f,0xeb]
> > +          vmovdqu8 %zmm19, %zmm29 {%k7}
> > +
> > +// CHECK: vmovdqu8 %zmm19, %zmm29 {%k7} {z}
> > +// CHECK:  encoding: [0x62,0x21,0x7f,0xcf,0x6f,0xeb]
> > +          vmovdqu8 %zmm19, %zmm29 {%k7} {z}
> > +
> > +// CHECK: vmovdqu8 (%rcx), %zmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x48,0x6f,0x29]
> > +          vmovdqu8 (%rcx), %zmm29
> > +
> > +// CHECK: vmovdqu8 291(%rax,%r14,8), %zmm29
> > +// CHECK:  encoding:
> [0x62,0x21,0x7f,0x48,0x6f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu8 291(%rax,%r14,8), %zmm29
> > +
> > +// CHECK: vmovdqu8 8128(%rdx), %zmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x48,0x6f,0x6a,0x7f]
> > +          vmovdqu8 8128(%rdx), %zmm29
> > +
> > +// CHECK: vmovdqu8 8192(%rdx), %zmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x48,0x6f,0xaa,0x00,0x20,0x00,0x00]
> > +          vmovdqu8 8192(%rdx), %zmm29
> > +
> > +// CHECK: vmovdqu8 -8192(%rdx), %zmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x48,0x6f,0x6a,0x80]
> > +          vmovdqu8 -8192(%rdx), %zmm29
> > +
> > +// CHECK: vmovdqu8 -8256(%rdx), %zmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x48,0x6f,0xaa,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu8 -8256(%rdx), %zmm29
> > +
> > +// CHECK: vmovdqu16 %zmm18, %zmm17
> > +// CHECK:  encoding: [0x62,0xa1,0xff,0x48,0x6f,0xca]
> > +          vmovdqu16 %zmm18, %zmm17
> > +
> > +// CHECK: vmovdqu16 %zmm18, %zmm17 {%k3}
> > +// CHECK:  encoding: [0x62,0xa1,0xff,0x4b,0x6f,0xca]
> > +          vmovdqu16 %zmm18, %zmm17 {%k3}
> > +
> > +// CHECK: vmovdqu16 %zmm18, %zmm17 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0xa1,0xff,0xcb,0x6f,0xca]
> > +          vmovdqu16 %zmm18, %zmm17 {%k3} {z}
> > +
> > +// CHECK: vmovdqu16 (%rcx), %zmm17
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x48,0x6f,0x09]
> > +          vmovdqu16 (%rcx), %zmm17
> > +
> > +// CHECK: vmovdqu16 291(%rax,%r14,8), %zmm17
> > +// CHECK:  encoding:
> [0x62,0xa1,0xff,0x48,0x6f,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu16 291(%rax,%r14,8), %zmm17
> > +
> > +// CHECK: vmovdqu16 8128(%rdx), %zmm17
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x48,0x6f,0x4a,0x7f]
> > +          vmovdqu16 8128(%rdx), %zmm17
> > +
> > +// CHECK: vmovdqu16 8192(%rdx), %zmm17
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x48,0x6f,0x8a,0x00,0x20,0x00,0x00]
> > +          vmovdqu16 8192(%rdx), %zmm17
> > +
> > +// CHECK: vmovdqu16 -8192(%rdx), %zmm17
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x48,0x6f,0x4a,0x80]
> > +          vmovdqu16 -8192(%rdx), %zmm17
> > +
> > +// CHECK: vmovdqu16 -8256(%rdx), %zmm17
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x48,0x6f,0x8a,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu16 -8256(%rdx), %zmm17
> > +
> > // CHECK: kandq  %k7, %k5, %k5
> > // CHECK:  encoding: [0xc4,0xe1,0xd4,0x41,0xef]
> >           kandq  %k7, %k5, %k5
> > @@ -127,3 +199,59 @@
> > // CHECK: kmovd  %k5, %r13d
> > // CHECK:  encoding: [0xc5,0x7b,0x93,0xed]
> >           kmovd  %k5, %r13d
> > +
> > +// CHECK: vmovdqu8 %zmm18, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x48,0x7f,0x11]
> > +          vmovdqu8 %zmm18, (%rcx)
> > +
> > +// CHECK: vmovdqu8 %zmm18, (%rcx) {%k3}
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x4b,0x7f,0x11]
> > +          vmovdqu8 %zmm18, (%rcx) {%k3}
> > +
> > +// CHECK: vmovdqu8 %zmm18, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7f,0x48,0x7f,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu8 %zmm18, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu8 %zmm18, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x48,0x7f,0x52,0x7f]
> > +          vmovdqu8 %zmm18, 8128(%rdx)
> > +
> > +// CHECK: vmovdqu8 %zmm18, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x48,0x7f,0x92,0x00,0x20,0x00,0x00]
> > +          vmovdqu8 %zmm18, 8192(%rdx)
> > +
> > +// CHECK: vmovdqu8 %zmm18, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x48,0x7f,0x52,0x80]
> > +          vmovdqu8 %zmm18, -8192(%rdx)
> > +
> > +// CHECK: vmovdqu8 %zmm18, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x48,0x7f,0x92,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu8 %zmm18, -8256(%rdx)
> > +
> > +// CHECK: vmovdqu16 %zmm28, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x48,0x7f,0x21]
> > +          vmovdqu16 %zmm28, (%rcx)
> > +
> > +// CHECK: vmovdqu16 %zmm28, (%rcx) {%k6}
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x4e,0x7f,0x21]
> > +          vmovdqu16 %zmm28, (%rcx) {%k6}
> > +
> > +// CHECK: vmovdqu16 %zmm28, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0xff,0x48,0x7f,0xa4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu16 %zmm28, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu16 %zmm28, 8128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x48,0x7f,0x62,0x7f]
> > +          vmovdqu16 %zmm28, 8128(%rdx)
> > +
> > +// CHECK: vmovdqu16 %zmm28, 8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x48,0x7f,0xa2,0x00,0x20,0x00,0x00]
> > +          vmovdqu16 %zmm28, 8192(%rdx)
> > +
> > +// CHECK: vmovdqu16 %zmm28, -8192(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x48,0x7f,0x62,0x80]
> > +          vmovdqu16 %zmm28, -8192(%rdx)
> > +
> > +// CHECK: vmovdqu16 %zmm28, -8256(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x48,0x7f,0xa2,0xc0,0xdf,0xff,0xff]
> > +          vmovdqu16 %zmm28, -8256(%rdx)
> >
> > Added: llvm/trunk/test/MC/X86/x86-64-avx512bw_vl.s
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/test/MC/X86/x86-64-avx512bw_vl.s?rev=214719&view=auto
> >
> ==============================================================================
> > --- llvm/trunk/test/MC/X86/x86-64-avx512bw_vl.s (added)
> > +++ llvm/trunk/test/MC/X86/x86-64-avx512bw_vl.s Mon Aug  4 09:35:15 2014
> > @@ -0,0 +1,257 @@
> > +// RUN: llvm-mc -triple x86_64-unknown-unknown -mcpu=skx
>  --show-encoding %s | FileCheck %s
> > +
> > +// CHECK: vmovdqu8 %xmm23, %xmm26
> > +// CHECK:  encoding: [0x62,0x21,0x7f,0x08,0x6f,0xd7]
> > +          vmovdqu8 %xmm23, %xmm26
> > +
> > +// CHECK: vmovdqu8 %xmm23, %xmm26 {%k2}
> > +// CHECK:  encoding: [0x62,0x21,0x7f,0x0a,0x6f,0xd7]
> > +          vmovdqu8 %xmm23, %xmm26 {%k2}
> > +
> > +// CHECK: vmovdqu8 %xmm23, %xmm26 {%k2} {z}
> > +// CHECK:  encoding: [0x62,0x21,0x7f,0x8a,0x6f,0xd7]
> > +          vmovdqu8 %xmm23, %xmm26 {%k2} {z}
> > +
> > +// CHECK: vmovdqu8 (%rcx), %xmm26
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x08,0x6f,0x11]
> > +          vmovdqu8 (%rcx), %xmm26
> > +
> > +// CHECK: vmovdqu8 291(%rax,%r14,8), %xmm26
> > +// CHECK:  encoding:
> [0x62,0x21,0x7f,0x08,0x6f,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu8 291(%rax,%r14,8), %xmm26
> > +
> > +// CHECK: vmovdqu8 2032(%rdx), %xmm26
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x08,0x6f,0x52,0x7f]
> > +          vmovdqu8 2032(%rdx), %xmm26
> > +
> > +// CHECK: vmovdqu8 2048(%rdx), %xmm26
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x08,0x6f,0x92,0x00,0x08,0x00,0x00]
> > +          vmovdqu8 2048(%rdx), %xmm26
> > +
> > +// CHECK: vmovdqu8 -2048(%rdx), %xmm26
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x08,0x6f,0x52,0x80]
> > +          vmovdqu8 -2048(%rdx), %xmm26
> > +
> > +// CHECK: vmovdqu8 -2064(%rdx), %xmm26
> > +// CHECK:  encoding: [0x62,0x61,0x7f,0x08,0x6f,0x92,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu8 -2064(%rdx), %xmm26
> > +
> > +// CHECK: vmovdqu8 %ymm29, %ymm18
> > +// CHECK:  encoding: [0x62,0x81,0x7f,0x28,0x6f,0xd5]
> > +          vmovdqu8 %ymm29, %ymm18
> > +
> > +// CHECK: vmovdqu8 %ymm29, %ymm18 {%k7}
> > +// CHECK:  encoding: [0x62,0x81,0x7f,0x2f,0x6f,0xd5]
> > +          vmovdqu8 %ymm29, %ymm18 {%k7}
> > +
> > +// CHECK: vmovdqu8 %ymm29, %ymm18 {%k7} {z}
> > +// CHECK:  encoding: [0x62,0x81,0x7f,0xaf,0x6f,0xd5]
> > +          vmovdqu8 %ymm29, %ymm18 {%k7} {z}
> > +
> > +// CHECK: vmovdqu8 (%rcx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x6f,0x11]
> > +          vmovdqu8 (%rcx), %ymm18
> > +
> > +// CHECK: vmovdqu8 291(%rax,%r14,8), %ymm18
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7f,0x28,0x6f,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu8 291(%rax,%r14,8), %ymm18
> > +
> > +// CHECK: vmovdqu8 4064(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x6f,0x52,0x7f]
> > +          vmovdqu8 4064(%rdx), %ymm18
> > +
> > +// CHECK: vmovdqu8 4096(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x6f,0x92,0x00,0x10,0x00,0x00]
> > +          vmovdqu8 4096(%rdx), %ymm18
> > +
> > +// CHECK: vmovdqu8 -4096(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x6f,0x52,0x80]
> > +          vmovdqu8 -4096(%rdx), %ymm18
> > +
> > +// CHECK: vmovdqu8 -4128(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x6f,0x92,0xe0,0xef,0xff,0xff]
> > +          vmovdqu8 -4128(%rdx), %ymm18
> > +
> > +// CHECK: vmovdqu16 %xmm24, %xmm29
> > +// CHECK:  encoding: [0x62,0x01,0xff,0x08,0x6f,0xe8]
> > +          vmovdqu16 %xmm24, %xmm29
> > +
> > +// CHECK: vmovdqu16 %xmm24, %xmm29 {%k6}
> > +// CHECK:  encoding: [0x62,0x01,0xff,0x0e,0x6f,0xe8]
> > +          vmovdqu16 %xmm24, %xmm29 {%k6}
> > +
> > +// CHECK: vmovdqu16 %xmm24, %xmm29 {%k6} {z}
> > +// CHECK:  encoding: [0x62,0x01,0xff,0x8e,0x6f,0xe8]
> > +          vmovdqu16 %xmm24, %xmm29 {%k6} {z}
> > +
> > +// CHECK: vmovdqu16 (%rcx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x08,0x6f,0x29]
> > +          vmovdqu16 (%rcx), %xmm29
> > +
> > +// CHECK: vmovdqu16 291(%rax,%r14,8), %xmm29
> > +// CHECK:  encoding:
> [0x62,0x21,0xff,0x08,0x6f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu16 291(%rax,%r14,8), %xmm29
> > +
> > +// CHECK: vmovdqu16 2032(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x08,0x6f,0x6a,0x7f]
> > +          vmovdqu16 2032(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu16 2048(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x08,0x6f,0xaa,0x00,0x08,0x00,0x00]
> > +          vmovdqu16 2048(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu16 -2048(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x08,0x6f,0x6a,0x80]
> > +          vmovdqu16 -2048(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu16 -2064(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x08,0x6f,0xaa,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu16 -2064(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu16 %ymm24, %ymm23
> > +// CHECK:  encoding: [0x62,0x81,0xff,0x28,0x6f,0xf8]
> > +          vmovdqu16 %ymm24, %ymm23
> > +
> > +// CHECK: vmovdqu16 %ymm24, %ymm23 {%k3}
> > +// CHECK:  encoding: [0x62,0x81,0xff,0x2b,0x6f,0xf8]
> > +          vmovdqu16 %ymm24, %ymm23 {%k3}
> > +
> > +// CHECK: vmovdqu16 %ymm24, %ymm23 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0x81,0xff,0xab,0x6f,0xf8]
> > +          vmovdqu16 %ymm24, %ymm23 {%k3} {z}
> > +
> > +// CHECK: vmovdqu16 (%rcx), %ymm23
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x28,0x6f,0x39]
> > +          vmovdqu16 (%rcx), %ymm23
> > +
> > +// CHECK: vmovdqu16 291(%rax,%r14,8), %ymm23
> > +// CHECK:  encoding:
> [0x62,0xa1,0xff,0x28,0x6f,0xbc,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu16 291(%rax,%r14,8), %ymm23
> > +
> > +// CHECK: vmovdqu16 4064(%rdx), %ymm23
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x28,0x6f,0x7a,0x7f]
> > +          vmovdqu16 4064(%rdx), %ymm23
> > +
> > +// CHECK: vmovdqu16 4096(%rdx), %ymm23
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x28,0x6f,0xba,0x00,0x10,0x00,0x00]
> > +          vmovdqu16 4096(%rdx), %ymm23
> > +
> > +// CHECK: vmovdqu16 -4096(%rdx), %ymm23
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x28,0x6f,0x7a,0x80]
> > +          vmovdqu16 -4096(%rdx), %ymm23
> > +
> > +// CHECK: vmovdqu16 -4128(%rdx), %ymm23
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x28,0x6f,0xba,0xe0,0xef,0xff,0xff]
> > +          vmovdqu16 -4128(%rdx), %ymm23
> > +
> > +// CHECK: vmovdqu8 %xmm17, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x08,0x7f,0x09]
> > +          vmovdqu8 %xmm17, (%rcx)
> > +
> > +// CHECK: vmovdqu8 %xmm17, (%rcx) {%k4}
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x0c,0x7f,0x09]
> > +          vmovdqu8 %xmm17, (%rcx) {%k4}
> > +
> > +// CHECK: vmovdqu8 %xmm17, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7f,0x08,0x7f,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu8 %xmm17, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu8 %xmm17, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x08,0x7f,0x4a,0x7f]
> > +          vmovdqu8 %xmm17, 2032(%rdx)
> > +
> > +// CHECK: vmovdqu8 %xmm17, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x08,0x7f,0x8a,0x00,0x08,0x00,0x00]
> > +          vmovdqu8 %xmm17, 2048(%rdx)
> > +
> > +// CHECK: vmovdqu8 %xmm17, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x08,0x7f,0x4a,0x80]
> > +          vmovdqu8 %xmm17, -2048(%rdx)
> > +
> > +// CHECK: vmovdqu8 %xmm17, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x08,0x7f,0x8a,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu8 %xmm17, -2064(%rdx)
> > +
> > +// CHECK: vmovdqu8 %ymm21, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x7f,0x29]
> > +          vmovdqu8 %ymm21, (%rcx)
> > +
> > +// CHECK: vmovdqu8 %ymm21, (%rcx) {%k1}
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x29,0x7f,0x29]
> > +          vmovdqu8 %ymm21, (%rcx) {%k1}
> > +
> > +// CHECK: vmovdqu8 %ymm21, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7f,0x28,0x7f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu8 %ymm21, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu8 %ymm21, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x7f,0x6a,0x7f]
> > +          vmovdqu8 %ymm21, 4064(%rdx)
> > +
> > +// CHECK: vmovdqu8 %ymm21, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x7f,0xaa,0x00,0x10,0x00,0x00]
> > +          vmovdqu8 %ymm21, 4096(%rdx)
> > +
> > +// CHECK: vmovdqu8 %ymm21, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x7f,0x6a,0x80]
> > +          vmovdqu8 %ymm21, -4096(%rdx)
> > +
> > +// CHECK: vmovdqu8 %ymm21, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7f,0x28,0x7f,0xaa,0xe0,0xef,0xff,0xff]
> > +          vmovdqu8 %ymm21, -4128(%rdx)
> > +
> > +// CHECK: vmovdqu16 %xmm23, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x08,0x7f,0x39]
> > +          vmovdqu16 %xmm23, (%rcx)
> > +
> > +// CHECK: vmovdqu16 %xmm23, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x0f,0x7f,0x39]
> > +          vmovdqu16 %xmm23, (%rcx) {%k7}
> > +
> > +// CHECK: vmovdqu16 %xmm23, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xff,0x08,0x7f,0xbc,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu16 %xmm23, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu16 %xmm23, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x08,0x7f,0x7a,0x7f]
> > +          vmovdqu16 %xmm23, 2032(%rdx)
> > +
> > +// CHECK: vmovdqu16 %xmm23, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x08,0x7f,0xba,0x00,0x08,0x00,0x00]
> > +          vmovdqu16 %xmm23, 2048(%rdx)
> > +
> > +// CHECK: vmovdqu16 %xmm23, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x08,0x7f,0x7a,0x80]
> > +          vmovdqu16 %xmm23, -2048(%rdx)
> > +
> > +// CHECK: vmovdqu16 %xmm23, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xff,0x08,0x7f,0xba,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu16 %xmm23, -2064(%rdx)
> > +
> > +// CHECK: vmovdqu16 %ymm29, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x28,0x7f,0x29]
> > +          vmovdqu16 %ymm29, (%rcx)
> > +
> > +// CHECK: vmovdqu16 %ymm29, (%rcx) {%k6}
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x2e,0x7f,0x29]
> > +          vmovdqu16 %ymm29, (%rcx) {%k6}
> > +
> > +// CHECK: vmovdqu16 %ymm29, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0xff,0x28,0x7f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu16 %ymm29, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu16 %ymm29, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x28,0x7f,0x6a,0x7f]
> > +          vmovdqu16 %ymm29, 4064(%rdx)
> > +
> > +// CHECK: vmovdqu16 %ymm29, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x28,0x7f,0xaa,0x00,0x10,0x00,0x00]
> > +          vmovdqu16 %ymm29, 4096(%rdx)
> > +
> > +// CHECK: vmovdqu16 %ymm29, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x28,0x7f,0x6a,0x80]
> > +          vmovdqu16 %ymm29, -4096(%rdx)
> > +
> > +// CHECK: vmovdqu16 %ymm29, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xff,0x28,0x7f,0xaa,0xe0,0xef,0xff,0xff]
> > +          vmovdqu16 %ymm29, -4128(%rdx)
> >
> > Added: llvm/trunk/test/MC/X86/x86-64-avx512f_vl.s
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/test/MC/X86/x86-64-avx512f_vl.s?rev=214719&view=auto
> >
> ==============================================================================
> > --- llvm/trunk/test/MC/X86/x86-64-avx512f_vl.s (added)
> > +++ llvm/trunk/test/MC/X86/x86-64-avx512f_vl.s Mon Aug  4 09:35:15 2014
> > @@ -0,0 +1,1025 @@
> > +// RUN: llvm-mc -triple x86_64-unknown-unknown -mcpu=skx
> --show-encoding %s | FileCheck %s
> > +
> > +// CHECK: vmovapd %xmm21, %xmm21
> > +// CHECK:  encoding: [0x62,0xa1,0xfd,0x08,0x28,0xed]
> > +          vmovapd %xmm21, %xmm21
> > +
> > +// CHECK: vmovapd %xmm21, %xmm21 {%k3}
> > +// CHECK:  encoding: [0x62,0xa1,0xfd,0x0b,0x28,0xed]
> > +          vmovapd %xmm21, %xmm21 {%k3}
> > +
> > +// CHECK: vmovapd %xmm21, %xmm21 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0xa1,0xfd,0x8b,0x28,0xed]
> > +          vmovapd %xmm21, %xmm21 {%k3} {z}
> > +
> > +// CHECK: vmovapd (%rcx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x28,0x29]
> > +          vmovapd (%rcx), %xmm21
> > +
> > +// CHECK: vmovapd 291(%rax,%r14,8), %xmm21
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x08,0x28,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovapd 291(%rax,%r14,8), %xmm21
> > +
> > +// CHECK: vmovapd 2032(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x28,0x6a,0x7f]
> > +          vmovapd 2032(%rdx), %xmm21
> > +
> > +// CHECK: vmovapd 2048(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x28,0xaa,0x00,0x08,0x00,0x00]
> > +          vmovapd 2048(%rdx), %xmm21
> > +
> > +// CHECK: vmovapd -2048(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x28,0x6a,0x80]
> > +          vmovapd -2048(%rdx), %xmm21
> > +
> > +// CHECK: vmovapd -2064(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x28,0xaa,0xf0,0xf7,0xff,0xff]
> > +          vmovapd -2064(%rdx), %xmm21
> > +
> > +// CHECK: vmovapd %ymm17, %ymm18
> > +// CHECK:  encoding: [0x62,0xa1,0xfd,0x28,0x28,0xd1]
> > +          vmovapd %ymm17, %ymm18
> > +
> > +// CHECK: vmovapd %ymm17, %ymm18 {%k2}
> > +// CHECK:  encoding: [0x62,0xa1,0xfd,0x2a,0x28,0xd1]
> > +          vmovapd %ymm17, %ymm18 {%k2}
> > +
> > +// CHECK: vmovapd %ymm17, %ymm18 {%k2} {z}
> > +// CHECK:  encoding: [0x62,0xa1,0xfd,0xaa,0x28,0xd1]
> > +          vmovapd %ymm17, %ymm18 {%k2} {z}
> > +
> > +// CHECK: vmovapd (%rcx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x28,0x11]
> > +          vmovapd (%rcx), %ymm18
> > +
> > +// CHECK: vmovapd 291(%rax,%r14,8), %ymm18
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x28,0x28,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovapd 291(%rax,%r14,8), %ymm18
> > +
> > +// CHECK: vmovapd 4064(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x28,0x52,0x7f]
> > +          vmovapd 4064(%rdx), %ymm18
> > +
> > +// CHECK: vmovapd 4096(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x28,0x92,0x00,0x10,0x00,0x00]
> > +          vmovapd 4096(%rdx), %ymm18
> > +
> > +// CHECK: vmovapd -4096(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x28,0x52,0x80]
> > +          vmovapd -4096(%rdx), %ymm18
> > +
> > +// CHECK: vmovapd -4128(%rdx), %ymm18
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x28,0x92,0xe0,0xef,0xff,0xff]
> > +          vmovapd -4128(%rdx), %ymm18
> > +
> > +// CHECK: vmovaps %xmm29, %xmm22
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x08,0x28,0xf5]
> > +          vmovaps %xmm29, %xmm22
> > +
> > +// CHECK: vmovaps %xmm29, %xmm22 {%k1}
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x09,0x28,0xf5]
> > +          vmovaps %xmm29, %xmm22 {%k1}
> > +
> > +// CHECK: vmovaps %xmm29, %xmm22 {%k1} {z}
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x89,0x28,0xf5]
> > +          vmovaps %xmm29, %xmm22 {%k1} {z}
> > +
> > +// CHECK: vmovaps (%rcx), %xmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x28,0x31]
> > +          vmovaps (%rcx), %xmm22
> > +
> > +// CHECK: vmovaps 291(%rax,%r14,8), %xmm22
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7c,0x08,0x28,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovaps 291(%rax,%r14,8), %xmm22
> > +
> > +// CHECK: vmovaps 2032(%rdx), %xmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x28,0x72,0x7f]
> > +          vmovaps 2032(%rdx), %xmm22
> > +
> > +// CHECK: vmovaps 2048(%rdx), %xmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x28,0xb2,0x00,0x08,0x00,0x00]
> > +          vmovaps 2048(%rdx), %xmm22
> > +
> > +// CHECK: vmovaps -2048(%rdx), %xmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x28,0x72,0x80]
> > +          vmovaps -2048(%rdx), %xmm22
> > +
> > +// CHECK: vmovaps -2064(%rdx), %xmm22
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x28,0xb2,0xf0,0xf7,0xff,0xff]
> > +          vmovaps -2064(%rdx), %xmm22
> > +
> > +// CHECK: vmovaps %ymm28, %ymm25
> > +// CHECK:  encoding: [0x62,0x01,0x7c,0x28,0x28,0xcc]
> > +          vmovaps %ymm28, %ymm25
> > +
> > +// CHECK: vmovaps %ymm28, %ymm25 {%k3}
> > +// CHECK:  encoding: [0x62,0x01,0x7c,0x2b,0x28,0xcc]
> > +          vmovaps %ymm28, %ymm25 {%k3}
> > +
> > +// CHECK: vmovaps %ymm28, %ymm25 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0x01,0x7c,0xab,0x28,0xcc]
> > +          vmovaps %ymm28, %ymm25 {%k3} {z}
> > +
> > +// CHECK: vmovaps (%rcx), %ymm25
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x28,0x09]
> > +          vmovaps (%rcx), %ymm25
> > +
> > +// CHECK: vmovaps 291(%rax,%r14,8), %ymm25
> > +// CHECK:  encoding:
> [0x62,0x21,0x7c,0x28,0x28,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovaps 291(%rax,%r14,8), %ymm25
> > +
> > +// CHECK: vmovaps 4064(%rdx), %ymm25
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x28,0x4a,0x7f]
> > +          vmovaps 4064(%rdx), %ymm25
> > +
> > +// CHECK: vmovaps 4096(%rdx), %ymm25
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x28,0x8a,0x00,0x10,0x00,0x00]
> > +          vmovaps 4096(%rdx), %ymm25
> > +
> > +// CHECK: vmovaps -4096(%rdx), %ymm25
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x28,0x4a,0x80]
> > +          vmovaps -4096(%rdx), %ymm25
> > +
> > +// CHECK: vmovaps -4128(%rdx), %ymm25
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x28,0x8a,0xe0,0xef,0xff,0xff]
> > +          vmovaps -4128(%rdx), %ymm25
> > +
> > +// CHECK: vmovdqa32 %xmm24, %xmm21
> > +// CHECK:  encoding: [0x62,0x81,0x7d,0x08,0x6f,0xe8]
> > +          vmovdqa32 %xmm24, %xmm21
> > +
> > +// CHECK: vmovdqa32 %xmm24, %xmm21 {%k6}
> > +// CHECK:  encoding: [0x62,0x81,0x7d,0x0e,0x6f,0xe8]
> > +          vmovdqa32 %xmm24, %xmm21 {%k6}
> > +
> > +// CHECK: vmovdqa32 %xmm24, %xmm21 {%k6} {z}
> > +// CHECK:  encoding: [0x62,0x81,0x7d,0x8e,0x6f,0xe8]
> > +          vmovdqa32 %xmm24, %xmm21 {%k6} {z}
> > +
> > +// CHECK: vmovdqa32 (%rcx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x08,0x6f,0x29]
> > +          vmovdqa32 (%rcx), %xmm21
> > +
> > +// CHECK: vmovdqa32 291(%rax,%r14,8), %xmm21
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7d,0x08,0x6f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa32 291(%rax,%r14,8), %xmm21
> > +
> > +// CHECK: vmovdqa32 2032(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x08,0x6f,0x6a,0x7f]
> > +          vmovdqa32 2032(%rdx), %xmm21
> > +
> > +// CHECK: vmovdqa32 2048(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x08,0x6f,0xaa,0x00,0x08,0x00,0x00]
> > +          vmovdqa32 2048(%rdx), %xmm21
> > +
> > +// CHECK: vmovdqa32 -2048(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x08,0x6f,0x6a,0x80]
> > +          vmovdqa32 -2048(%rdx), %xmm21
> > +
> > +// CHECK: vmovdqa32 -2064(%rdx), %xmm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7d,0x08,0x6f,0xaa,0xf0,0xf7,0xff,0xff]
> > +          vmovdqa32 -2064(%rdx), %xmm21
> > +
> > +// CHECK: vmovdqa32 %ymm28, %ymm24
> > +// CHECK:  encoding: [0x62,0x01,0x7d,0x28,0x6f,0xc4]
> > +          vmovdqa32 %ymm28, %ymm24
> > +
> > +// CHECK: vmovdqa32 %ymm28, %ymm24 {%k5}
> > +// CHECK:  encoding: [0x62,0x01,0x7d,0x2d,0x6f,0xc4]
> > +          vmovdqa32 %ymm28, %ymm24 {%k5}
> > +
> > +// CHECK: vmovdqa32 %ymm28, %ymm24 {%k5} {z}
> > +// CHECK:  encoding: [0x62,0x01,0x7d,0xad,0x6f,0xc4]
> > +          vmovdqa32 %ymm28, %ymm24 {%k5} {z}
> > +
> > +// CHECK: vmovdqa32 (%rcx), %ymm24
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x6f,0x01]
> > +          vmovdqa32 (%rcx), %ymm24
> > +
> > +// CHECK: vmovdqa32 291(%rax,%r14,8), %ymm24
> > +// CHECK:  encoding:
> [0x62,0x21,0x7d,0x28,0x6f,0x84,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa32 291(%rax,%r14,8), %ymm24
> > +
> > +// CHECK: vmovdqa32 4064(%rdx), %ymm24
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x6f,0x42,0x7f]
> > +          vmovdqa32 4064(%rdx), %ymm24
> > +
> > +// CHECK: vmovdqa32 4096(%rdx), %ymm24
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x6f,0x82,0x00,0x10,0x00,0x00]
> > +          vmovdqa32 4096(%rdx), %ymm24
> > +
> > +// CHECK: vmovdqa32 -4096(%rdx), %ymm24
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x6f,0x42,0x80]
> > +          vmovdqa32 -4096(%rdx), %ymm24
> > +
> > +// CHECK: vmovdqa32 -4128(%rdx), %ymm24
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x6f,0x82,0xe0,0xef,0xff,0xff]
> > +          vmovdqa32 -4128(%rdx), %ymm24
> > +
> > +// CHECK: vmovdqa64 %xmm24, %xmm27
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0x08,0x6f,0xd8]
> > +          vmovdqa64 %xmm24, %xmm27
> > +
> > +// CHECK: vmovdqa64 %xmm24, %xmm27 {%k5}
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0x0d,0x6f,0xd8]
> > +          vmovdqa64 %xmm24, %xmm27 {%k5}
> > +
> > +// CHECK: vmovdqa64 %xmm24, %xmm27 {%k5} {z}
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0x8d,0x6f,0xd8]
> > +          vmovdqa64 %xmm24, %xmm27 {%k5} {z}
> > +
> > +// CHECK: vmovdqa64 (%rcx), %xmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x6f,0x19]
> > +          vmovdqa64 (%rcx), %xmm27
> > +
> > +// CHECK: vmovdqa64 291(%rax,%r14,8), %xmm27
> > +// CHECK:  encoding:
> [0x62,0x21,0xfd,0x08,0x6f,0x9c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa64 291(%rax,%r14,8), %xmm27
> > +
> > +// CHECK: vmovdqa64 2032(%rdx), %xmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x6f,0x5a,0x7f]
> > +          vmovdqa64 2032(%rdx), %xmm27
> > +
> > +// CHECK: vmovdqa64 2048(%rdx), %xmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x6f,0x9a,0x00,0x08,0x00,0x00]
> > +          vmovdqa64 2048(%rdx), %xmm27
> > +
> > +// CHECK: vmovdqa64 -2048(%rdx), %xmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x6f,0x5a,0x80]
> > +          vmovdqa64 -2048(%rdx), %xmm27
> > +
> > +// CHECK: vmovdqa64 -2064(%rdx), %xmm27
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x6f,0x9a,0xf0,0xf7,0xff,0xff]
> > +          vmovdqa64 -2064(%rdx), %xmm27
> > +
> > +// CHECK: vmovdqa64 %ymm29, %ymm30
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0x28,0x6f,0xf5]
> > +          vmovdqa64 %ymm29, %ymm30
> > +
> > +// CHECK: vmovdqa64 %ymm29, %ymm30 {%k3}
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0x2b,0x6f,0xf5]
> > +          vmovdqa64 %ymm29, %ymm30 {%k3}
> > +
> > +// CHECK: vmovdqa64 %ymm29, %ymm30 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0xab,0x6f,0xf5]
> > +          vmovdqa64 %ymm29, %ymm30 {%k3} {z}
> > +
> > +// CHECK: vmovdqa64 (%rcx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x6f,0x31]
> > +          vmovdqa64 (%rcx), %ymm30
> > +
> > +// CHECK: vmovdqa64 291(%rax,%r14,8), %ymm30
> > +// CHECK:  encoding:
> [0x62,0x21,0xfd,0x28,0x6f,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa64 291(%rax,%r14,8), %ymm30
> > +
> > +// CHECK: vmovdqa64 4064(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x6f,0x72,0x7f]
> > +          vmovdqa64 4064(%rdx), %ymm30
> > +
> > +// CHECK: vmovdqa64 4096(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x6f,0xb2,0x00,0x10,0x00,0x00]
> > +          vmovdqa64 4096(%rdx), %ymm30
> > +
> > +// CHECK: vmovdqa64 -4096(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x6f,0x72,0x80]
> > +          vmovdqa64 -4096(%rdx), %ymm30
> > +
> > +// CHECK: vmovdqa64 -4128(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x6f,0xb2,0xe0,0xef,0xff,0xff]
> > +          vmovdqa64 -4128(%rdx), %ymm30
> > +
> > +// CHECK: vmovdqu32 %xmm19, %xmm29
> > +// CHECK:  encoding: [0x62,0x21,0x7e,0x08,0x6f,0xeb]
> > +          vmovdqu32 %xmm19, %xmm29
> > +
> > +// CHECK: vmovdqu32 %xmm19, %xmm29 {%k6}
> > +// CHECK:  encoding: [0x62,0x21,0x7e,0x0e,0x6f,0xeb]
> > +          vmovdqu32 %xmm19, %xmm29 {%k6}
> > +
> > +// CHECK: vmovdqu32 %xmm19, %xmm29 {%k6} {z}
> > +// CHECK:  encoding: [0x62,0x21,0x7e,0x8e,0x6f,0xeb]
> > +          vmovdqu32 %xmm19, %xmm29 {%k6} {z}
> > +
> > +// CHECK: vmovdqu32 (%rcx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7e,0x08,0x6f,0x29]
> > +          vmovdqu32 (%rcx), %xmm29
> > +
> > +// CHECK: vmovdqu32 291(%rax,%r14,8), %xmm29
> > +// CHECK:  encoding:
> [0x62,0x21,0x7e,0x08,0x6f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu32 291(%rax,%r14,8), %xmm29
> > +
> > +// CHECK: vmovdqu32 2032(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7e,0x08,0x6f,0x6a,0x7f]
> > +          vmovdqu32 2032(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu32 2048(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7e,0x08,0x6f,0xaa,0x00,0x08,0x00,0x00]
> > +          vmovdqu32 2048(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu32 -2048(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7e,0x08,0x6f,0x6a,0x80]
> > +          vmovdqu32 -2048(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu32 -2064(%rdx), %xmm29
> > +// CHECK:  encoding: [0x62,0x61,0x7e,0x08,0x6f,0xaa,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu32 -2064(%rdx), %xmm29
> > +
> > +// CHECK: vmovdqu32 %ymm18, %ymm17
> > +// CHECK:  encoding: [0x62,0xa1,0x7e,0x28,0x6f,0xca]
> > +          vmovdqu32 %ymm18, %ymm17
> > +
> > +// CHECK: vmovdqu32 %ymm18, %ymm17 {%k3}
> > +// CHECK:  encoding: [0x62,0xa1,0x7e,0x2b,0x6f,0xca]
> > +          vmovdqu32 %ymm18, %ymm17 {%k3}
> > +
> > +// CHECK: vmovdqu32 %ymm18, %ymm17 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0xa1,0x7e,0xab,0x6f,0xca]
> > +          vmovdqu32 %ymm18, %ymm17 {%k3} {z}
> > +
> > +// CHECK: vmovdqu32 (%rcx), %ymm17
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x6f,0x09]
> > +          vmovdqu32 (%rcx), %ymm17
> > +
> > +// CHECK: vmovdqu32 291(%rax,%r14,8), %ymm17
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7e,0x28,0x6f,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu32 291(%rax,%r14,8), %ymm17
> > +
> > +// CHECK: vmovdqu32 4064(%rdx), %ymm17
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x6f,0x4a,0x7f]
> > +          vmovdqu32 4064(%rdx), %ymm17
> > +
> > +// CHECK: vmovdqu32 4096(%rdx), %ymm17
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x6f,0x8a,0x00,0x10,0x00,0x00]
> > +          vmovdqu32 4096(%rdx), %ymm17
> > +
> > +// CHECK: vmovdqu32 -4096(%rdx), %ymm17
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x6f,0x4a,0x80]
> > +          vmovdqu32 -4096(%rdx), %ymm17
> > +
> > +// CHECK: vmovdqu32 -4128(%rdx), %ymm17
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x6f,0x8a,0xe0,0xef,0xff,0xff]
> > +          vmovdqu32 -4128(%rdx), %ymm17
> > +
> > +// CHECK: vmovdqu64 %xmm19, %xmm24
> > +// CHECK:  encoding: [0x62,0x21,0xfe,0x08,0x6f,0xc3]
> > +          vmovdqu64 %xmm19, %xmm24
> > +
> > +// CHECK: vmovdqu64 %xmm19, %xmm24 {%k5}
> > +// CHECK:  encoding: [0x62,0x21,0xfe,0x0d,0x6f,0xc3]
> > +          vmovdqu64 %xmm19, %xmm24 {%k5}
> > +
> > +// CHECK: vmovdqu64 %xmm19, %xmm24 {%k5} {z}
> > +// CHECK:  encoding: [0x62,0x21,0xfe,0x8d,0x6f,0xc3]
> > +          vmovdqu64 %xmm19, %xmm24 {%k5} {z}
> > +
> > +// CHECK: vmovdqu64 (%rcx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x08,0x6f,0x01]
> > +          vmovdqu64 (%rcx), %xmm24
> > +
> > +// CHECK: vmovdqu64 291(%rax,%r14,8), %xmm24
> > +// CHECK:  encoding:
> [0x62,0x21,0xfe,0x08,0x6f,0x84,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu64 291(%rax,%r14,8), %xmm24
> > +
> > +// CHECK: vmovdqu64 2032(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x08,0x6f,0x42,0x7f]
> > +          vmovdqu64 2032(%rdx), %xmm24
> > +
> > +// CHECK: vmovdqu64 2048(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x08,0x6f,0x82,0x00,0x08,0x00,0x00]
> > +          vmovdqu64 2048(%rdx), %xmm24
> > +
> > +// CHECK: vmovdqu64 -2048(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x08,0x6f,0x42,0x80]
> > +          vmovdqu64 -2048(%rdx), %xmm24
> > +
> > +// CHECK: vmovdqu64 -2064(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x08,0x6f,0x82,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu64 -2064(%rdx), %xmm24
> > +
> > +// CHECK: vmovdqu64 %ymm21, %ymm29
> > +// CHECK:  encoding: [0x62,0x21,0xfe,0x28,0x6f,0xed]
> > +          vmovdqu64 %ymm21, %ymm29
> > +
> > +// CHECK: vmovdqu64 %ymm21, %ymm29 {%k3}
> > +// CHECK:  encoding: [0x62,0x21,0xfe,0x2b,0x6f,0xed]
> > +          vmovdqu64 %ymm21, %ymm29 {%k3}
> > +
> > +// CHECK: vmovdqu64 %ymm21, %ymm29 {%k3} {z}
> > +// CHECK:  encoding: [0x62,0x21,0xfe,0xab,0x6f,0xed]
> > +          vmovdqu64 %ymm21, %ymm29 {%k3} {z}
> > +
> > +// CHECK: vmovdqu64 (%rcx), %ymm29
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x28,0x6f,0x29]
> > +          vmovdqu64 (%rcx), %ymm29
> > +
> > +// CHECK: vmovdqu64 291(%rax,%r14,8), %ymm29
> > +// CHECK:  encoding:
> [0x62,0x21,0xfe,0x28,0x6f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu64 291(%rax,%r14,8), %ymm29
> > +
> > +// CHECK: vmovdqu64 4064(%rdx), %ymm29
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x28,0x6f,0x6a,0x7f]
> > +          vmovdqu64 4064(%rdx), %ymm29
> > +
> > +// CHECK: vmovdqu64 4096(%rdx), %ymm29
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x28,0x6f,0xaa,0x00,0x10,0x00,0x00]
> > +          vmovdqu64 4096(%rdx), %ymm29
> > +
> > +// CHECK: vmovdqu64 -4096(%rdx), %ymm29
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x28,0x6f,0x6a,0x80]
> > +          vmovdqu64 -4096(%rdx), %ymm29
> > +
> > +// CHECK: vmovdqu64 -4128(%rdx), %ymm29
> > +// CHECK:  encoding: [0x62,0x61,0xfe,0x28,0x6f,0xaa,0xe0,0xef,0xff,0xff]
> > +          vmovdqu64 -4128(%rdx), %ymm29
> > +
> > +// CHECK: vmovupd %xmm22, %xmm24
> > +// CHECK:  encoding: [0x62,0x21,0xfd,0x08,0x10,0xc6]
> > +          vmovupd %xmm22, %xmm24
> > +
> > +// CHECK: vmovupd %xmm22, %xmm24 {%k6}
> > +// CHECK:  encoding: [0x62,0x21,0xfd,0x0e,0x10,0xc6]
> > +          vmovupd %xmm22, %xmm24 {%k6}
> > +
> > +// CHECK: vmovupd %xmm22, %xmm24 {%k6} {z}
> > +// CHECK:  encoding: [0x62,0x21,0xfd,0x8e,0x10,0xc6]
> > +          vmovupd %xmm22, %xmm24 {%k6} {z}
> > +
> > +// CHECK: vmovupd (%rcx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x10,0x01]
> > +          vmovupd (%rcx), %xmm24
> > +
> > +// CHECK: vmovupd 291(%rax,%r14,8), %xmm24
> > +// CHECK:  encoding:
> [0x62,0x21,0xfd,0x08,0x10,0x84,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovupd 291(%rax,%r14,8), %xmm24
> > +
> > +// CHECK: vmovupd 2032(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x10,0x42,0x7f]
> > +          vmovupd 2032(%rdx), %xmm24
> > +
> > +// CHECK: vmovupd 2048(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x10,0x82,0x00,0x08,0x00,0x00]
> > +          vmovupd 2048(%rdx), %xmm24
> > +
> > +// CHECK: vmovupd -2048(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x10,0x42,0x80]
> > +          vmovupd -2048(%rdx), %xmm24
> > +
> > +// CHECK: vmovupd -2064(%rdx), %xmm24
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x08,0x10,0x82,0xf0,0xf7,0xff,0xff]
> > +          vmovupd -2064(%rdx), %xmm24
> > +
> > +// CHECK: vmovupd %ymm25, %ymm30
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0x28,0x10,0xf1]
> > +          vmovupd %ymm25, %ymm30
> > +
> > +// CHECK: vmovupd %ymm25, %ymm30 {%k7}
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0x2f,0x10,0xf1]
> > +          vmovupd %ymm25, %ymm30 {%k7}
> > +
> > +// CHECK: vmovupd %ymm25, %ymm30 {%k7} {z}
> > +// CHECK:  encoding: [0x62,0x01,0xfd,0xaf,0x10,0xf1]
> > +          vmovupd %ymm25, %ymm30 {%k7} {z}
> > +
> > +// CHECK: vmovupd (%rcx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x10,0x31]
> > +          vmovupd (%rcx), %ymm30
> > +
> > +// CHECK: vmovupd 291(%rax,%r14,8), %ymm30
> > +// CHECK:  encoding:
> [0x62,0x21,0xfd,0x28,0x10,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovupd 291(%rax,%r14,8), %ymm30
> > +
> > +// CHECK: vmovupd 4064(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x10,0x72,0x7f]
> > +          vmovupd 4064(%rdx), %ymm30
> > +
> > +// CHECK: vmovupd 4096(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x10,0xb2,0x00,0x10,0x00,0x00]
> > +          vmovupd 4096(%rdx), %ymm30
> > +
> > +// CHECK: vmovupd -4096(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x10,0x72,0x80]
> > +          vmovupd -4096(%rdx), %ymm30
> > +
> > +// CHECK: vmovupd -4128(%rdx), %ymm30
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x10,0xb2,0xe0,0xef,0xff,0xff]
> > +          vmovupd -4128(%rdx), %ymm30
> > +
> > +// CHECK: vmovups %xmm29, %xmm20
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x08,0x10,0xe5]
> > +          vmovups %xmm29, %xmm20
> > +
> > +// CHECK: vmovups %xmm29, %xmm20 {%k6}
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x0e,0x10,0xe5]
> > +          vmovups %xmm29, %xmm20 {%k6}
> > +
> > +// CHECK: vmovups %xmm29, %xmm20 {%k6} {z}
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x8e,0x10,0xe5]
> > +          vmovups %xmm29, %xmm20 {%k6} {z}
> > +
> > +// CHECK: vmovups (%rcx), %xmm20
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x10,0x21]
> > +          vmovups (%rcx), %xmm20
> > +
> > +// CHECK: vmovups 291(%rax,%r14,8), %xmm20
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7c,0x08,0x10,0xa4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovups 291(%rax,%r14,8), %xmm20
> > +
> > +// CHECK: vmovups 2032(%rdx), %xmm20
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x10,0x62,0x7f]
> > +          vmovups 2032(%rdx), %xmm20
> > +
> > +// CHECK: vmovups 2048(%rdx), %xmm20
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x10,0xa2,0x00,0x08,0x00,0x00]
> > +          vmovups 2048(%rdx), %xmm20
> > +
> > +// CHECK: vmovups -2048(%rdx), %xmm20
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x10,0x62,0x80]
> > +          vmovups -2048(%rdx), %xmm20
> > +
> > +// CHECK: vmovups -2064(%rdx), %xmm20
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x08,0x10,0xa2,0xf0,0xf7,0xff,0xff]
> > +          vmovups -2064(%rdx), %xmm20
> > +
> > +// CHECK: vmovups %ymm26, %ymm21
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x28,0x10,0xea]
> > +          vmovups %ymm26, %ymm21
> > +
> > +// CHECK: vmovups %ymm26, %ymm21 {%k6}
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0x2e,0x10,0xea]
> > +          vmovups %ymm26, %ymm21 {%k6}
> > +
> > +// CHECK: vmovups %ymm26, %ymm21 {%k6} {z}
> > +// CHECK:  encoding: [0x62,0x81,0x7c,0xae,0x10,0xea]
> > +          vmovups %ymm26, %ymm21 {%k6} {z}
> > +
> > +// CHECK: vmovups (%rcx), %ymm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x10,0x29]
> > +          vmovups (%rcx), %ymm21
> > +
> > +// CHECK: vmovups 291(%rax,%r14,8), %ymm21
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7c,0x28,0x10,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovups 291(%rax,%r14,8), %ymm21
> > +
> > +// CHECK: vmovups 4064(%rdx), %ymm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x10,0x6a,0x7f]
> > +          vmovups 4064(%rdx), %ymm21
> > +
> > +// CHECK: vmovups 4096(%rdx), %ymm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x10,0xaa,0x00,0x10,0x00,0x00]
> > +          vmovups 4096(%rdx), %ymm21
> > +
> > +// CHECK: vmovups -4096(%rdx), %ymm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x10,0x6a,0x80]
> > +          vmovups -4096(%rdx), %ymm21
> > +
> > +// CHECK: vmovups -4128(%rdx), %ymm21
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x10,0xaa,0xe0,0xef,0xff,0xff]
> > +          vmovups -4128(%rdx), %ymm21
> > +
> > +// CHECK: vmovapd %xmm22, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x29,0x31]
> > +          vmovapd %xmm22, (%rcx)
> > +
> > +// CHECK: vmovapd %xmm22, (%rcx) {%k2}
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x0a,0x29,0x31]
> > +          vmovapd %xmm22, (%rcx) {%k2}
> > +
> > +// CHECK: vmovapd %xmm22, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x08,0x29,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovapd %xmm22, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovapd %xmm22, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x29,0x72,0x7f]
> > +          vmovapd %xmm22, 2032(%rdx)
> > +
> > +// CHECK: vmovapd %xmm22, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x29,0xb2,0x00,0x08,0x00,0x00]
> > +          vmovapd %xmm22, 2048(%rdx)
> > +
> > +// CHECK: vmovapd %xmm22, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x29,0x72,0x80]
> > +          vmovapd %xmm22, -2048(%rdx)
> > +
> > +// CHECK: vmovapd %xmm22, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x29,0xb2,0xf0,0xf7,0xff,0xff]
> > +          vmovapd %xmm22, -2064(%rdx)
> > +
> > +// CHECK: vmovapd %ymm17, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x29,0x09]
> > +          vmovapd %ymm17, (%rcx)
> > +
> > +// CHECK: vmovapd %ymm17, (%rcx) {%k6}
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x2e,0x29,0x09]
> > +          vmovapd %ymm17, (%rcx) {%k6}
> > +
> > +// CHECK: vmovapd %ymm17, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x28,0x29,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovapd %ymm17, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovapd %ymm17, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x29,0x4a,0x7f]
> > +          vmovapd %ymm17, 4064(%rdx)
> > +
> > +// CHECK: vmovapd %ymm17, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x29,0x8a,0x00,0x10,0x00,0x00]
> > +          vmovapd %ymm17, 4096(%rdx)
> > +
> > +// CHECK: vmovapd %ymm17, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x29,0x4a,0x80]
> > +          vmovapd %ymm17, -4096(%rdx)
> > +
> > +// CHECK: vmovapd %ymm17, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x28,0x29,0x8a,0xe0,0xef,0xff,0xff]
> > +          vmovapd %ymm17, -4128(%rdx)
> > +
> > +// CHECK: vmovaps %xmm29, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x29,0x29]
> > +          vmovaps %xmm29, (%rcx)
> > +
> > +// CHECK: vmovaps %xmm29, (%rcx) {%k5}
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x0d,0x29,0x29]
> > +          vmovaps %xmm29, (%rcx) {%k5}
> > +
> > +// CHECK: vmovaps %xmm29, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0x7c,0x08,0x29,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovaps %xmm29, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovaps %xmm29, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x29,0x6a,0x7f]
> > +          vmovaps %xmm29, 2032(%rdx)
> > +
> > +// CHECK: vmovaps %xmm29, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x29,0xaa,0x00,0x08,0x00,0x00]
> > +          vmovaps %xmm29, 2048(%rdx)
> > +
> > +// CHECK: vmovaps %xmm29, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x29,0x6a,0x80]
> > +          vmovaps %xmm29, -2048(%rdx)
> > +
> > +// CHECK: vmovaps %xmm29, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x29,0xaa,0xf0,0xf7,0xff,0xff]
> > +          vmovaps %xmm29, -2064(%rdx)
> > +
> > +// CHECK: vmovaps %ymm28, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x29,0x21]
> > +          vmovaps %ymm28, (%rcx)
> > +
> > +// CHECK: vmovaps %ymm28, (%rcx) {%k6}
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x2e,0x29,0x21]
> > +          vmovaps %ymm28, (%rcx) {%k6}
> > +
> > +// CHECK: vmovaps %ymm28, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0x7c,0x28,0x29,0xa4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovaps %ymm28, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovaps %ymm28, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x29,0x62,0x7f]
> > +          vmovaps %ymm28, 4064(%rdx)
> > +
> > +// CHECK: vmovaps %ymm28, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x29,0xa2,0x00,0x10,0x00,0x00]
> > +          vmovaps %ymm28, 4096(%rdx)
> > +
> > +// CHECK: vmovaps %ymm28, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x29,0x62,0x80]
> > +          vmovaps %ymm28, -4096(%rdx)
> > +
> > +// CHECK: vmovaps %ymm28, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x28,0x29,0xa2,0xe0,0xef,0xff,0xff]
> > +          vmovaps %ymm28, -4128(%rdx)
> > +
> > +// CHECK: vmovdqa32 %xmm24, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x08,0x7f,0x01]
> > +          vmovdqa32 %xmm24, (%rcx)
> > +
> > +// CHECK: vmovdqa32 %xmm24, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x0f,0x7f,0x01]
> > +          vmovdqa32 %xmm24, (%rcx) {%k7}
> > +
> > +// CHECK: vmovdqa32 %xmm24, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0x7d,0x08,0x7f,0x84,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa32 %xmm24, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqa32 %xmm24, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x08,0x7f,0x42,0x7f]
> > +          vmovdqa32 %xmm24, 2032(%rdx)
> > +
> > +// CHECK: vmovdqa32 %xmm24, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x08,0x7f,0x82,0x00,0x08,0x00,0x00]
> > +          vmovdqa32 %xmm24, 2048(%rdx)
> > +
> > +// CHECK: vmovdqa32 %xmm24, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x08,0x7f,0x42,0x80]
> > +          vmovdqa32 %xmm24, -2048(%rdx)
> > +
> > +// CHECK: vmovdqa32 %xmm24, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x08,0x7f,0x82,0xf0,0xf7,0xff,0xff]
> > +          vmovdqa32 %xmm24, -2064(%rdx)
> > +
> > +// CHECK: vmovdqa32 %ymm29, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x7f,0x29]
> > +          vmovdqa32 %ymm29, (%rcx)
> > +
> > +// CHECK: vmovdqa32 %ymm29, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x2f,0x7f,0x29]
> > +          vmovdqa32 %ymm29, (%rcx) {%k7}
> > +
> > +// CHECK: vmovdqa32 %ymm29, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0x7d,0x28,0x7f,0xac,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa32 %ymm29, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqa32 %ymm29, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x7f,0x6a,0x7f]
> > +          vmovdqa32 %ymm29, 4064(%rdx)
> > +
> > +// CHECK: vmovdqa32 %ymm29, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x7f,0xaa,0x00,0x10,0x00,0x00]
> > +          vmovdqa32 %ymm29, 4096(%rdx)
> > +
> > +// CHECK: vmovdqa32 %ymm29, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x7f,0x6a,0x80]
> > +          vmovdqa32 %ymm29, -4096(%rdx)
> > +
> > +// CHECK: vmovdqa32 %ymm29, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7d,0x28,0x7f,0xaa,0xe0,0xef,0xff,0xff]
> > +          vmovdqa32 %ymm29, -4128(%rdx)
> > +
> > +// CHECK: vmovdqa64 %xmm17, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x7f,0x09]
> > +          vmovdqa64 %xmm17, (%rcx)
> > +
> > +// CHECK: vmovdqa64 %xmm17, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x0f,0x7f,0x09]
> > +          vmovdqa64 %xmm17, (%rcx) {%k7}
> > +
> > +// CHECK: vmovdqa64 %xmm17, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x08,0x7f,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa64 %xmm17, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqa64 %xmm17, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x7f,0x4a,0x7f]
> > +          vmovdqa64 %xmm17, 2032(%rdx)
> > +
> > +// CHECK: vmovdqa64 %xmm17, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x7f,0x8a,0x00,0x08,0x00,0x00]
> > +          vmovdqa64 %xmm17, 2048(%rdx)
> > +
> > +// CHECK: vmovdqa64 %xmm17, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x7f,0x4a,0x80]
> > +          vmovdqa64 %xmm17, -2048(%rdx)
> > +
> > +// CHECK: vmovdqa64 %xmm17, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x7f,0x8a,0xf0,0xf7,0xff,0xff]
> > +          vmovdqa64 %xmm17, -2064(%rdx)
> > +
> > +// CHECK: vmovdqa64 %ymm24, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x7f,0x01]
> > +          vmovdqa64 %ymm24, (%rcx)
> > +
> > +// CHECK: vmovdqa64 %ymm24, (%rcx) {%k2}
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x2a,0x7f,0x01]
> > +          vmovdqa64 %ymm24, (%rcx) {%k2}
> > +
> > +// CHECK: vmovdqa64 %ymm24, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0xfd,0x28,0x7f,0x84,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqa64 %ymm24, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqa64 %ymm24, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x7f,0x42,0x7f]
> > +          vmovdqa64 %ymm24, 4064(%rdx)
> > +
> > +// CHECK: vmovdqa64 %ymm24, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x7f,0x82,0x00,0x10,0x00,0x00]
> > +          vmovdqa64 %ymm24, 4096(%rdx)
> > +
> > +// CHECK: vmovdqa64 %ymm24, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x7f,0x42,0x80]
> > +          vmovdqa64 %ymm24, -4096(%rdx)
> > +
> > +// CHECK: vmovdqa64 %ymm24, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x7f,0x82,0xe0,0xef,0xff,0xff]
> > +          vmovdqa64 %ymm24, -4128(%rdx)
> > +
> > +// CHECK: vmovdqu32 %xmm17, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x08,0x7f,0x09]
> > +          vmovdqu32 %xmm17, (%rcx)
> > +
> > +// CHECK: vmovdqu32 %xmm17, (%rcx) {%k4}
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x0c,0x7f,0x09]
> > +          vmovdqu32 %xmm17, (%rcx) {%k4}
> > +
> > +// CHECK: vmovdqu32 %xmm17, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7e,0x08,0x7f,0x8c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu32 %xmm17, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu32 %xmm17, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x08,0x7f,0x4a,0x7f]
> > +          vmovdqu32 %xmm17, 2032(%rdx)
> > +
> > +// CHECK: vmovdqu32 %xmm17, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x08,0x7f,0x8a,0x00,0x08,0x00,0x00]
> > +          vmovdqu32 %xmm17, 2048(%rdx)
> > +
> > +// CHECK: vmovdqu32 %xmm17, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x08,0x7f,0x4a,0x80]
> > +          vmovdqu32 %xmm17, -2048(%rdx)
> > +
> > +// CHECK: vmovdqu32 %xmm17, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x08,0x7f,0x8a,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu32 %xmm17, -2064(%rdx)
> > +
> > +// CHECK: vmovdqu32 %ymm20, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x7f,0x21]
> > +          vmovdqu32 %ymm20, (%rcx)
> > +
> > +// CHECK: vmovdqu32 %ymm20, (%rcx) {%k1}
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x29,0x7f,0x21]
> > +          vmovdqu32 %ymm20, (%rcx) {%k1}
> > +
> > +// CHECK: vmovdqu32 %ymm20, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7e,0x28,0x7f,0xa4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu32 %ymm20, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu32 %ymm20, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x7f,0x62,0x7f]
> > +          vmovdqu32 %ymm20, 4064(%rdx)
> > +
> > +// CHECK: vmovdqu32 %ymm20, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x7f,0xa2,0x00,0x10,0x00,0x00]
> > +          vmovdqu32 %ymm20, 4096(%rdx)
> > +
> > +// CHECK: vmovdqu32 %ymm20, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x7f,0x62,0x80]
> > +          vmovdqu32 %ymm20, -4096(%rdx)
> > +
> > +// CHECK: vmovdqu32 %ymm20, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7e,0x28,0x7f,0xa2,0xe0,0xef,0xff,0xff]
> > +          vmovdqu32 %ymm20, -4128(%rdx)
> > +
> > +// CHECK: vmovdqu64 %xmm20, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x08,0x7f,0x21]
> > +          vmovdqu64 %xmm20, (%rcx)
> > +
> > +// CHECK: vmovdqu64 %xmm20, (%rcx) {%k6}
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x0e,0x7f,0x21]
> > +          vmovdqu64 %xmm20, (%rcx) {%k6}
> > +
> > +// CHECK: vmovdqu64 %xmm20, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfe,0x08,0x7f,0xa4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu64 %xmm20, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu64 %xmm20, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x08,0x7f,0x62,0x7f]
> > +          vmovdqu64 %xmm20, 2032(%rdx)
> > +
> > +// CHECK: vmovdqu64 %xmm20, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x08,0x7f,0xa2,0x00,0x08,0x00,0x00]
> > +          vmovdqu64 %xmm20, 2048(%rdx)
> > +
> > +// CHECK: vmovdqu64 %xmm20, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x08,0x7f,0x62,0x80]
> > +          vmovdqu64 %xmm20, -2048(%rdx)
> > +
> > +// CHECK: vmovdqu64 %xmm20, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x08,0x7f,0xa2,0xf0,0xf7,0xff,0xff]
> > +          vmovdqu64 %xmm20, -2064(%rdx)
> > +
> > +// CHECK: vmovdqu64 %ymm19, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x28,0x7f,0x19]
> > +          vmovdqu64 %ymm19, (%rcx)
> > +
> > +// CHECK: vmovdqu64 %ymm19, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x2f,0x7f,0x19]
> > +          vmovdqu64 %ymm19, (%rcx) {%k7}
> > +
> > +// CHECK: vmovdqu64 %ymm19, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfe,0x28,0x7f,0x9c,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovdqu64 %ymm19, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovdqu64 %ymm19, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x28,0x7f,0x5a,0x7f]
> > +          vmovdqu64 %ymm19, 4064(%rdx)
> > +
> > +// CHECK: vmovdqu64 %ymm19, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x28,0x7f,0x9a,0x00,0x10,0x00,0x00]
> > +          vmovdqu64 %ymm19, 4096(%rdx)
> > +
> > +// CHECK: vmovdqu64 %ymm19, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x28,0x7f,0x5a,0x80]
> > +          vmovdqu64 %ymm19, -4096(%rdx)
> > +
> > +// CHECK: vmovdqu64 %ymm19, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfe,0x28,0x7f,0x9a,0xe0,0xef,0xff,0xff]
> > +          vmovdqu64 %ymm19, -4128(%rdx)
> > +
> > +// CHECK: vmovupd %xmm22, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x11,0x31]
> > +          vmovupd %xmm22, (%rcx)
> > +
> > +// CHECK: vmovupd %xmm22, (%rcx) {%k7}
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x0f,0x11,0x31]
> > +          vmovupd %xmm22, (%rcx) {%k7}
> > +
> > +// CHECK: vmovupd %xmm22, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0xfd,0x08,0x11,0xb4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovupd %xmm22, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovupd %xmm22, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x11,0x72,0x7f]
> > +          vmovupd %xmm22, 2032(%rdx)
> > +
> > +// CHECK: vmovupd %xmm22, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x11,0xb2,0x00,0x08,0x00,0x00]
> > +          vmovupd %xmm22, 2048(%rdx)
> > +
> > +// CHECK: vmovupd %xmm22, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x11,0x72,0x80]
> > +          vmovupd %xmm22, -2048(%rdx)
> > +
> > +// CHECK: vmovupd %xmm22, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0xfd,0x08,0x11,0xb2,0xf0,0xf7,0xff,0xff]
> > +          vmovupd %xmm22, -2064(%rdx)
> > +
> > +// CHECK: vmovupd %ymm28, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x11,0x21]
> > +          vmovupd %ymm28, (%rcx)
> > +
> > +// CHECK: vmovupd %ymm28, (%rcx) {%k1}
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x29,0x11,0x21]
> > +          vmovupd %ymm28, (%rcx) {%k1}
> > +
> > +// CHECK: vmovupd %ymm28, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0xfd,0x28,0x11,0xa4,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovupd %ymm28, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovupd %ymm28, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x11,0x62,0x7f]
> > +          vmovupd %ymm28, 4064(%rdx)
> > +
> > +// CHECK: vmovupd %ymm28, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x11,0xa2,0x00,0x10,0x00,0x00]
> > +          vmovupd %ymm28, 4096(%rdx)
> > +
> > +// CHECK: vmovupd %ymm28, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x11,0x62,0x80]
> > +          vmovupd %ymm28, -4096(%rdx)
> > +
> > +// CHECK: vmovupd %ymm28, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0xfd,0x28,0x11,0xa2,0xe0,0xef,0xff,0xff]
> > +          vmovupd %ymm28, -4128(%rdx)
> > +
> > +// CHECK: vmovups %xmm26, (%rcx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x11,0x11]
> > +          vmovups %xmm26, (%rcx)
> > +
> > +// CHECK: vmovups %xmm26, (%rcx) {%k5}
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x0d,0x11,0x11]
> > +          vmovups %xmm26, (%rcx) {%k5}
> > +
> > +// CHECK: vmovups %xmm26, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0x21,0x7c,0x08,0x11,0x94,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovups %xmm26, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovups %xmm26, 2032(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x11,0x52,0x7f]
> > +          vmovups %xmm26, 2032(%rdx)
> > +
> > +// CHECK: vmovups %xmm26, 2048(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x11,0x92,0x00,0x08,0x00,0x00]
> > +          vmovups %xmm26, 2048(%rdx)
> > +
> > +// CHECK: vmovups %xmm26, -2048(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x11,0x52,0x80]
> > +          vmovups %xmm26, -2048(%rdx)
> > +
> > +// CHECK: vmovups %xmm26, -2064(%rdx)
> > +// CHECK:  encoding: [0x62,0x61,0x7c,0x08,0x11,0x92,0xf0,0xf7,0xff,0xff]
> > +          vmovups %xmm26, -2064(%rdx)
> > +
> > +// CHECK: vmovups %ymm23, (%rcx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x11,0x39]
> > +          vmovups %ymm23, (%rcx)
> > +
> > +// CHECK: vmovups %ymm23, (%rcx) {%k3}
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x2b,0x11,0x39]
> > +          vmovups %ymm23, (%rcx) {%k3}
> > +
> > +// CHECK: vmovups %ymm23, 291(%rax,%r14,8)
> > +// CHECK:  encoding:
> [0x62,0xa1,0x7c,0x28,0x11,0xbc,0xf0,0x23,0x01,0x00,0x00]
> > +          vmovups %ymm23, 291(%rax,%r14,8)
> > +
> > +// CHECK: vmovups %ymm23, 4064(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x11,0x7a,0x7f]
> > +          vmovups %ymm23, 4064(%rdx)
> > +
> > +// CHECK: vmovups %ymm23, 4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x11,0xba,0x00,0x10,0x00,0x00]
> > +          vmovups %ymm23, 4096(%rdx)
> > +
> > +// CHECK: vmovups %ymm23, -4096(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x11,0x7a,0x80]
> > +          vmovups %ymm23, -4096(%rdx)
> > +
> > +// CHECK: vmovups %ymm23, -4128(%rdx)
> > +// CHECK:  encoding: [0x62,0xe1,0x7c,0x28,0x11,0xba,0xe0,0xef,0xff,0xff]
> > +          vmovups %ymm23, -4128(%rdx)
> >
> > Modified: llvm/trunk/utils/TableGen/X86DisassemblerTables.cpp
> > URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/utils/TableGen/X86DisassemblerTables.cpp?rev=214719&r1=214718&r2=214719&view=diff
> >
> ==============================================================================
> > --- llvm/trunk/utils/TableGen/X86DisassemblerTables.cpp (original)
> > +++ llvm/trunk/utils/TableGen/X86DisassemblerTables.cpp Mon Aug  4
> 09:35:15 2014
> > @@ -205,8 +205,19 @@ static inline bool inheritsFrom(Instruct
> >   case IC_EVEX_XD_K:
> >     return inheritsFrom(child, IC_EVEX_W_XD_K) ||
> >            inheritsFrom(child, IC_EVEX_L_W_XD_K);
> > +  case IC_EVEX_K_B:
> > +  case IC_EVEX_KZ:
> > +    return false;
> > +  case IC_EVEX_XS_KZ:
> > +    return inheritsFrom(child, IC_EVEX_W_XS_KZ) ||
> > +           inheritsFrom(child, IC_EVEX_L_W_XS_KZ);
> > +  case IC_EVEX_XD_KZ:
> > +    return inheritsFrom(child, IC_EVEX_W_XD_KZ) ||
> > +           inheritsFrom(child, IC_EVEX_L_W_XD_KZ);
> > +  case IC_EVEX_KZ_B:
> >   case IC_EVEX_OPSIZE_K:
> >   case IC_EVEX_OPSIZE_B:
> > +  case IC_EVEX_OPSIZE_KZ:
> >     return false;
> >   case IC_EVEX_W_K:
> >   case IC_EVEX_W_XS_K:
> >
> >
> > _______________________________________________
> > llvm-commits mailing list
> > llvm-commits at cs.uiuc.edu
> > http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.llvm.org/pipermail/llvm-commits/attachments/20140805/c2419229/attachment.html>


More information about the llvm-commits mailing list