From 3a173dfc7240aed52a9ebaea6890a76cf54ec333 Mon Sep 17 00:00:00 2001 From: Chris Lattner Date: Sun, 3 Oct 2004 20:35:00 +0000 Subject: [PATCH] Add support to the instruction patterns for AT&T style output, which will hopefully lead to the death of the 'GasBugWorkaroundEmitter'. This also includes changes to wrap the whole file to 80 columns! Woot! :) Note that the AT&T style output has not been tested at all. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@16638 91177308-0d34-0410-b5e6-96231b3b80d8 --- lib/Target/X86/X86InstrInfo.td | 1532 ++++++++++++++++++++------------ 1 file changed, 963 insertions(+), 569 deletions(-) diff --git a/lib/Target/X86/X86InstrInfo.td b/lib/Target/X86/X86InstrInfo.td index 2cab50b62fe..b080b5e29d6 100644 --- a/lib/Target/X86/X86InstrInfo.td +++ b/lib/Target/X86/X86InstrInfo.td @@ -77,7 +77,8 @@ def CondMovFP : FPFormat<6>; def SpecialFP : FPFormat<7>; -class X86Inst opcod, Format f, ImmType i, dag ops, string AsmStr> : Instruction { +class X86Inst opcod, Format f, ImmType i, dag ops, string AsmStr> + : Instruction { let Namespace = "X86"; bits<8> Opcode = opcod; @@ -123,11 +124,14 @@ class DF { bits<4> Prefix = 10; } //===----------------------------------------------------------------------===// // Instruction templates... -class I o, Format f, dag ops, string asm> : X86Inst; - -class Ii8 o, Format f, dag ops, string asm> : X86Inst; -class Ii16 o, Format f, dag ops, string asm> : X86Inst; -class Ii32 o, Format f, dag ops, string asm> : X86Inst; +class I o, Format f, dag ops, string asm> + : X86Inst; +class Ii8 o, Format f, dag ops, string asm> + : X86Inst; +class Ii16 o, Format f, dag ops, string asm> + : X86Inst; +class Ii32 o, Format f, dag ops, string asm> + : X86Inst; //===----------------------------------------------------------------------===// // Instruction list... @@ -190,42 +194,61 @@ let isCall = 1 in def LEAVE : I<0xC9, RawFrm, (ops), "leave">, Imp<[EBP,ESP],[EBP,ESP]>; def POP32r : I<0x58, AddRegFrm, - (ops R32:$reg), "pop $reg">, Imp<[ESP],[ESP]>; + (ops R32:$reg), "pop{l} $reg">, Imp<[ESP],[ESP]>; -let isTwoAddress = 1 in // R32 = bswap R32 +let isTwoAddress = 1 in // R32 = bswap R32 def BSWAP32r : I<0xC8, AddRegFrm, - (ops R32:$dst, R32:$src), "bswap $dst">, TB; + (ops R32:$dst, R32:$src), "bswap{l} $dst">, TB; def XCHG8rr : I<0x86, MRMDestReg, // xchg R8, R8 - (ops R8:$src1, R8:$src2), "xchg $src1, $src2">; + (ops R8:$src1, R8:$src2), + "xchg{b} {$src2|$src1}, {$src1|$src2}">; def XCHG16rr : I<0x87, MRMDestReg, // xchg R16, R16 - (ops R16:$src1, R16:$src2), "xchg $src1, $src2">, OpSize; + (ops R16:$src1, R16:$src2), + "xchg{w} {$src2|$src1}, {$src1|$src2}">, OpSize; def XCHG32rr : I<0x87, MRMDestReg, // xchg R32, R32 - (ops R32:$src1, R32:$src2), "xchg $src1, $src2">; - -def XCHG8mr : I<0x86, MRMDestMem, (ops i8mem:$src1, R8:$src2), "xchg $src1, $src2">; -def XCHG16mr : I<0x87, MRMDestMem, (ops i16mem:$src1, R16:$src2), "xchg $src1, $src2">, OpSize; -def XCHG32mr : I<0x87, MRMDestMem, (ops i32mem:$src1, R32:$src2), "xchg $src1, $src2">; -def XCHG8rm : I<0x86, MRMSrcMem , (ops R8:$src1, i8mem:$src2), "xchg $src1, $src2">; -def XCHG16rm : I<0x87, MRMSrcMem , (ops R16:$src1, i16mem:$src2), "xchg $src1, $src2">, OpSize; -def XCHG32rm : I<0x87, MRMSrcMem , (ops R32:$src1, i32mem:$src2), "xchg $src1, $src2">; - -def LEA16r : I<0x8D, MRMSrcMem, (ops R16:$dst, i32mem:$src), "lea $dst, $src">, OpSize; -def LEA32r : I<0x8D, MRMSrcMem, (ops R32:$dst, i32mem:$src), "lea $dst, $src">; - - -def REP_MOVSB : I<0xA4, RawFrm, (ops), "rep movsb">, + (ops R32:$src1, R32:$src2), + "xchg{l} {$src2|$src1}, {$src1|$src2}">; + +def XCHG8mr : I<0x86, MRMDestMem, + (ops i8mem:$src1, R8:$src2), + "xchg{b} {$src2|$src1}, {$src1|$src2}">; +def XCHG16mr : I<0x87, MRMDestMem, + (ops i16mem:$src1, R16:$src2), + "xchg{w} {$src2|$src1}, {$src1|$src2}">, OpSize; +def XCHG32mr : I<0x87, MRMDestMem, + (ops i32mem:$src1, R32:$src2), + "xchg{l} {$src2|$src1}, {$src1|$src2}">; +def XCHG8rm : I<0x86, MRMSrcMem, + (ops R8:$src1, i8mem:$src2), + "xchg{b} {$src2|$src1}, {$src1|$src2}">; +def XCHG16rm : I<0x87, MRMSrcMem, + (ops R16:$src1, i16mem:$src2), + "xchg{w} {$src2|$src1}, {$src1|$src2}">, OpSize; +def XCHG32rm : I<0x87, MRMSrcMem, + (ops R32:$src1, i32mem:$src2), + "xchg{l} {$src2|$src1}, {$src1|$src2}">; + +def LEA16r : I<0x8D, MRMSrcMem, + (ops R16:$dst, i32mem:$src), + "lea{w} {$src|$dst}, {$dst|$src}">, OpSize; +def LEA32r : I<0x8D, MRMSrcMem, + (ops R32:$dst, i32mem:$src), + "lea{l} {$src|$dst}, {$dst|$src}">; + + +def REP_MOVSB : I<0xA4, RawFrm, (ops), "{rep;movsb|rep movsb}">, Imp<[ECX,EDI,ESI], [ECX,EDI,ESI]>, REP; -def REP_MOVSW : I<0xA5, RawFrm, (ops), "rep movsw">, +def REP_MOVSW : I<0xA5, RawFrm, (ops), "{rep;movsw|rep movsw}">, Imp<[ECX,EDI,ESI], [ECX,EDI,ESI]>, REP, OpSize; -def REP_MOVSD : I<0xA5, RawFrm, (ops), "rep movsd">, +def REP_MOVSD : I<0xA5, RawFrm, (ops), "{rep;movsd|rep movsd}">, Imp<[ECX,EDI,ESI], [ECX,EDI,ESI]>, REP; -def REP_STOSB : I<0xAA, RawFrm, (ops), "rep stosb">, +def REP_STOSB : I<0xAA, RawFrm, (ops), "{rep;stosb|rep stosb}">, Imp<[AL,ECX,EDI], [ECX,EDI]>, REP; -def REP_STOSW : I<0xAB, RawFrm, (ops), "rep stosw">, +def REP_STOSW : I<0xAB, RawFrm, (ops), "{rep;stosw|rep stosw}">, Imp<[AX,ECX,EDI], [ECX,EDI]>, REP, OpSize; -def REP_STOSD : I<0xAB, RawFrm, (ops), "rep stosd">, +def REP_STOSD : I<0xAB, RawFrm, (ops), "{rep;stosd|rep stosd}">, Imp<[EAX,ECX,EDI], [ECX,EDI]>, REP; @@ -233,98 +256,122 @@ def REP_STOSD : I<0xAB, RawFrm, (ops), "rep stosd">, // Input/Output Instructions... // def IN8rr : I<0xEC, RawFrm, (ops), - "in %AL, %DX">, Imp<[DX], [AL]>; + "in{b} {%DX, %AL|AL, DX}">, Imp<[DX], [AL]>; def IN16rr : I<0xED, RawFrm, (ops), - "in %AX, %DX">, Imp<[DX], [AX]>, OpSize; + "in{w} {%DX, %AX|AX, DX}">, Imp<[DX], [AX]>, OpSize; def IN32rr : I<0xED, RawFrm, (ops), - "in %EAX, %DX">, Imp<[DX],[EAX]>; + "in{l} {%DX, %EAX|EAX, DX}">, Imp<[DX],[EAX]>; def IN8ri : Ii16<0xE4, RawFrm, (ops i16imm:$port), - "in %AL, $port">, Imp<[], [AL]>; + "in{b} {$port, %AL|AL, $port}">, Imp<[], [AL]>; def IN16ri : Ii16<0xE5, RawFrm, (ops i16imm:$port), - "in %AX, $port">, Imp<[], [AX]>, OpSize; + "in{w} {$port, %AX|AX, $port}">, Imp<[], [AX]>, OpSize; def IN32ri : Ii16<0xE5, RawFrm, (ops i16imm:$port), - "in %EAX, $port">, Imp<[],[EAX]>; + "in{l} {$port, %EAX|EAX, $port}">, Imp<[],[EAX]>; def OUT8rr : I<0xEE, RawFrm, (ops), - "out %DX, %AL">, Imp<[DX, AL], []>; + "out{b} {%AL, %DX|DX, AL}">, Imp<[DX, AL], []>; def OUT16rr : I<0xEF, RawFrm, (ops), - "out %DX, %AX">, Imp<[DX, AX], []>, OpSize; + "out{w} {%AX, %DX|DX, AX}">, Imp<[DX, AX], []>, OpSize; def OUT32rr : I<0xEF, RawFrm, (ops), - "out %DX, %EAX">, Imp<[DX, EAX], []>; + "out{l} {%EAX, %DX|DX, EAX}">, Imp<[DX, EAX], []>; def OUT8ir : Ii16<0xE6, RawFrm, (ops i16imm:$port), - "out $port, %AL">, Imp<[AL], []>; + "out{b} {%AL, $port|$port, AL}">, Imp<[AL], []>; def OUT16ir : Ii16<0xE7, RawFrm, (ops i16imm:$port), - "out $port, %AX">, Imp<[AX], []>, OpSize; + "out{w} {%AX, $port|$port, AX}">, Imp<[AX], []>, OpSize; def OUT32ir : Ii16<0xE7, RawFrm, (ops i16imm:$port), - "out $port, %EAX">, Imp<[EAX], []>; + "out{l} {%EAX, $port|$port, %EAX}">, Imp<[EAX], []>; //===----------------------------------------------------------------------===// // Move Instructions... // -def MOV8rr : I<0x88, MRMDestReg, (ops R8 :$dst, R8 :$src), "mov $dst, $src">; -def MOV16rr : I<0x89, MRMDestReg, (ops R16:$dst, R16:$src), "mov $dst, $src">, OpSize; -def MOV32rr : I<0x89, MRMDestReg, (ops R32:$dst, R32:$src), "mov $dst, $src">; -def MOV8ri : Ii8 <0xB0, AddRegFrm, (ops R8 :$dst, i8imm :$src), "mov $dst, $src">; -def MOV16ri : Ii16<0xB8, AddRegFrm, (ops R16:$dst, i16imm:$src), "mov $dst, $src">, OpSize; -def MOV32ri : Ii32<0xB8, AddRegFrm, (ops R32:$dst, i32imm:$src), "mov $dst, $src">; -def MOV8mi : Ii8 <0xC6, MRM0m, (ops i8mem :$dst, i8imm :$src), "mov $dst, $src">; -def MOV16mi : Ii16<0xC7, MRM0m, (ops i16mem:$dst, i16imm:$src), "mov $dst, $src">, OpSize; -def MOV32mi : Ii32<0xC7, MRM0m, (ops i32mem:$dst, i32imm:$src), "mov $dst, $src">; - -def MOV8rm : I<0x8A, MRMSrcMem, (ops R8 :$dst, i8mem :$src), "mov $dst, $src">; -def MOV16rm : I<0x8B, MRMSrcMem, (ops R16:$dst, i16mem:$src), "mov $dst, $src">, OpSize; -def MOV32rm : I<0x8B, MRMSrcMem, (ops R32:$dst, i32mem:$src), "mov $dst, $src">; - -def MOV8mr : I<0x88, MRMDestMem, (ops i8mem :$dst, R8 :$src), "mov $dst, $src">; -def MOV16mr : I<0x89, MRMDestMem, (ops i16mem:$dst, R16:$src), "mov $dst, $src">, OpSize; -def MOV32mr : I<0x89, MRMDestMem, (ops i32mem:$dst, R32:$src), "mov $dst, $src">; +def MOV8rr : I<0x88, MRMDestReg, (ops R8 :$dst, R8 :$src), + "mov{b} {$src, $dst|$dst, $src}">; +def MOV16rr : I<0x89, MRMDestReg, (ops R16:$dst, R16:$src), + "mov{w} {$src, $dst|$dst, $src}">, OpSize; +def MOV32rr : I<0x89, MRMDestReg, (ops R32:$dst, R32:$src), + "mov{l} {$src, $dst|$dst, $src}">; +def MOV8ri : Ii8 <0xB0, AddRegFrm, (ops R8 :$dst, i8imm :$src), + "mov{b} {$src, $dst|$dst, $src}">; +def MOV16ri : Ii16<0xB8, AddRegFrm, (ops R16:$dst, i16imm:$src), + "mov{w} {$src, $dst|$dst, $src}">, OpSize; +def MOV32ri : Ii32<0xB8, AddRegFrm, (ops R32:$dst, i32imm:$src), + "mov{l} {$src, $dst|$dst, $src}">; +def MOV8mi : Ii8 <0xC6, MRM0m, (ops i8mem :$dst, i8imm :$src), + "mov{b} {$src, $dst|$dst, $src}">; +def MOV16mi : Ii16<0xC7, MRM0m, (ops i16mem:$dst, i16imm:$src), + "mov{w} {$src, $dst|$dst, $src}">, OpSize; +def MOV32mi : Ii32<0xC7, MRM0m, (ops i32mem:$dst, i32imm:$src), + "mov{l} {$src, $dst|$dst, $src}">; + +def MOV8rm : I<0x8A, MRMSrcMem, (ops R8 :$dst, i8mem :$src), + "mov{b} {$src, $dst|$dst, $src}">; +def MOV16rm : I<0x8B, MRMSrcMem, (ops R16:$dst, i16mem:$src), + "mov{w} {$src, $dst|$dst, $src}">, OpSize; +def MOV32rm : I<0x8B, MRMSrcMem, (ops R32:$dst, i32mem:$src), + "mov{l} {$src, $dst|$dst, $src}">; + +def MOV8mr : I<0x88, MRMDestMem, (ops i8mem :$dst, R8 :$src), + "mov{b} {$src, $dst|$dst, $src}">; +def MOV16mr : I<0x89, MRMDestMem, (ops i16mem:$dst, R16:$src), + "mov{w} {$src, $dst|$dst, $src}">, OpSize; +def MOV32mr : I<0x89, MRMDestMem, (ops i32mem:$dst, R32:$src), + "mov{l} {$src, $dst|$dst, $src}">; //===----------------------------------------------------------------------===// // Fixed-Register Multiplication and Division Instructions... // // Extra precision multiplication -def MUL8r : I<0xF6, MRM4r, (ops R8:$src), "mul $src">, +def MUL8r : I<0xF6, MRM4r, (ops R8:$src), "mul{b} $src">, Imp<[AL],[AX]>; // AL,AH = AL*R8 -def MUL16r : I<0xF7, MRM4r, (ops R16:$src), "mul $src">, +def MUL16r : I<0xF7, MRM4r, (ops R16:$src), "mul{w} $src">, Imp<[AX],[AX,DX]>, OpSize; // AX,DX = AX*R16 -def MUL32r : I<0xF7, MRM4r, (ops R32:$src), "mul $src">, +def MUL32r : I<0xF7, MRM4r, (ops R32:$src), "mul{l} $src">, Imp<[EAX],[EAX,EDX]>; // EAX,EDX = EAX*R32 def MUL8m : I<0xF6, MRM4m, (ops i8mem :$src), - "mul $src">, Imp<[AL],[AX]>; // AL,AH = AL*[mem8] + "mul{b} $src">, Imp<[AL],[AX]>; // AL,AH = AL*[mem8] def MUL16m : I<0xF7, MRM4m, (ops i16mem:$src), - "mul $src">, Imp<[AX],[AX,DX]>, OpSize; // AX,DX = AX*[mem16] + "mul{w} $src">, Imp<[AX],[AX,DX]>, OpSize; // AX,DX = AX*[mem16] def MUL32m : I<0xF7, MRM4m, (ops i32mem:$src), - "mul $src">, Imp<[EAX],[EAX,EDX]>; // EAX,EDX = EAX*[mem32] + "mul{l} $src">, Imp<[EAX],[EAX,EDX]>; // EAX,EDX = EAX*[mem32] // unsigned division/remainder -def DIV8r : I<0xF6, MRM6r, (ops R8:$src), "div $src">, - Imp<[AX],[AX]>; // AX/r8 = AL,AH -def DIV16r : I<0xF7, MRM6r, (ops R16:$src), "div $src">, - Imp<[AX,DX],[AX,DX]>, OpSize; // DX:AX/r16 = AX,DX -def DIV32r : I<0xF7, MRM6r, (ops R32:$src), "div $src">, - Imp<[EAX,EDX],[EAX,EDX]>; // EDX:EAX/r32 = EAX,EDX -def DIV8m : I<0xF6, MRM6m, (ops i8mem:$src), "div $src">, Imp<[AX],[AX]>; // AX/[mem8] = AL,AH -def DIV16m : I<0xF7, MRM6m, (ops i16mem:$src), "div $src">, Imp<[AX,DX],[AX,DX]>, OpSize; // DX:AX/[mem16] = AX,DX -def DIV32m : I<0xF7, MRM6m, (ops i32mem:$src), "div $src">, Imp<[EAX,EDX],[EAX,EDX]>; // EDX:EAX/[mem32] = EAX,EDX +def DIV8r : I<0xF6, MRM6r, (ops R8:$src), // AX/r8 = AL,AH + "div{b} $src">, Imp<[AX],[AX]>; +def DIV16r : I<0xF7, MRM6r, (ops R16:$src), // DX:AX/r16 = AX,DX + "div{w} $src">, Imp<[AX,DX],[AX,DX]>, OpSize; +def DIV32r : I<0xF7, MRM6r, (ops R32:$src), // EDX:EAX/r32 = EAX,EDX + "div{l} $src">, Imp<[EAX,EDX],[EAX,EDX]>; +def DIV8m : I<0xF6, MRM6m, (ops i8mem:$src), // AX/[mem8] = AL,AH + "div{b} $src">, Imp<[AX],[AX]>; +def DIV16m : I<0xF7, MRM6m, (ops i16mem:$src), // DX:AX/[mem16] = AX,DX + "div{w} $src">, Imp<[AX,DX],[AX,DX]>, OpSize; +def DIV32m : I<0xF7, MRM6m, (ops i32mem:$src), // EDX:EAX/[mem32] = EAX,EDX + "div{l} $src">, Imp<[EAX,EDX],[EAX,EDX]>; // Signed division/remainder. -def IDIV8r : I<0xF6, MRM7r, (ops R8:$src), "idiv $src">, - Imp<[AX],[AX]>; // AX/r8 = AL,AH -def IDIV16r: I<0xF7, MRM7r, (ops R16:$src), "idiv $src">, - Imp<[AX,DX],[AX,DX]>, OpSize; // DX:AX/r16 = AX,DX -def IDIV32r: I<0xF7, MRM7r, (ops R32:$src), "idiv $src">, - Imp<[EAX,EDX],[EAX,EDX]>; // EDX:EAX/r32 = EAX,EDX -def IDIV8m : I<0xF6, MRM7m, (ops i8mem:$src), "idiv $src">, Imp<[AX],[AX]>; // AX/[mem8] = AL,AH -def IDIV16m: I<0xF7, MRM7m, (ops i16mem:$src), "idiv $src">, Imp<[AX,DX],[AX,DX]>, OpSize; // DX:AX/[mem16] = AX,DX -def IDIV32m: I<0xF7, MRM7m, (ops i32mem:$src), "idiv $src">, Imp<[EAX,EDX],[EAX,EDX]>; // EDX:EAX/[mem32] = EAX,EDX +def IDIV8r : I<0xF6, MRM7r, (ops R8:$src), // AX/r8 = AL,AH + "idiv{b} $src">, Imp<[AX],[AX]>; +def IDIV16r: I<0xF7, MRM7r, (ops R16:$src), // DX:AX/r16 = AX,DX + "idiv{w} $src">, Imp<[AX,DX],[AX,DX]>, OpSize; +def IDIV32r: I<0xF7, MRM7r, (ops R32:$src), // EDX:EAX/r32 = EAX,EDX + "idiv{l} $src">, Imp<[EAX,EDX],[EAX,EDX]>; +def IDIV8m : I<0xF6, MRM7m, (ops i8mem:$src), // AX/[mem8] = AL,AH + "idiv{b} $src">, Imp<[AX],[AX]>; +def IDIV16m: I<0xF7, MRM7m, (ops i16mem:$src), // DX:AX/[mem16] = AX,DX + "idiv{w} $src">, Imp<[AX,DX],[AX,DX]>, OpSize; +def IDIV32m: I<0xF7, MRM7m, (ops i32mem:$src), // EDX:EAX/[mem32] = EAX,EDX + "idiv{l} $src">, Imp<[EAX,EDX],[EAX,EDX]>; // Sign-extenders for division. -def CBW : I<0x98, RawFrm, (ops), "cbw">, Imp<[AL],[AH]>; // AX = signext(AL) -def CWD : I<0x99, RawFrm, (ops), "cwd">, Imp<[AX],[DX]>; // DX:AX = signext(AX) -def CDQ : I<0x99, RawFrm, (ops), "cdq">, Imp<[EAX],[EDX]>; // EDX:EAX = signext(EAX) +def CBW : I<0x98, RawFrm, (ops), + "{cbtw|cbw}">, Imp<[AL],[AH]>; // AX = signext(AL) +def CWD : I<0x99, RawFrm, (ops), + "{cwtl|cwd}">, Imp<[AX],[DX]>; // DX:AX = signext(AX) +def CDQ : I<0x99, RawFrm, (ops), + "{cltd|cdq}">, Imp<[EAX],[EDX]>; // EDX:EAX = signext(EAX) //===----------------------------------------------------------------------===// @@ -333,469 +380,681 @@ def CDQ : I<0x99, RawFrm, (ops), "cdq">, Imp<[EAX],[EDX]>; // EDX:EAX = signext( let isTwoAddress = 1 in { // Conditional moves -def CMOVB16rr : I<0x42, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovb $dst, $src2">, TB, OpSize; // if , TB, OpSize; // if , TB; // if , TB; // if , TB, OpSize; // if >=u, R16 = R16 -def CMOVAE16rm: I<0x43, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovae $dst, $src2">, TB, OpSize; // if >=u, R16 = [mem16] -def CMOVAE32rr: I<0x43, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovae $dst, $src2">, TB; // if >=u, R32 = R32 -def CMOVAE32rm: I<0x43, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovae $dst, $src2">, TB; // if >=u, R32 = [mem32] - -def CMOVE16rr : I<0x44, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmove $dst, $src2">, TB, OpSize; // if ==, R16 = R16 -def CMOVE16rm : I<0x44, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmove $dst, $src2">, TB, OpSize; // if ==, R16 = [mem16] -def CMOVE32rr : I<0x44, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmove $dst, $src2">, TB; // if ==, R32 = R32 -def CMOVE32rm : I<0x44, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmove $dst, $src2">, TB; // if ==, R32 = [mem32] - -def CMOVNE16rr: I<0x45, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovne $dst, $src2">, TB, OpSize; // if !=, R16 = R16 -def CMOVNE16rm: I<0x45, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovne $dst, $src2">, TB, OpSize; // if !=, R16 = [mem16] -def CMOVNE32rr: I<0x45, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovne $dst, $src2">, TB; // if !=, R32 = R32 -def CMOVNE32rm: I<0x45, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovne $dst, $src2">, TB; // if !=, R32 = [mem32] - -def CMOVBE16rr: I<0x46, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovbe $dst, $src2">, TB, OpSize; // if <=u, R16 = R16 -def CMOVBE16rm: I<0x46, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovbe $dst, $src2">, TB, OpSize; // if <=u, R16 = [mem16] -def CMOVBE32rr: I<0x46, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovbe $dst, $src2">, TB; // if <=u, R32 = R32 -def CMOVBE32rm: I<0x46, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovbe $dst, $src2">, TB; // if <=u, R32 = [mem32] - -def CMOVA16rr : I<0x47, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmova $dst, $src2">, TB, OpSize; // if >u, R16 = R16 -def CMOVA16rm : I<0x47, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmova $dst, $src2">, TB, OpSize; // if >u, R16 = [mem16] -def CMOVA32rr : I<0x47, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmova $dst, $src2">, TB; // if >u, R32 = R32 -def CMOVA32rm : I<0x47, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmova $dst, $src2">, TB; // if >u, R32 = [mem32] - -def CMOVS16rr : I<0x48, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovs $dst, $src2">, TB, OpSize; // if signed, R16 = R16 -def CMOVS16rm : I<0x48, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovs $dst, $src2">, TB, OpSize; // if signed, R16 = [mem16] -def CMOVS32rr : I<0x48, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovs $dst, $src2">, TB; // if signed, R32 = R32 -def CMOVS32rm : I<0x48, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovs $dst, $src2">, TB; // if signed, R32 = [mem32] - -def CMOVNS16rr: I<0x49, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovns $dst, $src2">, TB, OpSize; // if !signed, R16 = R16 -def CMOVNS16rm: I<0x49, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovns $dst, $src2">, TB, OpSize; // if !signed, R16 = [mem16] -def CMOVNS32rr: I<0x49, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovns $dst, $src2">, TB; // if !signed, R32 = R32 -def CMOVNS32rm: I<0x49, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovns $dst, $src2">, TB; // if !signed, R32 = [mem32] - -def CMOVL16rr : I<0x4C, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovl $dst, $src2">, TB, OpSize; // if , TB, OpSize; // if , TB; // if , TB; // if , TB, OpSize; // if >=s, R16 = R16 -def CMOVGE16rm: I<0x4D, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovge $dst, $src2">, TB, OpSize; // if >=s, R16 = [mem16] -def CMOVGE32rr: I<0x4D, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovge $dst, $src2">, TB; // if >=s, R32 = R32 -def CMOVGE32rm: I<0x4D, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovge $dst, $src2">, TB; // if >=s, R32 = [mem32] - -def CMOVLE16rr: I<0x4E, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovle $dst, $src2">, TB, OpSize; // if <=s, R16 = R16 -def CMOVLE16rm: I<0x4E, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovle $dst, $src2">, TB, OpSize; // if <=s, R16 = [mem16] -def CMOVLE32rr: I<0x4E, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovle $dst, $src2">, TB; // if <=s, R32 = R32 -def CMOVLE32rm: I<0x4E, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovle $dst, $src2">, TB; // if <=s, R32 = [mem32] - -def CMOVG16rr : I<0x4F, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), - "cmovg $dst, $src2">, TB, OpSize; // if >s, R16 = R16 -def CMOVG16rm : I<0x4F, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), - "cmovg $dst, $src2">, TB, OpSize; // if >s, R16 = [mem16] -def CMOVG32rr : I<0x4F, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), - "cmovg $dst, $src2">, TB; // if >s, R32 = R32 -def CMOVG32rm : I<0x4F, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "cmovg $dst, $src2">, TB; // if >s, R32 = [mem32] +def CMOVB16rr : I<0x42, MRMSrcReg, // if , TB, OpSize; +def CMOVB16rm : I<0x42, MRMSrcMem, // if , TB, OpSize; +def CMOVB32rr : I<0x42, MRMSrcReg, // if , TB; +def CMOVB32rm : I<0x42, MRMSrcMem, // if , TB; + +def CMOVAE16rr: I<0x43, MRMSrcReg, // if >=u, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovae {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVAE16rm: I<0x43, MRMSrcMem, // if >=u, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovae {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVAE32rr: I<0x43, MRMSrcReg, // if >=u, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovae {$src2, $dst|$dst, $src2}">, TB; +def CMOVAE32rm: I<0x43, MRMSrcMem, // if >=u, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovae {$src2, $dst|$dst, $src2}">, TB; + +def CMOVE16rr : I<0x44, MRMSrcReg, // if ==, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmove {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVE16rm : I<0x44, MRMSrcMem, // if ==, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmove {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVE32rr : I<0x44, MRMSrcReg, // if ==, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmove {$src2, $dst|$dst, $src2}">, TB; +def CMOVE32rm : I<0x44, MRMSrcMem, // if ==, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmove {$src2, $dst|$dst, $src2}">, TB; + +def CMOVNE16rr: I<0x45, MRMSrcReg, // if !=, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovne {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVNE16rm: I<0x45, MRMSrcMem, // if !=, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovne {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVNE32rr: I<0x45, MRMSrcReg, // if !=, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovne {$src2, $dst|$dst, $src2}">, TB; +def CMOVNE32rm: I<0x45, MRMSrcMem, // if !=, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovne {$src2, $dst|$dst, $src2}">, TB; + +def CMOVBE16rr: I<0x46, MRMSrcReg, // if <=u, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovbe {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVBE16rm: I<0x46, MRMSrcMem, // if <=u, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovbe {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVBE32rr: I<0x46, MRMSrcReg, // if <=u, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovbe {$src2, $dst|$dst, $src2}">, TB; +def CMOVBE32rm: I<0x46, MRMSrcMem, // if <=u, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovbe {$src2, $dst|$dst, $src2}">, TB; + +def CMOVA16rr : I<0x47, MRMSrcReg, // if >u, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmova {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVA16rm : I<0x47, MRMSrcMem, // if >u, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmova {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVA32rr : I<0x47, MRMSrcReg, // if >u, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmova {$src2, $dst|$dst, $src2}">, TB; +def CMOVA32rm : I<0x47, MRMSrcMem, // if >u, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmova {$src2, $dst|$dst, $src2}">, TB; + +def CMOVS16rr : I<0x48, MRMSrcReg, // if signed, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovs {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVS16rm : I<0x48, MRMSrcMem, // if signed, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovs {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVS32rr : I<0x48, MRMSrcReg, // if signed, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovs {$src2, $dst|$dst, $src2}">, TB; +def CMOVS32rm : I<0x48, MRMSrcMem, // if signed, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovs {$src2, $dst|$dst, $src2}">, TB; + +def CMOVNS16rr: I<0x49, MRMSrcReg, // if !signed, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovns {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVNS16rm: I<0x49, MRMSrcMem, // if !signed, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovns {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVNS32rr: I<0x49, MRMSrcReg, // if !signed, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovns {$src2, $dst|$dst, $src2}">, TB; +def CMOVNS32rm: I<0x49, MRMSrcMem, // if !signed, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovns {$src2, $dst|$dst, $src2}">, TB; + +def CMOVL16rr : I<0x4C, MRMSrcReg, // if , TB, OpSize; +def CMOVL16rm : I<0x4C, MRMSrcMem, // if , TB, OpSize; +def CMOVL32rr : I<0x4C, MRMSrcReg, // if , TB; +def CMOVL32rm : I<0x4C, MRMSrcMem, // if , TB; + +def CMOVGE16rr: I<0x4D, MRMSrcReg, // if >=s, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovge {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVGE16rm: I<0x4D, MRMSrcMem, // if >=s, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovge {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVGE32rr: I<0x4D, MRMSrcReg, // if >=s, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovge {$src2, $dst|$dst, $src2}">, TB; +def CMOVGE32rm: I<0x4D, MRMSrcMem, // if >=s, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovge {$src2, $dst|$dst, $src2}">, TB; + +def CMOVLE16rr: I<0x4E, MRMSrcReg, // if <=s, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovle {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVLE16rm: I<0x4E, MRMSrcMem, // if <=s, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovle {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVLE32rr: I<0x4E, MRMSrcReg, // if <=s, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovle {$src2, $dst|$dst, $src2}">, TB; +def CMOVLE32rm: I<0x4E, MRMSrcMem, // if <=s, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovle {$src2, $dst|$dst, $src2}">, TB; + +def CMOVG16rr : I<0x4F, MRMSrcReg, // if >s, R16 = R16 + (ops R16:$dst, R16:$src1, R16:$src2), + "cmovg {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVG16rm : I<0x4F, MRMSrcMem, // if >s, R16 = [mem16] + (ops R16:$dst, R16:$src1, i16mem:$src2), + "cmovg {$src2, $dst|$dst, $src2}">, TB, OpSize; +def CMOVG32rr : I<0x4F, MRMSrcReg, // if >s, R32 = R32 + (ops R32:$dst, R32:$src1, R32:$src2), + "cmovg {$src2, $dst|$dst, $src2}">, TB; +def CMOVG32rm : I<0x4F, MRMSrcMem, // if >s, R32 = [mem32] + (ops R32:$dst, R32:$src1, i32mem:$src2), + "cmovg {$src2, $dst|$dst, $src2}">, TB; // unary instructions -def NEG8r : I<0xF6, MRM3r, (ops R8 :$dst, R8 :$src), "neg $dst">; -def NEG16r : I<0xF7, MRM3r, (ops R16:$dst, R16:$src), "neg $dst">, OpSize; -def NEG32r : I<0xF7, MRM3r, (ops R32:$dst, R32:$src), "neg $dst">; +def NEG8r : I<0xF6, MRM3r, (ops R8 :$dst, R8 :$src), "neg{b} $dst">; +def NEG16r : I<0xF7, MRM3r, (ops R16:$dst, R16:$src), "neg{w} $dst">, OpSize; +def NEG32r : I<0xF7, MRM3r, (ops R32:$dst, R32:$src), "neg{l} $dst">; let isTwoAddress = 0 in { - def NEG8m : I<0xF6, MRM3m, (ops i8mem :$dst), "neg $dst">; - def NEG16m : I<0xF7, MRM3m, (ops i16mem:$dst), "neg $dst">, OpSize; - def NEG32m : I<0xF7, MRM3m, (ops i32mem:$dst), "neg $dst">; + def NEG8m : I<0xF6, MRM3m, (ops i8mem :$dst), "neg{b} $dst">; + def NEG16m : I<0xF7, MRM3m, (ops i16mem:$dst), "neg{w} $dst">, OpSize; + def NEG32m : I<0xF7, MRM3m, (ops i32mem:$dst), "neg{l} $dst">; } -def NOT8r : I<0xF6, MRM2r, (ops R8 :$dst, R8 :$src), "not $dst">; -def NOT16r : I<0xF7, MRM2r, (ops R16:$dst, R16:$src), "not $dst">, OpSize; -def NOT32r : I<0xF7, MRM2r, (ops R32:$dst, R32:$src), "not $dst">; +def NOT8r : I<0xF6, MRM2r, (ops R8 :$dst, R8 :$src), "not{b} $dst">; +def NOT16r : I<0xF7, MRM2r, (ops R16:$dst, R16:$src), "not{w} $dst">, OpSize; +def NOT32r : I<0xF7, MRM2r, (ops R32:$dst, R32:$src), "not{l} $dst">; let isTwoAddress = 0 in { - def NOT8m : I<0xF6, MRM2m, (ops i8mem :$dst), "not $dst">; - def NOT16m : I<0xF7, MRM2m, (ops i16mem:$dst), "not $dst">, OpSize; - def NOT32m : I<0xF7, MRM2m, (ops i32mem:$dst), "not $dst">; + def NOT8m : I<0xF6, MRM2m, (ops i8mem :$dst), "not{b} $dst">; + def NOT16m : I<0xF7, MRM2m, (ops i16mem:$dst), "not{w} $dst">, OpSize; + def NOT32m : I<0xF7, MRM2m, (ops i32mem:$dst), "not{l} $dst">; } -def INC8r : I<0xFE, MRM0r, (ops R8 :$dst, R8 :$src), "inc $dst">; -def INC16r : I<0xFF, MRM0r, (ops R16:$dst, R16:$src), "inc $dst">, OpSize; -def INC32r : I<0xFF, MRM0r, (ops R32:$dst, R32:$src), "inc $dst">; +def INC8r : I<0xFE, MRM0r, (ops R8 :$dst, R8 :$src), "inc{b} $dst">; +def INC16r : I<0xFF, MRM0r, (ops R16:$dst, R16:$src), "inc{w} $dst">, OpSize; +def INC32r : I<0xFF, MRM0r, (ops R32:$dst, R32:$src), "inc{l} $dst">; let isTwoAddress = 0 in { - def INC8m : I<0xFE, MRM0m, (ops i8mem :$dst), "inc $dst">; - def INC16m : I<0xFF, MRM0m, (ops i16mem:$dst), "inc $dst">, OpSize; - def INC32m : I<0xFF, MRM0m, (ops i32mem:$dst), "inc $dst">; + def INC8m : I<0xFE, MRM0m, (ops i8mem :$dst), "inc{b} $dst">; + def INC16m : I<0xFF, MRM0m, (ops i16mem:$dst), "inc{w} $dst">, OpSize; + def INC32m : I<0xFF, MRM0m, (ops i32mem:$dst), "inc{l} $dst">; } -def DEC8r : I<0xFE, MRM1r, (ops R8 :$dst, R8 :$src), "dec $dst">; -def DEC16r : I<0xFF, MRM1r, (ops R16:$dst, R16:$src), "dec $dst">, OpSize; -def DEC32r : I<0xFF, MRM1r, (ops R32:$dst, R32:$src), "dec $dst">; +def DEC8r : I<0xFE, MRM1r, (ops R8 :$dst, R8 :$src), "dec{b} $dst">; +def DEC16r : I<0xFF, MRM1r, (ops R16:$dst, R16:$src), "dec{w} $dst">, OpSize; +def DEC32r : I<0xFF, MRM1r, (ops R32:$dst, R32:$src), "dec{l} $dst">; let isTwoAddress = 0 in { - def DEC8m : I<0xFE, MRM1m, (ops i8mem :$dst), "dec $dst">; - def DEC16m : I<0xFF, MRM1m, (ops i16mem:$dst), "dec $dst">, OpSize; - def DEC32m : I<0xFF, MRM1m, (ops i32mem:$dst), "dec $dst">; + def DEC8m : I<0xFE, MRM1m, (ops i8mem :$dst), "dec{b} $dst">; + def DEC16m : I<0xFF, MRM1m, (ops i16mem:$dst), "dec{w} $dst">, OpSize; + def DEC32m : I<0xFF, MRM1m, (ops i32mem:$dst), "dec{l} $dst">; } // Logical operators... -def AND8rr : I<0x20, MRMDestReg, (ops R8 :$dst, R8 :$src1, R8 :$src2), "and $dst, $src2">; -def AND16rr : I<0x21, MRMDestReg, (ops R16:$dst, R16:$src1, R16:$src2), "and $dst, $src2">, OpSize; -def AND32rr : I<0x21, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), "and $dst, $src2">; - -def AND8rm : I<0x22, MRMSrcMem , (ops R8 :$dst, R8 :$src1, i8mem :$src2), "and $dst, $src2">; -def AND16rm : I<0x23, MRMSrcMem , (ops R16:$dst, R16:$src1, i16mem:$src2), "and $dst, $src2">, OpSize; -def AND32rm : I<0x23, MRMSrcMem , (ops R32:$dst, R32:$src1, i32mem:$src2), "and $dst, $src2">; - -def AND8ri : Ii8 <0x80, MRM4r, (ops R8 :$dst, R8 :$src1, i8imm :$src2), - "and $dst, $src2">; -def AND16ri : Ii16<0x81, MRM4r, (ops R16:$dst, R16:$src1, i16imm:$src2), - "and $dst, $src2">, OpSize; -def AND32ri : Ii32<0x81, MRM4r, (ops R32:$dst, R32:$src1, i32imm:$src2), - "and $dst, $src2">; -def AND16ri8 : Ii8<0x83, MRM4r, (ops R16:$dst, R16:$src1, i8imm:$src2), - "and $dst, $src2" >, OpSize; -def AND32ri8 : Ii8<0x83, MRM4r, (ops R32:$dst, R32:$src1, i8imm:$src2), - "and $dst, $src2">; +def AND8rr : I<0x20, MRMDestReg, + (ops R8 :$dst, R8 :$src1, R8 :$src2), + "and{b} {$src2, $dst|$dst, $src2}">; +def AND16rr : I<0x21, MRMDestReg, + (ops R16:$dst, R16:$src1, R16:$src2), + "and{w} {$src2, $dst|$dst, $src2}">, OpSize; +def AND32rr : I<0x21, MRMDestReg, + (ops R32:$dst, R32:$src1, R32:$src2), + "and{l} {$src2, $dst|$dst, $src2}">; + +def AND8rm : I<0x22, MRMSrcMem, + (ops R8 :$dst, R8 :$src1, i8mem :$src2), + "and{b} {$src2, $dst|$dst, $src2}">; +def AND16rm : I<0x23, MRMSrcMem, + (ops R16:$dst, R16:$src1, i16mem:$src2), + "and{w} {$src2, $dst|$dst, $src2}">, OpSize; +def AND32rm : I<0x23, MRMSrcMem, + (ops R32:$dst, R32:$src1, i32mem:$src2), + "and{l} {$src2, $dst|$dst, $src2}">; + +def AND8ri : Ii8<0x80, MRM4r, + (ops R8 :$dst, R8 :$src1, i8imm :$src2), + "and{b} {$src2, $dst|$dst, $src2}">; +def AND16ri : Ii16<0x81, MRM4r, + (ops R16:$dst, R16:$src1, i16imm:$src2), + "and{w} {$src2, $dst|$dst, $src2}">, OpSize; +def AND32ri : Ii32<0x81, MRM4r, + (ops R32:$dst, R32:$src1, i32imm:$src2), + "and{l} {$src2, $dst|$dst, $src2}">; +def AND16ri8 : Ii8<0x83, MRM4r, + (ops R16:$dst, R16:$src1, i8imm:$src2), + "and{w} {$src2, $dst|$dst, $src2}" >, OpSize; +def AND32ri8 : Ii8<0x83, MRM4r, + (ops R32:$dst, R32:$src1, i8imm:$src2), + "and{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def AND8mr : I<0x20, MRMDestMem, (ops i8mem :$dst, R8 :$src), "and $dst, $src">; - def AND16mr : I<0x21, MRMDestMem, (ops i16mem:$dst, R16:$src), "and $dst, $src">, OpSize; - def AND32mr : I<0x21, MRMDestMem, (ops i32mem:$dst, R32:$src), "and $dst, $src">; - def AND8mi : Ii8 <0x80, MRM4m, (ops i8mem :$dst, i8imm :$src), "and $dst, $src">; - def AND16mi : Ii16<0x81, MRM4m, (ops i16mem:$dst, i16imm:$src), "and $dst, $src">, OpSize; - def AND32mi : Ii32<0x81, MRM4m, (ops i32mem:$dst, i32imm:$src), "and $dst, $src">; - def AND16mi8 : Ii8 <0x83, MRM4m, (ops i16mem:$dst, i8imm :$src), "and $dst, $src">, OpSize; - def AND32mi8 : Ii8 <0x83, MRM4m, (ops i32mem:$dst, i8imm :$src), "and $dst, $src">; + def AND8mr : I<0x20, MRMDestMem, + (ops i8mem :$dst, R8 :$src), + "and{b} {$src, $dst|$dst, $src}">; + def AND16mr : I<0x21, MRMDestMem, + (ops i16mem:$dst, R16:$src), + "and{w} {$src, $dst|$dst, $src}">, OpSize; + def AND32mr : I<0x21, MRMDestMem, + (ops i32mem:$dst, R32:$src), + "and{l} {$src, $dst|$dst, $src}">; + def AND8mi : Ii8<0x80, MRM4m, + (ops i8mem :$dst, i8imm :$src), + "and{b} {$src, $dst|$dst, $src}">; + def AND16mi : Ii16<0x81, MRM4m, + (ops i16mem:$dst, i16imm:$src), + "and{w} {$src, $dst|$dst, $src}">, OpSize; + def AND32mi : Ii32<0x81, MRM4m, + (ops i32mem:$dst, i32imm:$src), + "and{l} {$src, $dst|$dst, $src}">; + def AND16mi8 : Ii8<0x83, MRM4m, + (ops i16mem:$dst, i8imm :$src), + "and{w} {$src, $dst|$dst, $src}">, OpSize; + def AND32mi8 : Ii8<0x83, MRM4m, + (ops i32mem:$dst, i8imm :$src), + "and{l} {$src, $dst|$dst, $src}">; } def OR8rr : I<0x08, MRMDestReg, (ops R8 :$dst, R8 :$src1, R8 :$src2), - "or $dst, $src2">; + "or{b} {$src2, $dst|$dst, $src2}">; def OR16rr : I<0x09, MRMDestReg, (ops R16:$dst, R16:$src1, R16:$src2), - "or $dst, $src2">, OpSize; + "or{w} {$src2, $dst|$dst, $src2}">, OpSize; def OR32rr : I<0x09, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), - "or $dst, $src2">; + "or{l} {$src2, $dst|$dst, $src2}">; def OR8rm : I<0x0A, MRMSrcMem , (ops R8 :$dst, R8 :$src1, i8mem :$src2), - "or $dst, $src2">; + "or{b} {$src2, $dst|$dst, $src2}">; def OR16rm : I<0x0B, MRMSrcMem , (ops R16:$dst, R16:$src1, i16mem:$src2), - "or $dst, $src2">, OpSize; + "or{w} {$src2, $dst|$dst, $src2}">, OpSize; def OR32rm : I<0x0B, MRMSrcMem , (ops R32:$dst, R32:$src1, i32mem:$src2), - "or $dst, $src2">; + "or{l} {$src2, $dst|$dst, $src2}">; def OR8ri : Ii8 <0x80, MRM1r, (ops R8 :$dst, R8 :$src1, i8imm:$src2), - "or $dst, $src2">; + "or{b} {$src2, $dst|$dst, $src2}">; def OR16ri : Ii16<0x81, MRM1r, (ops R16:$dst, R16:$src1, i16imm:$src2), - "or $dst, $src2">, OpSize; + "or{w} {$src2, $dst|$dst, $src2}">, OpSize; def OR32ri : Ii32<0x81, MRM1r, (ops R32:$dst, R32:$src1, i32imm:$src2), - "or $dst, $src2">; + "or{l} {$src2, $dst|$dst, $src2}">; def OR16ri8 : Ii8<0x83, MRM1r, (ops R8 :$dst, R8 :$src1, i8imm:$src2), - "or $dst, $src2">, OpSize; + "or{w} {$src2, $dst|$dst, $src2}">, OpSize; def OR32ri8 : Ii8<0x83, MRM1r, (ops R32:$dst, R32:$src1, i8imm:$src2), - "or $dst, $src2">; + "or{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { def OR8mr : I<0x08, MRMDestMem, (ops i8mem:$dst, R8:$src), - "or $dst, $src">; + "or{b} $dst, $src">; def OR16mr : I<0x09, MRMDestMem, (ops i16mem:$dst, R16:$src), - "or $dst, $src">, OpSize; + "or{w} $dst, $src">, OpSize; def OR32mr : I<0x09, MRMDestMem, (ops i32mem:$dst, R32:$src), - "or $dst, $src">; + "or{l} $dst, $src">; def OR8mi : Ii8<0x80, MRM1m, (ops i8mem :$dst, i8imm:$src), - "or $dst, $src">; + "or{b} $dst, $src">; def OR16mi : Ii16<0x81, MRM1m, (ops i16mem:$dst, i16imm:$src), - "or $dst, $src">, OpSize; + "or{w} $dst, $src">, OpSize; def OR32mi : Ii32<0x81, MRM1m, (ops i32mem:$dst, i32imm:$src), - "or $dst, $src">; + "or{l} $dst, $src">; def OR16mi8 : Ii8<0x83, MRM1m, (ops i16mem:$dst, i8imm:$src), - "or $dst, $src">, OpSize; + "or{w} $dst, $src">, OpSize; def OR32mi8 : Ii8<0x83, MRM1m, (ops i32mem:$dst, i8imm:$src), - "or $dst, $src">; + "or{l} $dst, $src">; } -def XOR8rr : I<0x30, MRMDestReg, (ops R8 :$dst, R8 :$src1, R8 :$src2), "xor $dst, $src2">; -def XOR16rr : I<0x31, MRMDestReg, (ops R16:$dst, R16:$src1, R16:$src2), "xor $dst, $src2">, OpSize; -def XOR32rr : I<0x31, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), "xor $dst, $src2">; -def XOR8rm : I<0x32, MRMSrcMem , (ops R8 :$dst, R8:$src1, i8mem :$src2), "xor $dst, $src2">; -def XOR16rm : I<0x33, MRMSrcMem , (ops R16:$dst, R8:$src1, i16mem:$src2), "xor $dst, $src2">, OpSize; -def XOR32rm : I<0x33, MRMSrcMem , (ops R32:$dst, R8:$src1, i32mem:$src2), "xor $dst, $src2">; - -def XOR8ri : Ii8 <0x80, MRM6r, (ops R8:$dst, R8:$src1, i8imm:$src2), "xor $dst, $src2">; -def XOR16ri : Ii16 <0x81, MRM6r, (ops R16:$dst, R16:$src1, i16imm:$src2), "xor $dst, $src2">, OpSize; -def XOR32ri : Ii32 <0x81, MRM6r, (ops R32:$dst, R32:$src1, i32imm:$src2), "xor $dst, $src2">; -def XOR16ri8 : Ii8<0x83, MRM6r, (ops R16:$dst, R16:$src1, i8imm:$src2), - "xor $dst, $src2">, OpSize; -def XOR32ri8 : Ii8<0x83, MRM6r, (ops R32:$dst, R32:$src1, i8imm:$src2), - "xor $dst, $src2">; +def XOR8rr : I<0x30, MRMDestReg, + (ops R8 :$dst, R8 :$src1, R8 :$src2), + "xor{b} {$src2, $dst|$dst, $src2}">; +def XOR16rr : I<0x31, MRMDestReg, + (ops R16:$dst, R16:$src1, R16:$src2), + "xor{w} {$src2, $dst|$dst, $src2}">, OpSize; +def XOR32rr : I<0x31, MRMDestReg, + (ops R32:$dst, R32:$src1, R32:$src2), + "xor{l} {$src2, $dst|$dst, $src2}">; +def XOR8rm : I<0x32, MRMSrcMem , + (ops R8 :$dst, R8:$src1, i8mem :$src2), + "xor{b} {$src2, $dst|$dst, $src2}">; +def XOR16rm : I<0x33, MRMSrcMem , + (ops R16:$dst, R8:$src1, i16mem:$src2), + "xor{w} {$src2, $dst|$dst, $src2}">, OpSize; +def XOR32rm : I<0x33, MRMSrcMem , + (ops R32:$dst, R8:$src1, i32mem:$src2), + "xor{l} {$src2, $dst|$dst, $src2}">; + +def XOR8ri : Ii8<0x80, MRM6r, + (ops R8:$dst, R8:$src1, i8imm:$src2), + "xor{b} {$src2, $dst|$dst, $src2}">; +def XOR16ri : Ii16<0x81, MRM6r, + (ops R16:$dst, R16:$src1, i16imm:$src2), + "xor{w} {$src2, $dst|$dst, $src2}">, OpSize; +def XOR32ri : Ii32<0x81, MRM6r, + (ops R32:$dst, R32:$src1, i32imm:$src2), + "xor{l} {$src2, $dst|$dst, $src2}">; +def XOR16ri8 : Ii8<0x83, MRM6r, + (ops R16:$dst, R16:$src1, i8imm:$src2), + "xor{w} {$src2, $dst|$dst, $src2}">, OpSize; +def XOR32ri8 : Ii8<0x83, MRM6r, + (ops R32:$dst, R32:$src1, i8imm:$src2), + "xor{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def XOR8mr : I<0x30, MRMDestMem, (ops i8mem :$dst, R8 :$src), "xor $dst, $src">; - def XOR16mr : I<0x31, MRMDestMem, (ops i16mem:$dst, R16:$src), "xor $dst, $src">, OpSize; - def XOR32mr : I<0x31, MRMDestMem, (ops i32mem:$dst, R32:$src), "xor $dst, $src">; - def XOR8mi : Ii8 <0x80, MRM6m, (ops i8mem :$dst, i8imm :$src), "xor $dst, $src">; - def XOR16mi : Ii16<0x81, MRM6m, (ops i16mem:$dst, i16imm:$src), "xor $dst, $src">, OpSize; - def XOR32mi : Ii32<0x81, MRM6m, (ops i32mem:$dst, i32imm:$src), "xor $dst, $src">; - def XOR16mi8 : Ii8 <0x83, MRM6m, (ops i16mem:$dst, i8imm :$src), "xor $dst, $src">, OpSize; - def XOR32mi8 : Ii8 <0x83, MRM6m, (ops i32mem:$dst, i8imm :$src), "xor $dst, $src">; + def XOR8mr : I<0x30, MRMDestMem, + (ops i8mem :$dst, R8 :$src), + "xor{b} {$src, $dst|$dst, $src}">; + def XOR16mr : I<0x31, MRMDestMem, + (ops i16mem:$dst, R16:$src), + "xor{w} {$src, $dst|$dst, $src}">, OpSize; + def XOR32mr : I<0x31, MRMDestMem, + (ops i32mem:$dst, R32:$src), + "xor{l} {$src, $dst|$dst, $src}">; + def XOR8mi : Ii8<0x80, MRM6m, + (ops i8mem :$dst, i8imm :$src), + "xor{b} {$src, $dst|$dst, $src}">; + def XOR16mi : Ii16<0x81, MRM6m, + (ops i16mem:$dst, i16imm:$src), + "xor{w} {$src, $dst|$dst, $src}">, OpSize; + def XOR32mi : Ii32<0x81, MRM6m, + (ops i32mem:$dst, i32imm:$src), + "xor{l} {$src, $dst|$dst, $src}">; + def XOR16mi8 : Ii8<0x83, MRM6m, + (ops i16mem:$dst, i8imm :$src), + "xor{w} {$src, $dst|$dst, $src}">, OpSize; + def XOR32mi8 : Ii8<0x83, MRM6m, + (ops i32mem:$dst, i8imm :$src), + "xor{l} {$src, $dst|$dst, $src}">; } // Shift instructions // FIXME: provide shorter instructions when imm8 == 1 -def SHL8rCL : I<0xD2, MRM4r, (ops R8 :$dst, R8 :$src), "shl $dst, %CL">, Imp<[CL],[]>; -def SHL16rCL : I<0xD3, MRM4r, (ops R16:$dst, R16:$src), "shl $dst, %CL">, Imp<[CL],[]>, OpSize; -def SHL32rCL : I<0xD3, MRM4r, (ops R32:$dst, R32:$src), "shl $dst, %CL">, Imp<[CL],[]>; +def SHL8rCL : I<0xD2, MRM4r, (ops R8 :$dst, R8 :$src), + "shl{b} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; +def SHL16rCL : I<0xD3, MRM4r, (ops R16:$dst, R16:$src), + "shl{w} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>, OpSize; +def SHL32rCL : I<0xD3, MRM4r, (ops R32:$dst, R32:$src), + "shl{l} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; def SHL8ri : Ii8<0xC0, MRM4r, (ops R8 :$dst, R8 :$src1, i8imm:$src2), - "shl $dst, $src2">; + "shl{b} {$src2, $dst|$dst, $src2}">; def SHL16ri : Ii8<0xC1, MRM4r, (ops R16:$dst, R16:$src1, i8imm:$src2), - "shl $dst, $src2">, OpSize; + "shl{w} {$src2, $dst|$dst, $src2}">, OpSize; def SHL32ri : Ii8<0xC1, MRM4r, (ops R32:$dst, R32:$src1, i8imm:$src2), - "shl $dst, $src2">; + "shl{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def SHL8mCL : I<0xD2, MRM4m, (ops i8mem :$dst), "shl $dst, %CL">, Imp<[CL],[]>; - def SHL16mCL : I<0xD3, MRM4m, (ops i16mem:$dst), "shl $dst, %CL">, Imp<[CL],[]>, OpSize; - def SHL32mCL : I<0xD3, MRM4m, (ops i32mem:$dst), "shl $dst, %CL">, Imp<[CL],[]>; - def SHL8mi : Ii8<0xC0, MRM4m, (ops i8mem :$dst, i8imm:$src), "shl $dst, $src">; - def SHL16mi : Ii8<0xC1, MRM4m, (ops i16mem:$dst, i8imm:$src), "shl $dst, $src">, OpSize; - def SHL32mi : Ii8<0xC1, MRM4m, (ops i32mem:$dst, i8imm:$src), "shl $dst, $src">; + def SHL8mCL : I<0xD2, MRM4m, (ops i8mem :$dst), + "shl{b} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; + def SHL16mCL : I<0xD3, MRM4m, (ops i16mem:$dst), + "shl{w} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>, OpSize; + def SHL32mCL : I<0xD3, MRM4m, (ops i32mem:$dst), + "shl{l} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; + def SHL8mi : Ii8<0xC0, MRM4m, (ops i8mem :$dst, i8imm:$src), + "shl{b} {$src, $dst|$dst, $src}">; + def SHL16mi : Ii8<0xC1, MRM4m, (ops i16mem:$dst, i8imm:$src), + "shl{w} {$src, $dst|$dst, $src}">, OpSize; + def SHL32mi : Ii8<0xC1, MRM4m, (ops i32mem:$dst, i8imm:$src), + "shl{l} {$src, $dst|$dst, $src}">; } -def SHR8rCL : I<0xD2, MRM5r, (ops R8 :$dst, R8 :$src), "shr $dst, %CL">, Imp<[CL],[]>; -def SHR16rCL : I<0xD3, MRM5r, (ops R16:$dst, R16:$src), "shr $dst, %CL">, Imp<[CL],[]>, OpSize; -def SHR32rCL : I<0xD3, MRM5r, (ops R32:$dst, R32:$src), "shr $dst, %CL">, Imp<[CL],[]>; +def SHR8rCL : I<0xD2, MRM5r, (ops R8 :$dst, R8 :$src), + "shr{b} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; +def SHR16rCL : I<0xD3, MRM5r, (ops R16:$dst, R16:$src), + "shr{w} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>, OpSize; +def SHR32rCL : I<0xD3, MRM5r, (ops R32:$dst, R32:$src), + "shr{l} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; -def SHR8ri : Ii8 <0xC0, MRM5r, (ops R8:$dst, R8:$src1, i8imm:$src2), "shr $dst, $src2">; -def SHR16ri : Ii8 <0xC1, MRM5r, (ops R16:$dst, R16:$src1, i8imm:$src2), "shr $dst, $src2">, OpSize; -def SHR32ri : Ii8 <0xC1, MRM5r, (ops R32:$dst, R32:$src1, i8imm:$src2), "shr $dst, $src2">; +def SHR8ri : Ii8<0xC0, MRM5r, (ops R8:$dst, R8:$src1, i8imm:$src2), + "shr{b} {$src2, $dst|$dst, $src2}">; +def SHR16ri : Ii8<0xC1, MRM5r, (ops R16:$dst, R16:$src1, i8imm:$src2), + "shr{w} {$src2, $dst|$dst, $src2}">, OpSize; +def SHR32ri : Ii8<0xC1, MRM5r, (ops R32:$dst, R32:$src1, i8imm:$src2), + "shr{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def SHR8mCL : I<0xD2, MRM5m, (ops i8mem :$dst), "shr $dst, %CL">, Imp<[CL],[]>; - def SHR16mCL : I<0xD3, MRM5m, (ops i16mem:$dst), "shr $dst, %CL">, Imp<[CL],[]>, OpSize; - def SHR32mCL : I<0xD3, MRM5m, (ops i32mem:$dst), "shr $dst, %CL">, Imp<[CL],[]>; - def SHR8mi : Ii8<0xC0, MRM5m, (ops i8mem :$dst, i8imm:$src), "shr $dst, $src">; - def SHR16mi : Ii8<0xC1, MRM5m, (ops i16mem:$dst, i8imm:$src), "shr $dst, $src">, OpSize; - def SHR32mi : Ii8<0xC1, MRM5m, (ops i32mem:$dst, i8imm:$src), "shr $dst, $src">; + def SHR8mCL : I<0xD2, MRM5m, (ops i8mem :$dst), + "shr{b} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; + def SHR16mCL : I<0xD3, MRM5m, (ops i16mem:$dst), + "shr{w} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>, OpSize; + def SHR32mCL : I<0xD3, MRM5m, (ops i32mem:$dst), + "shr{l} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; + def SHR8mi : Ii8<0xC0, MRM5m, (ops i8mem :$dst, i8imm:$src), + "shr{b} {$src, $dst|$dst, $src}">; + def SHR16mi : Ii8<0xC1, MRM5m, (ops i16mem:$dst, i8imm:$src), + "shr{w} {$src, $dst|$dst, $src}">, OpSize; + def SHR32mi : Ii8<0xC1, MRM5m, (ops i32mem:$dst, i8imm:$src), + "shr{l} {$src, $dst|$dst, $src}">; } -def SAR8rCL : I<0xD2, MRM7r, (ops R8 :$dst, R8 :$src), "sar $dst, %CL">, Imp<[CL],[]>; -def SAR16rCL : I<0xD3, MRM7r, (ops R16:$dst, R16:$src), "sar $dst, %CL">, Imp<[CL],[]>, OpSize; -def SAR32rCL : I<0xD3, MRM7r, (ops R32:$dst, R32:$src), "sar $dst, %CL">, Imp<[CL],[]>; +def SAR8rCL : I<0xD2, MRM7r, (ops R8 :$dst, R8 :$src), + "sar{b} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; +def SAR16rCL : I<0xD3, MRM7r, (ops R16:$dst, R16:$src), + "sar{w} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>, OpSize; +def SAR32rCL : I<0xD3, MRM7r, (ops R32:$dst, R32:$src), + "sar{l} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; def SAR8ri : Ii8<0xC0, MRM7r, (ops R8 :$dst, R8 :$src1, i8imm:$src2), - "sar $dst, $src2">; + "sar{b} {$src2, $dst|$dst, $src2}">; def SAR16ri : Ii8<0xC1, MRM7r, (ops R16:$dst, R16:$src1, i8imm:$src2), - "sar $dst, $src2">, OpSize; + "sar{w} {$src2, $dst|$dst, $src2}">, OpSize; def SAR32ri : Ii8<0xC1, MRM7r, (ops R32:$dst, R32:$src1, i8imm:$src2), - "sar $dst, $src2">; + "sar{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def SAR8mCL : I<0xD2, MRM7m, (ops i8mem :$dst), "sar $dst, %CL">, Imp<[CL],[]>; - def SAR16mCL : I<0xD3, MRM7m, (ops i16mem:$dst), "sar $dst, %CL">, Imp<[CL],[]>, OpSize; - def SAR32mCL : I<0xD3, MRM7m, (ops i32mem:$dst), "sar $dst, %CL">, Imp<[CL],[]>; - def SAR8mi : Ii8<0xC0, MRM7m, (ops i8mem :$dst, i8imm:$src), "sar $dst, $src">; - def SAR16mi : Ii8<0xC1, MRM7m, (ops i16mem:$dst, i8imm:$src), "sar $dst, $src">, OpSize; - def SAR32mi : Ii8<0xC1, MRM7m, (ops i32mem:$dst, i8imm:$src), "sar $dst, $src">; + def SAR8mCL : I<0xD2, MRM7m, (ops i8mem :$dst), + "sar{b} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; + def SAR16mCL : I<0xD3, MRM7m, (ops i16mem:$dst), + "sar{w} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>, OpSize; + def SAR32mCL : I<0xD3, MRM7m, (ops i32mem:$dst), + "sar{l} {%CL, $dst|$dst, %CL}">, Imp<[CL],[]>; + def SAR8mi : Ii8<0xC0, MRM7m, (ops i8mem :$dst, i8imm:$src), + "sar{b} {$src, $dst|$dst, $src}">; + def SAR16mi : Ii8<0xC1, MRM7m, (ops i16mem:$dst, i8imm:$src), + "sar{w} {$src, $dst|$dst, $src}">, OpSize; + def SAR32mi : Ii8<0xC1, MRM7m, (ops i32mem:$dst, i8imm:$src), + "sar{l} {$src, $dst|$dst, $src}">; } def SHLD32rrCL : I<0xA5, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), - "shld $dst, $src2, %CL">, Imp<[CL],[]>, TB; + "shld{l} {%CL, $src2, $dst|$dst, $src2, %CL}">, + Imp<[CL],[]>, TB; def SHRD32rrCL : I<0xAD, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), - "shrd $dst, $src2, %CL">, Imp<[CL],[]>, TB; -def SHLD32rri8 : Ii8<0xA4, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2, i8imm:$src3), - "shld $dst, $src2, $src3">, TB; -def SHRD32rri8 : Ii8<0xAC, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2, i8imm:$src3), - "shrd $dst, $src2, $src3">, TB; + "shrd{l} {%CL, $src2, $dst|$dst, $src2, %CL}">, + Imp<[CL],[]>, TB; +def SHLD32rri8 : Ii8<0xA4, MRMDestReg, + (ops R32:$dst, R32:$src1, R32:$src2, i8imm:$src3), + "shld{l} {$src3, $src2, $dst|$dst, $src2, $src3}">, TB; +def SHRD32rri8 : Ii8<0xAC, MRMDestReg, + (ops R32:$dst, R32:$src1, R32:$src2, i8imm:$src3), + "shrd{l} {$src3, $src2, $dst|$dst, $src2, $src3}">, TB; let isTwoAddress = 0 in { def SHLD32mrCL : I<0xA5, MRMDestMem, (ops i32mem:$dst, R32:$src2), - "shld $dst, $src2, %CL">, Imp<[CL],[]>, TB; + "shld{l} {%CL, $src2, $dst|$dst, $src2, %CL}">, + Imp<[CL],[]>, TB; def SHRD32mrCL : I<0xAD, MRMDestMem, (ops i32mem:$dst, R32:$src2), - "shrd $dst, $src2, %CL">, Imp<[CL],[]>, TB; - def SHLD32mri8 : Ii8<0xA4, MRMDestMem, (ops i32mem:$dst, R32:$src2, i8imm:$src3), - "shld $dst, $src2, $src3">, TB; - def SHRD32mri8 : Ii8<0xAC, MRMDestMem, (ops i32mem:$dst, R32:$src2, i8imm:$src3), - "shrd $dst, $src2, $src3">, TB; + "shrd{l} {%CL, $src2, $dst|$dst, $src2, %CL}">, + Imp<[CL],[]>, TB; + def SHLD32mri8 : Ii8<0xA4, MRMDestMem, + (ops i32mem:$dst, R32:$src2, i8imm:$src3), + "shld{l} {$src3, $src2, $dst|$dst, $src2, $src3}">, TB; + def SHRD32mri8 : Ii8<0xAC, MRMDestMem, + (ops i32mem:$dst, R32:$src2, i8imm:$src3), + "shrd{l} {$src3, $src2, $dst|$dst, $src2, $src3}">, TB; } // Arithmetic... -def ADD8rr : I<0x00, MRMDestReg, (ops R8 :$dst, R8 :$src1, R8 :$src2), "add $dst, $src2">; -def ADD16rr : I<0x01, MRMDestReg, (ops R16:$dst, R16:$src1, R16:$src2), "add $dst, $src2">, OpSize; -def ADD32rr : I<0x01, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), "add $dst, $src2">; -def ADD8rm : I<0x02, MRMSrcMem, (ops R8 :$dst, R8 :$src1, i8mem :$src2), "add $dst, $src2">; -def ADD16rm : I<0x03, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), "add $dst, $src2">, OpSize; -def ADD32rm : I<0x03, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), "add $dst, $src2">; - -def ADD8ri : Ii8 <0x80, MRM0r, (ops R8:$dst, R8:$src1, i8imm:$src2), "add $dst, $src2">; -def ADD16ri : Ii16 <0x81, MRM0r, (ops R16:$dst, R16:$src1, i16imm:$src2), "add $dst, $src2">, OpSize; -def ADD32ri : Ii32 <0x81, MRM0r, (ops R32:$dst, R32:$src1, i32imm:$src2), "add $dst, $src2">; - -def ADD16ri8 : Ii8 <0x83, MRM0r, (ops R16:$dst, R16:$src1, i8imm:$src2), "add $dst, $src2">, OpSize; -def ADD32ri8 : Ii8 <0x83, MRM0r, (ops R32:$dst, R32:$src1, i8imm:$src2), "add $dst, $src2">; +def ADD8rr : I<0x00, MRMDestReg, (ops R8 :$dst, R8 :$src1, R8 :$src2), + "add{b} {$src2, $dst|$dst, $src2}">; +def ADD16rr : I<0x01, MRMDestReg, (ops R16:$dst, R16:$src1, R16:$src2), + "add{w} {$src2, $dst|$dst, $src2}">, OpSize; +def ADD32rr : I<0x01, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), + "add{l} {$src2, $dst|$dst, $src2}">; +def ADD8rm : I<0x02, MRMSrcMem, (ops R8 :$dst, R8 :$src1, i8mem :$src2), + "add{b} {$src2, $dst|$dst, $src2}">; +def ADD16rm : I<0x03, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), + "add{w} {$src2, $dst|$dst, $src2}">, OpSize; +def ADD32rm : I<0x03, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), + "add{l} {$src2, $dst|$dst, $src2}">; + +def ADD8ri : Ii8<0x80, MRM0r, (ops R8:$dst, R8:$src1, i8imm:$src2), + "add{b} {$src2, $dst|$dst, $src2}">; +def ADD16ri : Ii16<0x81, MRM0r, (ops R16:$dst, R16:$src1, i16imm:$src2), + "add{w} {$src2, $dst|$dst, $src2}">, OpSize; +def ADD32ri : Ii32<0x81, MRM0r, (ops R32:$dst, R32:$src1, i32imm:$src2), + "add{l} {$src2, $dst|$dst, $src2}">; + +def ADD16ri8 : Ii8<0x83, MRM0r, (ops R16:$dst, R16:$src1, i8imm:$src2), + "add{w} {$src2, $dst|$dst, $src2}">, OpSize; +def ADD32ri8 : Ii8<0x83, MRM0r, (ops R32:$dst, R32:$src1, i8imm:$src2), + "add{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def ADD8mr : I<0x00, MRMDestMem, (ops i8mem :$dst, R8 :$src2), "add $dst, $src2">; - def ADD16mr : I<0x01, MRMDestMem, (ops i16mem:$dst, R16:$src2), "add $dst, $src2">, OpSize; - def ADD32mr : I<0x01, MRMDestMem, (ops i32mem:$dst, R32:$src2), "add $dst, $src2">; - def ADD8mi : Ii8 <0x80, MRM0m, (ops i8mem :$dst, i8imm :$src2), "add $dst, $src2">; - def ADD16mi : Ii16<0x81, MRM0m, (ops i16mem:$dst, i16imm:$src2), "add $dst, $src2">, OpSize; - def ADD32mi : Ii32<0x81, MRM0m, (ops i32mem:$dst, i32imm:$src2), "add $dst, $src2">; - def ADD16mi8 : Ii8 <0x83, MRM0m, (ops i16mem:$dst, i8imm :$src2), "add $dst, $src2">, OpSize; - def ADD32mi8 : Ii8 <0x83, MRM0m, (ops i32mem:$dst, i8imm :$src2), "add $dst, $src2">; + def ADD8mr : I<0x00, MRMDestMem, (ops i8mem :$dst, R8 :$src2), + "add{b} {$src2, $dst|$dst, $src2}">; + def ADD16mr : I<0x01, MRMDestMem, (ops i16mem:$dst, R16:$src2), + "add{w} {$src2, $dst|$dst, $src2}">, OpSize; + def ADD32mr : I<0x01, MRMDestMem, (ops i32mem:$dst, R32:$src2), + "add{l} {$src2, $dst|$dst, $src2}">; + def ADD8mi : Ii8<0x80, MRM0m, (ops i8mem :$dst, i8imm :$src2), + "add{b} {$src2, $dst|$dst, $src2}">; + def ADD16mi : Ii16<0x81, MRM0m, (ops i16mem:$dst, i16imm:$src2), + "add{w} {$src2, $dst|$dst, $src2}">, OpSize; + def ADD32mi : Ii32<0x81, MRM0m, (ops i32mem:$dst, i32imm:$src2), + "add{l} {$src2, $dst|$dst, $src2}">; + def ADD16mi8 : Ii8<0x83, MRM0m, (ops i16mem:$dst, i8imm :$src2), + "add{w} {$src2, $dst|$dst, $src2}">, OpSize; + def ADD32mi8 : Ii8<0x83, MRM0m, (ops i32mem:$dst, i8imm :$src2), + "add{l} {$src2, $dst|$dst, $src2}">; } -def ADC32rr : I<0x11, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), "adc $dst, $src2">; -def ADC32rm : I<0x13, MRMSrcMem , (ops R32:$dst, R32:$src1, i32mem:$src2), "adc $dst, $src2">; -def ADC32ri : Ii32<0x81, MRM2r, (ops R32:$dst, R32:$src1, i32imm:$src2), "adc $dst, $src2">; -def ADC32ri8 : Ii8 <0x83, MRM2r, (ops R32:$dst, R32:$src1, i8imm:$src2), "adc $dst, $src2">; +def ADC32rr : I<0x11, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; +def ADC32rm : I<0x13, MRMSrcMem , (ops R32:$dst, R32:$src1, i32mem:$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; +def ADC32ri : Ii32<0x81, MRM2r, (ops R32:$dst, R32:$src1, i32imm:$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; +def ADC32ri8 : Ii8<0x83, MRM2r, (ops R32:$dst, R32:$src1, i8imm:$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def ADC32mr : I<0x11, MRMDestMem, (ops i32mem:$dst, R32:$src2), "adc $dst, $src2">; - def ADC32mi : Ii32<0x81, MRM2m, (ops i32mem:$dst, i32imm:$src2), "adc $dst, $src2">; - def ADC32mi8 : Ii8 <0x83, MRM2m, (ops i32mem:$dst, i8imm :$src2), "adc $dst, $src2">; + def ADC32mr : I<0x11, MRMDestMem, (ops i32mem:$dst, R32:$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; + def ADC32mi : Ii32<0x81, MRM2m, (ops i32mem:$dst, i32imm:$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; + def ADC32mi8 : Ii8<0x83, MRM2m, (ops i32mem:$dst, i8imm :$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; } -def SUB8rr : I<0x28, MRMDestReg, (ops R8 :$dst, R8 :$src1, R8 :$src2), "sub $dst, $src2">; -def SUB16rr : I<0x29, MRMDestReg, (ops R16:$dst, R16:$src1, R16:$src2), "sub $dst, $src2">, OpSize; -def SUB32rr : I<0x29, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), "sub $dst, $src2">; -def SUB8rm : I<0x2A, MRMSrcMem, (ops R8 :$dst, R8 :$src1, i8mem :$src2), "sub $dst, $src2">; -def SUB16rm : I<0x2B, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), "sub $dst, $src2">, OpSize; -def SUB32rm : I<0x2B, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), "sub $dst, $src2">; +def SUB8rr : I<0x28, MRMDestReg, (ops R8 :$dst, R8 :$src1, R8 :$src2), + "sub{b} {$src2, $dst|$dst, $src2}">; +def SUB16rr : I<0x29, MRMDestReg, (ops R16:$dst, R16:$src1, R16:$src2), + "sub{w} {$src2, $dst|$dst, $src2}">, OpSize; +def SUB32rr : I<0x29, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), + "sub{l} {$src2, $dst|$dst, $src2}">; +def SUB8rm : I<0x2A, MRMSrcMem, (ops R8 :$dst, R8 :$src1, i8mem :$src2), + "sub{b} {$src2, $dst|$dst, $src2}">; +def SUB16rm : I<0x2B, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), + "sub{w} {$src2, $dst|$dst, $src2}">, OpSize; +def SUB32rm : I<0x2B, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), + "sub{l} {$src2, $dst|$dst, $src2}">; def SUB8ri : Ii8 <0x80, MRM5r, (ops R8:$dst, R8:$src1, i8imm:$src2), - "sub $dst, $src2">; + "sub{b} {$src2, $dst|$dst, $src2}">; def SUB16ri : Ii16<0x81, MRM5r, (ops R16:$dst, R16:$src1, i16imm:$src2), - "sub $dst, $src2">, OpSize; + "sub{w} {$src2, $dst|$dst, $src2}">, OpSize; def SUB32ri : Ii32<0x81, MRM5r, (ops R32:$dst, R32:$src1, i32imm:$src2), - "sub $dst, $src2">; + "sub{l} {$src2, $dst|$dst, $src2}">; def SUB16ri8 : Ii8<0x83, MRM5r, (ops R16:$dst, R16:$src1, i8imm:$src2), - "sub $dst, $src2">, OpSize; + "sub{w} {$src2, $dst|$dst, $src2}">, OpSize; def SUB32ri8 : Ii8<0x83, MRM5r, (ops R32:$dst, R32:$src1, i8imm:$src2), - "sub $dst, $src2">; + "sub{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def SUB8mr : I<0x28, MRMDestMem, (ops i8mem :$dst, R8 :$src2), "sub $dst, $src2">; - def SUB16mr : I<0x29, MRMDestMem, (ops i16mem:$dst, R16:$src2), "sub $dst, $src2">, OpSize; - def SUB32mr : I<0x29, MRMDestMem, (ops i32mem:$dst, R32:$src2), "sub $dst, $src2">; - def SUB8mi : Ii8 <0x80, MRM5m, (ops i8mem :$dst, i8imm:$src2), "sub $dst, $src2">; - def SUB16mi : Ii16<0x81, MRM5m, (ops i16mem:$dst, i16imm:$src2), "sub $dst, $src2">, OpSize; - def SUB32mi : Ii32<0x81, MRM5m, (ops i32mem:$dst, i32imm:$src2), "sub $dst, $src2">; - def SUB16mi8 : Ii8 <0x83, MRM5m, (ops i16mem:$dst, i8imm :$src2), "sub $dst, $src2">, OpSize; - def SUB32mi8 : Ii8 <0x83, MRM5m, (ops i32mem:$dst, i8imm :$src2), "sub $dst, $src2">; + def SUB8mr : I<0x28, MRMDestMem, (ops i8mem :$dst, R8 :$src2), + "sub{b} {$src2, $dst|$dst, $src2}">; + def SUB16mr : I<0x29, MRMDestMem, (ops i16mem:$dst, R16:$src2), + "sub{w} {$src2, $dst|$dst, $src2}">, OpSize; + def SUB32mr : I<0x29, MRMDestMem, (ops i32mem:$dst, R32:$src2), + "sub{l} {$src2, $dst|$dst, $src2}">; + def SUB8mi : Ii8<0x80, MRM5m, (ops i8mem :$dst, i8imm:$src2), + "sub{b} {$src2, $dst|$dst, $src2}">; + def SUB16mi : Ii16<0x81, MRM5m, (ops i16mem:$dst, i16imm:$src2), + "sub{w} {$src2, $dst|$dst, $src2}">, OpSize; + def SUB32mi : Ii32<0x81, MRM5m, (ops i32mem:$dst, i32imm:$src2), + "sub{l} {$src2, $dst|$dst, $src2}">; + def SUB16mi8 : Ii8<0x83, MRM5m, (ops i16mem:$dst, i8imm :$src2), + "sub{w} {$src2, $dst|$dst, $src2}">, OpSize; + def SUB32mi8 : Ii8<0x83, MRM5m, (ops i32mem:$dst, i8imm :$src2), + "sub{l} {$src2, $dst|$dst, $src2}">; } -def SBB32rr : I<0x19, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), - "adc $dst, $src2">; +def SBB32rr : I<0x19, MRMDestReg, (ops R32:$dst, R32:$src1, R32:$src2), + "adc{l} {$src2, $dst|$dst, $src2}">; let isTwoAddress = 0 in { - def SBB32mr : I<0x19, MRMDestMem, (ops i32mem:$dst, R32:$src2), "sbb $dst, $src2">; - def SBB32mi : Ii32<0x81, MRM3m, (ops i32mem:$dst, i32imm:$src2), "sbb $dst, $src2">; - def SBB32mi8 : Ii8 <0x83, MRM3m, (ops i32mem:$dst, i8imm :$src2), "sbb $dst, $src2">; + def SBB32mr : I<0x19, MRMDestMem, (ops i32mem:$dst, R32:$src2), + "sbb{l} {$src2, $dst|$dst, $src2}">; + def SBB32mi : Ii32<0x81, MRM3m, (ops i32mem:$dst, i32imm:$src2), + "sbb{l} {$src2, $dst|$dst, $src2}">; + def SBB32mi8 : Ii8<0x83, MRM3m, (ops i32mem:$dst, i8imm :$src2), + "sbb{l} {$src2, $dst|$dst, $src2}">; } def SBB32rm : I<0x1B, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), - "sbb $dst, $src2">; + "sbb{l} {$src2, $dst|$dst, $src2}">; def SBB32ri : Ii32<0x81, MRM3r, (ops R32:$dst, R32:$src1, i32imm:$src2), - "sbb $dst, $src2">; + "sbb{l} {$src2, $dst|$dst, $src2}">; def SBB32ri8 : Ii8<0x83, MRM3r, (ops R32:$dst, R32:$src1, i8imm:$src2), - "sbb $dst, $src2">; + "sbb{l} {$src2, $dst|$dst, $src2}">; -def IMUL16rr : I<0xAF, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), "imul $dst, $src2">, TB, OpSize; -def IMUL32rr : I<0xAF, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), "imul $dst, $src2">, TB; -def IMUL16rm : I<0xAF, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), "imul $dst, $src2">, TB, OpSize; -def IMUL32rm : I<0xAF, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), "imul $dst, $src2">, TB ; +def IMUL16rr : I<0xAF, MRMSrcReg, (ops R16:$dst, R16:$src1, R16:$src2), + "imul{w} {$src2, $dst|$dst, $src2}">, TB, OpSize; +def IMUL32rr : I<0xAF, MRMSrcReg, (ops R32:$dst, R32:$src1, R32:$src2), + "imul{l} {$src2, $dst|$dst, $src2}">, TB; +def IMUL16rm : I<0xAF, MRMSrcMem, (ops R16:$dst, R16:$src1, i16mem:$src2), + "imul{w} {$src2, $dst|$dst, $src2}">, TB, OpSize; +def IMUL32rm : I<0xAF, MRMSrcMem, (ops R32:$dst, R32:$src1, i32mem:$src2), + "imul{l} {$src2, $dst|$dst, $src2}">, TB; } // end Two Address instructions // Suprisingly enough, these are not two address instructions! -def IMUL16rri : Ii16<0x69, MRMSrcReg, (ops R16:$dst, R16:$src1, i16imm:$src2), - "imul $dst, $src1, $src2">, OpSize; // R16 = R16*I16 -def IMUL32rri : Ii32<0x69, MRMSrcReg, (ops R32:$dst, R32:$src1, i32imm:$src2), - "imul $dst, $src1, $src2">; // R32 = R32*I32 -def IMUL16rri8 : Ii8<0x6B, MRMSrcReg, (ops R16:$dst, R16:$src1, i8imm:$src2), - "imul $dst, $src1, $src2">, OpSize; // R16 = R16*I8 -def IMUL32rri8 : Ii8<0x6B, MRMSrcReg, (ops R32:$dst, R32:$src1, i8imm:$src2), - "imul $dst, $src1, $src2">; // R32 = R32*I8 - -def IMUL16rmi : Ii16<0x69, MRMSrcMem, (ops R32:$dst, i16mem:$src1, i16imm:$src2), - "imul $dst, $src1, $src2">, OpSize; // R16 = [mem16]*I16 -def IMUL32rmi : Ii32<0x69, MRMSrcMem, (ops R32:$dst, i32mem:$src1, i32imm:$src2), - "imul $dst, $src1, $src2">; // R32 = [mem32]*I32 -def IMUL16rmi8 : Ii8<0x6B, MRMSrcMem, (ops R32:$dst, i16mem:$src1, i8imm :$src2), - "imul $dst, $src1, $src2">, OpSize; // R16 = [mem16]*I8 -def IMUL32rmi8 : Ii8<0x6B, MRMSrcMem, (ops R32:$dst, i32mem:$src1, i8imm: $src2), - "imul $dst, $src1, $src2">; // R32 = [mem32]*I8 +def IMUL16rri : Ii16<0x69, MRMSrcReg, // R16 = R16*I16 + (ops R16:$dst, R16:$src1, i16imm:$src2), + "imul{w} {$src2, $src1, $dst|$dst, $src1, $src2}">, + OpSize; +def IMUL32rri : Ii32<0x69, MRMSrcReg, // R32 = R32*I32 + (ops R32:$dst, R32:$src1, i32imm:$src2), + "imul{l} {$src2, $src1, $dst|$dst, $src1, $src2}">; +def IMUL16rri8 : Ii8<0x6B, MRMSrcReg, // R16 = R16*I8 + (ops R16:$dst, R16:$src1, i8imm:$src2), + "imul{w} {$src2, $src1, $dst|$dst, $src1, $src2}">, OpSize; +def IMUL32rri8 : Ii8<0x6B, MRMSrcReg, // R32 = R32*I8 + (ops R32:$dst, R32:$src1, i8imm:$src2), + "imul{l} {$src2, $src1, $dst|$dst, $src1, $src2}">; + +def IMUL16rmi : Ii16<0x69, MRMSrcMem, // R16 = [mem16]*I16 + (ops R32:$dst, i16mem:$src1, i16imm:$src2), + "imul{w} {$src2, $src1, $dst|$dst, $src1, $src2}">, OpSize; +def IMUL32rmi : Ii32<0x69, MRMSrcMem, // R32 = [mem32]*I32 + (ops R32:$dst, i32mem:$src1, i32imm:$src2), + "imul{l} {$src2, $src1, $dst|$dst, $src1, $src2}">; +def IMUL16rmi8 : Ii8<0x6B, MRMSrcMem, // R16 = [mem16]*I8 + (ops R32:$dst, i16mem:$src1, i8imm :$src2), + "imul{w} {$src2, $src1, $dst|$dst, $src1, $src2}">, OpSize; +def IMUL32rmi8 : Ii8<0x6B, MRMSrcMem, // R32 = [mem32]*I8 + (ops R32:$dst, i32mem:$src1, i8imm: $src2), + "imul{l} {$src2, $src1, $dst|$dst, $src1, $src2}">; //===----------------------------------------------------------------------===// // Test instructions are just like AND, except they don't generate a result. +// def TEST8rr : I<0x84, MRMDestReg, (ops R8:$src1, R8:$src2), - "test $src1, $src2">; + "test{b} {$src2, $src1|$src1, $src2}">; def TEST16rr : I<0x85, MRMDestReg, (ops R16:$src1, R16:$src2), - "test $src1, $src2">, OpSize; + "test{w} {$src2, $src1|$src1, $src2}">, OpSize; def TEST32rr : I<0x85, MRMDestReg, (ops R32:$src1, R32:$src2), - "test $src1, $src2">; + "test{l} {$src2, $src1|$src1, $src2}">; def TEST8mr : I<0x84, MRMDestMem, (ops i8mem :$src1, R8 :$src2), - "test $src1, $src2">; + "test{b} {$src2, $src1|$src1, $src2}">; def TEST16mr : I<0x85, MRMDestMem, (ops i16mem:$src1, R16:$src2), - "test $src1, $src2">, OpSize; + "test{w} {$src2, $src1|$src1, $src2}">, OpSize; def TEST32mr : I<0x85, MRMDestMem, (ops i32mem:$src1, R32:$src2), - "test $src1, $src2">; + "test{l} {$src2, $src1|$src1, $src2}">; def TEST8rm : I<0x84, MRMSrcMem, (ops R8 :$src1, i8mem :$src2), - "test $src1, $src2">; + "test{b} {$src2, $src1|$src1, $src2}">; def TEST16rm : I<0x85, MRMSrcMem, (ops R16:$src1, i16mem:$src2), - "test $src1, $src2">, OpSize; + "test{w} {$src2, $src1|$src1, $src2}">, OpSize; def TEST32rm : I<0x85, MRMSrcMem, (ops R32:$src1, i32mem:$src2), - "test $src1, $src2">; + "test{l} {$src2, $src1|$src1, $src2}">; def TEST8ri : Ii8 <0xF6, MRM0r, (ops R8:$dst, i8imm:$src), - "test $dst, $src">; // flags = R8 & imm8 + "test{b} $dst, $src">; // flags = R8 & imm8 def TEST16ri : Ii16<0xF7, MRM0r, (ops R16:$dst, i16imm:$src), - "test $dst, $src">, OpSize; // flags = R16 & imm16 + "test{w} $dst, $src">, OpSize; // flags = R16 & imm16 def TEST32ri : Ii32<0xF7, MRM0r, (ops R32:$dst, i32imm:$src), - "test $dst, $src">; // flags = R32 & imm32 + "test{l} $dst, $src">; // flags = R32 & imm32 def TEST8mi : Ii8 <0xF6, MRM0m, (ops i32mem:$dst, i8imm:$src), - "test $dst, $src">; // flags = [mem8] & imm8 + "test{b} $dst, $src">; // flags = [mem8] & imm8 def TEST16mi : Ii16<0xF7, MRM0m, (ops i16mem:$dst, i16imm:$src), - "test $dst, $src">, OpSize; // flags = [mem16] & imm16 + "test{w} $dst, $src">, OpSize; // flags = [mem16] & imm16 def TEST32mi : Ii32<0xF7, MRM0m, (ops i32mem:$dst, i32imm:$src), - "test $dst, $src">; // flags = [mem32] & imm32 + "test{l} $dst, $src">; // flags = [mem32] & imm32 @@ -803,64 +1062,132 @@ def TEST32mi : Ii32<0xF7, MRM0m, (ops i32mem:$dst, i32imm:$src), def SAHF : I<0x9E, RawFrm, (ops), "sahf">, Imp<[AH],[]>; // flags = AH def LAHF : I<0x9F, RawFrm, (ops), "lahf">, Imp<[],[AH]>; // AH = flags -def SETBr : I<0x92, MRM0r, (ops R8 :$dst), "setb $dst">, TB; // R8 = < unsign -def SETBm : I<0x92, MRM0m, (ops i8mem:$dst), "setb $dst">, TB; // [mem8] = < unsign -def SETAEr : I<0x93, MRM0r, (ops R8 :$dst), "setae $dst">, TB; // R8 = >= unsign -def SETAEm : I<0x93, MRM0m, (ops i8mem:$dst), "setae $dst">, TB; // [mem8] = >= unsign -def SETEr : I<0x94, MRM0r, (ops R8 :$dst), "sete $dst">, TB; // R8 = == -def SETEm : I<0x94, MRM0m, (ops i8mem:$dst), "sete $dst">, TB; // [mem8] = == -def SETNEr : I<0x95, MRM0r, (ops R8 :$dst), "setne $dst">, TB; // R8 = != -def SETNEm : I<0x95, MRM0m, (ops i8mem:$dst), "setne $dst">, TB; // [mem8] = != -def SETBEr : I<0x96, MRM0r, (ops R8 :$dst), "setbe $dst">, TB; // R8 = <= unsign -def SETBEm : I<0x96, MRM0m, (ops i8mem:$dst), "setbe $dst">, TB; // [mem8] = <= unsign -def SETAr : I<0x97, MRM0r, (ops R8 :$dst), "seta $dst">, TB; // R8 = > signed -def SETAm : I<0x97, MRM0m, (ops i8mem:$dst), "seta $dst">, TB; // [mem8] = > signed -def SETSr : I<0x98, MRM0r, (ops R8 :$dst), "sets $dst">, TB; // R8 = -def SETSm : I<0x98, MRM0m, (ops i8mem:$dst), "sets $dst">, TB; // [mem8] = -def SETNSr : I<0x99, MRM0r, (ops R8 :$dst), "setns $dst">, TB; // R8 = ! -def SETNSm : I<0x99, MRM0m, (ops i8mem:$dst), "setns $dst">, TB; // [mem8] = ! -def SETPr : I<0x9A, MRM0r, (ops R8 :$dst), "setp $dst">, TB; // R8 = parity -def SETPm : I<0x9A, MRM0m, (ops i8mem:$dst), "setp $dst">, TB; // [mem8] = parity -def SETLr : I<0x9C, MRM0r, (ops R8 :$dst), "setl $dst">, TB; // R8 = < signed -def SETLm : I<0x9C, MRM0m, (ops i8mem:$dst), "setl $dst">, TB; // [mem8] = < signed -def SETGEr : I<0x9D, MRM0r, (ops R8 :$dst), "setge $dst">, TB; // R8 = >= signed -def SETGEm : I<0x9D, MRM0m, (ops i8mem:$dst), "setge $dst">, TB; // [mem8] = >= signed -def SETLEr : I<0x9E, MRM0r, (ops R8 :$dst), "setle $dst">, TB; // R8 = <= signed -def SETLEm : I<0x9E, MRM0m, (ops i8mem:$dst), "setle $dst">, TB; // [mem8] = <= signed -def SETGr : I<0x9F, MRM0r, (ops R8 :$dst), "setg $dst">, TB; // R8 = < signed -def SETGm : I<0x9F, MRM0m, (ops i8mem:$dst), "setg $dst">, TB; // [mem8] = < signed +def SETBr : I<0x92, MRM0r, + (ops R8 :$dst), "setb $dst">, TB; // R8 = < unsign +def SETBm : I<0x92, MRM0m, + (ops i8mem:$dst), "setb $dst">, TB; // [mem8] = < unsign +def SETAEr : I<0x93, MRM0r, + (ops R8 :$dst), "setae $dst">, TB; // R8 = >= unsign +def SETAEm : I<0x93, MRM0m, + (ops i8mem:$dst), "setae $dst">, TB; // [mem8] = >= unsign +def SETEr : I<0x94, MRM0r, + (ops R8 :$dst), "sete $dst">, TB; // R8 = == +def SETEm : I<0x94, MRM0m, + (ops i8mem:$dst), "sete $dst">, TB; // [mem8] = == +def SETNEr : I<0x95, MRM0r, + (ops R8 :$dst), "setne $dst">, TB; // R8 = != +def SETNEm : I<0x95, MRM0m, + (ops i8mem:$dst), "setne $dst">, TB; // [mem8] = != +def SETBEr : I<0x96, MRM0r, + (ops R8 :$dst), "setbe $dst">, TB; // R8 = <= unsign +def SETBEm : I<0x96, MRM0m, + (ops i8mem:$dst), "setbe $dst">, TB; // [mem8] = <= unsign +def SETAr : I<0x97, MRM0r, + (ops R8 :$dst), "seta $dst">, TB; // R8 = > signed +def SETAm : I<0x97, MRM0m, + (ops i8mem:$dst), "seta $dst">, TB; // [mem8] = > signed +def SETSr : I<0x98, MRM0r, + (ops R8 :$dst), "sets $dst">, TB; // R8 = +def SETSm : I<0x98, MRM0m, + (ops i8mem:$dst), "sets $dst">, TB; // [mem8] = +def SETNSr : I<0x99, MRM0r, + (ops R8 :$dst), "setns $dst">, TB; // R8 = ! +def SETNSm : I<0x99, MRM0m, + (ops i8mem:$dst), "setns $dst">, TB; // [mem8] = ! +def SETPr : I<0x9A, MRM0r, + (ops R8 :$dst), "setp $dst">, TB; // R8 = parity +def SETPm : I<0x9A, MRM0m, + (ops i8mem:$dst), "setp $dst">, TB; // [mem8] = parity +def SETLr : I<0x9C, MRM0r, + (ops R8 :$dst), "setl $dst">, TB; // R8 = < signed +def SETLm : I<0x9C, MRM0m, + (ops i8mem:$dst), "setl $dst">, TB; // [mem8] = < signed +def SETGEr : I<0x9D, MRM0r, + (ops R8 :$dst), "setge $dst">, TB; // R8 = >= signed +def SETGEm : I<0x9D, MRM0m, + (ops i8mem:$dst), "setge $dst">, TB; // [mem8] = >= signed +def SETLEr : I<0x9E, MRM0r, + (ops R8 :$dst), "setle $dst">, TB; // R8 = <= signed +def SETLEm : I<0x9E, MRM0m, + (ops i8mem:$dst), "setle $dst">, TB; // [mem8] = <= signed +def SETGr : I<0x9F, MRM0r, + (ops R8 :$dst), "setg $dst">, TB; // R8 = < signed +def SETGm : I<0x9F, MRM0m, + (ops i8mem:$dst), "setg $dst">, TB; // [mem8] = < signed // Integer comparisons -def CMP8rr : I<0x38, MRMDestReg, (ops R8 :$src1, R8 :$src2), "cmp $src1, $src2">; -def CMP16rr : I<0x39, MRMDestReg, (ops R16:$src1, R16:$src2), "cmp $src1, $src2">, OpSize; -def CMP32rr : I<0x39, MRMDestReg, (ops R32:$src1, R32:$src2), "cmp $src1, $src2">; -def CMP8mr : I<0x38, MRMDestMem, (ops i8mem :$src1, R8 :$src2), "cmp $src1, $src2">; -def CMP16mr : I<0x39, MRMDestMem, (ops i16mem:$src1, R16:$src2), "cmp $src1, $src2">, OpSize; -def CMP32mr : I<0x39, MRMDestMem, (ops i32mem:$src1, R32:$src2), "cmp $src1, $src2">; -def CMP8rm : I<0x3A, MRMSrcMem , (ops R8 :$src1, i8mem :$src2), "cmp $src1, $src2">; -def CMP16rm : I<0x3B, MRMSrcMem , (ops R16:$src1, i16mem:$src2), "cmp $src1, $src2">, OpSize; -def CMP32rm : I<0x3B, MRMSrcMem , (ops R32:$src1, i32mem:$src2), "cmp $src1, $src2">; -def CMP8ri : Ii8 <0x80, MRM7r, (ops R16:$dst, i8imm:$src), "cmp $dst, $src">; -def CMP16ri : Ii16<0x81, MRM7r, (ops R16:$dst, i16imm:$src), "cmp $dst, $src">, OpSize; -def CMP32ri : Ii32<0x81, MRM7r, (ops R32:$dst, i32imm:$src), "cmp $dst, $src">; -def CMP8mi : Ii8 <0x80, MRM7m, (ops i8mem :$dst, i8imm :$src), "cmp $dst, $src">; -def CMP16mi : Ii16<0x81, MRM7m, (ops i16mem:$dst, i16imm:$src), "cmp $dst, $src">, OpSize; -def CMP32mi : Ii32<0x81, MRM7m, (ops i32mem:$dst, i32imm:$src), "cmp $dst, $src">; +def CMP8rr : I<0x38, MRMDestReg, + (ops R8 :$src1, R8 :$src2), + "cmp{b} {$src2, $src1|$src1, $src2}">; +def CMP16rr : I<0x39, MRMDestReg, + (ops R16:$src1, R16:$src2), + "cmp{w} {$src2, $src1|$src1, $src2}">, OpSize; +def CMP32rr : I<0x39, MRMDestReg, + (ops R32:$src1, R32:$src2), + "cmp{l} {$src2, $src1|$src1, $src2}">; +def CMP8mr : I<0x38, MRMDestMem, + (ops i8mem :$src1, R8 :$src2), + "cmp{b} {$src2, $src1|$src1, $src2}">; +def CMP16mr : I<0x39, MRMDestMem, + (ops i16mem:$src1, R16:$src2), + "cmp{w} {$src2, $src1|$src1, $src2}">, OpSize; +def CMP32mr : I<0x39, MRMDestMem, + (ops i32mem:$src1, R32:$src2), + "cmp{l} {$src2, $src1|$src1, $src2}">; +def CMP8rm : I<0x3A, MRMSrcMem, + (ops R8 :$src1, i8mem :$src2), + "cmp{b} {$src2, $src1|$src1, $src2}">; +def CMP16rm : I<0x3B, MRMSrcMem, + (ops R16:$src1, i16mem:$src2), + "cmp{w} {$src2, $src1|$src1, $src2}">, OpSize; +def CMP32rm : I<0x3B, MRMSrcMem, + (ops R32:$src1, i32mem:$src2), + "cmp{l} {$src2, $src1|$src1, $src2}">; +def CMP8ri : Ii8<0x80, MRM7r, + (ops R16:$src1, i8imm:$src2), + "cmp{b} {$src2, $src1|$src1, $src2}">; +def CMP16ri : Ii16<0x81, MRM7r, + (ops R16:$src1, i16imm:$src2), + "cmp{w} {$src2, $src1|$src1, $src2}">, OpSize; +def CMP32ri : Ii32<0x81, MRM7r, + (ops R32:$src1, i32imm:$src2), + "cmp{l} {$src2, $src1|$src1, $src2}">; +def CMP8mi : Ii8 <0x80, MRM7m, + (ops i8mem :$src1, i8imm :$src2), + "cmp{b} {$src2, $src1|$src1, $src2}">; +def CMP16mi : Ii16<0x81, MRM7m, + (ops i16mem:$src1, i16imm:$src2), + "cmp{w} {$src2, $src1|$src1, $src2}">, OpSize; +def CMP32mi : Ii32<0x81, MRM7m, + (ops i32mem:$src1, i32imm:$src2), + "cmp{l} {$src2, $src1|$src1, $src2}">; // Sign/Zero extenders -def MOVSX16rr8 : I<0xBE, MRMSrcReg, (ops R16:$dst, R8 :$src), "movsx $dst, $src">, TB, OpSize; -def MOVSX32rr8 : I<0xBE, MRMSrcReg, (ops R32:$dst, R8 :$src), "movsx $dst, $src">, TB; -def MOVSX32rr16: I<0xBF, MRMSrcReg, (ops R32:$dst, R16:$src), "movsx $dst, $src">, TB; -def MOVSX16rm8 : I<0xBE, MRMSrcMem, (ops R16:$dst, i8mem :$src), "movsx $dst, $src">, TB, OpSize; -def MOVSX32rm8 : I<0xBE, MRMSrcMem, (ops R32:$dst, i8mem :$src), "movsx $dst, $src">, TB; -def MOVSX32rm16: I<0xBF, MRMSrcMem, (ops R32:$dst, i16mem:$src), "movsx $dst, $src">, TB; - -def MOVZX16rr8 : I<0xB6, MRMSrcReg, (ops R16:$dst, R8 :$src), "movzx $dst, $src">, TB, OpSize; -def MOVZX32rr8 : I<0xB6, MRMSrcReg, (ops R32:$dst, R8 :$src), "movzx $dst, $src">, TB; -def MOVZX32rr16: I<0xB7, MRMSrcReg, (ops R32:$dst, R16:$src), "movzx $dst, $src">, TB; -def MOVZX16rm8 : I<0xB6, MRMSrcMem, (ops R16:$dst, i8mem :$src), "movzx $dst, $src">, TB, OpSize; -def MOVZX32rm8 : I<0xB6, MRMSrcMem, (ops R32:$dst, i8mem :$src), "movzx $dst, $src">, TB; -def MOVZX32rm16: I<0xB7, MRMSrcMem, (ops R32:$dst, i16mem:$src), "movzx $dst, $src">, TB; +def MOVSX16rr8 : I<0xBE, MRMSrcReg, (ops R16:$dst, R8 :$src), + "movs{bw|x} {$src, $dst|$dst, $src}">, TB, OpSize; +def MOVSX16rm8 : I<0xBE, MRMSrcMem, (ops R16:$dst, i8mem :$src), + "movs{bw|x} {$src, $dst|$dst, $src}">, TB, OpSize; +def MOVSX32rr8 : I<0xBE, MRMSrcReg, (ops R32:$dst, R8 :$src), + "movs{bl|x} {$src, $dst|$dst, $src}">, TB; +def MOVSX32rm8 : I<0xBE, MRMSrcMem, (ops R32:$dst, i8mem :$src), + "movs{bl|x} {$src, $dst|$dst, $src}">, TB; +def MOVSX32rr16: I<0xBF, MRMSrcReg, (ops R32:$dst, R16:$src), + "movs{wl|x} {$src, $dst|$dst, $src}">, TB; +def MOVSX32rm16: I<0xBF, MRMSrcMem, (ops R32:$dst, i16mem:$src), + "movs{wl|x} {$src, $dst|$dst, $src}">, TB; + +def MOVZX16rr8 : I<0xB6, MRMSrcReg, (ops R16:$dst, R8 :$src), + "movz{bw|x} {$src, $dst|$dst, $src}">, TB, OpSize; +def MOVZX16rm8 : I<0xB6, MRMSrcMem, (ops R16:$dst, i8mem :$src), + "movz{bw|x} {$src, $dst|$dst, $src}">, TB, OpSize; +def MOVZX32rr8 : I<0xB6, MRMSrcReg, (ops R32:$dst, R8 :$src), + "movz{bl|x} {$src, $dst|$dst, $src}">, TB; +def MOVZX32rm8 : I<0xB6, MRMSrcMem, (ops R32:$dst, i8mem :$src), + "movz{bl|x} {$src, $dst|$dst, $src}">, TB; +def MOVZX32rr16: I<0xB7, MRMSrcReg, (ops R32:$dst, R16:$src), + "movz{wl|x} {$src, $dst|$dst, $src}">, TB; +def MOVZX32rm16: I<0xB7, MRMSrcMem, (ops R32:$dst, i16mem:$src), + "movz{wl|x} {$src, $dst|$dst, $src}">, TB; //===----------------------------------------------------------------------===// @@ -879,11 +1206,16 @@ class FPI o, Format F, FPFormat fp, dag ops, string asm> // because they can be expanded by the fp spackifier into one of many different // forms of instructions for doing these operations. Until the stackifier runs, // we prefer to be abstract. -def FpMOV : FPI<0, Pseudo, SpecialFP, (ops RFP, RFP), "">; // f1 = fmov f2 -def FpADD : FPI<0, Pseudo, TwoArgFP , (ops RFP, RFP, RFP), "">; // f1 = fadd f2, f3 -def FpSUB : FPI<0, Pseudo, TwoArgFP , (ops RFP, RFP, RFP), "">; // f1 = fsub f2, f3 -def FpMUL : FPI<0, Pseudo, TwoArgFP , (ops RFP, RFP, RFP), "">; // f1 = fmul f2, f3 -def FpDIV : FPI<0, Pseudo, TwoArgFP , (ops RFP, RFP, RFP), "">; // f1 = fdiv f2, f3 +def FpMOV : FPI<0, Pseudo, SpecialFP, + (ops RFP, RFP), "">; // f1 = fmov f2 +def FpADD : FPI<0, Pseudo, TwoArgFP , + (ops RFP, RFP, RFP), "">; // f1 = fadd f2, f3 +def FpSUB : FPI<0, Pseudo, TwoArgFP , + (ops RFP, RFP, RFP), "">; // f1 = fsub f2, f3 +def FpMUL : FPI<0, Pseudo, TwoArgFP , + (ops RFP, RFP, RFP), "">; // f1 = fmul f2, f3 +def FpDIV : FPI<0, Pseudo, TwoArgFP , + (ops RFP, RFP, RFP), "">; // f1 = fdiv f2, f3 def FpGETRESULT : FPI<0, Pseudo, SpecialFP, (ops RFP), "">, Imp<[ST0], []>; // FPR = ST(0) @@ -891,63 +1223,95 @@ def FpGETRESULT : FPI<0, Pseudo, SpecialFP, (ops RFP), "">, def FpSETRESULT : FPI<0, Pseudo, SpecialFP, (ops RFP), "">, Imp<[], [ST0]>; // ST(0) = FPR -// FADD reg, mem: Before stackification, these are represented by: R1 = FADD* R2, [mem] -def FADD32m : FPI<0xD8, MRM0m, OneArgFPRW, (ops f32mem:$src), "fadd $src">; // ST(0) = ST(0) + [mem32real] -def FADD64m : FPI<0xDC, MRM0m, OneArgFPRW, (ops f64mem:$src), "fadd $src">; // ST(0) = ST(0) + [mem64real] +// FADD reg, mem: Before stackification, these are represented by: +// R1 = FADD* R2, [mem] +def FADD32m : FPI<0xD8, MRM0m, OneArgFPRW, // ST(0) = ST(0) + [mem32real] + (ops f32mem:$src), "fadd $src">; +def FADD64m : FPI<0xDC, MRM0m, OneArgFPRW, // ST(0) = ST(0) + [mem64real] + (ops f64mem:$src), "fadd $src">; + /* -def FIADD16m : FPI<0xDE, MRM0m, OneArgFPRW, (ops i16mem:$src), // ST(0) = ST(0) + [mem16int] - "fiadd $src">; -def FIADD32m : FPI<0xDA, MRM0m, OneArgFPRW, (ops i32mem:$src), // ST(0) = ST(0) + [mem32int] - "fiadd $src">; +def FIADD16m : FPI<0xDE, MRM0m, OneArgFPRW, // ST(0) = ST(0) + [mem16int] + (ops i16mem:$src), "fiadd $src">; +def FIADD32m : FPI<0xDA, MRM0m, OneArgFPRW, // ST(0) = ST(0) + [mem32int] + (ops i32mem:$src), "fiadd $src">; */ -// FMUL reg, mem: Before stackification, these are represented by: R1 = FMUL* R2, [mem] -def FMUL32m : FPI<0xD8, MRM1m, OneArgFPRW, (ops f32mem:$src), "fmul $src">; // ST(0) = ST(0) * [mem32real] -def FMUL64m : FPI<0xDC, MRM1m, OneArgFPRW, (ops f64mem:$src), "fmul $src">; // ST(0) = ST(0) * [mem64real] -//def FIMUL16m : FPI16m<"fimul", 0xDE, MRM1m, OneArgFPRW>; // ST(0) = ST(0) * [mem16int] -//def FIMUL32m : FPI32m<"fimul", 0xDA, MRM1m, OneArgFPRW>; // ST(0) = ST(0) * [mem32int] - -// FSUB reg, mem: Before stackification, these are represented by: R1 = FSUB* R2, [mem] -def FSUB32m : FPI<0xD8, MRM4m, OneArgFPRW, (ops f32mem:$src), "fsub $src">; // ST(0) = ST(0) - [mem32real] -def FSUB64m : FPI<0xDC, MRM4m, OneArgFPRW, (ops f64mem:$src), "fsub $src">; // ST(0) = ST(0) - [mem64real] -//def FISUB16m : FPI16m<"fisub", 0xDE, MRM4m, OneArgFPRW>; // ST(0) = ST(0) - [mem16int] -//def FISUB32m : FPI32m<"fisub", 0xDA, MRM4m, OneArgFPRW>; // ST(0) = ST(0) - [mem32int] - -// FSUBR reg, mem: Before stackification, these are represented by: R1 = FSUBR* R2, [mem] -// Note that the order of operands does not reflect the operation being performed. -def FSUBR32m : FPI<0xD8, MRM5m, OneArgFPRW, (ops f32mem:$src), "fsubr $src">; // ST(0) = [mem32real] - ST(0) -def FSUBR64m : FPI<0xDC, MRM5m, OneArgFPRW, (ops f64mem:$src), "fsubr $src">; // ST(0) = [mem64real] - ST(0) -//def FISUBR16m : FPI16m<"fisubr", 0xDE, MRM5m, OneArgFPRW>; // ST(0) = [mem16int] - ST(0) -//def FISUBR32m : FPI32m<"fisubr", 0xDA, MRM5m, OneArgFPRW>; // ST(0) = [mem32int] - ST(0) - -// FDIV reg, mem: Before stackification, these are represented by: R1 = FDIV* R2, [mem] -def FDIV32m : FPI<0xD8, MRM6m, OneArgFPRW, (ops f32mem:$src), "fdiv $src">; // ST(0) = ST(0) / [mem32real] -def FDIV64m : FPI<0xDC, MRM6m, OneArgFPRW, (ops f64mem:$src), "fdiv $src">; // ST(0) = ST(0) / [mem64real] -//def FIDIV16m : FPI16m<"fidiv", 0xDE, MRM6m, OneArgFPRW>; // ST(0) = ST(0) / [mem16int] -//def FIDIV32m : FPI32m<"fidiv", 0xDA, MRM6m, OneArgFPRW>; // ST(0) = ST(0) / [mem32int] - -// FDIVR reg, mem: Before stackification, these are represented by: R1 = FDIVR* R2, [mem] -// Note that the order of operands does not reflect the operation being performed. -def FDIVR32m : FPI<0xD8, MRM7m, OneArgFPRW, (ops f32mem:$src), "fdivr $src">; // ST(0) = [mem32real] / ST(0) -def FDIVR64m : FPI<0xDC, MRM7m, OneArgFPRW, (ops f64mem:$src), "fdivr $src">; // ST(0) = [mem64real] / ST(0) -//def FIDIVR16m : FPI16m<"fidivr", 0xDE, MRM7m, OneArgFPRW>; // ST(0) = [mem16int] / ST(0) -//def FIDIVR32m : FPI32m<"fidivr", 0xDA, MRM7m, OneArgFPRW>; // ST(0) = [mem32int] / ST(0) +// FMUL reg, mem: Before stackification, these are represented by: +// R1 = FMUL* R2, [mem] +def FMUL32m : FPI<0xD8, MRM1m, OneArgFPRW, // ST(0) = ST(0) * [mem32real] + (ops f32mem:$src), "fmul $src">; +def FMUL64m : FPI<0xDC, MRM1m, OneArgFPRW, // ST(0) = ST(0) * [mem64real] + (ops f64mem:$src), "fmul $src">; +// ST(0) = ST(0) * [mem16int] +//def FIMUL16m : FPI16m<"fimul", 0xDE, MRM1m, OneArgFPRW>; +// ST(0) = ST(0) * [mem32int] +//def FIMUL32m : FPI32m<"fimul", 0xDA, MRM1m, OneArgFPRW>; + +// FSUB reg, mem: Before stackification, these are represented by: +// R1 = FSUB* R2, [mem] +def FSUB32m : FPI<0xD8, MRM4m, OneArgFPRW, // ST(0) = ST(0) - [mem32real] + (ops f32mem:$src), "fsub $src">; +def FSUB64m : FPI<0xDC, MRM4m, OneArgFPRW, // ST(0) = ST(0) - [mem64real] + (ops f64mem:$src), "fsub $src">; +// ST(0) = ST(0) - [mem16int] +//def FISUB16m : FPI16m<"fisub", 0xDE, MRM4m, OneArgFPRW>; +// ST(0) = ST(0) - [mem32int] +//def FISUB32m : FPI32m<"fisub", 0xDA, MRM4m, OneArgFPRW>; + +// FSUBR reg, mem: Before stackification, these are represented by: +// R1 = FSUBR* R2, [mem] + +// Note that the order of operands does not reflect the operation being +// performed. +def FSUBR32m : FPI<0xD8, MRM5m, OneArgFPRW, // ST(0) = [mem32real] - ST(0) + (ops f32mem:$src), "fsubr $src">; +def FSUBR64m : FPI<0xDC, MRM5m, OneArgFPRW, // ST(0) = [mem64real] - ST(0) + (ops f64mem:$src), "fsubr $src">; +// ST(0) = [mem16int] - ST(0) +//def FISUBR16m : FPI16m<"fisubr", 0xDE, MRM5m, OneArgFPRW>; +// ST(0) = [mem32int] - ST(0) +//def FISUBR32m : FPI32m<"fisubr", 0xDA, MRM5m, OneArgFPRW>; + +// FDIV reg, mem: Before stackification, these are represented by: +// R1 = FDIV* R2, [mem] +def FDIV32m : FPI<0xD8, MRM6m, OneArgFPRW, // ST(0) = ST(0) / [mem32real] + (ops f32mem:$src), "fdiv $src">; +def FDIV64m : FPI<0xDC, MRM6m, OneArgFPRW, // ST(0) = ST(0) / [mem64real] + (ops f64mem:$src), "fdiv $src">; +// ST(0) = ST(0) / [mem16int] +//def FIDIV16m : FPI16m<"fidiv", 0xDE, MRM6m, OneArgFPRW>; +// ST(0) = ST(0) / [mem32int] +//def FIDIV32m : FPI32m<"fidiv", 0xDA, MRM6m, OneArgFPRW>; + +// FDIVR reg, mem: Before stackification, these are represented by: +// R1 = FDIVR* R2, [mem] +// Note that the order of operands does not reflect the operation being +// performed. +def FDIVR32m : FPI<0xD8, MRM7m, OneArgFPRW, // ST(0) = [mem32real] / ST(0) + (ops f32mem:$src), "fdivr $src">; +def FDIVR64m : FPI<0xDC, MRM7m, OneArgFPRW, // ST(0) = [mem64real] / ST(0) + (ops f64mem:$src), "fdivr $src">; +// ST(0) = [mem16int] / ST(0) +//def FIDIVR16m : FPI16m<"fidivr", 0xDE, MRM7m, OneArgFPRW>; +// ST(0) = [mem32int] / ST(0) +//def FIDIVR32m : FPI32m<"fidivr", 0xDA, MRM7m, OneArgFPRW>; // Floating point cmovs... let isTwoAddress = 1, Uses = [ST0], Defs = [ST0] in { def FCMOVB : FPI<0xC0, AddRegFrm, CondMovFP, - (ops RST:$op), "fcmovb %ST(0), $op">, DA; + (ops RST:$op), "fcmovb {$op, %ST(0)|%ST(0), $op}">, DA; def FCMOVBE : FPI<0xD0, AddRegFrm, CondMovFP, - (ops RST:$op), "fcmovbe %ST(0), $op">, DA; + (ops RST:$op), "fcmovbe {$op, %ST(0)|%ST(0), $op}">, DA; def FCMOVE : FPI<0xC8, AddRegFrm, CondMovFP, - (ops RST:$op), "fcmove %ST(0), $op">, DA; + (ops RST:$op), "fcmove {$op, %ST(0)|%ST(0), $op}">, DA; def FCMOVAE : FPI<0xC0, AddRegFrm, CondMovFP, - (ops RST:$op), "fcmovae %ST(0), $op">, DB; + (ops RST:$op), "fcmovae {$op, %ST(0)|%ST(0), $op}">, DB; def FCMOVA : FPI<0xD0, AddRegFrm, CondMovFP, - (ops RST:$op), "fcmova %ST(0), $op">, DB; + (ops RST:$op), "fcmova {$op, %ST(0)|%ST(0), $op}">, DB; def FCMOVNE : FPI<0xC8, AddRegFrm, CondMovFP, - (ops RST:$op), "fcmovne %ST(0), $op">, DB; + (ops RST:$op), "fcmovne {$op, %ST(0)|%ST(0), $op}">, DB; } // Floating point loads & stores... @@ -973,7 +1337,8 @@ def FISTP16m : FPI<0xDF, MRM3m , NotFP , (ops i16mem:$op), "fistp $op">; def FISTP32m : FPI<0xDB, MRM3m , NotFP , (ops i32mem:$op), "fistp $op">; def FISTP64m : FPI<0xDF, MRM7m , OneArgFP, (ops i64mem:$op), "fistpll $op">; -def FXCH : FPI<0xC8, AddRegFrm, NotFP, (ops RST:$op), "fxch $op">, D9; // fxch ST(i), ST(0) +def FXCH : FPI<0xC8, AddRegFrm, NotFP, + (ops RST:$op), "fxch $op">, D9; // fxch ST(i), ST(0) // Floating point constant loads... def FLD0 : FPI<0xEE, RawFrm, ZeroArgFP, (ops), "fldz">, D9; @@ -981,61 +1346,90 @@ def FLD1 : FPI<0xE8, RawFrm, ZeroArgFP, (ops), "fld1">, D9; // Unary operations... -def FCHS : FPI<0xE0, RawFrm, OneArgFPRW, (ops), "fchs">, D9; // f1 = fchs f2 -def FTST : FPI<0xE4, RawFrm, OneArgFP, (ops), "ftst">, D9; // ftst ST(0) +def FCHS : FPI<0xE0, RawFrm, OneArgFPRW, (ops), "fchs">, D9; // f1 = fchs f2 +def FTST : FPI<0xE4, RawFrm, OneArgFP, (ops), "ftst">, D9; // ftst ST(0) // Binary arithmetic operations... -class FPST0rInst o, dag ops, string asm> : I, D8 { +class FPST0rInst o, dag ops, string asm> + : I, D8 { list Uses = [ST0]; list Defs = [ST0]; } -class FPrST0Inst o, dag ops, string asm> : I, DC { +class FPrST0Inst o, dag ops, string asm> + : I, DC { list Uses = [ST0]; } -class FPrST0PInst o, dag ops, string asm> : I, DE { +class FPrST0PInst o, dag ops, string asm> + : I, DE { list Uses = [ST0]; } -def FADDST0r : FPST0rInst <0xC0, (ops RST:$op), "fadd $op">; -def FADDrST0 : FPrST0Inst <0xC0, (ops RST:$op), "fadd $op, %ST(0)">; -def FADDPrST0 : FPrST0PInst<0xC0, (ops RST:$op), "faddp $op">; - -def FSUBRST0r : FPST0rInst <0xE8, (ops RST:$op), "fsubr $op">; -def FSUBrST0 : FPrST0Inst <0xE8, (ops RST:$op), "fsub $op, %ST(0)">; -def FSUBPrST0 : FPrST0PInst<0xE8, (ops RST:$op), "fsubp $op">; - -def FSUBST0r : FPST0rInst <0xE0, (ops RST:$op), "fsub $op">; -def FSUBRrST0 : FPrST0Inst <0xE0, (ops RST:$op), "fsubr $op, %ST(0)">; -def FSUBRPrST0 : FPrST0PInst<0xE0, (ops RST:$op), "fsubrp $op">; - -def FMULST0r : FPST0rInst <0xC8, (ops RST:$op), "fmul $op">; -def FMULrST0 : FPrST0Inst <0xC8, (ops RST:$op), "fmul $op, %ST(0)">; -def FMULPrST0 : FPrST0PInst<0xC8, (ops RST:$op), "fmulp $op">; - -def FDIVRST0r : FPST0rInst <0xF8, (ops RST:$op), "fdivr $op">; -def FDIVrST0 : FPrST0Inst <0xF8, (ops RST:$op), "fdiv $op, %ST(0)">; -def FDIVPrST0 : FPrST0PInst<0xF8, (ops RST:$op), "fdivp $op">; - -def FDIVST0r : FPST0rInst <0xF0, (ops RST:$op), "fdiv $op">; // ST(0) = ST(0) / ST(i) -def FDIVRrST0 : FPrST0Inst <0xF0, (ops RST:$op), "fdivr $op, %ST(0)">; // ST(i) = ST(0) / ST(i) -def FDIVRPrST0 : FPrST0PInst<0xF0, (ops RST:$op), "fdivrp $op">; // ST(i) = ST(0) / ST(i), pop +def FADDST0r : FPST0rInst <0xC0, (ops RST:$op), + "fadd $op">; +def FADDrST0 : FPrST0Inst <0xC0, (ops RST:$op), + "fadd {%ST(0), $op|$op, %ST(0)}">; +def FADDPrST0 : FPrST0PInst<0xC0, (ops RST:$op), + "faddp $op">; + +def FSUBRST0r : FPST0rInst <0xE8, (ops RST:$op), + "fsubr $op">; +def FSUBrST0 : FPrST0Inst <0xE8, (ops RST:$op), + "fsub {%ST(0), $op|$op, %ST(0)}">; +def FSUBPrST0 : FPrST0PInst<0xE8, (ops RST:$op), + "fsubp $op">; + +def FSUBST0r : FPST0rInst <0xE0, (ops RST:$op), + "fsub $op">; +def FSUBRrST0 : FPrST0Inst <0xE0, (ops RST:$op), + "fsubr {%ST(0), $op|$op, %ST(0)}">; +def FSUBRPrST0 : FPrST0PInst<0xE0, (ops RST:$op), + "fsubrp $op">; + +def FMULST0r : FPST0rInst <0xC8, (ops RST:$op), + "fmul $op">; +def FMULrST0 : FPrST0Inst <0xC8, (ops RST:$op), + "fmul {%ST(0), $op|$op, %ST(0)}">; +def FMULPrST0 : FPrST0PInst<0xC8, (ops RST:$op), + "fmulp $op">; + +def FDIVRST0r : FPST0rInst <0xF8, (ops RST:$op), + "fdivr $op">; +def FDIVrST0 : FPrST0Inst <0xF8, (ops RST:$op), + "fdiv {%ST(0), $op|$op, %ST(0)}">; +def FDIVPrST0 : FPrST0PInst<0xF8, (ops RST:$op), + "fdivp $op">; + +def FDIVST0r : FPST0rInst <0xF0, (ops RST:$op), // ST(0) = ST(0) / ST(i) + "fdiv $op">; +def FDIVRrST0 : FPrST0Inst <0xF0, (ops RST:$op), // ST(i) = ST(0) / ST(i) + "fdivr {%ST(0), $op|$op, %ST(0)}">; +def FDIVRPrST0 : FPrST0PInst<0xF0, (ops RST:$op), // ST(i) = ST(0) / ST(i), pop + "fdivrp $op">; // Floating point compares -def FUCOMr : FPI<0xE0, AddRegFrm, CompareFP, (ops RST:$reg), - "fucom $reg">, DD, Imp<[ST0],[]>; // FPSW = compare ST(0) with ST(i) -def FUCOMPr : I<0xE8, AddRegFrm, (ops RST:$reg), - "fucomp $reg">, DD, Imp<[ST0],[]>; // FPSW = compare ST(0) with ST(i), pop -def FUCOMPPr : I<0xE9, RawFrm, (ops), - "fucompp">, DA, Imp<[ST0],[]>; // compare ST(0) with ST(1), pop, pop - -def FUCOMIr : FPI<0xE8, AddRegFrm, CompareFP, (ops RST:$reg), - "fucomi %ST(0), $reg">, DB, Imp<[ST0],[]>; // CC = compare ST(0) with ST(i) -def FUCOMIPr : I<0xE8, AddRegFrm, (ops RST:$reg), - "fucomip %ST(0), $reg">, DF, Imp<[ST0],[]>; // CC = compare ST(0) with ST(i), pop +def FUCOMr : FPI<0xE0, AddRegFrm, CompareFP, // FPSW = cmp ST(0) with ST(i) + (ops RST:$reg), + "fucom $reg">, DD, Imp<[ST0],[]>; +def FUCOMPr : I<0xE8, AddRegFrm, + (ops RST:$reg), // FPSW = cmp ST(0) with ST(i), pop + "fucomp $reg">, DD, Imp<[ST0],[]>; +def FUCOMPPr : I<0xE9, RawFrm, + (ops), // cmp ST(0) with ST(1), pop, pop + "fucompp">, DA, Imp<[ST0],[]>; + +def FUCOMIr : FPI<0xE8, AddRegFrm, CompareFP, // CC = cmp ST(0) with ST(i) + (ops RST:$reg), + "fucomi {$reg, %ST(0)|%ST(0), $reg}">, DB, Imp<[ST0],[]>; +def FUCOMIPr : I<0xE8, AddRegFrm, // CC = cmp ST(0) with ST(i), pop + (ops RST:$reg), + "fucomip {$reg, %ST(0)|%ST(0), $reg}">, DF, Imp<[ST0],[]>; // Floating point flag ops -def FNSTSW8r : I<0xE0, RawFrm, (ops), "fnstsw">, DF, Imp<[],[AX]>; // AX = fp flags +def FNSTSW8r : I<0xE0, RawFrm, // AX = fp flags + (ops), "fnstsw">, DF, Imp<[],[AX]>; -def FNSTCW16m : I<0xD9, MRM7m, (ops i16mem:$dst), "fnstcw $dst">; // [mem16] = X87 control world -def FLDCW16m : I<0xD9, MRM5m, (ops i16mem:$dst), "fldcw $dst">; // X87 control world = [mem16] +def FNSTCW16m : I<0xD9, MRM7m, // [mem16] = X87 control world + (ops i16mem:$dst), "fnstcw $dst">; +def FLDCW16m : I<0xD9, MRM5m, // X87 control world = [mem16] + (ops i16mem:$dst), "fldcw $dst">; -- 2.34.1