X-Git-Url: http://demsky.eecs.uci.edu/git/?a=blobdiff_plain;f=test%2FCodeGen%2FAArch64%2Fldst-unscaledimm.ll;h=a3625975822c000c3934f78524c6f0c6888993e6;hb=44d5618b02667e8b939453d1e87cea7374b8cef1;hp=af4d7d721fe6e326c7f49a0225e4600e61672289;hpb=198d8baafbfdfcf5a5e219602a5d94ed263973b4;p=oota-llvm.git diff --git a/test/CodeGen/AArch64/ldst-unscaledimm.ll b/test/CodeGen/AArch64/ldst-unscaledimm.ll index af4d7d721fe..a3625975822 100644 --- a/test/CodeGen/AArch64/ldst-unscaledimm.ll +++ b/test/CodeGen/AArch64/ldst-unscaledimm.ll @@ -16,32 +16,32 @@ define void @ldst_8bit() { ; No architectural support for loads to 16-bit or 8-bit since we ; promote i8 during lowering. - %addr_8bit = load i8** @varptr + %addr_8bit = load i8*, i8** @varptr ; match a sign-extending load 8-bit -> 32-bit %addr_sext32 = getelementptr i8, i8* %addr_8bit, i64 -256 - %val8_sext32 = load volatile i8* %addr_sext32 + %val8_sext32 = load volatile i8, i8* %addr_sext32 %val32_signed = sext i8 %val8_sext32 to i32 store volatile i32 %val32_signed, i32* @var_32bit ; CHECK: ldursb {{w[0-9]+}}, [{{x[0-9]+}}, #-256] ; match a zero-extending load volatile 8-bit -> 32-bit %addr_zext32 = getelementptr i8, i8* %addr_8bit, i64 -12 - %val8_zext32 = load volatile i8* %addr_zext32 + %val8_zext32 = load volatile i8, i8* %addr_zext32 %val32_unsigned = zext i8 %val8_zext32 to i32 store volatile i32 %val32_unsigned, i32* @var_32bit ; CHECK: ldurb {{w[0-9]+}}, [{{x[0-9]+}}, #-12] ; match an any-extending load volatile 8-bit -> 32-bit %addr_anyext = getelementptr i8, i8* %addr_8bit, i64 -1 - %val8_anyext = load volatile i8* %addr_anyext + %val8_anyext = load volatile i8, i8* %addr_anyext %newval8 = add i8 %val8_anyext, 1 store volatile i8 %newval8, i8* @var_8bit ; CHECK: ldurb {{w[0-9]+}}, [{{x[0-9]+}}, #-1] ; match a sign-extending load volatile 8-bit -> 64-bit %addr_sext64 = getelementptr i8, i8* %addr_8bit, i64 -5 - %val8_sext64 = load volatile i8* %addr_sext64 + %val8_sext64 = load volatile i8, i8* %addr_sext64 %val64_signed = sext i8 %val8_sext64 to i64 store volatile i64 %val64_signed, i64* @var_64bit ; CHECK: ldursb {{x[0-9]+}}, [{{x[0-9]+}}, #-5] @@ -50,21 +50,21 @@ define void @ldst_8bit() { ; This uses the fact that ldrb w0, [x0] will zero out the high 32-bits ; of x0 so it's identical to load volatileing to 32-bits. %addr_zext64 = getelementptr i8, i8* %addr_8bit, i64 -9 - %val8_zext64 = load volatile i8* %addr_zext64 + %val8_zext64 = load volatile i8, i8* %addr_zext64 %val64_unsigned = zext i8 %val8_zext64 to i64 store volatile i64 %val64_unsigned, i64* @var_64bit ; CHECK: ldurb {{w[0-9]+}}, [{{x[0-9]+}}, #-9] ; truncating store volatile 32-bits to 8-bits %addr_trunc32 = getelementptr i8, i8* %addr_8bit, i64 -256 - %val32 = load volatile i32* @var_32bit + %val32 = load volatile i32, i32* @var_32bit %val8_trunc32 = trunc i32 %val32 to i8 store volatile i8 %val8_trunc32, i8* %addr_trunc32 ; CHECK: sturb {{w[0-9]+}}, [{{x[0-9]+}}, #-256] ; truncating store volatile 64-bits to 8-bits %addr_trunc64 = getelementptr i8, i8* %addr_8bit, i64 -1 - %val64 = load volatile i64* @var_64bit + %val64 = load volatile i64, i64* @var_64bit %val8_trunc64 = trunc i64 %val64 to i8 store volatile i8 %val8_trunc64, i8* %addr_trunc64 ; CHECK: sturb {{w[0-9]+}}, [{{x[0-9]+}}, #-1] @@ -77,12 +77,12 @@ define void @ldst_16bit() { ; No architectural support for loads to 16-bit or 16-bit since we ; promote i16 during lowering. - %addr_8bit = load i8** @varptr + %addr_8bit = load i8*, i8** @varptr ; match a sign-extending load 16-bit -> 32-bit %addr8_sext32 = getelementptr i8, i8* %addr_8bit, i64 -256 %addr_sext32 = bitcast i8* %addr8_sext32 to i16* - %val16_sext32 = load volatile i16* %addr_sext32 + %val16_sext32 = load volatile i16, i16* %addr_sext32 %val32_signed = sext i16 %val16_sext32 to i32 store volatile i32 %val32_signed, i32* @var_32bit ; CHECK: ldursh {{w[0-9]+}}, [{{x[0-9]+}}, #-256] @@ -90,7 +90,7 @@ define void @ldst_16bit() { ; match a zero-extending load volatile 16-bit -> 32-bit. With offset that would be unaligned. %addr8_zext32 = getelementptr i8, i8* %addr_8bit, i64 15 %addr_zext32 = bitcast i8* %addr8_zext32 to i16* - %val16_zext32 = load volatile i16* %addr_zext32 + %val16_zext32 = load volatile i16, i16* %addr_zext32 %val32_unsigned = zext i16 %val16_zext32 to i32 store volatile i32 %val32_unsigned, i32* @var_32bit ; CHECK: ldurh {{w[0-9]+}}, [{{x[0-9]+}}, #15] @@ -98,7 +98,7 @@ define void @ldst_16bit() { ; match an any-extending load volatile 16-bit -> 32-bit %addr8_anyext = getelementptr i8, i8* %addr_8bit, i64 -1 %addr_anyext = bitcast i8* %addr8_anyext to i16* - %val16_anyext = load volatile i16* %addr_anyext + %val16_anyext = load volatile i16, i16* %addr_anyext %newval16 = add i16 %val16_anyext, 1 store volatile i16 %newval16, i16* @var_16bit ; CHECK: ldurh {{w[0-9]+}}, [{{x[0-9]+}}, #-1] @@ -106,7 +106,7 @@ define void @ldst_16bit() { ; match a sign-extending load volatile 16-bit -> 64-bit %addr8_sext64 = getelementptr i8, i8* %addr_8bit, i64 -5 %addr_sext64 = bitcast i8* %addr8_sext64 to i16* - %val16_sext64 = load volatile i16* %addr_sext64 + %val16_sext64 = load volatile i16, i16* %addr_sext64 %val64_signed = sext i16 %val16_sext64 to i64 store volatile i64 %val64_signed, i64* @var_64bit ; CHECK: ldursh {{x[0-9]+}}, [{{x[0-9]+}}, #-5] @@ -116,7 +116,7 @@ define void @ldst_16bit() { ; of x0 so it's identical to load volatileing to 32-bits. %addr8_zext64 = getelementptr i8, i8* %addr_8bit, i64 9 %addr_zext64 = bitcast i8* %addr8_zext64 to i16* - %val16_zext64 = load volatile i16* %addr_zext64 + %val16_zext64 = load volatile i16, i16* %addr_zext64 %val64_unsigned = zext i16 %val16_zext64 to i64 store volatile i64 %val64_unsigned, i64* @var_64bit ; CHECK: ldurh {{w[0-9]+}}, [{{x[0-9]+}}, #9] @@ -124,7 +124,7 @@ define void @ldst_16bit() { ; truncating store volatile 32-bits to 16-bits %addr8_trunc32 = getelementptr i8, i8* %addr_8bit, i64 -256 %addr_trunc32 = bitcast i8* %addr8_trunc32 to i16* - %val32 = load volatile i32* @var_32bit + %val32 = load volatile i32, i32* @var_32bit %val16_trunc32 = trunc i32 %val32 to i16 store volatile i16 %val16_trunc32, i16* %addr_trunc32 ; CHECK: sturh {{w[0-9]+}}, [{{x[0-9]+}}, #-256] @@ -132,7 +132,7 @@ define void @ldst_16bit() { ; truncating store volatile 64-bits to 16-bits %addr8_trunc64 = getelementptr i8, i8* %addr_8bit, i64 -1 %addr_trunc64 = bitcast i8* %addr8_trunc64 to i16* - %val64 = load volatile i64* @var_64bit + %val64 = load volatile i64, i64* @var_64bit %val16_trunc64 = trunc i64 %val64 to i16 store volatile i16 %val16_trunc64, i16* %addr_trunc64 ; CHECK: sturh {{w[0-9]+}}, [{{x[0-9]+}}, #-1] @@ -143,12 +143,12 @@ define void @ldst_16bit() { define void @ldst_32bit() { ; CHECK-LABEL: ldst_32bit: - %addr_8bit = load i8** @varptr + %addr_8bit = load i8*, i8** @varptr ; Straight 32-bit load/store %addr32_8_noext = getelementptr i8, i8* %addr_8bit, i64 1 %addr32_noext = bitcast i8* %addr32_8_noext to i32* - %val32_noext = load volatile i32* %addr32_noext + %val32_noext = load volatile i32, i32* %addr32_noext store volatile i32 %val32_noext, i32* %addr32_noext ; CHECK: ldur {{w[0-9]+}}, [{{x[0-9]+}}, #1] ; CHECK: stur {{w[0-9]+}}, [{{x[0-9]+}}, #1] @@ -156,7 +156,7 @@ define void @ldst_32bit() { ; Zero-extension to 64-bits %addr32_8_zext = getelementptr i8, i8* %addr_8bit, i64 -256 %addr32_zext = bitcast i8* %addr32_8_zext to i32* - %val32_zext = load volatile i32* %addr32_zext + %val32_zext = load volatile i32, i32* %addr32_zext %val64_unsigned = zext i32 %val32_zext to i64 store volatile i64 %val64_unsigned, i64* @var_64bit ; CHECK: ldur {{w[0-9]+}}, [{{x[0-9]+}}, #-256] @@ -165,7 +165,7 @@ define void @ldst_32bit() { ; Sign-extension to 64-bits %addr32_8_sext = getelementptr i8, i8* %addr_8bit, i64 -12 %addr32_sext = bitcast i8* %addr32_8_sext to i32* - %val32_sext = load volatile i32* %addr32_sext + %val32_sext = load volatile i32, i32* %addr32_sext %val64_signed = sext i32 %val32_sext to i64 store volatile i64 %val64_signed, i64* @var_64bit ; CHECK: ldursw {{x[0-9]+}}, [{{x[0-9]+}}, #-12] @@ -177,7 +177,7 @@ define void @ldst_32bit() { %addr32_8_trunc = getelementptr i8, i8* %addr_8bit, i64 -20 %addr32_trunc = bitcast i8* %addr32_8_trunc to i32* - %val64_trunc = load volatile i64* %addr64_trunc + %val64_trunc = load volatile i64, i64* %addr64_trunc %val32_trunc = trunc i64 %val64_trunc to i32 store volatile i32 %val32_trunc, i32* %addr32_trunc ; CHECK: ldur {{x[0-9]+}}, [{{x[0-9]+}}, #255] @@ -189,11 +189,11 @@ define void @ldst_32bit() { define void @ldst_float() { ; CHECK-LABEL: ldst_float: - %addr_8bit = load i8** @varptr + %addr_8bit = load i8*, i8** @varptr %addrfp_8 = getelementptr i8, i8* %addr_8bit, i64 -5 %addrfp = bitcast i8* %addrfp_8 to float* - %valfp = load volatile float* %addrfp + %valfp = load volatile float, float* %addrfp ; CHECK: ldur {{s[0-9]+}}, [{{x[0-9]+}}, #-5] ; CHECK-NOFP-NOT: ldur {{s[0-9]+}}, @@ -207,11 +207,11 @@ define void @ldst_float() { define void @ldst_double() { ; CHECK-LABEL: ldst_double: - %addr_8bit = load i8** @varptr + %addr_8bit = load i8*, i8** @varptr %addrfp_8 = getelementptr i8, i8* %addr_8bit, i64 4 %addrfp = bitcast i8* %addrfp_8 to double* - %valfp = load volatile double* %addrfp + %valfp = load volatile double, double* %addrfp ; CHECK: ldur {{d[0-9]+}}, [{{x[0-9]+}}, #4] ; CHECK-NOFP-NOT: ldur {{d[0-9]+}},