; CHECK: 7c: 5d
; CHECK: 7d: c3
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 0, i32 0)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 1, i32 1)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 2, i32 2)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 3, i32 3)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 4, i32 4)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 5, i32 5)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 6, i32 6)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 7, i32 7)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 8, i32 8)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 9, i32 9)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 10, i32 10)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 11, i32 11)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 12, i32 12)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 13, i32 13)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 14, i32 14)
- tail call void (i32, i32, ...)* @llvm.experimental.stackmap(i32 15, i32 15)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 0, i32 0)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 1, i32 1)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 2, i32 2)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 3, i32 3)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 4, i32 4)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 5, i32 5)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 6, i32 6)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 7, i32 7)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 8, i32 8)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 9, i32 9)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 10, i32 10)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 11, i32 11)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 12, i32 12)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 13, i32 13)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 14, i32 14)
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 15, i32 15)
+; Add an extra stackmap with a zero-length shadow to thwart the shadow
+; optimization. This will force all 15 bytes of the previous shadow to be
+; padded with nops.
+ tail call void (i64, i32, ...)* @llvm.experimental.stackmap(i64 16, i32 0)
ret void
}
-declare void @llvm.experimental.stackmap(i32, i32, ...)
+declare void @llvm.experimental.stackmap(i64, i32, ...)