llvm.org GIT mirror llvm / a813b38 test / CodeGen / SystemZ / vec-abs-04.ll
a813b38

Tree @a813b38 (Download .tar.gz)

vec-abs-04.ll @a813b38raw · history · blame

; Test v2i64 absolute.
;
; RUN: llc < %s -mtriple=s390x-linux-gnu -mcpu=z13 | FileCheck %s

; Test with slt.
define <2 x i64> @f1(<2 x i64> %val) {
; CHECK-LABEL: f1:
; CHECK: vlpg %v24, %v24
; CHECK: br %r14
  %cmp = icmp slt <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %neg, <2 x i64> %val
  ret <2 x i64> %ret
}

; Test with sle.
define <2 x i64> @f2(<2 x i64> %val) {
; CHECK-LABEL: f2:
; CHECK: vlpg %v24, %v24
; CHECK: br %r14
  %cmp = icmp sle <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %neg, <2 x i64> %val
  ret <2 x i64> %ret
}

; Test with sgt.
define <2 x i64> @f3(<2 x i64> %val) {
; CHECK-LABEL: f3:
; CHECK: vlpg %v24, %v24
; CHECK: br %r14
  %cmp = icmp sgt <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %val, <2 x i64> %neg
  ret <2 x i64> %ret
}

; Test with sge.
define <2 x i64> @f4(<2 x i64> %val) {
; CHECK-LABEL: f4:
; CHECK: vlpg %v24, %v24
; CHECK: br %r14
  %cmp = icmp sge <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %val, <2 x i64> %neg
  ret <2 x i64> %ret
}

; Test that negative absolute uses VLPG too.  There is no vector equivalent
; of LOAD NEGATIVE.
define <2 x i64> @f5(<2 x i64> %val) {
; CHECK-LABEL: f5:
; CHECK: vlpg [[REG:%v[0-9]+]], %v24
; CHECK: vlcg %v24, [[REG]]
; CHECK: br %r14
  %cmp = icmp slt <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %abs = select <2 x i1> %cmp, <2 x i64> %neg, <2 x i64> %val
  %ret = sub <2 x i64> zeroinitializer, %abs
  ret <2 x i64> %ret
}

; Try another form of negative absolute (slt version).
define <2 x i64> @f6(<2 x i64> %val) {
; CHECK-LABEL: f6:
; CHECK: vlpg [[REG:%v[0-9]+]], %v24
; CHECK: vlcg %v24, [[REG]]
; CHECK: br %r14
  %cmp = icmp slt <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %val, <2 x i64> %neg
  ret <2 x i64> %ret
}

; Test with sle.
define <2 x i64> @f7(<2 x i64> %val) {
; CHECK-LABEL: f7:
; CHECK: vlpg [[REG:%v[0-9]+]], %v24
; CHECK: vlcg %v24, [[REG]]
; CHECK: br %r14
  %cmp = icmp sle <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %val, <2 x i64> %neg
  ret <2 x i64> %ret
}

; Test with sgt.
define <2 x i64> @f8(<2 x i64> %val) {
; CHECK-LABEL: f8:
; CHECK: vlpg [[REG:%v[0-9]+]], %v24
; CHECK: vlcg %v24, [[REG]]
; CHECK: br %r14
  %cmp = icmp sgt <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %neg, <2 x i64> %val
  ret <2 x i64> %ret
}

; Test with sge.
define <2 x i64> @f9(<2 x i64> %val) {
; CHECK-LABEL: f9:
; CHECK: vlpg [[REG:%v[0-9]+]], %v24
; CHECK: vlcg %v24, [[REG]]
; CHECK: br %r14
  %cmp = icmp sge <2 x i64> %val, zeroinitializer
  %neg = sub <2 x i64> zeroinitializer, %val
  %ret = select <2 x i1> %cmp, <2 x i64> %neg, <2 x i64> %val
  ret <2 x i64> %ret
}

; Test with an SRA-based boolean vector.
define <2 x i64> @f10(<2 x i64> %val) {
; CHECK-LABEL: f10:
; CHECK: vlpg %v24, %v24
; CHECK: br %r14
  %shr = ashr <2 x i64> %val, <i64 63, i64 63>
  %neg = sub <2 x i64> zeroinitializer, %val
  %and1 = and <2 x i64> %shr, %neg
  %not = xor <2 x i64> %shr, <i64 -1, i64 -1>
  %and2 = and <2 x i64> %not, %val
  %ret = or <2 x i64> %and1, %and2
  ret <2 x i64> %ret
}

; ...and again in reverse
define <2 x i64> @f11(<2 x i64> %val) {
; CHECK-LABEL: f11:
; CHECK: vlpg [[REG:%v[0-9]+]], %v24
; CHECK: vlcg %v24, [[REG]]
; CHECK: br %r14
  %shr = ashr <2 x i64> %val, <i64 63, i64 63>
  %and1 = and <2 x i64> %shr, %val
  %not = xor <2 x i64> %shr, <i64 -1, i64 -1>
  %neg = sub <2 x i64> zeroinitializer, %val
  %and2 = and <2 x i64> %not, %neg
  %ret = or <2 x i64> %and1, %and2
  ret <2 x i64> %ret
}