Tree @release_60 (Download .tar.gz)
- ..
- GlobalISel
- 32-bit-local-address-space.ll
- add-debug.ll
- add.i16.ll
- add.ll
- add.v2i16.ll
- add_i128.ll
- add_i64.ll
- addrspacecast-captured.ll
- addrspacecast-constantexpr.ll
- addrspacecast.ll
- adjust-writemask-invalid-copy.ll
- alignbit-pat.ll
- always-uniform.ll
- amdgcn.bitcast.ll
- amdgcn.private-memory.ll
- amdgpu-alias-analysis.ll
- amdgpu-codegenprepare-fdiv.ll
- amdgpu-codegenprepare-i16-to-i32.ll
- amdgpu-inline.ll
- amdgpu-shader-calling-convention.ll
- amdgpu.private-memory.ll
- amdgpu.work-item-intrinsics.deprecated.ll
- amdpal-cs.ll
- amdpal-es.ll
- amdpal-gs.ll
- amdpal-hs.ll
- amdpal-ls.ll
- amdpal-ps.ll
- amdpal-psenable.ll
- amdpal-vs.ll
- amdpal.ll
- and-gcn.ll
- and.ll
- annotate-kernel-features-hsa-call.ll
- annotate-kernel-features-hsa.ll
- annotate-kernel-features.ll
- anonymous-gv.ll
- any_extend_vector_inreg.ll
- anyext.ll
- array-ptr-calc-i32.ll
- array-ptr-calc-i64.ll
- ashr.v2i16.ll
- atomic_cmp_swap_local.ll
- atomic_load_add.ll
- atomic_load_sub.ll
- attr-amdgpu-flat-work-group-size.ll
- attr-amdgpu-num-sgpr.ll
- attr-amdgpu-num-vgpr.ll
- attr-amdgpu-waves-per-eu.ll
- attr-unparseable.ll
- barrier-elimination.ll
- basic-branch.ll
- basic-call-return.ll
- basic-loop.ll
- bfe-combine.ll
- bfe-patterns.ll
- bfe_uint.ll
- bfi_int.ll
- bfm.ll
- big_alu.ll
- bitcast-vector-extract.ll
- bitreverse-inline-immediates.ll
- bitreverse.ll
- br_cc.f16.ll
- branch-condition-and.ll
- branch-relax-bundle.ll
- branch-relax-spill.ll
- branch-relaxation.ll
- branch-uniformity.ll
- break-smem-soft-clauses.mir
- break-vmem-soft-clauses.mir
- bswap.ll
- bug-vopc-commute.ll
- build_vector.ll
- byval-frame-setup.ll
- call-argument-types.ll
- call-encoding.ll
- call-graph-register-usage.ll
- call-preserved-registers.ll
- call-return-types.ll
- call_fs.ll
- callee-frame-setup.ll
- callee-special-input-sgprs.ll
- callee-special-input-vgprs.ll
- calling-conventions.ll
- captured-frame-index.ll
- cayman-loop-bug.ll
- cf-loop-on-constant.ll
- cf-stack-bug.ll
- cf_end.ll
- cgp-addressing-modes-flat.ll
- cgp-addressing-modes.ll
- cgp-bitfield-extract.ll
- clamp-modifier.ll
- clamp-omod-special-case.mir
- clamp.ll
- cluster-flat-loads-postra.mir
- cluster-flat-loads.mir
- cndmask-no-def-vcc.ll
- coalescer-subrange-crash.ll
- coalescer-subreg-join.mir
- coalescer_distribute.ll
- coalescer_remat.ll
- codegen-prepare-addrmode-sext.ll
- collapse-endcf.ll
- combine-and-sext-bool.ll
- combine-cond-add-sub.ll
- combine-ftrunc.ll
- combine_vloads.ll
- commute-compares.ll
- commute-shifts.ll
- commute_modifiers.ll
- complex-folding.ll
- concat_vectors.ll
- constant-fold-imm-immreg.mir
- constant-fold-mi-operands.ll
- control-flow-fastregalloc.ll
- control-flow-optnone.ll
- convergent-inlineasm.ll
- copy-illegal-type.ll
- copy-to-reg.ll
- ctlz.ll
- ctlz_zero_undef.ll
- ctpop.ll
- ctpop16.ll
- ctpop64.ll
- cttz_zero_undef.ll
- cube.ll
- cvt_f32_ubyte.ll
- cvt_flr_i32_f32.ll
- cvt_rpi_i32_f32.ll
- dagcomb-shuffle-vecextend-non2.ll
- dagcombine-reassociate-bug.ll
- dagcombiner-bug-illegal-vec4-int-to-fp.ll
- dead_copy.mir
- debug-value.ll
- debug.ll
- debugger-emit-prologue.ll
- debugger-insert-nops.ll
- debugger-reserve-regs.ll
- default-fp-mode.ll
- detect-dead-lanes.mir
- disconnected-predset-break-bug.ll
- drop-mem-operand-move-smrd.ll
- ds-combine-large-stride.ll
- ds-negative-offset-addressing-mode-loop.ll
- ds-sub-offset.ll
- ds_read2.ll
- ds_read2_offset_order.ll
- ds_read2_superreg.ll
- ds_read2st64.ll
- ds_write2.ll
- ds_write2st64.ll
- dynamic_stackalloc.ll
- early-if-convert-cost.ll
- early-if-convert.ll
- early-inline-alias.ll
- early-inline.ll
- elf-header.ll
- elf-notes.ll
- elf.ll
- elf.r600.ll
- else.ll
- empty-function.ll
- enable-no-signed-zeros-fp-math.ll
- endcf-loop-header.ll
- endpgm-dce.mir
- enqueue-kernel.ll
- env-amdgiz.ll
- env-amdgizcl.ll
- exceed-max-sgprs.ll
- extend-bit-ops-i16.ll
- extload-align.ll
- extload-private.ll
- extload.ll
- extract-vector-elt-build-vector-combine.ll
- extract_vector_elt-f16.ll
- extract_vector_elt-f64.ll
- extract_vector_elt-i16.ll
- extract_vector_elt-i64.ll
- extract_vector_elt-i8.ll
- extractelt-to-trunc.ll
- fabs.f16.ll
- fabs.f64.ll
- fabs.ll
- fadd-fma-fmul-combine.ll
- fadd.f16.ll
- fadd.ll
- fadd64.ll
- fcanonicalize-elimination.ll
- fcanonicalize.f16.ll
- fcanonicalize.ll
- fceil.ll
- fceil64.ll
- fcmp-cnd.ll
- fcmp-cnde-int-args.ll
- fcmp.f16.ll
- fcmp.ll
- fcmp64.ll
- fconst64.ll
- fcopysign.f16.ll
- fcopysign.f32.ll
- fcopysign.f64.ll
- fdiv.f16.ll
- fdiv.f64.ll
- fdiv.ll
- fence-amdgiz.ll
- fence-barrier.ll
- fetch-limits.r600.ll
- fetch-limits.r700+.ll
- ffloor.f64.ll
- ffloor.ll
- fix-vgpr-copies.mir
- fix-wwm-liveness.mir
- flat-address-space.ll
- flat-for-global-subtarget-feature.ll
- flat-load-clustering.mir
- flat-scratch-reg.ll
- flat_atomics.ll
- flat_atomics_i64.ll
- floor.ll
- fma-combine.ll
- fma.f64.ll
- fma.ll
- fmad.ll
- fmax.ll
- fmax3.f64.ll
- fmax3.ll
- fmax_legacy.f64.ll
- fmax_legacy.ll
- fmaxnum.f64.ll
- fmaxnum.ll
- fmed3.ll
- fmin.ll
- fmin3.ll
- fmin_fmax_legacy.amdgcn.ll
- fmin_legacy.f64.ll
- fmin_legacy.ll
- fminnum.f64.ll
- fminnum.ll
- fmul-2-combine-multi-use.ll
- fmul.f16.ll
- fmul.ll
- fmul64.ll
- fmuladd.f16.ll
- fmuladd.f32.ll
- fmuladd.f64.ll
- fmuladd.v2f16.ll
- fnearbyint.ll
- fneg-combines.ll
- fneg-fabs.f16.ll
- fneg-fabs.f64.ll
- fneg-fabs.ll
- fneg.f16.ll
- fneg.f64.ll
- fneg.ll
- fold-cndmask.mir
- fold-fmul-to-neg-abs.ll
- fold-immediate-output-mods.mir
- fold-operands-order.mir
- fp-classify.ll
- fp16_to_fp32.ll
- fp16_to_fp64.ll
- fp32_to_fp16.ll
- fp_to_sint.f64.ll
- fp_to_sint.ll
- fp_to_uint.f64.ll
- fp_to_uint.ll
- fpext-free.ll
- fpext.f16.ll
- fpext.ll
- fptosi.f16.ll
- fptoui.f16.ll
- fptrunc.f16.ll
- fptrunc.ll
- fract.f64.ll
- fract.ll
- frame-index-amdgiz.ll
- frame-index-elimination.ll
- frem.ll
- fsqrt.f64.ll
- fsqrt.ll
- fsub.f16.ll
- fsub.ll
- fsub64.ll
- ftrunc.f64.ll
- ftrunc.ll
- function-args.ll
- function-returns.ll
- gep-address-space.ll
- global-constant.ll
- global-directive.ll
- global-extload-i16.ll
- global-smrd-unknown.ll
- global-variable-relocs.ll
- global_atomics.ll
- global_atomics_i64.ll
- global_smrd.ll
- global_smrd_cfg.ll
- gv-const-addrspace.ll
- gv-offset-folding.ll
- half.ll
- hazard-inlineasm.mir
- hazard.mir
- hoist-cond.ll
- hsa-default-device.ll
- hsa-fp-mode.ll
- hsa-func-align.ll
- hsa-func.ll
- hsa-globals.ll
- hsa-group-segment.ll
- hsa-metadata-deduce-ro-arg.ll
- hsa-metadata-enqueu-kernel.ll
- hsa-metadata-from-llvm-ir-full.ll
- hsa-metadata-images.ll
- hsa-metadata-invalid-ocl-version-1.ll
- hsa-metadata-invalid-ocl-version-2.ll
- hsa-metadata-invalid-ocl-version-3.ll
- hsa-metadata-kernel-code-props.ll
- hsa-metadata-kernel-debug-props.ll
- hsa-note-no-func.ll
- hsa.ll
- huge-private-buffer.ll
- i1-copy-implicit-def.ll
- i1-copy-phi.ll
- i8-to-double-to-float.ll
- icmp-select-sete-reverse-args.ll
- icmp.i16.ll
- icmp64.ll
- illegal-sgpr-to-vgpr-copy.ll
- image-attributes.ll
- image-resource-id.ll
- imm.ll
- imm16.ll
- immv216.ll
- indirect-addressing-si-noopt.ll
- indirect-addressing-si.ll
- indirect-private-64.ll
- infer-addrpace-pipeline.ll
- infinite-loop-evergreen.ll
- infinite-loop.ll
- inline-asm.ll
- inline-attr.ll
- inline-calls.ll
- inline-constraints.ll
- inlineasm-16.ll
- inlineasm-illegal-type.ll
- inlineasm-packed.ll
- InlineAsmCrash.ll
- input-mods.ll
- insert-skips-kill-uncond.mir
- insert-waits-callee.mir
- insert-waits-exp.mir
- insert_subreg.ll
- insert_vector_elt.ll
- insert_vector_elt.v2i16.ll
- inserted-wait-states.mir
- internalize.ll
- invalid-addrspacecast.ll
- invariant-load-no-alias-store.ll
- invert-br-undef-vcc.mir
- ipra.ll
- jump-address.ll
- kcache-fold.ll
- kernarg-stack-alignment.ll
- kernel-args.ll
- knownbits-recursion.ll
- large-alloca-compute.ll
- large-alloca-graphics.ll
- large-constant-initializer.ll
- large-work-group-promote-alloca.ll
- lds-alignment.ll
- lds-initializer.ll
- lds-m0-init-in-loop.ll
- lds-oqap-crash.ll
- lds-output-queue.ll
- lds-size.ll
- lds-zero-initializer.ll
- legalizedag-bug-expand-setcc.ll
- limit-coalesce.mir
- lit.local.cfg
- literals.ll
- liveness.mir
- llvm.amdgcn.alignb.ll
- llvm.amdgcn.atomic.dec.ll
- llvm.amdgcn.atomic.inc.ll
- llvm.amdgcn.buffer.atomic.ll
- llvm.amdgcn.buffer.load.format.ll
- llvm.amdgcn.buffer.load.ll
- llvm.amdgcn.buffer.store.format.ll
- llvm.amdgcn.buffer.store.ll
- llvm.amdgcn.buffer.wbinvl1.ll
- llvm.amdgcn.buffer.wbinvl1.sc.ll
- llvm.amdgcn.buffer.wbinvl1.vol.ll
- llvm.amdgcn.class.f16.ll
- llvm.amdgcn.class.ll
- llvm.amdgcn.cos.f16.ll
- llvm.amdgcn.cos.ll
- llvm.amdgcn.cubeid.ll
- llvm.amdgcn.cubema.ll
- llvm.amdgcn.cubesc.ll
- llvm.amdgcn.cubetc.ll
- llvm.amdgcn.cvt.pk.i16.ll
- llvm.amdgcn.cvt.pk.u16.ll
- llvm.amdgcn.cvt.pknorm.i16.ll
- llvm.amdgcn.cvt.pknorm.u16.ll
- llvm.amdgcn.cvt.pkrtz.ll
- llvm.amdgcn.dispatch.id.ll
- llvm.amdgcn.dispatch.ptr.ll
- llvm.amdgcn.div.fixup.f16.ll
- llvm.amdgcn.div.fixup.ll
- llvm.amdgcn.div.fmas.ll
- llvm.amdgcn.div.scale.ll
- llvm.amdgcn.ds.bpermute.ll
- llvm.amdgcn.ds.permute.ll
- llvm.amdgcn.ds.swizzle.ll
- llvm.amdgcn.exp.compr.ll
- llvm.amdgcn.exp.ll
- llvm.amdgcn.fcmp.ll
- llvm.amdgcn.fdiv.fast.ll
- llvm.amdgcn.fmed3.f16.ll
- llvm.amdgcn.fmed3.ll
- llvm.amdgcn.fmul.legacy.ll
- llvm.amdgcn.fract.f16.ll
- llvm.amdgcn.fract.ll
- llvm.amdgcn.frexp.exp.f16.ll
- llvm.amdgcn.frexp.exp.ll
- llvm.amdgcn.frexp.mant.f16.ll
- llvm.amdgcn.frexp.mant.ll
- llvm.amdgcn.groupstaticsize.ll
- llvm.amdgcn.icmp.ll
- llvm.amdgcn.image.atomic.ll
- llvm.amdgcn.image.gather4.ll
- llvm.amdgcn.image.getlod.ll
- llvm.amdgcn.image.ll
- llvm.amdgcn.image.sample.ll
- llvm.amdgcn.image.sample.o.ll
- llvm.amdgcn.implicit.buffer.ptr.hsa.ll
- llvm.amdgcn.implicit.buffer.ptr.ll
- llvm.amdgcn.implicitarg.ptr.ll
- llvm.amdgcn.init.exec.ll
- llvm.amdgcn.interp.ll
- llvm.amdgcn.kernarg.segment.ptr.ll
- llvm.amdgcn.kill.ll
- llvm.amdgcn.ldexp.f16.ll
- llvm.amdgcn.ldexp.ll
- llvm.amdgcn.lerp.ll
- llvm.amdgcn.log.clamp.ll
- llvm.amdgcn.mbcnt.ll
- llvm.amdgcn.mov.dpp.ll
- llvm.amdgcn.mqsad.pk.u16.u8.ll
- llvm.amdgcn.mqsad.u32.u8.ll
- llvm.amdgcn.msad.u8.ll
- llvm.amdgcn.ps.live.ll
- llvm.amdgcn.qsad.pk.u16.u8.ll
- llvm.amdgcn.queue.ptr.ll
- llvm.amdgcn.rcp.f16.ll
- llvm.amdgcn.rcp.legacy.ll
- llvm.amdgcn.rcp.ll
- llvm.amdgcn.readfirstlane.ll
- llvm.amdgcn.readlane.ll
- llvm.amdgcn.rsq.clamp.ll
- llvm.amdgcn.rsq.f16.ll
- llvm.amdgcn.rsq.legacy.ll
- llvm.amdgcn.rsq.ll
- llvm.amdgcn.s.barrier.ll
- llvm.amdgcn.s.dcache.inv.ll
- llvm.amdgcn.s.dcache.inv.vol.ll
- llvm.amdgcn.s.dcache.wb.ll
- llvm.amdgcn.s.dcache.wb.vol.ll
- llvm.amdgcn.s.decperflevel.ll
- llvm.amdgcn.s.getpc.ll
- llvm.amdgcn.s.getreg.ll
- llvm.amdgcn.s.incperflevel.ll
- llvm.amdgcn.s.memrealtime.ll
- llvm.amdgcn.s.memtime.ll
- llvm.amdgcn.s.sleep.ll
- llvm.amdgcn.s.waitcnt.ll
- llvm.amdgcn.sad.hi.u8.ll
- llvm.amdgcn.sad.u16.ll
- llvm.amdgcn.sad.u8.ll
- llvm.amdgcn.sbfe.ll
- llvm.amdgcn.sendmsg.ll
- llvm.amdgcn.set.inactive.ll
- llvm.amdgcn.sffbh.ll
- llvm.amdgcn.sin.f16.ll
- llvm.amdgcn.sin.ll
- llvm.amdgcn.tbuffer.load.ll
- llvm.amdgcn.tbuffer.store.ll
- llvm.amdgcn.trig.preop.ll
- llvm.amdgcn.ubfe.ll
- llvm.amdgcn.unreachable.ll
- llvm.amdgcn.update.dpp.ll
- llvm.amdgcn.wave.barrier.ll
- llvm.amdgcn.workgroup.id.ll
- llvm.amdgcn.workitem.id.ll
- llvm.amdgcn.wqm.vote.ll
- llvm.AMDGPU.kill.ll
- llvm.amdgpu.kilp.ll
- llvm.ceil.f16.ll
- llvm.cos.f16.ll
- llvm.cos.ll
- llvm.dbg.value.ll
- llvm.exp2.f16.ll
- llvm.exp2.ll
- llvm.floor.f16.ll
- llvm.fma.f16.ll
- llvm.fmuladd.f16.ll
- llvm.log.f16.ll
- llvm.log.ll
- llvm.log10.f16.ll
- llvm.log10.ll
- llvm.log2.f16.ll
- llvm.log2.ll
- llvm.maxnum.f16.ll
- llvm.memcpy.ll
- llvm.minnum.f16.ll
- llvm.pow.ll
- llvm.r600.cube.ll
- llvm.r600.dot4.ll
- llvm.r600.group.barrier.ll
- llvm.r600.read.local.size.ll
- llvm.r600.recipsqrt.clamped.ll
- llvm.r600.recipsqrt.ieee.ll
- llvm.r600.tex.ll
- llvm.rint.f16.ll
- llvm.rint.f64.ll
- llvm.rint.ll
- llvm.round.f64.ll
- llvm.round.ll
- llvm.SI.load.dword.ll
- llvm.SI.tbuffer.store.ll
- llvm.sin.f16.ll
- llvm.sin.ll
- llvm.sqrt.f16.ll
- llvm.trunc.f16.ll
- load-constant-f64.ll
- load-constant-i1.ll
- load-constant-i16.ll
- load-constant-i32.ll
- load-constant-i64.ll
- load-constant-i8.ll
- load-global-f32.ll
- load-global-f64.ll
- load-global-i1.ll
- load-global-i16.ll
- load-global-i32.ll
- load-global-i64.ll
- load-global-i8.ll
- load-hi16.ll
- load-input-fold.ll
- load-lo16.ll
- load-local-f32.ll
- load-local-f64.ll
- load-local-i1.ll
- load-local-i16.ll
- load-local-i32.ll
- load-local-i64.ll
- load-local-i8.ll
- load-private-double16-amdgiz.ll
- load-weird-sizes.ll
- local-64.ll
- local-atomics.ll
- local-atomics64.ll
- local-memory.amdgcn.ll
- local-memory.ll
- local-memory.r600.ll
- local-stack-slot-offset.ll
- loop-address.ll
- loop-idiom.ll
- loop_break.ll
- lower-mem-intrinsics.ll
- lower-range-metadata-intrinsic-call.ll
- lshl64-to-32.ll
- lshr.v2i16.ll
- macro-fusion-cluster-vcc-uses.mir
- mad-combine.ll
- mad-mix-hi.ll
- mad-mix-lo.ll
- mad-mix.ll
- mad24-get-global-id.ll
- mad_64_32.ll
- mad_int24.ll
- mad_uint24.ll
- madak.ll
- madmk.ll
- max-literals.ll
- max.i16.ll
- max.ll
- max3.ll
- mem-builtins.ll
- memory-legalizer-atomic-cmpxchg.ll
- memory-legalizer-atomic-fence.ll
- memory-legalizer-atomic-rmw.ll
- memory-legalizer-invalid-syncscope.ll
- memory-legalizer-load.ll
- memory-legalizer-store-infinite-loop.ll
- memory-legalizer-store.ll
- merge-load-store.mir
- merge-m0.mir
- merge-store-crash.ll
- merge-store-usedef.ll
- merge-stores.ll
- mesa_regression.ll
- min.ll
- min3.ll
- misched-killflags.mir
- missing-store.ll
- move-addr64-rsrc-dead-subreg-writes.ll
- move-to-valu-atomicrmw.ll
- move-to-valu-worklist.ll
- movreld-bug.ll
- movrels-bug.mir
- mubuf-offset-private.ll
- mubuf-shader-vgpr.ll
- mubuf.ll
- mul.ll
- mul_int24.ll
- mul_uint24-amdgcn.ll
- mul_uint24-r600.ll
- multi-divergent-exit-region.ll
- multilevel-break.ll
- nested-calls.ll
- nested-loop-conditions.ll
- no-hsa-graphics-shaders.ll
- no-initializer-constant-addrspace.ll
- no-shrink-extloads.ll
- nop-data.ll
- not-scalarize-volatile-load.ll
- nullptr.ll
- omod.ll
- opencl-image-metadata.ll
- operand-folding.ll
- operand-spacing.ll
- opt-sgpr-to-vgpr-copy.mir
- optimize-if-exec-masking.mir
- or.ll
- over-max-lds-size.ll
- pack.v2f16.ll
- pack.v2i16.ll
- packed-op-sel.ll
- packetizer.ll
- parallelandifcollapse.ll
- parallelorifcollapse.ll
- partial-sgpr-to-vgpr-spills.ll
- partially-dead-super-register-immediate.ll
- predicate-dp4.ll
- predicates.ll
- private-access-no-objects.ll
- private-element-size.ll
- private-memory-atomics.ll
- private-memory-r600.ll
- promote-alloca-addrspacecast.ll
- promote-alloca-array-aggregate.ll
- promote-alloca-array-allocation.ll
- promote-alloca-bitcast-function.ll
- promote-alloca-calling-conv.ll
- promote-alloca-globals.ll
- promote-alloca-invariant-markers.ll
- promote-alloca-lifetime.ll
- promote-alloca-mem-intrinsics.ll
- promote-alloca-no-opts.ll
- promote-alloca-padding-size-estimate.ll
- promote-alloca-stored-pointer-value.ll
- promote-alloca-to-lds-icmp.ll
- promote-alloca-to-lds-phi.ll
- promote-alloca-to-lds-select.ll
- promote-alloca-unhandled-intrinsic.ll
- promote-alloca-volatile.ll
- pv-packing.ll
- pv.ll
- r600-constant-array-fixup.ll
- r600-encoding.ll
- r600-export-fix.ll
- r600-infinite-loop-bug-while-reorganizing-vector.ll
- r600-legalize-umax-bug.ll
- r600.alu-limits.ll
- r600.amdgpu-alias-analysis.ll
- r600.bitcast.ll
- r600.global_atomics.ll
- r600.private-memory.ll
- r600.work-item-intrinsics.ll
- r600cfg.ll
- rcp-pattern.ll
- read-register-invalid-subtarget.ll
- read-register-invalid-type-i32.ll
- read-register-invalid-type-i64.ll
- read_register.ll
- readcyclecounter.ll
- readlane_exec0.mir
- README
- reduce-load-width-alignment.ll
- reduce-saveexec.mir
- reduce-store-width-alignment.ll
- reg-coalescer-sched-crash.ll
- regcoal-subrange-join.mir
- regcoalesce-dbg.mir
- regcoalesce-prune.mir
- register-count-comments.ll
- rename-disconnected-bug.ll
- rename-independent-subregs-mac-operands.mir
- rename-independent-subregs.mir
- reorder-stores.ll
- ret.ll
- ret_jump.ll
- rewrite-out-arguments-address-space.ll
- rewrite-out-arguments.ll
- rotl.i64.ll
- rotl.ll
- rotr.i64.ll
- rotr.ll
- rsq.ll
- rv7x0_count3.ll
- s_addk_i32.ll
- s_movk_i32.ll
- s_mulk_i32.ll
- sad.ll
- saddo.ll
- salu-to-valu.ll
- sampler-resource-id.ll
- scalar-store-cache-flush.mir
- scalar_to_vector.ll
- sched-crash-dbg-value.mir
- schedule-fs-loop-nested-if.ll
- schedule-fs-loop-nested.ll
- schedule-fs-loop.ll
- schedule-global-loads.ll
- schedule-if-2.ll
- schedule-if.ll
- schedule-ilp.ll
- schedule-kernel-arg-loads.ll
- schedule-regpressure-limit.ll
- schedule-regpressure-limit2.ll
- schedule-regpressure.mir
- schedule-vs-if-nested-loop-failure.ll
- schedule-vs-if-nested-loop.ll
- scheduler-subrange-crash.ll
- scratch-buffer.ll
- scratch-simple.ll
- sdiv.ll
- sdivrem24.ll
- sdivrem64.ll
- sdwa-gfx9.mir
- sdwa-peephole-instr.mir
- sdwa-peephole.ll
- sdwa-preserve.mir
- sdwa-scalar-ops.mir
- sdwa-vop2-64bit.mir
- select-fabs-fneg-extract-legacy.ll
- select-fabs-fneg-extract.ll
- select-i1.ll
- select-opt.ll
- select-vectors.ll
- select.f16.ll
- select.ll
- select64.ll
- selectcc-cnd.ll
- selectcc-cnde-int.ll
- selectcc-icmp-select-float.ll
- selectcc-opt.ll
- selectcc.ll
- selected-stack-object.ll
- sendmsg-m0-hazard.mir
- set-dx10.ll
- setcc-equivalent.ll
- setcc-fneg-constant.ll
- setcc-opt.ll
- setcc-sext.ll
- setcc.ll
- setcc64.ll
- seto.ll
- setuo.ll
- sext-eliminate.ll
- sext-in-reg-failure-r600.ll
- sext-in-reg.ll
- sgpr-control-flow.ll
- sgpr-copy-duplicate-operand.ll
- sgpr-copy.ll
- sgprcopies.ll
- shared-op-cycle.ll
- shift-and-i128-ubfe.ll
- shift-and-i64-ubfe.ll
- shift-i64-opts.ll
- shl-add-to-add-shl.ll
- shl.ll
- shl.v2i16.ll
- shl_add_constant.ll
- shl_add_ptr.ll
- shrink-add-sub-constant.ll
- shrink-carry.mir
- shrink-vop3-carry-out.mir
- si-annotate-cf-noloop.ll
- si-annotate-cf-unreachable.ll
- si-annotate-cf.ll
- si-annotate-cfg-loop-assert.ll
- si-fix-sgpr-copies.mir
- si-instr-info-correct-implicit-operands.ll
- si-lod-bias.ll
- si-lower-control-flow-kill.ll
- si-lower-control-flow-unreachable-block.ll
- si-scheduler.ll
- si-sgpr-spill.ll
- si-spill-cf.ll
- si-spill-sgpr-stack.ll
- si-triv-disjoint-mem-access.ll
- si-vector-hang.ll
- sibling-call.ll
- sign_extend.ll
- simplify-libcalls.ll
- simplifydemandedbits-recursion.ll
- sint_to_fp.f64.ll
- sint_to_fp.i64.ll
- sint_to_fp.ll
- sitofp.f16.ll
- skip-if-dead.ll
- smed3.ll
- sminmax.ll
- sminmax.v2i16.ll
- smrd-vccz-bug.ll
- smrd.ll
- sopk-compares.ll
- spill-alloc-sgpr-init-bug.ll
- spill-cfg-position.ll
- spill-empty-live-interval.mir
- spill-m0.ll
- spill-scavenge-offset.ll
- spill-to-smem-m0.ll
- spill-wide-sgpr.ll
- split-scalar-i64-add.ll
- split-smrd.ll
- split-vector-memoperand-offsets.ll
- splitkit.mir
- sra.ll
- srem.ll
- srl.ll
- ssubo.ll
- stack-size-overflow.ll
- stack-slot-color-sgpr-vgpr-spills.mir
- store-barrier.ll
- store-global.ll
- store-hi16.ll
- store-local.ll
- store-private.ll
- store-v3i64.ll
- store-vector-ptrs.ll
- store-weird-sizes.ll
- store_typed.ll
- stress-calls.ll
- structurize.ll
- structurize1.ll
- sub.i16.ll
- sub.ll
- sub.v2i16.ll
- subreg-coalescer-crash.ll
- subreg-coalescer-undef-use.ll
- subreg-eliminate-dead.ll
- subreg-intervals.mir
- subreg_interference.mir
- swizzle-export.ll
- syncscopes.ll
- tail-call-cgp.ll
- target-cpu.ll
- tex-clause-antidep.ll
- texture-input-merge.ll
- trap.ll
- trunc-bitcast-vector.ll
- trunc-cmp-constant.ll
- trunc-store-f64-to-f16.ll
- trunc-store-i1.ll
- trunc-store.ll
- trunc-vector-store-assertion-failure.ll
- trunc.ll
- tti-unroll-prefs.ll
- twoaddr-mad.mir
- uaddo.ll
- udiv.ll
- udivrem.ll
- udivrem24.ll
- udivrem64.ll
- uint_to_fp.f64.ll
- uint_to_fp.i64.ll
- uint_to_fp.ll
- uitofp.f16.ll
- umed3.ll
- unaligned-load-store.ll
- undefined-physreg-sgpr-spill.mir
- undefined-subreg-liverange.ll
- unhandled-loop-condition-assertion.ll
- uniform-branch-intrinsic-cond.ll
- uniform-cfg.ll
- uniform-crash.ll
- uniform-loop-inside-nonuniform.ll
- uniform-PHI.ll
- unify-metadata.ll
- unigine-liveness-crash.ll
- unknown-processor.ll
- unroll.ll
- unsupported-calls.ll
- unsupported-cc.ll
- urem.ll
- use-sgpr-multiple-times.ll
- usubo.ll
- v1i64-kernel-arg.ll
- v_cndmask.ll
- v_cvt_pk_u8_f32.ll
- v_mac.ll
- v_mac_f16.ll
- v_madak_f16.ll
- valu-i1.ll
- vccz-corrupt-bug-workaround.mir
- vector-alloca.ll
- vector-extract-insert.ll
- vectorize-global-local.ll
- vertex-fetch-encoding.ll
- vgpr-spill-emergency-stack-slot-compute.ll
- vgpr-spill-emergency-stack-slot.ll
- vi-removed-intrinsics.ll
- vop-shrink-frame-index.mir
- vop-shrink-non-ssa.mir
- vop-shrink.ll
- vselect.ll
- vselect64.ll
- vtx-fetch-branch.ll
- vtx-schedule.ll
- wait.ll
- waitcnt-flat.ll
- waitcnt-looptest.ll
- waitcnt-permute.mir
- waitcnt.mir
- widen-vselect-and-mask.ll
- widen_extending_scalar_loads.ll
- wqm.ll
- wqm.mir
- write-register-vgpr-into-sgpr.ll
- write_register.ll
- wrong-transalu-pos-fix.ll
- xfail.r600.bitcast.ll
- xnor.ll
- xor.ll
- zero_extend.ll
- zext-i64-bit-operand.ll
- zext-lid.ll
private-memory-r600.ll @release_60 — raw · history · blame
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 | ; RUN: llc -march=r600 -mtriple=r600---amdgiz -mcpu=redwood < %s | FileCheck %s -check-prefix=R600 -check-prefix=FUNC
; RUN: opt -S -mtriple=r600-unknown-unknown-amdgiz -mcpu=redwood -amdgpu-promote-alloca < %s | FileCheck -check-prefix=OPT %s
target datalayout = "A5"
declare i32 @llvm.r600.read.tidig.x() nounwind readnone
; FUNC-LABEL: {{^}}mova_same_clause:
; R600: LDS_WRITE
; R600: LDS_WRITE
; R600: LDS_READ
; R600: LDS_READ
; OPT: call i32 @llvm.r600.read.local.size.y(), !range !0
; OPT: call i32 @llvm.r600.read.local.size.z(), !range !0
; OPT: call i32 @llvm.r600.read.tidig.x(), !range !1
; OPT: call i32 @llvm.r600.read.tidig.y(), !range !1
; OPT: call i32 @llvm.r600.read.tidig.z(), !range !1
define amdgpu_kernel void @mova_same_clause(i32 addrspace(1)* nocapture %out, i32 addrspace(1)* nocapture %in) #0 {
entry:
%stack = alloca [5 x i32], align 4, addrspace(5)
%0 = load i32, i32 addrspace(1)* %in, align 4
%arrayidx1 = getelementptr inbounds [5 x i32], [5 x i32] addrspace(5)* %stack, i32 0, i32 %0
store i32 4, i32 addrspace(5)* %arrayidx1, align 4
%arrayidx2 = getelementptr inbounds i32, i32 addrspace(1)* %in, i32 1
%1 = load i32, i32 addrspace(1)* %arrayidx2, align 4
%arrayidx3 = getelementptr inbounds [5 x i32], [5 x i32] addrspace(5)* %stack, i32 0, i32 %1
store i32 5, i32 addrspace(5)* %arrayidx3, align 4
%arrayidx10 = getelementptr inbounds [5 x i32], [5 x i32] addrspace(5)* %stack, i32 0, i32 0
%2 = load i32, i32 addrspace(5)* %arrayidx10, align 4
store i32 %2, i32 addrspace(1)* %out, align 4
%arrayidx12 = getelementptr inbounds [5 x i32], [5 x i32] addrspace(5)* %stack, i32 0, i32 1
%3 = load i32, i32 addrspace(5)* %arrayidx12
%arrayidx13 = getelementptr inbounds i32, i32 addrspace(1)* %out, i32 1
store i32 %3, i32 addrspace(1)* %arrayidx13
ret void
}
; This test checks that the stack offset is calculated correctly for structs.
; All register loads/stores should be optimized away, so there shouldn't be
; any MOVA instructions.
;
; XXX: This generated code has unnecessary MOVs, we should be able to optimize
; this.
; FUNC-LABEL: {{^}}multiple_structs:
; R600-NOT: MOVA_INT
%struct.point = type { i32, i32 }
define amdgpu_kernel void @multiple_structs(i32 addrspace(1)* %out) #0 {
entry:
%a = alloca %struct.point, addrspace(5)
%b = alloca %struct.point, addrspace(5)
%a.x.ptr = getelementptr inbounds %struct.point, %struct.point addrspace(5)* %a, i32 0, i32 0
%a.y.ptr = getelementptr inbounds %struct.point, %struct.point addrspace(5)* %a, i32 0, i32 1
%b.x.ptr = getelementptr inbounds %struct.point, %struct.point addrspace(5)* %b, i32 0, i32 0
%b.y.ptr = getelementptr inbounds %struct.point, %struct.point addrspace(5)* %b, i32 0, i32 1
store i32 0, i32 addrspace(5)* %a.x.ptr
store i32 1, i32 addrspace(5)* %a.y.ptr
store i32 2, i32 addrspace(5)* %b.x.ptr
store i32 3, i32 addrspace(5)* %b.y.ptr
%a.indirect.ptr = getelementptr inbounds %struct.point, %struct.point addrspace(5)* %a, i32 0, i32 0
%b.indirect.ptr = getelementptr inbounds %struct.point, %struct.point addrspace(5)* %b, i32 0, i32 0
%a.indirect = load i32, i32 addrspace(5)* %a.indirect.ptr
%b.indirect = load i32, i32 addrspace(5)* %b.indirect.ptr
%0 = add i32 %a.indirect, %b.indirect
store i32 %0, i32 addrspace(1)* %out
ret void
}
; Test direct access of a private array inside a loop. The private array
; loads and stores should be lowered to copies, so there shouldn't be any
; MOVA instructions.
; FUNC-LABEL: {{^}}direct_loop:
; R600-NOT: MOVA_INT
define amdgpu_kernel void @direct_loop(i32 addrspace(1)* %out, i32 addrspace(1)* %in) #0 {
entry:
%prv_array_const = alloca [2 x i32], addrspace(5)
%prv_array = alloca [2 x i32], addrspace(5)
%a = load i32, i32 addrspace(1)* %in
%b_src_ptr = getelementptr inbounds i32, i32 addrspace(1)* %in, i32 1
%b = load i32, i32 addrspace(1)* %b_src_ptr
%a_dst_ptr = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %prv_array_const, i32 0, i32 0
store i32 %a, i32 addrspace(5)* %a_dst_ptr
%b_dst_ptr = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %prv_array_const, i32 0, i32 1
store i32 %b, i32 addrspace(5)* %b_dst_ptr
br label %for.body
for.body:
%inc = phi i32 [0, %entry], [%count, %for.body]
%x_ptr = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %prv_array_const, i32 0, i32 0
%x = load i32, i32 addrspace(5)* %x_ptr
%y_ptr = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %prv_array, i32 0, i32 0
%y = load i32, i32 addrspace(5)* %y_ptr
%xy = add i32 %x, %y
store i32 %xy, i32 addrspace(5)* %y_ptr
%count = add i32 %inc, 1
%done = icmp eq i32 %count, 4095
br i1 %done, label %for.end, label %for.body
for.end:
%value_ptr = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %prv_array, i32 0, i32 0
%value = load i32, i32 addrspace(5)* %value_ptr
store i32 %value, i32 addrspace(1)* %out
ret void
}
; FUNC-LABEL: {{^}}short_array:
; R600: MOVA_INT
define amdgpu_kernel void @short_array(i32 addrspace(1)* %out, i32 %index) #0 {
entry:
%0 = alloca [2 x i16], addrspace(5)
%1 = getelementptr inbounds [2 x i16], [2 x i16] addrspace(5)* %0, i32 0, i32 0
%2 = getelementptr inbounds [2 x i16], [2 x i16] addrspace(5)* %0, i32 0, i32 1
store i16 0, i16 addrspace(5)* %1
store i16 1, i16 addrspace(5)* %2
%3 = getelementptr inbounds [2 x i16], [2 x i16] addrspace(5)* %0, i32 0, i32 %index
%4 = load i16, i16 addrspace(5)* %3
%5 = sext i16 %4 to i32
store i32 %5, i32 addrspace(1)* %out
ret void
}
; FUNC-LABEL: {{^}}char_array:
; R600: MOVA_INT
define amdgpu_kernel void @char_array(i32 addrspace(1)* %out, i32 %index) #0 {
entry:
%0 = alloca [2 x i8], addrspace(5)
%1 = getelementptr inbounds [2 x i8], [2 x i8] addrspace(5)* %0, i32 0, i32 0
%2 = getelementptr inbounds [2 x i8], [2 x i8] addrspace(5)* %0, i32 0, i32 1
store i8 0, i8 addrspace(5)* %1
store i8 1, i8 addrspace(5)* %2
%3 = getelementptr inbounds [2 x i8], [2 x i8] addrspace(5)* %0, i32 0, i32 %index
%4 = load i8, i8 addrspace(5)* %3
%5 = sext i8 %4 to i32
store i32 %5, i32 addrspace(1)* %out
ret void
}
; Make sure we don't overwrite workitem information with private memory
; FUNC-LABEL: {{^}}work_item_info:
; R600-NOT: MOV T0.X
; Additional check in case the move ends up in the last slot
; R600-NOT: MOV * TO.X
define amdgpu_kernel void @work_item_info(i32 addrspace(1)* %out, i32 %in) #0 {
entry:
%0 = alloca [2 x i32], addrspace(5)
%1 = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %0, i32 0, i32 0
%2 = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %0, i32 0, i32 1
store i32 0, i32 addrspace(5)* %1
store i32 1, i32 addrspace(5)* %2
%3 = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %0, i32 0, i32 %in
%4 = load i32, i32 addrspace(5)* %3
%5 = call i32 @llvm.r600.read.tidig.x()
%6 = add i32 %4, %5
store i32 %6, i32 addrspace(1)* %out
ret void
}
; Test that two stack objects are not stored in the same register
; The second stack object should be in T3.X
; FUNC-LABEL: {{^}}no_overlap:
; R600_CHECK: MOV
; R600_CHECK: [[CHAN:[XYZW]]]+
; R600-NOT: [[CHAN]]+
define amdgpu_kernel void @no_overlap(i32 addrspace(1)* %out, i32 %in) #0 {
entry:
%0 = alloca [3 x i8], align 1, addrspace(5)
%1 = alloca [2 x i8], align 1, addrspace(5)
%2 = getelementptr inbounds [3 x i8], [3 x i8] addrspace(5)* %0, i32 0, i32 0
%3 = getelementptr inbounds [3 x i8], [3 x i8] addrspace(5)* %0, i32 0, i32 1
%4 = getelementptr inbounds [3 x i8], [3 x i8] addrspace(5)* %0, i32 0, i32 2
%5 = getelementptr inbounds [2 x i8], [2 x i8] addrspace(5)* %1, i32 0, i32 0
%6 = getelementptr inbounds [2 x i8], [2 x i8] addrspace(5)* %1, i32 0, i32 1
store i8 0, i8 addrspace(5)* %2
store i8 1, i8 addrspace(5)* %3
store i8 2, i8 addrspace(5)* %4
store i8 1, i8 addrspace(5)* %5
store i8 0, i8 addrspace(5)* %6
%7 = getelementptr inbounds [3 x i8], [3 x i8] addrspace(5)* %0, i32 0, i32 %in
%8 = getelementptr inbounds [2 x i8], [2 x i8] addrspace(5)* %1, i32 0, i32 %in
%9 = load i8, i8 addrspace(5)* %7
%10 = load i8, i8 addrspace(5)* %8
%11 = add i8 %9, %10
%12 = sext i8 %11 to i32
store i32 %12, i32 addrspace(1)* %out
ret void
}
define amdgpu_kernel void @char_array_array(i32 addrspace(1)* %out, i32 %index) #0 {
entry:
%alloca = alloca [2 x [2 x i8]], addrspace(5)
%gep0 = getelementptr inbounds [2 x [2 x i8]], [2 x [2 x i8]] addrspace(5)* %alloca, i32 0, i32 0, i32 0
%gep1 = getelementptr inbounds [2 x [2 x i8]], [2 x [2 x i8]] addrspace(5)* %alloca, i32 0, i32 0, i32 1
store i8 0, i8 addrspace(5)* %gep0
store i8 1, i8 addrspace(5)* %gep1
%gep2 = getelementptr inbounds [2 x [2 x i8]], [2 x [2 x i8]] addrspace(5)* %alloca, i32 0, i32 0, i32 %index
%load = load i8, i8 addrspace(5)* %gep2
%sext = sext i8 %load to i32
store i32 %sext, i32 addrspace(1)* %out
ret void
}
define amdgpu_kernel void @i32_array_array(i32 addrspace(1)* %out, i32 %index) #0 {
entry:
%alloca = alloca [2 x [2 x i32]], addrspace(5)
%gep0 = getelementptr inbounds [2 x [2 x i32]], [2 x [2 x i32]] addrspace(5)* %alloca, i32 0, i32 0, i32 0
%gep1 = getelementptr inbounds [2 x [2 x i32]], [2 x [2 x i32]] addrspace(5)* %alloca, i32 0, i32 0, i32 1
store i32 0, i32 addrspace(5)* %gep0
store i32 1, i32 addrspace(5)* %gep1
%gep2 = getelementptr inbounds [2 x [2 x i32]], [2 x [2 x i32]] addrspace(5)* %alloca, i32 0, i32 0, i32 %index
%load = load i32, i32 addrspace(5)* %gep2
store i32 %load, i32 addrspace(1)* %out
ret void
}
define amdgpu_kernel void @i64_array_array(i64 addrspace(1)* %out, i32 %index) #0 {
entry:
%alloca = alloca [2 x [2 x i64]], addrspace(5)
%gep0 = getelementptr inbounds [2 x [2 x i64]], [2 x [2 x i64]] addrspace(5)* %alloca, i32 0, i32 0, i32 0
%gep1 = getelementptr inbounds [2 x [2 x i64]], [2 x [2 x i64]] addrspace(5)* %alloca, i32 0, i32 0, i32 1
store i64 0, i64 addrspace(5)* %gep0
store i64 1, i64 addrspace(5)* %gep1
%gep2 = getelementptr inbounds [2 x [2 x i64]], [2 x [2 x i64]] addrspace(5)* %alloca, i32 0, i32 0, i32 %index
%load = load i64, i64 addrspace(5)* %gep2
store i64 %load, i64 addrspace(1)* %out
ret void
}
%struct.pair32 = type { i32, i32 }
define amdgpu_kernel void @struct_array_array(i32 addrspace(1)* %out, i32 %index) #0 {
entry:
%alloca = alloca [2 x [2 x %struct.pair32]], addrspace(5)
%gep0 = getelementptr inbounds [2 x [2 x %struct.pair32]], [2 x [2 x %struct.pair32]] addrspace(5)* %alloca, i32 0, i32 0, i32 0, i32 1
%gep1 = getelementptr inbounds [2 x [2 x %struct.pair32]], [2 x [2 x %struct.pair32]] addrspace(5)* %alloca, i32 0, i32 0, i32 1, i32 1
store i32 0, i32 addrspace(5)* %gep0
store i32 1, i32 addrspace(5)* %gep1
%gep2 = getelementptr inbounds [2 x [2 x %struct.pair32]], [2 x [2 x %struct.pair32]] addrspace(5)* %alloca, i32 0, i32 0, i32 %index, i32 0
%load = load i32, i32 addrspace(5)* %gep2
store i32 %load, i32 addrspace(1)* %out
ret void
}
define amdgpu_kernel void @struct_pair32_array(i32 addrspace(1)* %out, i32 %index) #0 {
entry:
%alloca = alloca [2 x %struct.pair32], addrspace(5)
%gep0 = getelementptr inbounds [2 x %struct.pair32], [2 x %struct.pair32] addrspace(5)* %alloca, i32 0, i32 0, i32 1
%gep1 = getelementptr inbounds [2 x %struct.pair32], [2 x %struct.pair32] addrspace(5)* %alloca, i32 0, i32 1, i32 0
store i32 0, i32 addrspace(5)* %gep0
store i32 1, i32 addrspace(5)* %gep1
%gep2 = getelementptr inbounds [2 x %struct.pair32], [2 x %struct.pair32] addrspace(5)* %alloca, i32 0, i32 %index, i32 0
%load = load i32, i32 addrspace(5)* %gep2
store i32 %load, i32 addrspace(1)* %out
ret void
}
define amdgpu_kernel void @select_private(i32 addrspace(1)* %out, i32 %in) nounwind {
entry:
%tmp = alloca [2 x i32], addrspace(5)
%tmp1 = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %tmp, i32 0, i32 0
%tmp2 = getelementptr inbounds [2 x i32], [2 x i32] addrspace(5)* %tmp, i32 0, i32 1
store i32 0, i32 addrspace(5)* %tmp1
store i32 1, i32 addrspace(5)* %tmp2
%cmp = icmp eq i32 %in, 0
%sel = select i1 %cmp, i32 addrspace(5)* %tmp1, i32 addrspace(5)* %tmp2
%load = load i32, i32 addrspace(5)* %sel
store i32 %load, i32 addrspace(1)* %out
ret void
}
; AMDGPUPromoteAlloca does not know how to handle ptrtoint. When it
; finds one, it should stop trying to promote.
; FUNC-LABEL: ptrtoint:
; SI-NOT: ds_write
; SI: buffer_store_dword v{{[0-9]+}}, v{{[0-9]+}}, s[{{[0-9]+:[0-9]+}}], s{{[0-9]+}} offen
; SI: buffer_load_dword v{{[0-9]+}}, v{{[0-9]+}}, s[{{[0-9]+:[0-9]+}}], s{{[0-9]+}} offen ;
define amdgpu_kernel void @ptrtoint(i32 addrspace(1)* %out, i32 %a, i32 %b) #0 {
%alloca = alloca [16 x i32], addrspace(5)
%tmp0 = getelementptr inbounds [16 x i32], [16 x i32] addrspace(5)* %alloca, i32 0, i32 %a
store i32 5, i32 addrspace(5)* %tmp0
%tmp1 = ptrtoint [16 x i32] addrspace(5)* %alloca to i32
%tmp2 = add i32 %tmp1, 5
%tmp3 = inttoptr i32 %tmp2 to i32 addrspace(5)*
%tmp4 = getelementptr inbounds i32, i32 addrspace(5)* %tmp3, i32 %b
%tmp5 = load i32, i32 addrspace(5)* %tmp4
store i32 %tmp5, i32 addrspace(1)* %out
ret void
}
; OPT: !0 = !{i32 0, i32 257}
; OPT: !1 = !{i32 0, i32 256}
attributes #0 = { nounwind "amdgpu-waves-per-eu"="1,2" }
|