//===-- AMDGPUGIsel.td - AMDGPU GlobalISel Patterns---------*- tablegen -*-===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// // This files contains patterns that should only be used by GlobalISel. For // example patterns for V_* instructions that have S_* equivalents. // SelectionDAG does not support selecting V_* instructions. //===----------------------------------------------------------------------===// include "AMDGPU.td" include "AMDGPUCombine.td" def sd_vsrc0 : ComplexPattern<i32, 1, "">; def gi_vsrc0 : GIComplexOperandMatcher<s32, "selectVSRC0">, GIComplexPatternEquiv<sd_vsrc0>; def sd_vcsrc : ComplexPattern<i32, 1, "">; def gi_vcsrc : GIComplexOperandMatcher<s32, "selectVCSRC">, GIComplexPatternEquiv<sd_vcsrc>; def gi_vop3mods0 : GIComplexOperandMatcher<s32, "selectVOP3Mods0">, GIComplexPatternEquiv<VOP3Mods0>; def gi_vop3mods : GIComplexOperandMatcher<s32, "selectVOP3Mods">, GIComplexPatternEquiv<VOP3Mods>; def gi_vop3_no_mods : GIComplexOperandMatcher<s32, "selectVOP3NoMods">, GIComplexPatternEquiv<VOP3NoMods>; def gi_vop3mods_nnan : GIComplexOperandMatcher<s32, "selectVOP3Mods_nnan">, GIComplexPatternEquiv<VOP3Mods_nnan>; def gi_vop3omods : GIComplexOperandMatcher<s32, "selectVOP3OMods">, GIComplexPatternEquiv<VOP3OMods>; def gi_vop3pmods : GIComplexOperandMatcher<s32, "selectVOP3PMods">, GIComplexPatternEquiv<VOP3PMods>; def gi_vop3pmodsdot : GIComplexOperandMatcher<s32, "selectVOP3PModsDOT">, GIComplexPatternEquiv<VOP3PModsDOT>; def gi_dotiuvop3pmods : GIComplexOperandMatcher<s32, "selectDotIUVOP3PMods">, GIComplexPatternEquiv<DotIUVOP3PMods>; def gi_wmmaopselvop3pmods : GIComplexOperandMatcher<s32, "selectWMMAOpSelVOP3PMods">, GIComplexPatternEquiv<WMMAOpSelVOP3PMods>; def gi_vop3opselmods : GIComplexOperandMatcher<s32, "selectVOP3OpSelMods">, GIComplexPatternEquiv<VOP3OpSelMods>; def gi_vinterpmods : GIComplexOperandMatcher<s32, "selectVINTERPMods">, GIComplexPatternEquiv<VINTERPMods>; def gi_vinterpmods_hi : GIComplexOperandMatcher<s32, "selectVINTERPModsHi">, GIComplexPatternEquiv<VINTERPModsHi>; // FIXME: Why do we have both VOP3OpSel and VOP3OpSelMods? def gi_vop3opsel : GIComplexOperandMatcher<s32, "selectVOP3OpSelMods">, GIComplexPatternEquiv<VOP3OpSel>; def gi_smrd_imm : GIComplexOperandMatcher<s64, "selectSmrdImm">, GIComplexPatternEquiv<SMRDImm>; def gi_smrd_imm32 : GIComplexOperandMatcher<s64, "selectSmrdImm32">, GIComplexPatternEquiv<SMRDImm32>; def gi_smrd_sgpr : GIComplexOperandMatcher<s64, "selectSmrdSgpr">, GIComplexPatternEquiv<SMRDSgpr>; def gi_smrd_sgpr_imm : GIComplexOperandMatcher<s64, "selectSmrdSgprImm">, GIComplexPatternEquiv<SMRDSgprImm>; def gi_flat_offset : GIComplexOperandMatcher<s64, "selectFlatOffset">, GIComplexPatternEquiv<FlatOffset>; def gi_global_offset : GIComplexOperandMatcher<s64, "selectGlobalOffset">, GIComplexPatternEquiv<GlobalOffset>; def gi_global_saddr : GIComplexOperandMatcher<s64, "selectGlobalSAddr">, GIComplexPatternEquiv<GlobalSAddr>; def gi_mubuf_scratch_offset : GIComplexOperandMatcher<s32, "selectMUBUFScratchOffset">, GIComplexPatternEquiv<MUBUFScratchOffset>; def gi_mubuf_scratch_offen : GIComplexOperandMatcher<s32, "selectMUBUFScratchOffen">, GIComplexPatternEquiv<MUBUFScratchOffen>; def gi_flat_scratch_offset : GIComplexOperandMatcher<s32, "selectScratchOffset">, GIComplexPatternEquiv<ScratchOffset>; def gi_flat_scratch_saddr : GIComplexOperandMatcher<s32, "selectScratchSAddr">, GIComplexPatternEquiv<ScratchSAddr>; def gi_flat_scratch_svaddr : GIComplexOperandMatcher<s32, "selectScratchSVAddr">, GIComplexPatternEquiv<ScratchSVAddr>; def gi_ds_1addr_1offset : GIComplexOperandMatcher<s32, "selectDS1Addr1Offset">, GIComplexPatternEquiv<DS1Addr1Offset>; def gi_ds_64bit_4byte_aligned : GIComplexOperandMatcher<s64, "selectDS64Bit4ByteAligned">, GIComplexPatternEquiv<DS64Bit4ByteAligned>; def gi_ds_128bit_8byte_aligned : GIComplexOperandMatcher<s64, "selectDS128Bit8ByteAligned">, GIComplexPatternEquiv<DS128Bit8ByteAligned>; def gi_mubuf_addr64 : GIComplexOperandMatcher<s64, "selectMUBUFAddr64">, GIComplexPatternEquiv<MUBUFAddr64>; def gi_mubuf_offset : GIComplexOperandMatcher<s64, "selectMUBUFOffset">, GIComplexPatternEquiv<MUBUFOffset>; def gi_smrd_buffer_imm : GIComplexOperandMatcher<s64, "selectSMRDBufferImm">, GIComplexPatternEquiv<SMRDBufferImm>; def gi_smrd_buffer_imm32 : GIComplexOperandMatcher<s64, "selectSMRDBufferImm32">, GIComplexPatternEquiv<SMRDBufferImm32>; // Separate load nodes are defined to glue m0 initialization in // SelectionDAG. The GISel selector can just insert m0 initialization // directly before selecting a glue-less load, so hide this // distinction. def : GINodeEquiv<G_LOAD, AMDGPUld_glue> { let CheckMMOIsNonAtomic = 1; let IfSignExtend = G_SEXTLOAD; let IfZeroExtend = G_ZEXTLOAD; } def : GINodeEquiv<G_STORE, AMDGPUst_glue> { let CheckMMOIsNonAtomic = 1; } def : GINodeEquiv<G_LOAD, AMDGPUatomic_ld_glue> { bit CheckMMOIsAtomic = 1; } def : GINodeEquiv<G_STORE, AMDGPUatomic_st_glue> { bit CheckMMOIsAtomic = 1; } def : GINodeEquiv<G_ATOMIC_CMPXCHG, atomic_cmp_swap_glue>; def : GINodeEquiv<G_ATOMICRMW_XCHG, atomic_swap_glue>; def : GINodeEquiv<G_ATOMICRMW_ADD, atomic_load_add_glue>; def : GINodeEquiv<G_ATOMICRMW_SUB, atomic_load_sub_glue>; def : GINodeEquiv<G_ATOMICRMW_AND, atomic_load_and_glue>; def : GINodeEquiv<G_ATOMICRMW_OR, atomic_load_or_glue>; def : GINodeEquiv<G_ATOMICRMW_XOR, atomic_load_xor_glue>; def : GINodeEquiv<G_ATOMICRMW_MIN, atomic_load_min_glue>; def : GINodeEquiv<G_ATOMICRMW_MAX, atomic_load_max_glue>; def : GINodeEquiv<G_ATOMICRMW_UMIN, atomic_load_umin_glue>; def : GINodeEquiv<G_ATOMICRMW_UMAX, atomic_load_umax_glue>; def : GINodeEquiv<G_ATOMICRMW_FADD, atomic_load_fadd_glue>; def : GINodeEquiv<G_AMDGPU_FFBH_U32, AMDGPUffbh_u32_impl>; def : GINodeEquiv<G_AMDGPU_FFBL_B32, AMDGPUffbl_b32_impl>; def : GINodeEquiv<G_AMDGPU_FMIN_LEGACY, AMDGPUfmin_legacy>; def : GINodeEquiv<G_AMDGPU_FMAX_LEGACY, AMDGPUfmax_legacy>; def : GINodeEquiv<G_AMDGPU_RCP_IFLAG, AMDGPUrcp_iflag>; def : GINodeEquiv<G_AMDGPU_CVT_F32_UBYTE0, AMDGPUcvt_f32_ubyte0>; def : GINodeEquiv<G_AMDGPU_CVT_F32_UBYTE1, AMDGPUcvt_f32_ubyte1>; def : GINodeEquiv<G_AMDGPU_CVT_F32_UBYTE2, AMDGPUcvt_f32_ubyte2>; def : GINodeEquiv<G_AMDGPU_CVT_F32_UBYTE3, AMDGPUcvt_f32_ubyte3>; def : GINodeEquiv<G_AMDGPU_CVT_PK_I16_I32, AMDGPUpk_i16_i32_impl>; def : GINodeEquiv<G_AMDGPU_SMED3, AMDGPUsmed3>; def : GINodeEquiv<G_AMDGPU_UMED3, AMDGPUumed3>; def : GINodeEquiv<G_AMDGPU_FMED3, AMDGPUfmed3_impl>; def : GINodeEquiv<G_AMDGPU_CLAMP, AMDGPUclamp>; def : GINodeEquiv<G_AMDGPU_ATOMIC_CMPXCHG, AMDGPUatomic_cmp_swap>; def : GINodeEquiv<G_AMDGPU_BUFFER_LOAD, SIbuffer_load>; def : GINodeEquiv<G_AMDGPU_BUFFER_LOAD_USHORT, SIbuffer_load_ushort>; def : GINodeEquiv<G_AMDGPU_BUFFER_LOAD_UBYTE, SIbuffer_load_ubyte>; def : GINodeEquiv<G_AMDGPU_BUFFER_LOAD_SSHORT, SIbuffer_load_short>; def : GINodeEquiv<G_AMDGPU_BUFFER_LOAD_SBYTE, SIbuffer_load_byte>; def : GINodeEquiv<G_AMDGPU_BUFFER_LOAD_FORMAT, SIbuffer_load_format>; def : GINodeEquiv<G_AMDGPU_BUFFER_LOAD_FORMAT_D16, SIbuffer_load_format_d16>; def : GINodeEquiv<G_AMDGPU_TBUFFER_LOAD_FORMAT, SItbuffer_load>; def : GINodeEquiv<G_AMDGPU_TBUFFER_LOAD_FORMAT_D16, SItbuffer_load_d16>; def : GINodeEquiv<G_AMDGPU_BUFFER_STORE, SIbuffer_store>; def : GINodeEquiv<G_AMDGPU_BUFFER_STORE_SHORT, SIbuffer_store_short>; def : GINodeEquiv<G_AMDGPU_BUFFER_STORE_BYTE, SIbuffer_store_byte>; def : GINodeEquiv<G_AMDGPU_BUFFER_STORE_FORMAT, SIbuffer_store_format>; def : GINodeEquiv<G_AMDGPU_BUFFER_STORE_FORMAT_D16, SIbuffer_store_format_d16>; def : GINodeEquiv<G_AMDGPU_TBUFFER_STORE_FORMAT, SItbuffer_store>; def : GINodeEquiv<G_AMDGPU_TBUFFER_STORE_FORMAT_D16, SItbuffer_store_d16>; // FIXME: Check MMO is atomic def : GINodeEquiv<G_AMDGPU_ATOMIC_INC, SIatomic_inc>; def : GINodeEquiv<G_AMDGPU_ATOMIC_DEC, SIatomic_dec>; def : GINodeEquiv<G_AMDGPU_ATOMIC_INC, atomic_inc_glue>; def : GINodeEquiv<G_AMDGPU_ATOMIC_DEC, atomic_dec_glue>; def : GINodeEquiv<G_AMDGPU_ATOMIC_FMIN, SIatomic_fmin>; def : GINodeEquiv<G_AMDGPU_ATOMIC_FMAX, SIatomic_fmax>; def : GINodeEquiv<G_AMDGPU_ATOMIC_FMIN, atomic_load_fmin_glue>; def : GINodeEquiv<G_AMDGPU_ATOMIC_FMAX, atomic_load_fmax_glue>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_SWAP, SIbuffer_atomic_swap>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_ADD, SIbuffer_atomic_add>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_SUB, SIbuffer_atomic_sub>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_SMIN, SIbuffer_atomic_smin>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_UMIN, SIbuffer_atomic_umin>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_SMAX, SIbuffer_atomic_smax>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_UMAX, SIbuffer_atomic_umax>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_AND, SIbuffer_atomic_and>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_OR, SIbuffer_atomic_or>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_XOR, SIbuffer_atomic_xor>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_INC, SIbuffer_atomic_inc>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_DEC, SIbuffer_atomic_dec>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_FADD, SIbuffer_atomic_fadd>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_FMIN, SIbuffer_atomic_fmin>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_FMAX, SIbuffer_atomic_fmax>; def : GINodeEquiv<G_AMDGPU_BUFFER_ATOMIC_CMPSWAP, SIbuffer_atomic_cmpswap>; def : GINodeEquiv<G_AMDGPU_S_BUFFER_LOAD, SIsbuffer_load>; def : GINodeEquiv<G_FPTRUNC_ROUND_UPWARD, SIfptrunc_round_upward>; def : GINodeEquiv<G_FPTRUNC_ROUND_DOWNWARD, SIfptrunc_round_downward>; class GISelSop2Pat < SDPatternOperator node, Instruction inst, ValueType dst_vt, ValueType src0_vt = dst_vt, ValueType src1_vt = src0_vt> : GCNPat < (dst_vt (node (src0_vt SReg_32:$src0), (src1_vt SReg_32:$src1))), (inst src0_vt:$src0, src1_vt:$src1) >; class GISelVop2Pat < SDPatternOperator node, Instruction inst, ValueType dst_vt, ValueType src0_vt = dst_vt, ValueType src1_vt = src0_vt> : GCNPat < (dst_vt (node (src0_vt (sd_vsrc0 src0_vt:$src0)), (src1_vt VGPR_32:$src1))), (inst src0_vt:$src0, src1_vt:$src1) >; class GISelVop2CommutePat < SDPatternOperator node, Instruction inst, ValueType dst_vt, ValueType src0_vt = dst_vt, ValueType src1_vt = src0_vt> : GCNPat < (dst_vt (node (src1_vt VGPR_32:$src1), (src0_vt (sd_vsrc0 src0_vt:$src0)))), (inst src0_vt:$src0, src1_vt:$src1) >; class GISelVop3Pat2 < SDPatternOperator node, Instruction inst, ValueType dst_vt, ValueType src0_vt = dst_vt, ValueType src1_vt = src0_vt> : GCNPat < (dst_vt (node (src0_vt (sd_vcsrc src0_vt:$src0)), (src1_vt (sd_vcsrc src1_vt:$src1)))), (inst src0_vt:$src0, src1_vt:$src1) >; class GISelVop3Pat2CommutePat < SDPatternOperator node, Instruction inst, ValueType dst_vt, ValueType src0_vt = dst_vt, ValueType src1_vt = src0_vt> : GCNPat < (dst_vt (node (src0_vt (sd_vcsrc src0_vt:$src0)), (src1_vt (sd_vcsrc src1_vt:$src1)))), (inst src0_vt:$src1, src1_vt:$src0) >; class GISelVop3Pat2ModsPat < SDPatternOperator node, Instruction inst, ValueType dst_vt, ValueType src0_vt = dst_vt, ValueType src1_vt = src0_vt> : GCNPat < (dst_vt (node (src0_vt (VOP3Mods0 src0_vt:$src0, i32:$src0_modifiers, i1:$clamp, i32:$omods)), (src1_vt (VOP3Mods src1_vt:$src1, i32:$src1_modifiers)))), (inst i32:$src0_modifiers, src0_vt:$src0, i32:$src1_modifiers, src1_vt:$src1, $clamp, $omods) >; multiclass GISelVop2IntrPat < SDPatternOperator node, Instruction inst, ValueType dst_vt, ValueType src_vt = dst_vt> { def : GISelVop2Pat <node, inst, dst_vt, src_vt>; // FIXME: Intrinsics aren't marked as commutable, so we need to add an explicit // pattern to handle commuting. This is another reason why legalizing to a // generic machine instruction may be better that matching the intrinsic // directly. def : GISelVop2CommutePat <node, inst, dst_vt, src_vt>; } // Since GlobalISel is more flexible then SelectionDAG, I think we can get // away with adding patterns for integer types and not legalizing all // loads and stores to vector types. This should help simplify the load/store // legalization. foreach Ty = [i64, p0, p1, p4] in { defm : SMRD_Pattern <"S_LOAD_DWORDX2", Ty>; } def gi_as_i32timm : GICustomOperandRenderer<"renderTruncTImm">, GISDNodeXFormEquiv<as_i32timm>; def gi_as_i16timm : GICustomOperandRenderer<"renderTruncTImm">, GISDNodeXFormEquiv<as_i16timm>; def gi_as_i8timm : GICustomOperandRenderer<"renderTruncTImm">, GISDNodeXFormEquiv<as_i8timm>; def gi_as_i1timm : GICustomOperandRenderer<"renderTruncTImm">, GISDNodeXFormEquiv<as_i1timm>; def gi_NegateImm : GICustomOperandRenderer<"renderNegateImm">, GISDNodeXFormEquiv<NegateImm>; def gi_bitcast_fpimm_to_i32 : GICustomOperandRenderer<"renderBitcastImm">, GISDNodeXFormEquiv<bitcast_fpimm_to_i32>; def gi_IMMPopCount : GICustomOperandRenderer<"renderPopcntImm">, GISDNodeXFormEquiv<IMMPopCount>; def gi_extract_cpol : GICustomOperandRenderer<"renderExtractCPol">, GISDNodeXFormEquiv<extract_cpol>; def gi_extract_swz : GICustomOperandRenderer<"renderExtractSWZ">, GISDNodeXFormEquiv<extract_swz>; def gi_set_glc : GICustomOperandRenderer<"renderSetGLC">, GISDNodeXFormEquiv<set_glc>; def gi_frameindex_to_targetframeindex : GICustomOperandRenderer<"renderFrameIndex">, GISDNodeXFormEquiv<frameindex_to_targetframeindex>;