; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -mtriple=arm64-apple-ios -mattr=+strict-align < %s | FileCheck %s ; Small (16 bytes here) unaligned memcpy() should be a function call if ; strict-alignment is turned on. define void @t0(i8* %out, i8* %in) { ; CHECK-LABEL: t0: ; CHECK: ; %bb.0: ; %entry ; CHECK-NEXT: stp x29, x30, [sp, #-16]! ; 16-byte Folded Spill ; CHECK-NEXT: .cfi_def_cfa_offset 16 ; CHECK-NEXT: .cfi_offset w30, -8 ; CHECK-NEXT: .cfi_offset w29, -16 ; CHECK-NEXT: mov w2, #16 ; CHECK-NEXT: bl _memcpy ; CHECK-NEXT: ldp x29, x30, [sp], #16 ; 16-byte Folded Reload ; CHECK-NEXT: ret entry: call void @llvm.memcpy.p0i8.p0i8.i64(i8* %out, i8* %in, i64 16, i1 false) ret void } ; Small (16 bytes here) aligned memcpy() should be inlined even if ; strict-alignment is turned on. define void @t1(i8* align 8 %out, i8* align 8 %in) { ; CHECK-LABEL: t1: ; CHECK: ; %bb.0: ; %entry ; CHECK-NEXT: ldp x9, x8, [x1] ; CHECK-NEXT: stp x9, x8, [x0] ; CHECK-NEXT: ret entry: call void @llvm.memcpy.p0i8.p0i8.i64(i8* align 8 %out, i8* align 8 %in, i64 16, i1 false) ret void } ; Tiny (4 bytes here) unaligned memcpy() should be inlined with byte sized ; loads and stores if strict-alignment is turned on. define void @t2(i8* %out, i8* %in) { ; CHECK-LABEL: t2: ; CHECK: ; %bb.0: ; %entry ; CHECK-NEXT: ldrb w8, [x1, #3] ; CHECK-NEXT: ldrb w9, [x1, #2] ; CHECK-NEXT: ldrb w10, [x1, #1] ; CHECK-NEXT: ldrb w11, [x1] ; CHECK-NEXT: strb w8, [x0, #3] ; CHECK-NEXT: strb w9, [x0, #2] ; CHECK-NEXT: strb w10, [x0, #1] ; CHECK-NEXT: strb w11, [x0] ; CHECK-NEXT: ret entry: call void @llvm.memcpy.p0i8.p0i8.i64(i8* %out, i8* %in, i64 4, i1 false) ret void } declare void @llvm.memcpy.p0i8.p0i8.i64(i8* nocapture, i8* nocapture readonly, i64, i1)