[Libclc-dev] [PATCH 1/3] Add vload* for addrspace(2) and use as constant load for R600

Tom Stellard tom at stellard.net
Fri Aug 9 20:35:47 PDT 2013


On Fri, Aug 09, 2013 at 02:42:56PM -0500, Aaron Watry wrote:
> Signed-off-by: Aaron Watry <awatry at gmail.com>

For the series:

Reviewed-by: Tom Stellard <thomas.stellard at amd.com>

> ---
>  generic/lib/shared/vload_impl.ll  | 33 ++++++++++++++++++++++++++++++++-
>  generic/lib/shared/vstore_impl.ll |  3 ++-
>  r600/lib/shared/vload.cl          | 10 ++++++++--
>  3 files changed, 42 insertions(+), 4 deletions(-)
> 
> diff --git a/generic/lib/shared/vload_impl.ll b/generic/lib/shared/vload_impl.ll
> index 2e70e5f..2416aaf 100644
> --- a/generic/lib/shared/vload_impl.ll
> +++ b/generic/lib/shared/vload_impl.ll
> @@ -1,4 +1,5 @@
> -; This provides optimized implementations of vload4/8/16 for 32-bit int/uint
> +; This provides optimized implementations of vload2/3/4/8/16 for 32-bit int/uint
> +; The address spaces get mapped to data types in target-specific usages
>  
>  define <2 x i32> @__clc_vload2_i32__addr1(i32 addrspace(1)* nocapture %addr) nounwind readonly alwaysinline {
>    %1 = bitcast i32 addrspace(1)* %addr to <2 x i32> addrspace(1)*
> @@ -30,6 +31,36 @@ define <16 x i32> @__clc_vload16_i32__addr1(i32 addrspace(1)* nocapture %addr) n
>    ret <16 x i32> %2
>  }
>  
> +define <2 x i32> @__clc_vload2_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline {
> +  %1 = bitcast i32 addrspace(2)* %addr to <2 x i32> addrspace(2)*
> +  %2 = load <2 x i32> addrspace(2)* %1, align 4, !tbaa !3
> +  ret <2 x i32> %2
> +}
> +
> +define <3 x i32> @__clc_vload3_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline {
> +  %1 = bitcast i32 addrspace(2)* %addr to <3 x i32> addrspace(2)*
> +  %2 = load <3 x i32> addrspace(2)* %1, align 4, !tbaa !3
> +  ret <3 x i32> %2
> +}
> +
> +define <4 x i32> @__clc_vload4_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline {
> +  %1 = bitcast i32 addrspace(2)* %addr to <4 x i32> addrspace(2)*
> +  %2 = load <4 x i32> addrspace(2)* %1, align 4, !tbaa !3
> +  ret <4 x i32> %2
> +}
> +
> +define <8 x i32> @__clc_vload8_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline {
> +  %1 = bitcast i32 addrspace(2)* %addr to <8 x i32> addrspace(2)*
> +  %2 = load <8 x i32> addrspace(2)* %1, align 4, !tbaa !3
> +  ret <8 x i32> %2
> +}
> +
> +define <16 x i32> @__clc_vload16_i32__addr2(i32 addrspace(2)* nocapture %addr) nounwind readonly alwaysinline {
> +  %1 = bitcast i32 addrspace(2)* %addr to <16 x i32> addrspace(2)*
> +  %2 = load <16 x i32> addrspace(2)* %1, align 4, !tbaa !3
> +  ret <16 x i32> %2
> +}
> +
>  !1 = metadata !{metadata !"char", metadata !5}
>  !2 = metadata !{metadata !"short", metadata !5}
>  !3 = metadata !{metadata !"int", metadata !5}
> diff --git a/generic/lib/shared/vstore_impl.ll b/generic/lib/shared/vstore_impl.ll
> index 388bce2..9e2a37b 100644
> --- a/generic/lib/shared/vstore_impl.ll
> +++ b/generic/lib/shared/vstore_impl.ll
> @@ -1,4 +1,5 @@
> -; This provides optimized implementations of vstore4/8/16 for 32-bit int/uint
> +; This provides optimized implementations of vstore2/3/4/8/16 for 32-bit int/uint
> +; The address spaces get mapped to data types in target-specific usages
>  
>  define void @__clc_vstore2_i32__addr1(<2 x i32> %vec, i32 addrspace(1)* nocapture %addr) nounwind alwaysinline {
>    %1 = bitcast i32 addrspace(1)* %addr to <2 x i32> addrspace(1)*
> diff --git a/r600/lib/shared/vload.cl b/r600/lib/shared/vload.cl
> index 6144dde..60502d1 100644
> --- a/r600/lib/shared/vload.cl
> +++ b/r600/lib/shared/vload.cl
> @@ -48,10 +48,8 @@ VLOAD_TYPES()
>  
>  VLOAD_VECTORIZE(int, __private)
>  VLOAD_VECTORIZE(int, __local)
> -VLOAD_VECTORIZE(int, __constant)
>  VLOAD_VECTORIZE(uint, __private)
>  VLOAD_VECTORIZE(uint, __local)
> -VLOAD_VECTORIZE(uint, __constant)
>  
>  _CLC_OVERLOAD _CLC_DEF int3 vload3(size_t offset, const global int *x) {
>    return (int3)(vload2(0, &x[3*offset]), x[3*offset+2]);
> @@ -59,6 +57,12 @@ _CLC_OVERLOAD _CLC_DEF int3 vload3(size_t offset, const global int *x) {
>  _CLC_OVERLOAD _CLC_DEF uint3 vload3(size_t offset, const global uint *x) {
>    return (uint3)(vload2(0, &x[3*offset]), x[3*offset+2]);
>  }
> +_CLC_OVERLOAD _CLC_DEF int3 vload3(size_t offset, const constant int *x) {
> +  return (int3)(vload2(0, &x[3*offset]), x[3*offset+2]);
> +}
> +_CLC_OVERLOAD _CLC_DEF uint3 vload3(size_t offset, const constant uint *x) {
> +  return (uint3)(vload2(0, &x[3*offset]), x[3*offset+2]);
> +}
>  
>  //We only define functions for typeN vloadN(), and then just bitcast the result for unsigned types
>  #define _CLC_VLOAD_ASM_DECL(PRIM_TYPE,LLVM_SCALAR_TYPE,ADDR_SPACE,ADDR_SPACE_ID) \
> @@ -83,9 +87,11 @@ _CLC_DECL PRIM_TYPE##16 __clc_vload16_##LLVM_SCALAR_TYPE##__addr##ADDR_SPACE_ID
>  
>  #define _CLC_VLOAD_ASM_OVERLOAD_ADDR_SPACES(PRIM_TYPE,S_PRIM_TYPE,LLVM_TYPE) \
>    _CLC_VLOAD_ASM_OVERLOAD_SIZES(PRIM_TYPE, S_PRIM_TYPE, LLVM_TYPE, global, 1) \
> +  _CLC_VLOAD_ASM_OVERLOAD_SIZES(PRIM_TYPE, S_PRIM_TYPE, LLVM_TYPE, constant, 2) \
>  
>  #define _CLC_VLOAD_ASM_OVERLOADS() \
>    _CLC_VLOAD_ASM_DECL(int,i32,__global,1) \
> +  _CLC_VLOAD_ASM_DECL(int,i32,__constant,2) \
>    _CLC_VLOAD_ASM_OVERLOAD_ADDR_SPACES(int,int,i32) \
>    _CLC_VLOAD_ASM_OVERLOAD_ADDR_SPACES(uint,int,i32) \
>  
> -- 
> 1.8.1.2
> 
> 
> _______________________________________________
> Libclc-dev mailing list
> Libclc-dev at pcc.me.uk
> http://www.pcc.me.uk/cgi-bin/mailman/listinfo/libclc-dev




More information about the Libclc-dev mailing list