<div dir="ltr">Peter,<div><br></div><div>I'm glad to hear that it works now.</div><div><br></div><div>As for texture lookups, I don't have any good news -- they are still unsupported.</div><div><br></div><div>--Artem<br></div><div><br></div></div><div class="gmail_extra"><br><div class="gmail_quote">On Tue, May 17, 2016 at 1:19 PM, Peter Steinbach <span dir="ltr"><<a href="mailto:steinbac@mpi-cbg.de" target="_blank">steinbac@mpi-cbg.de</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">Hi Artem,<br>
<br>
the described "bug" is gone with llvm/clang trunk. Are there any news on texture memory support with cuda clang yet?<br>
<br>
Thanks a bunch -<br>
P<span class=""><br>
<br>
On 17.05.2016 00:28, Artem Belevich wrote:<br>
</span><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><span class="">
Peter,<br>
<br>
I can't reproduce the problem with recent clang. I've tried compiling<br>
the same smaple code with cuda-7.0 and 7.5.<br>
clang version 3.9.0 (trunk 268962) (llvm/trunk 268980)<br>
<br>
Could you tell me what was clang version you used? If you still see the<br></span>
problem, please file a clang bug on <a href="http://llvm.org/bugs" rel="noreferrer" target="_blank">llvm.org/bugs</a> <<a href="http://llvm.org/bugs" rel="noreferrer" target="_blank">http://llvm.org/bugs</a>>.<span class=""><br>
<br>
--Artem<br>
<br>
<br>
% clang++ -I../../common/inc --cuda-gpu-arch=sm_35 simpleCUFFT.cu<br>
-L/usr/local/cuda-7.5/lib64 -lcufft -lcudart -o simpleCUFFT-clang<br>
% LD_LIBRARY_PATH=/usr/local/cuda-7.5/lib64 ./simpleCUFFT-clang<br>
[simpleCUFFT] is starting...<br>
GPU Device 0: "Tesla K40c" with compute capability 3.5<br>
<br>
Transforming signal cufftExecC2C<br>
Launching ComplexPointwiseMulAndScale<<< >>><br>
Transforming signal back cufftExecC2C<br>
<br>
<br>
On Tue, Apr 5, 2016 at 2:12 PM, Jingyue Wu <<a href="mailto:jingyue@google.com" target="_blank">jingyue@google.com</a><br></span><span class="">
<mailto:<a href="mailto:jingyue@google.com" target="_blank">jingyue@google.com</a>>> wrote:<br>
<br>
Would you mind upload your simpleCUFFT.cu code? It looks related to<br>
device code generation because building was successful.<br>
<br>
On Tue, Apr 5, 2016 at 1:56 PM, Peter Steinbach <<a href="mailto:steinbac@mpi-cbg.de" target="_blank">steinbac@mpi-cbg.de</a><br></span><div><div class="h5">
<mailto:<a href="mailto:steinbac@mpi-cbg.de" target="_blank">steinbac@mpi-cbg.de</a>>> wrote:<br>
<br>
Hi guys,<br>
<br>
first of all, please accept my apologies for contacting you by<br>
mail. I was a bit lost, which mailing list to choose from as<br>
pointed to by<br>
<a href="http://llvm.org/docs/CompileCudaWithLLVM.html" rel="noreferrer" target="_blank">http://llvm.org/docs/CompileCudaWithLLVM.html</a><br>
and the subsequent<br>
<a href="http://llvm.org/docs/#mailing-lists" rel="noreferrer" target="_blank">http://llvm.org/docs/#mailing-lists</a><br>
Feel free to deflect this request to the relevant mailing list<br>
or bug tracker.<br>
<br>
In any case, I am very interested in using GPUCC in favor of<br>
NVCC for a multitude of reasons (C++1X, compilation speed, ...).<br>
I started to "port" my favorite samples from the nvidia SDK.<br>
With clang 3.8, samples-7.0.28/7_CUDALibraries/simpleCUFFT as<br>
compiled with clang produces an error at runtime! Here is what I<br>
see with a K20c:<br>
<br>
$ clang++ --cuda-path=/sw/apps/cuda/7.0.28 -I../../common/inc<br>
-m64 --cuda-gpu-arch=sm_35 --cuda-gpu-arch=sm_35 -o<br>
simpleCUFFT.o -c simpleCUFFT.cu<br>
$ clang++ --cuda-path=/sw/apps/cuda/7.0.28<br>
-L/sw/apps/cuda/7.0.28/lib64 -lcudart -ldl -lrt -pthread -m64<br>
-o simpleCUFFT.llvm simpleCUFFT.o -lcufft<br>
$ ./simpleCUFFT.llvm<br>
[simpleCUFFT] is starting...<br>
GPU Device 0: "Tesla K20c" with compute capability 3.5<br>
<br>
Transforming signal cufftExecC2C<br>
Launching ComplexPointwiseMulAndScale<<< >>><br>
simpleCUFFT.cu(132) : getLastCudaError() CUDA error : Kernel<br>
execution failed [ ComplexPointwiseMulAndScale ] : (8) invalid<br>
device function.<br>
<br>
The same source code works just fine with nvcc 7.0.<br>
Any help would be appreciated.<br>
<br>
Best,<br>
Peter<br>
<br>
PS. From random comments, I had the feeling that you are looking<br>
at the SHOC benchmarks with gpucc. If so, please comment on:<br>
<a href="https://github.com/vetter/shoc/issues/48" rel="noreferrer" target="_blank">https://github.com/vetter/shoc/issues/48</a><br>
I don't wanna do work that is either pointless (support for<br>
textures) or was already done. ;)<br>
--<br>
Peter Steinbach, Dr. rer. nat.<br>
HPC Developer, Scientific Computing Facility<br>
<br>
Max Planck Institute of Molecular Cell Biology and Genetics<br>
Pfotenhauerstr. 108<br>
01307 Dresden<br>
Germany<br>
<br>
<br></div></div>
phone <a href="tel:%2B49%20351%20210%202882" value="+493512102882" target="_blank">+49 351 210 2882</a> <tel:%2B49%20351%20210%202882><br>
fax <a href="tel:%2B49%20351%20210%201689" value="+493512101689" target="_blank">+49 351 210 1689</a> <tel:%2B49%20351%20210%201689><br>
<a href="http://www.mpi-cbg.de" rel="noreferrer" target="_blank">www.mpi-cbg.de</a> <<a href="http://www.mpi-cbg.de" rel="noreferrer" target="_blank">http://www.mpi-cbg.de</a>><br>
<br>
<br>
<br>
<br>
</blockquote><div class="HOEnZb"><div class="h5">
<br>
-- <br>
Peter Steinbach, Dr. rer. nat.<br>
HPC Developer, Scientific Computing Facility<br>
<br>
Max Planck Institute of Molecular Cell Biology and Genetics<br>
Pfotenhauerstr. 108<br>
01307 Dresden<br>
Germany<br>
<br>
<br>
phone <a href="tel:%2B49%20351%20210%202882" value="+493512102882" target="_blank">+49 351 210 2882</a><br>
fax <a href="tel:%2B49%20351%20210%201689" value="+493512101689" target="_blank">+49 351 210 1689</a><br>
<a href="http://www.mpi-cbg.de" rel="noreferrer" target="_blank">www.mpi-cbg.de</a><br>
</div></div></blockquote></div><br><br clear="all"><div><br></div>-- <br><div class="gmail_signature"><div dir="ltr">--Artem Belevich</div></div>
</div>