[LLVMdev] [RFC] OpenMP offload infrastructure

John Leidel (jleidel) jleidel at micron.com
Mon Aug 11 11:59:48 PDT 2014


Samuel, thanks for the response.  I have a few short responses below.  

cheers
john


John D. Leidel
Software Compiler Development Manager
Micron Technology, Inc.
jleidel at micron.com
office: 972-521-5271
cell: 214-578-8510

On Aug 11, 2014, at 11:51 AM, Samuel Antão <samuelfantao at gmail.com> wrote:

> Hi John,
> 
> Thank you for the comments. I am addressing some of them bellow.
> 
> Regards,
> Samuel
> 
> 
> 2014-08-11 9:36 GMT-04:00 John Leidel (jleidel) <jleidel at micron.com>:
> Sergey [et.al], thanks for putting this proposal together.  Overall, this looks like a pretty solid approach to providing relatively hardware agnostic omp target functionality.  I had several comments/questions as summarized below:
> 
> Pros:
> - We [local colleagues and myself] like the concise target API.  We’re big fans of KISS development principles.
> - We believe this provides a good basis for future work in heterogeneous OMP support
> 
> Comments/Questions:
> - There doesn’t seem to be any mention of how mutable each runtime function is with respect to its target execution region.  The core OMP spec document notes in several places that certain user-visible runtime calls have “implementation defined” behavior depending upon where/how they’re used.  For example, what happens if the host runtime issues a __tgt_target_data_update() while the target is currently executing (__tgt_rtl_run_target_region() )?  Is this implementation defined?  I’m certainly ok with that answer, but I believe we need to explicitly state what the behavior is.
> 
> In my view the user-visible OpenMP calls that apply to target regions depend on the state kept in libtarget.so, and are therefore device-type independent. What is device dependent is how the OpenMP terminology is mapped. For example, get_num_teams() would operate on top of the state kept in libtarget.so but how the device interpret a team is device dependent and deviced by the target dependent runtime.
>  
> A different issue is how the RTL implementation for calls that are common for target and host (i.e. the kmpc_ calls) should be implemented. I think it is a good idea to have some flexibility in the codegen to tune the generation of these calls if the default interface is not suitable for a given target. But in general, the kmpc_ library implementation should be known to the toolchain of that target so it can properly drive the linking.  
> 
> About the specific example you mentioned. If I understand it correctly, following the current version of the spec tgt_rtl_run_target_region() has to be blocking so libtarget.so would have to wait for the update to be issued.  The actions in libtarget.so would have to be sequential exactly has the codegeneration expects. If for some reason these constraints change in future specs, both codegeneration and libtarget.so implementation would have to be made consistent.

I’ll echo back my understanding of your statements.  OpenMP user calls are well defined and apply to the calling target region (which is what I expect from my interactions with the subcommittee).  The ‘kmpc_*’ library call implementations (especially the ability to generate these calls) is “implementation defined.”  I believe this is the best path in order to allow for high performance implementations for orthogonal target architectures.  Finally, your statements regarding my example indicate that a target execution region is instantiated sequentially (eg, blocking) with respect to the calling construct (thread/task).  This was also my assumption.  I wanted to make sure others interpreted the document/spec the same way.    

> 
> 
> - I noticed that Alexandre Eichenberger was one of the authors.  Has he mentioned any support/compatibility with the profiling interfaces he (JMC, et.al.) proposed?  How does one integrate the proposed profiling runtime logic with a target region (specifically the dispatch & data movement interfaces)?  This would be very handy.
> 
> - I don’t see any mention of an interface to query the physical details of a device.  I know this strays a bit from the notion of portability, but it would be nice to have a simple interface (similar to ‘omp_get_max_threads’).  I stop short of querying information as detailed as provided by hwloc, but it would be nice for the user to have the ability to query the targets and see which ones are appropriate for execution.  This would essentially provide you the ability to build different implementations of a kernel and make a runtime decision on which one to execute.  EG,
> if( /* target of some specific type present */ ){
>     /* use the omp target interface */
> }else{
>    /* use the normal worksharing or tasking interfaces */
> }
> 
> (I realize this is more of an OMP spec question)
> 
> I agree this is more of an OMP spec issue. The fact we are addressing different device-types is already an extension to the spec which poses some issues. One of them, somehow related with this, is how the device ids are mapped to device types. Should this depend on flags passed to the compiler ( e.g. omptargets=A,B with ids 0-1 assigned to A and 2-3 to B given that the runtime identified in the system two devices of each), or should it depend on the environment? In the current proposal, libtarget.so abstracts a single target made of several targets, do we want to let the user prioritize which exact device to use? Should this be decided at compile time or runtime? 

If my memory serves me, the original OMP 4.0 spec for target execution/data regions is based upon the notion of a host + 1 target, rather than a host + `N` different targets.  This was definitely the right decision as we have to crawl before we can walk.  You bring up an interesting point with respect to the priority of the devices discovered/dispatched.  I fear this is a rather complex issue (although an interesting one).  

>  
> 
> - It would be nice to define a runtime and/or environment mechanism that permits the user to enable/disable specific targets.  For example, if a system had four GPUs, but you only wanted to enable two, it would be convenient to do so using an environment variable.  I realize that one could do this using actual runtime calls in the code with some amount of intelligence, but this somewhat defeats the purpose of portability.  Again, this is more related to the 4.x spec, but it does have implications in the lower-level runtime.
> 
> 
> I think this can be solved by the target dependent RTL alone by returning the number of available devices to libtarget.so based on some env variable specified by the RTL.
>  
> 
> cheers
> john
> 
> 
> On Aug 8, 2014, at 5:22 PM, Sergey Ostanevich <sergos.gnu at gmail.com> wrote:
> 
> > Hello everybody!
> >
> > I would like to present a proposal for implementation of OpenMP
> > offloading in LLVM. It was created by a list of authors and covers the
> > runtime part at most and at a very high level. I believe it will be
> > good to have input from community at this early stage before moving
> > deeper in details.
> >
> > The driver part is intentionally not touched, since we have no clear
> > vision on how one can use 3rd party compiler for target code
> > generation and incorporate its results into the final host link phase.
> > I hope to hear from you more on this.
> >
> > I invite you to take part in discussion of the document. Critics,
> > proposals, updates - all are welcome!
> >
> > Thank you,
> > Sergey Ostanevich
> > Open Source Compilers
> > Intel Corporation
> > <offload-proposal.pdf>_______________________________________________
> > LLVM Developers mailing list
> > LLVMdev at cs.uiuc.edu         http://llvm.cs.uiuc.edu
> > http://lists.cs.uiuc.edu/mailman/listinfo/llvmdev
> 
> 
> _______________________________________________
> LLVM Developers mailing list
> LLVMdev at cs.uiuc.edu         http://llvm.cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvmdev





More information about the llvm-dev mailing list