[LLVMdev] Proposal: Generic auto-vectorization and parallelization approach for LLVM and Polly

Tobias Grosser grosser at fim.uni-passau.de
Sat Jan 8 16:07:16 PST 2011


On 01/08/2011 06:52 PM, Renato Golin wrote:
> On 8 January 2011 18:27, Tobias Grosser<grosser at fim.uni-passau.de>  wrote:
>> OK. First of all to agree on a name, we decided to call the Polyhedral
>> analysis we develop PoLLy, as in Polly the parrot. ;-) Maybe it was a
>> misleading choice?
>
> I never realised... ;) Polly it is!
>
>
>> In general as I explained I agree that a three stage approach is useful,
>> for the reasons you explained, however it is more overhead (and just
>> implementation work) than the one we use now. I currently do not have the
>> time to implement the proposed approach. In case anybody is interested to
>> work on patches, I am happy to support this.
>
> Good. If it's just a matter of time (not design), things can be left
> ready for future implementation without breaking the current model. I
> thought there was a fundamental flaw with the three-stage design (and
> was eager to learn it).
Not at all. It was just soo much easier to create vector instructions 
right inside of Polly that I did not start another large project.

>> the only change
>> in polly is, that it either generates N scalar instructions per original
>> instruction or one vector instruction (if N is the number of loop iterations
>> which is equivalent to the vector width). So vectorization in Polly was very
>> easy to implement and already works reasonable well.
>
> Ok, this comes with another current change in LLVM: OpenCL. I explain.
>
> OpenCL has very large (and odd) vector sizes, that if implemented to
> vectorized units (like SSE or NEON), need to be legalised.
>
> Such a pass should be target specific and polly could make use of
> that. If polly always generate vector code (instead of reason if the
> number of unrolled operations are the same as the current target being
> compiled into), the later legalisation pass can deal with the odd
> sized vectors and transform into multiples of legal vector + some
> surplus of the module as normal instructions.
>
> Also, if the target doesn't have vector units, there could be a
> generic (or not) transformation to cpu instructions (if there isn't
> one already), so that makes your polly pass completely target
> agnostic.

I believe LLVM already has lowering of wide LLVM-IR vectors to a set of 
operations of vectors of target specific wide. This is done in the back 
end and works correct for all cases I tested. The problem with this is 
just performance. They fastest loop nest is different on architectures 
with different vector width and size of vector registers. A generic 
polyhedral optimizer will therefore not generate incorrect code, however 
the proposed transformations are wrong. If Polly e.g. always generates 
vector operations to be 32 floats width, this would result in not very 
pleasant code, but might still be lowered to something well performing 
on cpus supporting AVX. Matching the target vector width in our 
heuristics will obviously give the best performance. So to get optimal 
performance Polly needs to take target data into account.

Talking about OpenCL. The lowering you described for the large vector 
instructions sounds reasonable. Optimal code would however probably 
produced by revisiting the whole loop structure and generating one that 
is performance wise optimal for the target architecture.

>> Why are target specific vectorization passes needed to generate vector
>> instructions from Polly? The only target specific information I currently
>> see is the vector width, which a generic vectorization pass can obtain from
>> the target data information. Could you explain for which features target
>> specific vectorization would be needed?
>
> Not target specific, generic vectors. See above.
>
>
>> I have read this and the look interesting. I suppose they are created out of
>> the box, if a pass generates LLVM-IR vector instructions?
>
> Yup. It's pretty neat. SSE is probably similar, but with NEON, a
> pattern-match is done when the variable type is a vector.
>
> So, a multiplication followed by an addition in the right way is
> transformed into a multiply-and-add NEON instruction.
>
> An example (in a completely wrong IR, just to make a point):
>
> %a =<4 x i32>
> %b =<4 x i32>
> %c =<4 x i32>
> %mul = mul %b, %c
> %acc = add %mul, %a
>
> gets transformed into:
>
> VMLA.I32 q0, q1, q2
>
> Multiplying vectors (of the correct size) gets into VMUL, adding gets
> to VADD and so on...

Yes. I have seen this also for powerpc and hope we can soon use the AVX 
multiply and add instruction too.

Cheers
Tobi



More information about the llvm-dev mailing list