[Openmp-commits] [openmp] r273293 - Performance improvement: accessing thread struct as opposed to team struct
Jonathan Peyton via Openmp-commits
openmp-commits at lists.llvm.org
Tue Jun 21 11:30:16 PDT 2016
Author: jlpeyton
Date: Tue Jun 21 13:30:15 2016
New Revision: 273293
URL: http://llvm.org/viewvc/llvm-project?rev=273293&view=rev
Log:
Performance improvement: accessing thread struct as opposed to team struct
Replaced readings of nproc from team structure with ones from
thread structure to improve performance.
Patch by Andrey Churbanov.
Differential Revision: http://reviews.llvm.org/D21559
Modified:
openmp/trunk/runtime/src/kmp_dispatch.cpp
Modified: openmp/trunk/runtime/src/kmp_dispatch.cpp
URL: http://llvm.org/viewvc/llvm-project/openmp/trunk/runtime/src/kmp_dispatch.cpp?rev=273293&r1=273292&r2=273293&view=diff
==============================================================================
--- openmp/trunk/runtime/src/kmp_dispatch.cpp (original)
+++ openmp/trunk/runtime/src/kmp_dispatch.cpp Tue Jun 21 13:30:15 2016
@@ -738,7 +738,7 @@ __kmp_dispatch_init(
}
/* guided analytical not safe for too many threads */
- if ( team->t.t_nproc > 1<<20 && schedule == kmp_sch_guided_analytical_chunked ) {
+ if ( schedule == kmp_sch_guided_analytical_chunked && th->th.th_team_nproc > 1<<20 ) {
schedule = kmp_sch_guided_iterative_chunked;
KMP_WARNING( DispatchManyThreads );
}
@@ -836,7 +836,7 @@ __kmp_dispatch_init(
#if ( KMP_STATIC_STEAL_ENABLED && KMP_ARCH_X86_64 )
case kmp_sch_static_steal:
{
- T nproc = team->t.t_nproc;
+ T nproc = th->th.th_team_nproc;
T ntc, init;
KD_TRACE(100, ("__kmp_dispatch_init: T#%d kmp_sch_static_steal case\n", gtid ) );
@@ -869,7 +869,7 @@ __kmp_dispatch_init(
#endif
case kmp_sch_static_balanced:
{
- T nproc = team->t.t_nproc;
+ T nproc = th->th.th_team_nproc;
T init, limit;
KD_TRACE(100, ("__kmp_dispatch_init: T#%d kmp_sch_static_balanced case\n",
@@ -933,7 +933,7 @@ __kmp_dispatch_init(
} // case
case kmp_sch_guided_iterative_chunked :
{
- T nproc = team->t.t_nproc;
+ T nproc = th->th.th_team_nproc;
KD_TRACE(100,("__kmp_dispatch_init: T#%d kmp_sch_guided_iterative_chunked case\n",gtid));
if ( nproc > 1 ) {
@@ -956,7 +956,7 @@ __kmp_dispatch_init(
break;
case kmp_sch_guided_analytical_chunked:
{
- T nproc = team->t.t_nproc;
+ T nproc = th->th.th_team_nproc;
KD_TRACE(100, ("__kmp_dispatch_init: T#%d kmp_sch_guided_analytical_chunked case\n", gtid));
if ( nproc > 1 ) {
@@ -1074,8 +1074,8 @@ __kmp_dispatch_init(
break;
case kmp_sch_static_greedy:
KD_TRACE(100,("__kmp_dispatch_init: T#%d kmp_sch_static_greedy case\n",gtid));
- pr->u.p.parm1 = ( team -> t.t_nproc > 1 ) ?
- ( tc + team->t.t_nproc - 1 ) / team->t.t_nproc :
+ pr->u.p.parm1 = ( th->th.th_team_nproc > 1 ) ?
+ ( tc + th->th.th_team_nproc - 1 ) / th->th.th_team_nproc :
tc;
break;
case kmp_sch_static_chunked :
@@ -1095,7 +1095,7 @@ __kmp_dispatch_init(
parm1 = chunk;
/* F : size of the first cycle */
- parm2 = ( tc / (2 * team->t.t_nproc) );
+ parm2 = ( tc / (2 * th->th.th_team_nproc) );
if ( parm2 < 1 ) {
parm2 = 1;
@@ -1793,7 +1793,7 @@ __kmp_dispatch_next(
if ( p_st != NULL ) *p_st = incr;
- pr->u.p.count += team->t.t_nproc;
+ pr->u.p.count += th->th.th_team_nproc;
if ( incr == 1 ) {
*p_lb = start + init;
@@ -1963,8 +1963,8 @@ __kmp_dispatch_next(
trip = pr->u.p.tc;
- KMP_DEBUG_ASSERT(team->t.t_nproc > 1);
- KMP_DEBUG_ASSERT((2UL * chunkspec + 1) * (UT)team->t.t_nproc < trip);
+ KMP_DEBUG_ASSERT(th->th.th_team_nproc > 1);
+ KMP_DEBUG_ASSERT((2UL * chunkspec + 1) * (UT)th->th.th_team_nproc < trip);
while(1) { /* this while loop is a safeguard against unexpected zero chunk sizes */
chunkIdx = test_then_inc_acq< ST >((volatile ST *) & sh->u.s.iteration );
@@ -2135,7 +2135,7 @@ __kmp_dispatch_next(
}
#endif
- if ( (ST)num_done == team->t.t_nproc-1 ) {
+ if ( (ST)num_done == th->th.th_team_nproc - 1 ) {
/* NOTE: release this buffer to be reused */
KMP_MB(); /* Flush all pending memory write invalidates. */
More information about the Openmp-commits
mailing list