Lines Matching full:nproc

107                              typename traits_t<T>::unsigned_t nchunks, T nproc,  in __kmp_initialize_self_buffer()  argument
119 T num_procs_with_ecore = nproc - num_procs_with_pcore; in __kmp_initialize_self_buffer()
151 small_chunk = nchunks / nproc; // chunks per thread in __kmp_initialize_self_buffer()
152 extras = nchunks % nproc; in __kmp_initialize_self_buffer()
175 // ub (upper bound), and st (stride). nproc is the number of threads relevant
178 // the function within the group of nproc threads. It will have a value
179 // between 0 and nproc - 1. This is often just the thread id within a team, but
192 T nproc, T tid) { in __kmp_dispatch_init_algorithm() argument
210 "schedule:%%d chunk:%%%s nproc:%%%s tid:%%%s\n", in __kmp_dispatch_init_algorithm()
214 KD_TRACE(10, (buff, gtid, pr, lb, ub, st, schedule, chunk, nproc, tid)); in __kmp_dispatch_init_algorithm()
330 if (schedule == kmp_sch_guided_analytical_chunked && nproc > 1 << 20) { in __kmp_dispatch_init_algorithm()
440 if (nproc > 1 && ntc >= nproc) { in __kmp_dispatch_init_algorithm()
491 float p_ratio = (float)num_procs_with_pcore / nproc; in __kmp_dispatch_init_algorithm()
492 float e_ratio = (float)num_procs_with_ecore / nproc; in __kmp_dispatch_init_algorithm()
495 (((multiplier * num_procs_with_pcore) / nproc) + e_ratio); in __kmp_dispatch_init_algorithm()
521 KMP_DEBUG_ASSERT(nproc == num_procs_with_pcore + num_procs_with_ecore); in __kmp_dispatch_init_algorithm()
551 small_chunk = ntc / nproc; in __kmp_dispatch_init_algorithm()
552 extras = ntc % nproc; in __kmp_dispatch_init_algorithm()
569 // nproc (just a heuristics, could be optimized later on). in __kmp_dispatch_init_algorithm()
570 pr->u.p.parm3 = nproc; in __kmp_dispatch_init_algorithm()
571 pr->u.p.parm4 = (id + 1) % nproc; // remember neighbour tid in __kmp_dispatch_init_algorithm()
592 if (nproc > 1) { in __kmp_dispatch_init_algorithm()
595 if (tc < nproc) { in __kmp_dispatch_init_algorithm()
606 T small_chunk = tc / nproc; in __kmp_dispatch_init_algorithm()
607 T extras = tc % nproc; in __kmp_dispatch_init_algorithm()
610 pr->u.p.parm1 = (id == nproc - 1); in __kmp_dispatch_init_algorithm()
653 T nth = nproc; in __kmp_dispatch_init_algorithm()
672 if (nproc > 1) { in __kmp_dispatch_init_algorithm()
673 if ((2L * chunk + 1) * nproc >= tc) { in __kmp_dispatch_init_algorithm()
679 pr->u.p.parm2 = guided_int_param * nproc * (chunk + 1); in __kmp_dispatch_init_algorithm()
681 guided_flt_param / (double)nproc; // may occupy parm3 and parm4 in __kmp_dispatch_init_algorithm()
702 if (nproc > 1) { in __kmp_dispatch_init_algorithm()
703 if ((2L * chunk + 1) * nproc >= tc) { in __kmp_dispatch_init_algorithm()
708 /* commonly used term: (2 nproc - 1)/(2 nproc) */ in __kmp_dispatch_init_algorithm()
726 long double target = ((long double)chunk * 2 + 1) * nproc / tc; in __kmp_dispatch_init_algorithm()
732 /* commonly used term: (2 nproc - 1)/(2 nproc) */ in __kmp_dispatch_init_algorithm()
733 x = 1.0 - 0.5 / (double)nproc; in __kmp_dispatch_init_algorithm()
828 pr->u.p.parm1 = (nproc > 1) ? (tc + nproc - 1) / nproc : tc; in __kmp_dispatch_init_algorithm()
857 parm2 = (tc / (2 * nproc)); in __kmp_dispatch_init_algorithm()
1327 typename traits_t<T>::signed_t *p_st, T nproc, in __kmp_dispatch_next_algorithm() argument
1344 KMP_DEBUG_ASSERT(tid >= 0 && tid < nproc); in __kmp_dispatch_next_algorithm()
1351 "sh:%%p nproc:%%%s tid:%%%s\n", in __kmp_dispatch_next_algorithm()
1353 KD_TRACE(10, (buff, gtid, pr, sh, nproc, tid)); in __kmp_dispatch_next_algorithm()
1405 T oldVictimId = victimId ? victimId - 1 : nproc - 1; in __kmp_dispatch_next_algorithm()
1411 victimId = (victimId + 1) % nproc; in __kmp_dispatch_next_algorithm()
1417 continue; // try once more (nproc attempts in total) in __kmp_dispatch_next_algorithm()
1427 __kmp_initialize_self_buffer<T>(team, id, pr, nchunks, nproc, in __kmp_dispatch_next_algorithm()
1434 pr->u.p.parm4 = (id + 1) % nproc; // remember neighbour tid in __kmp_dispatch_next_algorithm()
1456 pr->u.p.parm4 = (victimId + 1) % nproc; // shift start victim tid in __kmp_dispatch_next_algorithm()
1465 pr->u.p.parm4 = (victimId + 1) % nproc; // shift start victim tid in __kmp_dispatch_next_algorithm()
1548 T oldVictimId = victimId ? victimId - 1 : nproc - 1; in __kmp_dispatch_next_algorithm()
1554 victimId = (victimId + 1) % nproc; in __kmp_dispatch_next_algorithm()
1560 continue; // try once more (nproc attempts in total) in __kmp_dispatch_next_algorithm()
1570 __kmp_initialize_self_buffer<T>(team, id, pr, nchunks, nproc, in __kmp_dispatch_next_algorithm()
1581 pr->u.p.parm4 = (id + 1) % nproc; // remember neighbour tid in __kmp_dispatch_next_algorithm()
1606 pr->u.p.parm4 = (victimId + 1) % nproc; // shift start victim id in __kmp_dispatch_next_algorithm()
1732 pr->u.p.count += nproc; in __kmp_dispatch_next_algorithm()
1814 pr->u.p.parm2) { // compare with K*nproc*(chunk+1), K=2 by default in __kmp_dispatch_next_algorithm()
1835 *(double *)&pr->u.p.parm3); // divide by K*nproc in __kmp_dispatch_next_algorithm()
1882 // compare with K*nproc*(chunk+1), K=2 by default in __kmp_dispatch_next_algorithm()
1903 // divide by K*nproc in __kmp_dispatch_next_algorithm()
1954 KMP_DEBUG_ASSERT(nproc > 1); in __kmp_dispatch_next_algorithm()
1955 KMP_DEBUG_ASSERT((2UL * chunkspec + 1) * (UT)nproc < trip); in __kmp_dispatch_next_algorithm()