reference, declaration → definition definition → references, declarations, derived classes, virtual overrides reference to multiple definitions → definitions unreferenced |
2133 if (threadInfo[num_avail][i] != UINT_MAX) { 2162 if (threadInfo[num_avail][osIdIndex] != UINT_MAX) 2172 threadInfo[num_avail][osIdIndex] = val; 2195 if (threadInfo[num_avail][pkgIdIndex] != UINT_MAX) 2197 threadInfo[num_avail][pkgIdIndex] = val; 2207 if (threadInfo[num_avail][coreIdIndex] != UINT_MAX) 2209 threadInfo[num_avail][coreIdIndex] = val; 2220 if (threadInfo[num_avail][threadIdIndex] != UINT_MAX) 2222 threadInfo[num_avail][threadIdIndex] = val; 2233 if (threadInfo[num_avail][nodeIdIndex + level] != UINT_MAX) 2235 threadInfo[num_avail][nodeIdIndex + level] = val; 2255 if ((int)num_avail == __kmp_xproc) { 2263 if (threadInfo[num_avail][osIdIndex] == UINT_MAX) { 2275 if (!KMP_CPU_ISSET(threadInfo[num_avail][osIdIndex], 2277 INIT_PROC_INFO(threadInfo[num_avail]); 2283 num_avail++; 2284 KMP_ASSERT(num_avail <= num_records); 2285 INIT_PROC_INFO(threadInfo[num_avail]); 2314 KMP_ASSERT(num_avail > 0); 2315 KMP_ASSERT(num_avail <= num_records); 2316 if (num_avail == 1) { 2371 qsort(threadInfo, num_avail, sizeof(*threadInfo), 2416 for (i = 1; i < num_avail; i++) { 2576 for (i = 0; i < num_avail; ++i) { // fill the os indices 2611 *address2os = (AddrUnsPair *)__kmp_allocate(sizeof(AddrUnsPair) * num_avail); 2616 for (i = 0; i < num_avail; ++i) { 2671 __kmp_affinity_print_topology(*address2os, num_avail, depth, pkgLevel,