Lines Matching refs:hw_thread
408 kmp_hw_thread_t &hw_thread = hw_threads[idx]; in _remove_radix1_layers() local
410 hw_thread.ids[d] = hw_thread.ids[d + 1]; in _remove_radix1_layers()
463 kmp_hw_thread_t &hw_thread = hw_threads[i]; in _gather_enumeration_information() local
465 int id = hw_thread.ids[layer]; in _gather_enumeration_information()
480 if (hw_thread.attrs.is_core_eff_valid() && in _gather_enumeration_information()
481 hw_thread.attrs.core_eff >= num_core_efficiencies) { in _gather_enumeration_information()
484 num_core_efficiencies = hw_thread.attrs.core_eff + 1; in _gather_enumeration_information()
486 if (hw_thread.attrs.is_core_type_valid()) { in _gather_enumeration_information()
489 if (hw_thread.attrs.get_core_type() == core_types[j]) { in _gather_enumeration_information()
496 core_types[num_core_types++] = hw_thread.attrs.get_core_type(); in _gather_enumeration_information()
504 previous_id[layer] = hw_thread.ids[layer]; in _gather_enumeration_information()
527 kmp_hw_thread_t &hw_thread = hw_threads[i]; in _get_ncores_with_attr() local
528 if (!find_all && hw_thread.ids[above_level] != previous_id[above_level]) { in _get_ncores_with_attr()
531 current = hw_thread.attrs.contains(attr); in _get_ncores_with_attr()
534 if (hw_thread.ids[level] != previous_id[level]) { in _get_ncores_with_attr()
535 if (hw_thread.attrs.contains(attr)) in _get_ncores_with_attr()
542 previous_id[level] = hw_thread.ids[level]; in _get_ncores_with_attr()
567 kmp_hw_thread_t &hw_thread = hw_threads[i]; in _set_sub_ids() local
570 if (hw_thread.ids[j] != previous_id[j]) { in _set_sub_ids()
580 previous_id[j] = hw_thread.ids[j]; in _set_sub_ids()
584 hw_thread.sub_ids[j] = sub_id[j]; in _set_sub_ids()
1277 kmp_hw_thread_t &hw_thread = hw_threads[i]; in filter_hw_subset() local
1282 if (hw_thread.sub_ids[level] != prev_sub_ids[level]) { in filter_hw_subset()
1290 core_eff_sub_ids[get_core_eff_index(hw_thread)]++; in filter_hw_subset()
1292 core_type_sub_ids[get_core_type_index(hw_thread)]++; in filter_hw_subset()
1296 core_eff_sub_ids[get_core_eff_index(hw_thread)] = 0; in filter_hw_subset()
1298 core_type_sub_ids[get_core_type_index(hw_thread)] = 0; in filter_hw_subset()
1305 prev_sub_ids[level] = hw_thread.sub_ids[level]; in filter_hw_subset()
1322 kmp_hw_core_type_t core_type = hw_thread.attrs.get_core_type(); in filter_hw_subset()
1323 int core_eff = hw_thread.attrs.get_core_eff(); in filter_hw_subset()
1341 sub_id = core_type_sub_ids[get_core_type_index(hw_thread)]; in filter_hw_subset()
1343 sub_id = core_eff_sub_ids[get_core_eff_index(hw_thread)]; in filter_hw_subset()
1356 sub_id = hw_thread.sub_ids[level]; in filter_hw_subset()
1366 KMP_CPU_CLR(hw_thread.os_id, filtered_mask); in filter_hw_subset()
1902 kmp_hw_thread_t &hw_thread = __kmp_topology->at(hw_thread_index); in __kmp_affinity_create_hwloc_map() local
1904 hw_thread.clear(); in __kmp_affinity_create_hwloc_map()
1905 hw_thread.ids[index] = pu->logical_index; in __kmp_affinity_create_hwloc_map()
1906 hw_thread.os_id = pu->os_index; in __kmp_affinity_create_hwloc_map()
1912 if (hwloc_bitmap_isset(cpukinds[i].mask, hw_thread.os_id)) { in __kmp_affinity_create_hwloc_map()
1918 hw_thread.attrs.set_core_type(cpukinds[cpukind_index].core_type); in __kmp_affinity_create_hwloc_map()
1919 hw_thread.attrs.set_core_eff(cpukinds[cpukind_index].efficiency); in __kmp_affinity_create_hwloc_map()
1943 hw_thread.ids[index] = memory->logical_index; in __kmp_affinity_create_hwloc_map()
1944 hw_thread.ids[index + 1] = sub_id; in __kmp_affinity_create_hwloc_map()
1956 hw_thread.ids[index] = obj->logical_index; in __kmp_affinity_create_hwloc_map()
1957 hw_thread.ids[index + 1] = sub_id; in __kmp_affinity_create_hwloc_map()
2018 kmp_hw_thread_t &hw_thread = __kmp_topology->at(avail_ct); in __kmp_affinity_create_flat_map() local
2019 hw_thread.clear(); in __kmp_affinity_create_flat_map()
2020 hw_thread.os_id = i; in __kmp_affinity_create_flat_map()
2021 hw_thread.ids[0] = i; in __kmp_affinity_create_flat_map()
2022 hw_thread.ids[1] = 0; in __kmp_affinity_create_flat_map()
2023 hw_thread.ids[2] = 0; in __kmp_affinity_create_flat_map()
2066 kmp_hw_thread_t &hw_thread = __kmp_topology->at(avail_ct++); in __kmp_affinity_create_proc_group_map() local
2067 hw_thread.clear(); in __kmp_affinity_create_proc_group_map()
2068 hw_thread.os_id = i; in __kmp_affinity_create_proc_group_map()
2069 hw_thread.ids[0] = i / BITS_PER_GROUP; in __kmp_affinity_create_proc_group_map()
2070 hw_thread.ids[1] = hw_thread.ids[2] = i % BITS_PER_GROUP; in __kmp_affinity_create_proc_group_map()
2473 kmp_hw_thread_t &hw_thread = __kmp_topology->at(i); in __kmp_affinity_create_apicid_map() local
2474 hw_thread.clear(); in __kmp_affinity_create_apicid_map()
2477 hw_thread.ids[idx++] = threadInfo[i].pkgId; in __kmp_affinity_create_apicid_map()
2480 hw_thread.ids[idx++] = threadInfo[i].coreId; in __kmp_affinity_create_apicid_map()
2483 hw_thread.ids[idx++] = threadInfo[i].threadId; in __kmp_affinity_create_apicid_map()
2485 hw_thread.os_id = os; in __kmp_affinity_create_apicid_map()
2776 kmp_hw_thread_t &hw_thread = __kmp_topology->at(hw_thread_index); in __kmp_affinity_create_x2apicid_map() local
2783 hw_thread.clear(); in __kmp_affinity_create_x2apicid_map()
2784 hw_thread.os_id = proc; in __kmp_affinity_create_x2apicid_map()
2787 hw_thread.ids[idx] = apic_id & my_levels[j].mask; in __kmp_affinity_create_x2apicid_map()
2789 hw_thread.ids[idx] >>= my_levels[j - 1].mask_width; in __kmp_affinity_create_x2apicid_map()
2798 hw_thread.attrs.set_core_type(type); in __kmp_affinity_create_x2apicid_map()
2799 hw_thread.attrs.set_core_eff(efficiency); in __kmp_affinity_create_x2apicid_map()
3571 kmp_hw_thread_t &hw_thread = __kmp_topology->at(i); in __kmp_affinity_create_cpuinfo_map() local
3572 hw_thread.clear(); in __kmp_affinity_create_cpuinfo_map()
3573 hw_thread.os_id = os; in __kmp_affinity_create_cpuinfo_map()
3581 hw_thread.ids[pkgLevel] = threadInfo[i][src_index]; in __kmp_affinity_create_cpuinfo_map()
3583 hw_thread.ids[coreLevel] = threadInfo[i][src_index]; in __kmp_affinity_create_cpuinfo_map()
3585 hw_thread.ids[threadLevel] = threadInfo[i][src_index]; in __kmp_affinity_create_cpuinfo_map()
4269 const kmp_hw_thread_t &hw_thread = __kmp_topology->at(i); in __kmp_affinity_find_core_level() local
4271 if (hw_thread.ids[j] > 0) { in __kmp_affinity_find_core_level()
4337 const kmp_hw_thread_t &hw_thread = __kmp_topology->at(osid_idx); in __kmp_affinity_get_mask_topology_info() local
4340 int id = hw_thread.sub_ids[level]; in __kmp_affinity_get_mask_topology_info()
4354 attrs.core_type = hw_thread.attrs.get_core_type(); in __kmp_affinity_get_mask_topology_info()
4355 attrs.core_eff = hw_thread.attrs.get_core_eff(); in __kmp_affinity_get_mask_topology_info()
4359 if (attrs.core_type != hw_thread.attrs.get_core_type()) in __kmp_affinity_get_mask_topology_info()
4361 if (attrs.core_eff != hw_thread.attrs.get_core_eff()) in __kmp_affinity_get_mask_topology_info()
4409 for (int hw_thread = 0; hw_thread < num_hw_threads; ++hw_thread) { in __kmp_affinity_get_topology_info() local
4410 int os_id = __kmp_topology->at(hw_thread).os_id; in __kmp_affinity_get_topology_info()
4412 __kmp_osid_to_hwthread_map[os_id] = hw_thread; in __kmp_affinity_get_topology_info()