From 196d16f2b18465b85228383e92a30f4b5f2b113c Mon Sep 17 00:00:00 2001 From: Olivier Houchard Date: Wed, 31 Dec 2025 02:41:38 +0100 Subject: [PATCH] MINOR: cpu-topo: Rename variables to better fit their usage Rename "visited_tsid" and "visited_ccx" to "touse_tsid" and "touse_ccx". They are not there to remember which tsid/ccx we alreaday visited, contrarily to visited_ccx_set and visited_cl_set, they are there to know which tsid/ccx we should use, so make that clear. --- src/cpu_topo.c | 72 +++++++++++++++++++++++++------------------------- 1 file changed, 36 insertions(+), 36 deletions(-) diff --git a/src/cpu_topo.c b/src/cpu_topo.c index c79092c77..57cdbc448 100644 --- a/src/cpu_topo.c +++ b/src/cpu_topo.c @@ -1077,8 +1077,8 @@ static int find_next_cpu_ccx(int start, int l3id) static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin, int gmax, char **err) { struct hap_cpuset node_cpu_set; - struct hap_cpuset visited_tsid; - struct hap_cpuset visited_ccx; + struct hap_cpuset touse_tsid; + struct hap_cpuset touse_ccx; int first_node_id = -1; int second_node_id = -1; int cpu; @@ -1119,8 +1119,8 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin * and make a CPU set of them. */ ha_cpuset_zero(&node_cpu_set); - ha_cpuset_zero(&visited_tsid); - ha_cpuset_zero(&visited_ccx); + ha_cpuset_zero(&touse_tsid); + ha_cpuset_zero(&touse_ccx); for (cpu = cpu_count = 0; cpu <= cpu_topo_lastcpu; cpu++) { if (ha_cpu_topo[cpu].no_id != first_node_id) ha_cpu_topo[cpu].st |= HA_CPU_F_IGNORED; @@ -1128,9 +1128,9 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin ha_cpuset_set(&node_cpu_set, ha_cpu_topo[cpu].idx); cpu_count++; - ha_cpuset_set(&visited_ccx, ha_cpu_topo[cpu].ca_id[3]); - if (!(cpu_policy_conf.flags & CPU_POLICY_ONE_THREAD_PER_CORE) || !ha_cpuset_isset(&visited_tsid, ha_cpu_topo[cpu].ts_id)) { - ha_cpuset_set(&visited_tsid, ha_cpu_topo[cpu].ts_id); + ha_cpuset_set(&touse_ccx, ha_cpu_topo[cpu].ca_id[3]); + if (!(cpu_policy_conf.flags & CPU_POLICY_ONE_THREAD_PER_CORE) || !ha_cpuset_isset(&touse_tsid, ha_cpu_topo[cpu].ts_id)) { + ha_cpuset_set(&touse_tsid, ha_cpu_topo[cpu].ts_id); thr_count++; } } @@ -1145,7 +1145,7 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin int corenb = 0; ha_cpuset_zero(&thrset); - tsid = ha_cpuset_ffs(&visited_tsid) - 1; + tsid = ha_cpuset_ffs(&touse_tsid) - 1; if (tsid != -1) { int next_try = 0; int got_cpu; @@ -1158,7 +1158,7 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin ha_cpuset_set(&thrset, ha_cpu_topo[got_cpu].idx); } } - ha_cpuset_clr(&visited_tsid, tsid); + ha_cpuset_clr(&touse_tsid, tsid); } if (cpu_policy_conf.flags & CPU_POLICY_ONE_THREAD_PER_CORE) same_core = 1; @@ -1185,7 +1185,7 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin int next_cpu = 0; int tsid; - tsid = ha_cpuset_ffs(&visited_tsid) - 1; + tsid = ha_cpuset_ffs(&touse_tsid) - 1; got_cpu = find_next_cpu_tsid(0, tsid); while ((got_cpu = find_next_cpu_tsid(next_cpu, tsid)) != -1) { if (!(ha_cpu_topo[got_cpu].st & HA_CPU_F_EXCL_MASK)) @@ -1196,7 +1196,7 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin if (got_cpu != -1) { ha_cpuset_set(&thrset, ha_cpu_topo[got_cpu].idx); } - ha_cpuset_clr(&visited_tsid, tsid); + ha_cpuset_clr(&touse_tsid, tsid); } else { int tid = ha_cpuset_ffs(&node_cpu_set) - 1; @@ -1218,14 +1218,14 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin int next_try = 0; if (same_ccx == 0) { - int l3id = ha_cpuset_ffs(&visited_ccx) - 1; + int l3id = ha_cpuset_ffs(&touse_ccx) - 1; ha_cpuset_zero(&thrset); while ((got_cpu = find_next_cpu_ccx(next_try, l3id)) != -1) { next_try = got_cpu + 1; same_ccx++; ha_cpuset_set(&thrset, ha_cpu_topo[got_cpu].idx); } - ha_cpuset_clr(&visited_ccx, l3id); + ha_cpuset_clr(&touse_ccx, l3id); } BUG_ON(same_ccx == 0); if (ha_cpuset_ffs(&thrset) != 0) @@ -1258,7 +1258,7 @@ static int cpu_policy_first_usable_node(int policy, int tmin, int tmax, int gmin } static void -cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct hap_cpuset visited_tsid, struct hap_cpuset visited_ccx) +cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct hap_cpuset touse_tsid, struct hap_cpuset touse_ccx) { struct hap_cpuset thrset; int nb_grp; @@ -1304,7 +1304,7 @@ cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct /* * Find the next available core, and assign the thread to it */ - tsid = ha_cpuset_ffs(&visited_tsid) - 1; + tsid = ha_cpuset_ffs(&touse_tsid) - 1; if (tsid != -1) { int next_try = 0; int got_cpu; @@ -1315,7 +1315,7 @@ cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct corenb++; } } - ha_cpuset_clr(&visited_tsid, tsid); + ha_cpuset_clr(&touse_tsid, tsid); ha_cpuset_assign(&cpu_map[global.nbtgroups].thread[thr], &thrset); } if (cpu_policy_conf.flags & CPU_POLICY_ONE_THREAD_PER_CORE) @@ -1335,7 +1335,7 @@ cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct int next_cpu = 0; int tsid; - tsid = ha_cpuset_ffs(&visited_tsid) - 1; + tsid = ha_cpuset_ffs(&touse_tsid) - 1; while ((got_cpu = find_next_cpu_tsid(next_cpu, tsid)) != -1) { if (!(ha_cpu_topo[got_cpu].st & HA_CPU_F_EXCL_MASK)) break; @@ -1344,7 +1344,7 @@ cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct if (got_cpu != -1) { ha_cpuset_set(&thrset, ha_cpu_topo[got_cpu].idx); - ha_cpuset_clr(&visited_tsid, tsid); + ha_cpuset_clr(&touse_tsid, tsid); } } else { int tid = ha_cpuset_ffs(&node_cpu_set) - 1; @@ -1359,7 +1359,7 @@ cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct } else if (cpu_policy_conf.affinity & CPU_AFFINITY_PER_CCX) { if (same_core == 0) { - int l3id = ha_cpuset_ffs(&visited_ccx) - 1; + int l3id = ha_cpuset_ffs(&touse_ccx) - 1; int got_cpu; int next_try = 0; ha_cpuset_zero(&thrset); @@ -1369,7 +1369,7 @@ cpu_policy_assign_threads(int cpu_count, struct hap_cpuset node_cpu_set, struct same_core++; ha_cpuset_set(&thrset, ha_cpu_topo[got_cpu].idx); } - ha_cpuset_clr(&visited_ccx, l3id); + ha_cpuset_clr(&touse_ccx, l3id); } BUG_ON(same_core == 0); if (ha_cpuset_ffs(&thrset) != 0) @@ -1400,8 +1400,8 @@ static int cpu_policy_group_by_cluster(int policy, int tmin, int tmax, int gmin, { struct hap_cpuset visited_cl_set; struct hap_cpuset node_cpu_set; - struct hap_cpuset visited_tsid; - struct hap_cpuset visited_ccx; + struct hap_cpuset touse_tsid; + struct hap_cpuset touse_ccx; int cpu, cpu_start; int cpu_count; int cid; @@ -1424,8 +1424,8 @@ static int cpu_policy_group_by_cluster(int policy, int tmin, int tmax, int gmin, while (global.nbtgroups < MAX_TGROUPS && global.nbthread < MAX_THREADS) { ha_cpuset_zero(&node_cpu_set); - ha_cpuset_zero(&visited_tsid); - ha_cpuset_zero(&visited_ccx); + ha_cpuset_zero(&touse_tsid); + ha_cpuset_zero(&touse_ccx); cid = -1; cpu_count = 0; for (cpu = cpu_start; cpu <= cpu_topo_lastcpu; cpu++) { @@ -1445,10 +1445,10 @@ static int cpu_policy_group_by_cluster(int policy, int tmin, int tmax, int gmin, /* make a mask of all of this cluster's CPUs */ ha_cpuset_set(&node_cpu_set, ha_cpu_topo[cpu].idx); - ha_cpuset_set(&visited_ccx, ha_cpu_topo[cpu].ca_id[3]); - if (!ha_cpuset_isset(&visited_tsid, ha_cpu_topo[cpu].ts_id)) { + ha_cpuset_set(&touse_ccx, ha_cpu_topo[cpu].ca_id[3]); + if (!ha_cpuset_isset(&touse_tsid, ha_cpu_topo[cpu].ts_id)) { cpu_count++; - ha_cpuset_set(&visited_tsid, ha_cpu_topo[cpu].ts_id); + ha_cpuset_set(&touse_tsid, ha_cpu_topo[cpu].ts_id); } else if (!(cpu_policy_conf.flags & CPU_POLICY_ONE_THREAD_PER_CORE)) cpu_count++; } @@ -1462,7 +1462,7 @@ static int cpu_policy_group_by_cluster(int policy, int tmin, int tmax, int gmin, ha_cpuset_set(&visited_cl_set, cid); - cpu_policy_assign_threads(cpu_count, node_cpu_set, visited_tsid, visited_ccx); + cpu_policy_assign_threads(cpu_count, node_cpu_set, touse_tsid, touse_ccx); } if (global.nbthread) @@ -1486,8 +1486,8 @@ static int cpu_policy_group_by_ccx(int policy, int tmin, int tmax, int gmin, int { struct hap_cpuset visited_ccx_set; struct hap_cpuset node_cpu_set; - struct hap_cpuset visited_tsid; - struct hap_cpuset visited_ccx; /* List of CCXs we'll currently use */ + struct hap_cpuset touse_tsid; + struct hap_cpuset touse_ccx; /* List of CCXs we'll currently use */ int cpu, cpu_start; int cpu_count; int l3id; @@ -1510,8 +1510,8 @@ static int cpu_policy_group_by_ccx(int policy, int tmin, int tmax, int gmin, int while (global.nbtgroups < MAX_TGROUPS && global.nbthread < MAX_THREADS) { ha_cpuset_zero(&node_cpu_set); - ha_cpuset_zero(&visited_tsid); - ha_cpuset_zero(&visited_ccx); + ha_cpuset_zero(&touse_tsid); + ha_cpuset_zero(&touse_ccx); l3id = -1; cpu_count = 0; for (cpu = cpu_start; cpu <= cpu_topo_lastcpu; cpu++) { @@ -1531,10 +1531,10 @@ static int cpu_policy_group_by_ccx(int policy, int tmin, int tmax, int gmin, int /* make a mask of all of this cluster's CPUs */ ha_cpuset_set(&node_cpu_set, ha_cpu_topo[cpu].idx); - ha_cpuset_set(&visited_ccx, ha_cpu_topo[cpu].ca_id[3]); - if (!ha_cpuset_isset(&visited_tsid, ha_cpu_topo[cpu].ts_id)) { + ha_cpuset_set(&touse_ccx, ha_cpu_topo[cpu].ca_id[3]); + if (!ha_cpuset_isset(&touse_tsid, ha_cpu_topo[cpu].ts_id)) { cpu_count++; - ha_cpuset_set(&visited_tsid, ha_cpu_topo[cpu].ts_id); + ha_cpuset_set(&touse_tsid, ha_cpu_topo[cpu].ts_id); } else if (!(cpu_policy_conf.flags & CPU_POLICY_ONE_THREAD_PER_CORE)) cpu_count++; } @@ -1548,7 +1548,7 @@ static int cpu_policy_group_by_ccx(int policy, int tmin, int tmax, int gmin, int ha_cpuset_set(&visited_ccx_set, l3id); - cpu_policy_assign_threads(cpu_count, node_cpu_set, visited_tsid, visited_ccx); + cpu_policy_assign_threads(cpu_count, node_cpu_set, touse_tsid, touse_ccx); } if (global.nbthread)