Lines Matching refs:i

271 	int i, ncpu_min;  in mpo_update_tunables()  local
292 for (i = 0; i < n_lgrpnodes; i++) { in mpo_update_tunables()
293 int ncpu = mpo_lgroup[i].ncpu; in mpo_update_tunables()
310 int n_cpunodes, i; in cpuid_to_cpunode() local
324 for (i = 0; i < n_cpunodes; i++) { in cpuid_to_cpunode()
325 if (md_get_prop_val(md, cpunodes[i], PROP_LG_CPU_ID, in cpuid_to_cpunode()
329 foundnode = cpunodes[i]; in cpuid_to_cpunode()
389 int i; in mpo_cpu_add() local
401 i = mpo_cpu_to_lgroup(md, cpunode); in mpo_cpu_add()
402 if (i == -1) in mpo_cpu_add()
405 mpo_cpu[cpuid].lgrp_index = i; in mpo_cpu_add()
406 mpo_cpu[cpuid].home = mpo_lgroup[i].addr_match >> home_mask_shift; in mpo_cpu_add()
407 mpo_lgroup[i].ncpu++; in mpo_cpu_add()
418 int i; in mpo_cpu_remove() local
423 i = mpo_cpu[cpuid].lgrp_index; in mpo_cpu_remove()
424 mpo_lgroup[i].ncpu--; in mpo_cpu_remove()
460 int i, j, result; in lgrp_update() local
476 for (i = 0; i < n_lgrpnodes; i++) { in lgrp_update()
477 mpo_lgroup[i].node = lgrpnodes[i]; in lgrp_update()
478 mpo_lgroup[i].id = i; in lgrp_update()
479 mpo_lgroup[i].ncpu = 0; in lgrp_update()
480 result = get_int(md, lgrpnodes[i], PROP_LG_MASK, in lgrp_update()
481 &mpo_lgroup[i].addr_mask); in lgrp_update()
482 result |= get_int(md, lgrpnodes[i], PROP_LG_MATCH, in lgrp_update()
483 &mpo_lgroup[i].addr_match); in lgrp_update()
489 mpo_lgroup[i].addr_mask = 0; in lgrp_update()
490 mpo_lgroup[i].addr_match = 0; in lgrp_update()
495 result = get_int(md, lgrpnodes[i], PROP_LG_LATENCY, in lgrp_update()
496 &mpo_lgroup[i].latency); in lgrp_update()
498 mpo_lgroup[i].latency = 0; in lgrp_update()
515 for (i = 0; i < n_lgrpnodes; i++) { in lgrp_update()
516 if (mpo_lgroup[0].addr_mask != mpo_lgroup[i].addr_mask) { in lgrp_update()
531 for (i = 0; i < n_lgrpnodes; i++) { in lgrp_update()
532 j = md_alloc_scan_dag(md, mpo_lgroup[i].node, in lgrp_update()
546 for (i = 0; i < n_lgrpnodes; i++) in lgrp_update()
547 mpo_lgroup[i].node = MDE_INVAL_ELEM_COOKIE; in lgrp_update()
567 uint64_t i, k, stripe, stride; in lgrp_traverse() local
662 for (i = 0; i < n_lgrpnodes; i++) { in lgrp_traverse()
664 mpo_lgroup[i].addr_match >> home_mask_shift); in lgrp_traverse()
693 for (i = 1; i < n_lgrpnodes; i++) { in lgrp_traverse()
694 if (mpo_lgroup[i].latency < lower_latency) { in lgrp_traverse()
695 lower_latency = mpo_lgroup[i].latency; in lgrp_traverse()
697 if (mpo_lgroup[i].latency > higher_latency) { in lgrp_traverse()
698 higher_latency = mpo_lgroup[i].latency; in lgrp_traverse()
706 for (i = 0; i < NCPU; i++) { in lgrp_traverse()
707 mpo_cpu[i].home = 0; in lgrp_traverse()
708 mpo_cpu[i].lgrp_index = -1; in lgrp_traverse()
712 for (i = 0; i < n_cpunodes; i++) { in lgrp_traverse()
715 result = get_int(md, cpunodes[i], PROP_LG_CPU_ID, &k); in lgrp_traverse()
722 o = mpo_cpu_to_lgroup(md, cpunodes[i]); in lgrp_traverse()
999 int i, mnode; in plat_pfn_to_mem_node() local
1014 for (i = 0; i < n_mblocks; i++) { in plat_pfn_to_mem_node()
1040 int i; in plat_rapfn_to_papfn() local
1053 for (mb = &mpo_mblock[0], i = 0; i < n_mblocks; i++, mb++) { in plat_rapfn_to_papfn()
1105 int i; in plat_mem_node_iterator_init() local
1147 i = 0; in plat_mem_node_iterator_init()
1150 i = it->mi_last_mblock; in plat_mem_node_iterator_init()
1152 mem_stripes[i * max_locality_groups + mnode].physmax); in plat_mem_node_iterator_init()
1153 if (++i == n_mblocks) { in plat_mem_node_iterator_init()
1165 for (; i < n_mblocks; i++) { in plat_mem_node_iterator_init()
1166 if (P2PHASE(mpo_mblock[i].ra_to_pa, szcpagesize)) in plat_mem_node_iterator_init()
1168 ms = &mem_stripes[i * max_locality_groups + mnode]; in plat_mem_node_iterator_init()
1174 if (i == n_mblocks) { in plat_mem_node_iterator_init()
1175 it->mi_last_mblock = i - 1; in plat_mem_node_iterator_init()
1180 it->mi_last_mblock = i; in plat_mem_node_iterator_init()
1182 mblock = &mpo_mblock[i]; in plat_mem_node_iterator_init()
1197 ASSERT((pfn + szcpgcnt - 1) <= mpo_mblock[i].end_pfn); in plat_mem_node_iterator_init()
1215 int i, npages; in plat_mem_node_intersect_range() local
1251 for (i = mnode; i < n_mem_stripes; i += max_locality_groups) { in plat_mem_node_intersect_range()
1252 ms = &mem_stripes[i]; in plat_mem_node_intersect_range()
1324 int i, max_szc; in valid_pages() local
1364 for (i = 0; i < n_mblocks; i++) { in valid_pages()
1399 if (i + 1 < n_mblocks && in valid_pages()
1432 int i, j; in fix_interleave() local
1436 for (i = 0; i < n_lgrpnodes; i++) { in fix_interleave()
1437 if ((mpo_lgroup[i].addr_mask & PAGEOFFSET) != 0) { in fix_interleave()
1439 mask = mpo_lgroup[i].addr_mask; in fix_interleave()
1441 mpo_lgroup[j++] = mpo_lgroup[i]; in fix_interleave()
1563 uint64_t i, j; in mblock_update() local
1567 for (i = 0, j = 0; j < mc->mc_nmblocks; j++) { in mblock_update()
1571 &mblock[i].base); in mblock_update()
1580 &mblock[i].size); in mblock_update()
1589 PROP_LG_RA_PA_OFFSET, &mblock[i].ra_to_pa); in mblock_update()
1593 mblock[i].ra_to_pa = 0; in mblock_update()
1596 "ra_to_pa = %lx\n", i, in mblock_update()
1597 mblock[i].base, in mblock_update()
1598 mblock[i].size, in mblock_update()
1599 mblock[i].ra_to_pa); in mblock_update()
1602 if (mblock[i].base > mblock[i].base + mblock[i].size) { in mblock_update()
1606 mblock[i].base, mblock[i].size); in mblock_update()
1612 if (mblock[i].size != 0) { in mblock_update()
1613 uint64_t base = mblock[i].base; in mblock_update()
1614 uint64_t end = base + mblock[i].size; in mblock_update()
1616 mblock[i].base_pfn = btop(base); in mblock_update()
1617 mblock[i].end_pfn = btop(end - 1); in mblock_update()
1618 i++; in mblock_update()
1622 if (i == 0) { in mblock_update()
1629 ASSERT(i <= mc->mc_nmblocks); in mblock_update()
1630 mc->mc_nmblocks = i; in mblock_update()
1690 int i, j; in mblock_update_del() local
1710 for (i = 0; i < nmblocks; i++) { in mblock_update_del()
1712 base = btop(mblock[i].base); in mblock_update_del()
1713 end = base + btop(mblock[i].size) - 1; in mblock_update_del()
1728 for (j = i; j < nmblocks - 1; j++) in mblock_update_del()
1734 for (j = nmblocks - 1; j >= i; j--) in mblock_update_del()
1736 mblock[i].size = ptob(ubase - base); in mblock_update_del()
1737 mblock[i].end_pfn = ubase - 1; in mblock_update_del()
1738 mblock[i + 1].base = ptob(uend + 1); in mblock_update_del()
1739 mblock[i + 1].size = ptob(end - uend); in mblock_update_del()
1740 mblock[i + 1].base_pfn = uend + 1; in mblock_update_del()
1745 " i=%d base=0x%lx end=0x%lx", i, base, end); in mblock_update_del()
1746 mblock[i].base = ptob(uend + 1); in mblock_update_del()
1747 mblock[i].size -= ptob(uend - ubase + 1); in mblock_update_del()
1749 mblock[i].base_pfn = base; in mblock_update_del()
1750 mblock[i].end_pfn = end; in mblock_update_del()
1755 " i=%d base=0x%lx end=0x%lx", i, base, end); in mblock_update_del()
1756 mblock[i].size -= ptob(uend - ubase + 1); in mblock_update_del()
1758 mblock[i].base_pfn = base; in mblock_update_del()
1759 mblock[i].end_pfn = end; in mblock_update_del()
1778 int i, mnode; in mstripe_update() local
1796 for (i = 0; i < nmblocks; i++) { in mstripe_update()
1797 mc->mc_stripes[i].exists = 1; in mstripe_update()
1798 mc->mc_stripes[i].physbase = mblock[i].base_pfn; in mstripe_update()
1799 mc->mc_stripes[i].physmax = mblock[i].end_pfn; in mstripe_update()
1810 for (i = 0; i < nmblocks; i++) { in mstripe_update()
1811 base = mblock[i].base; in mstripe_update()
1812 end = base + mblock[i].size; in mstripe_update()
1813 ra_to_pa = mblock[i].ra_to_pa; in mstripe_update()
1853 ms = &mc->mc_stripes[i * max_locality_groups + mnode]; in mstripe_update()
1889 int i, j, mnode, found; in mnode_update() local
1906 for (i = 0; i < mc->mc_nmblocks; i++) { in mnode_update()
1909 j = i * max_locality_groups + mnode; in mnode_update()
1953 for (i = 0; i < max_mem_nodes; i++) { in mnode_update()
1954 if (!mem_node_config[i].exists) in mnode_update()
1956 if (ubase >= mem_node_config[i].physbase && in mnode_update()
1957 ubase <= mem_node_config[i].physmax) in mnode_update()
1959 if (uend >= mem_node_config[i].physbase && in mnode_update()
1960 uend <= mem_node_config[i].physmax) in mnode_update()