Lines Matching +full:0 +full:xfe400000

107 { return !(pmd_val(pmd) & 0xFFFFFFF); }  in srmmu_pmd_none()
121 #define MSI_MBUS_ARBEN 0xe0001008 /* MBus Arbiter Enable register */
126 #define MSI_ASYNC_MODE 0x80000000 /* Operate the MSI asynchronously */
130 __asm__ __volatile__ ("lda [%0] %1, %%g3\n\t" in msi_set_sync()
132 "sta %%g3, [%0] %1\n\t" : : in msi_set_sync()
154 printk(KERN_ERR "Size 0x%x too small for nocache request\n", in __srmmu_get_nocache()
159 printk(KERN_ERR "Size 0x%x unaligned in nocache request\n", in __srmmu_get_nocache()
186 memset(tmp, 0, size); in srmmu_get_nocache()
198 printk("Vaddr %lx is smaller than nocache base 0x%lx\n", in srmmu_free_nocache()
203 printk("Vaddr %lx is bigger than nocache end 0x%lx\n", in srmmu_free_nocache()
208 printk("Size 0x%x is not a power of 2\n", size); in srmmu_free_nocache()
212 printk("Size 0x%x is too small\n", size); in srmmu_free_nocache()
216 printk("Vaddr %lx is not aligned to size 0x%x\n", vaddr, size); in srmmu_free_nocache()
232 unsigned long total = 0; in probe_memory()
235 for (i = 0; sp_banks[i].num_bytes; i++) in probe_memory()
283 panic("%s: Failed to allocate %lu bytes align=0x%x\n", in srmmu_nocache_init()
285 memset(srmmu_nocache_pool, 0, srmmu_nocache_size); in srmmu_nocache_init()
296 memset(__nocache_fix(srmmu_swapper_pg_dir), 0, SRMMU_PGD_TABLE_SIZE); in srmmu_nocache_init()
332 pgd_t *init = pgd_offset_k(0); in get_pgd_fast()
333 memset(pgd, 0, USER_PTRS_PER_PGD * sizeof(pgd_t)); in get_pgd_fast()
354 if ((ptep = pte_alloc_one_kernel(mm)) == 0) in pte_alloc_one()
458 for (ctx = 0; ctx < numctx; ctx++) { in sparc_context_init()
467 for (ctx = 0; ctx < numctx; ctx++) in sparc_context_init()
524 while (len != 0) { in srmmu_mapiorange()
554 while (len != 0) { in srmmu_unmapiorange()
591 #if 0 /* P3: deadwood to debug precise flushes on Swift. */
604 __asm__ __volatile__("sta %%g0, [%0] %1\n\t" : :
613 __asm__ __volatile__("sta %%g0, [%0] %1\n\t" : :
697 memset(__nocache_fix(pmdp), 0, SRMMU_PMD_TABLE_SIZE); in srmmu_early_allocate_ptable_skeleton()
705 memset(__nocache_fix(ptep), 0, PTE_SIZE); in srmmu_early_allocate_ptable_skeleton()
708 if (start > (0xffffffffUL - PMD_SIZE)) in srmmu_early_allocate_ptable_skeleton()
731 memset(pmdp, 0, SRMMU_PMD_TABLE_SIZE); in srmmu_allocate_ptable_skeleton()
740 memset(ptep, 0, PTE_SIZE); in srmmu_allocate_ptable_skeleton()
743 if (start > (0xffffffffUL - PMD_SIZE)) in srmmu_allocate_ptable_skeleton()
757 __asm__ __volatile__("lda [%1] %2, %0\n\t" : in srmmu_probe()
759 "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); in srmmu_probe()
781 int what; /* 0 = normal-pte, 1 = pmd-level pte, 2 = pgd-level pte */ in srmmu_inherit_prom_mappings()
784 if (start == 0) in srmmu_inherit_prom_mappings()
786 if (start == 0xfef00000) in srmmu_inherit_prom_mappings()
796 what = 0; in srmmu_inherit_prom_mappings()
822 memset(__nocache_fix(pmdp), 0, SRMMU_PMD_TABLE_SIZE); in srmmu_inherit_prom_mappings()
835 memset(__nocache_fix(ptep), 0, PTE_SIZE); in srmmu_inherit_prom_mappings()
883 if (phys_base > 0) { in map_kernel()
887 for (i = 0; sp_banks[i].num_bytes != 0; i++) { in map_kernel()
914 num_contexts = 0; in srmmu_paging_init()
915 while (cpunode != 0) { in srmmu_paging_init()
918 num_contexts = prom_getintdefault(cpunode, "mmu-nctx", 0x8); in srmmu_paging_init()
930 pages_avail = 0; in srmmu_paging_init()
935 srmmu_inherit_prom_mappings(0xfe400000, (LINUX_OPPROM_ENDVM - PAGE_SIZE)); in srmmu_paging_init()
942 for (i = 0; i < num_contexts; i++) in srmmu_paging_init()
977 unsigned long max_zone_pfn[MAX_NR_ZONES] = { 0 }; in srmmu_paging_init()
1003 return 0; in init_new_context()
1034 int cpu = 0; in init_vac_layout()
1035 unsigned long max_size = 0; in init_vac_layout()
1036 unsigned long min_line_size = 0x10000000; in init_vac_layout()
1040 while ((nd = prom_getsibling(nd)) != 0) { in init_vac_layout()
1069 if (nd == 0) { in init_vac_layout()
1094 #if 0 /* XXX I think this is bad news... -DaveM */ in poke_hypersparc()
1170 #define SWIFT_MASKID_ADDR 0x10003018
1175 __asm__ __volatile__("lda [%1] %2, %0\n\t" in init_swift()
1176 "srl %0, 0x18, %0\n\t" : in init_swift()
1181 case 0x11: in init_swift()
1182 case 0x20: in init_swift()
1183 case 0x23: in init_swift()
1184 case 0x30: in init_swift()
1205 case 0x25: in init_swift()
1206 case 0x31: in init_swift()
1221 flush_page_for_dma_global = 0; in init_swift()
1341 case 0: /* No SE cache */ in poke_turbosparc()
1512 viking_mxcc_present = 0; in init_viking()
1526 flush_page_for_dma_global = 0; in init_viking()
1551 hwbug_bitmask = 0; in get_srmmu_type()
1554 mod_typ = (mreg & 0xf0000000) >> 28; in get_srmmu_type()
1555 mod_rev = (mreg & 0x0f000000) >> 24; in get_srmmu_type()
1556 psr_typ = (psr >> 28) & 0xf; in get_srmmu_type()
1557 psr_vers = (psr >> 24) & 0xf; in get_srmmu_type()
1572 case 0: in get_srmmu_type()
1591 if (psr_typ == 0 && psr_vers == 5) { in get_srmmu_type()
1597 if (psr_typ == 0 && psr_vers == 4) { in get_srmmu_type()
1603 while ((cpunode = prom_getsibling(cpunode)) != 0) { in get_srmmu_type()
1621 ((psr_vers == 0) || in get_srmmu_type()
1622 ((psr_vers == 1) && (mod_typ == 0) && (mod_rev == 0)))) { in get_srmmu_type()