tag | line | file | source code |
ptep | 13 | arch/sparc/mm/loadmmu.c | void (*set_pte)(pte_t *ptep, pte_t entry); |
ptep | 106 | arch/sparc/mm/srmmu.c | int srmmu_pte_inuse(pte_t *ptep) { return mem_map[MAP_NR(ptep)] != 1; } |
ptep | 107 | arch/sparc/mm/srmmu.c | void srmmu_pte_clear(pte_t *ptep) { pte_val(*ptep) = 0; } |
ptep | 108 | arch/sparc/mm/srmmu.c | void srmmu_pte_reuse(pte_t *ptep) |
ptep | 110 | arch/sparc/mm/srmmu.c | if(!(mem_map[MAP_NR(ptep)] & MAP_PAGE_RESERVED)) |
ptep | 111 | arch/sparc/mm/srmmu.c | mem_map[MAP_NR(ptep)]++; |
ptep | 199 | arch/sparc/mm/srmmu.c | srmmu_pmd_set(pmd_t * pmdp, pte_t * ptep) |
ptep | 201 | arch/sparc/mm/srmmu.c | unsigned long page = (unsigned long) ptep; |
ptep | 431 | arch/sparc/mm/srmmu.c | srmmu_set_pte(pte_t *ptep, pte_t entry) |
ptep | 434 | arch/sparc/mm/srmmu.c | *ptep = entry; |
ptep | 461 | arch/sparc/mm/srmmu.c | pte_t *ptep; |
ptep | 465 | arch/sparc/mm/srmmu.c | ptep = srmmu_pte_offset(pmdp, virt_addr); |
ptep | 466 | arch/sparc/mm/srmmu.c | pte_val(*ptep) = (physaddr >> SRMMU_PTE_PPN_PADDR_SHIFT) & SRMMU_PTE_PPN_MASK; |
ptep | 469 | arch/sparc/mm/srmmu.c | pte_val(*ptep) |= (SRMMU_ACC_S_RDWREXEC | SRMMU_ET_PTE); |
ptep | 471 | arch/sparc/mm/srmmu.c | pte_val(*ptep) |= (SRMMU_ACC_S_RDEXEC | SRMMU_ET_PTE); |
ptep | 473 | arch/sparc/mm/srmmu.c | pte_val(*ptep) |= (bus_type << 28); |
ptep | 474 | arch/sparc/mm/srmmu.c | pte_val(*ptep) &= ~(SRMMU_PTE_C_MASK); /* Make sure cacheable bit is off. */ |
ptep | 638 | arch/sparc/mm/srmmu.c | pte_t *ptep = 0; |
ptep | 665 | arch/sparc/mm/srmmu.c | ptep = srmmu_init_alloc(&mempool, |
ptep | 667 | arch/sparc/mm/srmmu.c | srmmu_pmd_set(pmdp, ptep); |
ptep | 670 | arch/sparc/mm/srmmu.c | ptep = srmmu_pte_offset(pmdp, vaddr); |
ptep | 671 | arch/sparc/mm/srmmu.c | *ptep = srmmu_mk_pte(vaddr, SRMMU_PAGE_KERNEL); |
ptep | 685 | arch/sparc/mm/srmmu.c | ptep = srmmu_init_alloc(&mempool, |
ptep | 687 | arch/sparc/mm/srmmu.c | srmmu_pmd_set(pmdp, ptep); |
ptep | 704 | arch/sparc/mm/srmmu.c | ptep = srmmu_init_alloc(&mempool, |
ptep | 706 | arch/sparc/mm/srmmu.c | srmmu_pmd_set(pmdp, ptep); |
ptep | 708 | arch/sparc/mm/srmmu.c | ptep = srmmu_pte_offset(pmdp, vaddr); |
ptep | 710 | arch/sparc/mm/srmmu.c | *ptep++ = srmmu_mk_pte((unsigned int) start, SRMMU_PAGE_KERNEL); |
ptep | 712 | arch/sparc/mm/srmmu.c | *ptep++ = srmmu_mk_pte((unsigned int) srmmu_init_alloc(&mempool, PAGE_SIZE), |
ptep | 715 | arch/sparc/mm/srmmu.c | *ptep++ = srmmu_mk_pte((unsigned int) srmmu_init_alloc(&mempool, PAGE_SIZE), |
ptep | 718 | arch/sparc/mm/srmmu.c | *ptep = srmmu_mk_pte((unsigned int) srmmu_init_alloc(&mempool, PAGE_SIZE), |
ptep | 757 | arch/sparc/mm/srmmu.c | ptep = srmmu_init_alloc(&mempool, |
ptep | 759 | arch/sparc/mm/srmmu.c | srmmu_pmd_set(pmdp, ptep); |
ptep | 762 | arch/sparc/mm/srmmu.c | ptep = srmmu_pte_offset(pmdp, vaddr); |
ptep | 763 | arch/sparc/mm/srmmu.c | pte_val(*ptep) = prom_pte; |
ptep | 804 | arch/sparc/mm/srmmu.c | pte_t *ptep; |
ptep | 119 | arch/sparc/mm/sun4c.c | int sun4c_pte_inuse(pte_t *ptep) { return mem_map[MAP_NR(ptep)] != 1; } |
ptep | 120 | arch/sparc/mm/sun4c.c | void sun4c_pte_clear(pte_t *ptep) { pte_val(*ptep) = 0; } |
ptep | 121 | arch/sparc/mm/sun4c.c | void sun4c_pte_reuse(pte_t *ptep) |
ptep | 123 | arch/sparc/mm/sun4c.c | if(!(mem_map[MAP_NR(ptep)] & MAP_PAGE_RESERVED)) |
ptep | 124 | arch/sparc/mm/sun4c.c | mem_map[MAP_NR(ptep)]++; |
ptep | 190 | arch/sparc/mm/sun4c.c | sun4c_pgd_set(pgd_t * pgdp, pte_t * ptep) |
ptep | 193 | arch/sparc/mm/sun4c.c | pgd_val(*pgdp) |= (((((unsigned long) ptep)) >> |
ptep | 240 | arch/sparc/mm/sun4c.c | sun4c_pmd_set(pmd_t * pmdp, pte_t * ptep) |
ptep | 243 | arch/sparc/mm/sun4c.c | pmd_val(*pmdp) |= ((((unsigned long) ptep) >> (_SUN4C_PGD_PAGE_SHIFT)) & _SUN4C_PGD_PFN_MASK); |
ptep | 367 | arch/sparc/mm/sun4c.c | sun4c_set_pte(pte_t *ptep, pte_t entry) |
ptep | 370 | arch/sparc/mm/sun4c.c | *ptep = entry; |
ptep | 444 | arch/sparc/mm/sun4c.c | pte_t *ptep; |
ptep | 494 | arch/sparc/mm/sun4c.c | ptep = sun4c_pte_offset(pmdp, vaddr); |
ptep | 495 | arch/sparc/mm/sun4c.c | *ptep = sun4c_mk_pte(vaddr, SUN4C_PAGE_KERNEL); |
ptep | 157 | include/asm-alpha/pgtable.h | extern inline void pmd_set(pmd_t * pmdp, pte_t * ptep) |
ptep | 158 | include/asm-alpha/pgtable.h | { pmd_val(*pmdp) = _PAGE_TABLE | ((((unsigned long) ptep) - PAGE_OFFSET) << (32-PAGE_SHIFT)); } |
ptep | 174 | include/asm-alpha/pgtable.h | extern inline int pte_inuse(pte_t *ptep) { return mem_map[MAP_NR(ptep)].reserved || mem_map[MAP_NR(ptep)].count != 1; } |
ptep | 175 | include/asm-alpha/pgtable.h | extern inline void pte_clear(pte_t *ptep) { pte_val(*ptep) = 0; } |
ptep | 176 | include/asm-alpha/pgtable.h | extern inline void pte_reuse(pte_t * ptep) |
ptep | 178 | include/asm-alpha/pgtable.h | if (!mem_map[MAP_NR(ptep)].reserved) |
ptep | 179 | include/asm-alpha/pgtable.h | mem_map[MAP_NR(ptep)].count++; |
ptep | 146 | include/asm-i386/pgtable.h | extern inline int pte_inuse(pte_t *ptep) { return mem_map[MAP_NR(ptep)].reserved || mem_map[MAP_NR(ptep)].count != 1; } |
ptep | 147 | include/asm-i386/pgtable.h | extern inline void pte_clear(pte_t *ptep) { pte_val(*ptep) = 0; } |
ptep | 148 | include/asm-i386/pgtable.h | extern inline void pte_reuse(pte_t * ptep) |
ptep | 150 | include/asm-i386/pgtable.h | if (!mem_map[MAP_NR(ptep)].reserved) |
ptep | 151 | include/asm-i386/pgtable.h | mem_map[MAP_NR(ptep)].count++; |
ptep | 190 | include/asm-mips/page.h | extern inline void pte_clear(pte_t *ptep) { pte_val(*ptep) = 0; } |
ptep | 234 | include/asm-mips/page.h | extern inline void pgd_set(pgd_t * pgdp, pte_t * ptep) |
ptep | 235 | include/asm-mips/page.h | { pgd_val(*pgdp) = _PAGE_TABLE | (unsigned long) ptep; } |
ptep | 89 | include/asm-sparc/page.h | extern void (*set_pte)(pte_t *ptep, pte_t entry); |
ptep | 169 | mm/filemap.c | static inline int filemap_sync_pte(pte_t * ptep, struct vm_area_struct *vma, |
ptep | 172 | mm/filemap.c | pte_t pte = *ptep; |
ptep | 181 | mm/filemap.c | set_pte(ptep, pte_mkclean(pte)); |
ptep | 187 | mm/filemap.c | pte_clear(ptep); |